Topic: "inference-platform"
bentoml/BentoML
The easiest way to serve AI apps and models - Build Model Inference APIs, Job queues, LLM apps, Multi-model pipelines, and more!
Language: Python - Size: 95.5 MB - Last synced at: 1 day ago - Pushed at: 1 day ago - Stars: 7,684 - Forks: 837

InftyAI/llmaz
βΈοΈ Easy, advanced inference platform for large language models on Kubernetes. π Star to support our work!
Language: Go - Size: 9.79 MB - Last synced at: 5 days ago - Pushed at: 7 days ago - Stars: 162 - Forks: 25
