GitHub topics: llm-server
onnx/turnkeyml
Local LLM Server with NPU Acceleration
Language: Python - Size: 1.9 MB - Last synced at: 8 days ago - Pushed at: 8 days ago - Stars: 173 - Forks: 23

pikocloud/pikobrain
Function-calling API for LLM from multiple providers
Language: Go - Size: 408 KB - Last synced at: about 1 month ago - Pushed at: 9 months ago - Stars: 5 - Forks: 0

fcn94/llm_stream_endpoint
Simple LLM Rest API using Rust, Warp and Candle. Dedicated for quantized version of either phi-2 ( default) , Mistral, or Llama. Work using CPU or CUDA
Language: Rust - Size: 74.2 KB - Last synced at: 10 days ago - Pushed at: about 1 year ago - Stars: 4 - Forks: 1
