An open API service providing repository metadata for many open source software ecosystems.

GitHub topics: offline-llm

GGUFloader/gguf-loader

Run Mistral, LLaMA, and DeepSeek locally on Windows with zero setup — no Python required.

Language: Python - Size: 108 MB - Last synced at: 1 day ago - Pushed at: 1 day ago - Stars: 6 - Forks: 6

khoj-ai/khoj

Your AI second brain. Self-hostable. Get answers from the web or your docs. Build custom agents, schedule automations, do deep research. Turn any online or local LLM into your personal, autonomous AI (gpt, claude, gemini, llama, qwen, mistral). Get started - free.

Language: Python - Size: 111 MB - Last synced at: 5 days ago - Pushed at: 5 days ago - Stars: 30,602 - Forks: 1,749

reydiego24/Faster-Local-Voice-AI

Optimize your voice AI experience with Faster-Local-Voice-AI. Achieve low-latency STT and TTS on Ubuntu, all offline and fully configurable. 🚀💻

Language: Python - Size: 287 KB - Last synced at: 6 days ago - Pushed at: 6 days ago - Stars: 1 - Forks: 0

ineelhere/shiny.ollama

Chat offline with open-source LLMs like deepseek-r1, nemotron, qwen, llama and more all through a simple R package powered by Shiny and Ollama. 🚀

Language: R - Size: 4.93 MB - Last synced at: 2 days ago - Pushed at: 5 months ago - Stars: 20 - Forks: 0

yashboss1111/LLMux

A lightweight local LLM chat with a web UI and a C‑based server that runs any LLM chat executable as a child and communicates via pipes

Language: C - Size: 1.36 MB - Last synced at: 13 days ago - Pushed at: 13 days ago - Stars: 1 - Forks: 0

cyberguard-ai/local-llm-server

A containerized, offline-capable LLM API powered by Ollama. Automatically pulls models and serves them via a REST API. Perfect for homelab, personal AI assistants, and portable deployments.

Language: Python - Size: 28.3 KB - Last synced at: 19 days ago - Pushed at: 19 days ago - Stars: 0 - Forks: 0

karanIPS/claude-deep-research

Claude Deep Research config for Claude Code.

Size: 10.7 KB - Last synced at: 19 days ago - Pushed at: 19 days ago - Stars: 10 - Forks: 0

2LKitlab/obsidian-local-ai

Offline AI assistant plugin for Obsidian using encrypted local LLM models.

Language: Python - Size: 7.81 KB - Last synced at: 28 days ago - Pushed at: 3 months ago - Stars: 4 - Forks: 0

dieharders/obrew-studio-server

Obrew Studio - Server: A self-hostable machine learning engine. Build agents and schedule workflows private to you.

Language: Python - Size: 138 MB - Last synced at: 17 days ago - Pushed at: 3 months ago - Stars: 11 - Forks: 1

lurkydismal/LLMux

A lightweight local LLM chat with a web UI and a C‑based server that runs any LLM chat executable as a child and communicates via pipes

Language: C - Size: 1.36 MB - Last synced at: about 1 month ago - Pushed at: 2 months ago - Stars: 0 - Forks: 0

umutcamliyurt/TextCloak

A tool for concealing writing style using LLM

Language: Python - Size: 17.6 KB - Last synced at: 3 months ago - Pushed at: 11 months ago - Stars: 16 - Forks: 0

isaccanedo/khoj

🍳 Your AI second brain. Self-hostable. Get answers from the web or your docs. Build custom agents, schedule automations, do deep research. Turn any online or local LLM into your personal, autonomous AI (gpt, claude, gemini, llama, qwen, mistral). Get started - free

Language: Python - Size: 67.8 MB - Last synced at: 5 months ago - Pushed at: 6 months ago - Stars: 1 - Forks: 0

JKL404/Offline-Chat-Application

A private, free, offline-first chat application powered by Open Source AI models like DeepSeek, Llama, Mistral, etc. through Ollama.

Language: JavaScript - Size: 2.42 MB - Last synced at: 4 months ago - Pushed at: 6 months ago - Stars: 3 - Forks: 0

kpcyrd/summarize-cli

Attempt to summarize text from `stdin`, using a large language model (locally and offline), to `stdout`

Language: Rust - Size: 25.4 KB - Last synced at: 4 days ago - Pushed at: almost 2 years ago - Stars: 9 - Forks: 3