An open API service providing repository metadata for many open source software ecosystems.

Topic: "gguf-models"

pollockjj/ComfyUI-MultiGPU Fork of neuratech-ai/ComfyUI-MultiGPU

This custom_node for ComfyUI adds one-click "Virtual VRAM" for any GGUF UNet and CLIP loader, managing the offload of layers to DRAM or VRAM to maximize the latent space of your card. Also includes nodes for directly loading entire components (UNet, CLIP, VAE) onto the device you choose. Includes 16 examples covering common use cases.

Language: Python - Size: 14.1 MB - Last synced at: about 2 months ago - Pushed at: about 2 months ago - Stars: 212 - Forks: 9

ashioyajotham/fingpt_trader

An algorithmic trading system based on FinGPT, demonstrating new applications of large pre-trained Language Models in quantitative finance.

Language: Python - Size: 879 KB - Last synced at: 21 days ago - Pushed at: 21 days ago - Stars: 11 - Forks: 4

controlecidadao/samantha_ia

Experimental interface environment for open source LLM, designed to democratize the use of AI. Powered by llama-cpp, llama-cpp-python and Gradio.

Language: Python - Size: 23.1 MB - Last synced at: 17 days ago - Pushed at: 17 days ago - Stars: 10 - Forks: 1

HelpingAI/inferno

Run Llama 3.3, DeepSeek-R1, Phi-4, Gemma 3, Mistral Small 3.1, and other state-of-the-art language models locally with scorching-fast performance. Inferno provides an intuitive CLI and an OpenAI/Ollama-compatible API, putting the inferno of AI innovation directly in your hands.

Language: Python - Size: 145 KB - Last synced at: 23 days ago - Pushed at: 23 days ago - Stars: 6 - Forks: 0

mili-tan/Onllama.GGUFLinkOut

Create out symbolic links for the GGUF Models in Ollama Blobs. for use in other applications such as Llama.cpp/Jan/LMStudio etc. / 将 Ollama GGUF 模型文件软链接出,以便其他应用使用。

Language: C# - Size: 15.6 KB - Last synced at: about 2 months ago - Pushed at: 4 months ago - Stars: 5 - Forks: 1

LuisMiSanVe/GGUF-to-PyTorchTensor

Simple Python Script that converts the Weight of a GGUF Model to a PyTorch Tensor

Language: Python - Size: 28.3 KB - Last synced at: 2 months ago - Pushed at: 4 months ago - Stars: 1 - Forks: 0

mahctf/run-llama-server

This repository contains a Bash script that simplifies launching `llama-server` and manages GPU resources efficiently. It scans for models, sets memory limits, and adjusts GPU layers based on available VRAM, ensuring a smooth experience for users. 🐱💻✨

Language: Shell - Size: 119 KB - Last synced at: 4 days ago - Pushed at: 4 days ago - Stars: 0 - Forks: 0

nemmusu/run-llama-server

This is a Bash script to automatically launch llama-server, detects available .gguf models, and selects GPU layers based on your free VRAM.

Language: Shell - Size: 117 KB - Last synced at: 8 days ago - Pushed at: 8 days ago - Stars: 0 - Forks: 0

echenim/hf-batch-downloader

Automate bulk downloads of Hugging Face LLMs with retry logic, manifest export, checksum validation, and usage reporting. Ideal for managing GGUF models at scale.

Language: Python - Size: 5.86 KB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 0 - Forks: 0

testli-ai/outlines-llama-cpp-python-streaming-output

This repository demonstrates how to use outlines and llama-cpp-python for structured JSON generation with streaming output, integrating llama.cpp for local model inference and outlines for schema-based text generation.

Language: Python - Size: 98.6 KB - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 0 - Forks: 0