Python bindings for llama.cpp
Run Local LLMs on Any Device. Open-source
Personal AI, On Personal Devices
Interface for OuteTTS models
Claude Code, but it runs on your Mac for free
Run a full local LLM stack with one command using Docker
Your Personal AI Assistant; easy to install, deploy on local or coud
Inference Llama 2 in one file of pure C
An easy-to-understand framework for LLM samplers
Oobabooga - The definitive Web UI for local AI, with powerful features
Qwen3 is the large language model series developed by Qwen team
Powerful Android AI agent with tools, automation, and Linux shell
Performance-optimized AI inference on your GPUs
Towards Human-Sounding Speech
GLM-4 series: Open Multilingual Multimodal Chat LMs
Run GGUF models easily with a UI or API. One File. Zero Install.
Run any Llama 2 locally with gradio UI on GPU or CPU from anywhere
Chinese LLaMA & Alpaca large language model + local CPU/GPU training
JetBrains’ 4B parameter code model for completions