llamacpp
Here are 632 public repositories matching this topic...
Language:All
Sort:Most stars
Jan is an open source alternative to ChatGPT that runs 100% offline on your computer.
- Updated
Dec 17, 2025 - TypeScript
Your AI second brain. Self-hostable. Get answers from the web or your docs. Build custom agents, schedule automations, do deep research. Turn any online or local LLM into your personal, autonomous AI (gpt, claude, gemini, llama, qwen, mistral). Get started - free.
- Updated
Dec 8, 2025 - Python
Unified framework for building enterprise RAG pipelines with small, specialized models
- Updated
Jul 24, 2025 - Python
Swap GPT for any LLM by changing a single line of code. Xinference lets you run open-source, speech, and multimodal models on cloud, on-prem, or your laptop — all through one unified, production-ready inference API.
- Updated
Dec 17, 2025 - Python
Private & local AI personal knowledge management app for high entropy people.
- Updated
May 13, 2025 - JavaScript
A web interface for chatting with Alpaca through llama.cpp. Fully dockerized, with an easy to use API.
- Updated
Nov 21, 2025 - Svelte
State of the Art Natural Language Processing
- Updated
Dec 17, 2025 - Scala
Kernels & AI inference engine for mobile devices.
- Updated
Dec 17, 2025 - C++
Production ready toolkit to run AI locally
- Updated
Dec 17, 2025 - Kotlin
The most no-nonsense, locally or API-hosted AI code completion plugin for Visual Studio Code - like GitHub Copilot but 100% free.
- Updated
Aug 7, 2025 - TypeScript
⚡ Python-free Rust inference server — OpenAI-API compatible. GGUF + SafeTensors, hot model swap, auto-discovery, single binary. FREE now, FREE forever.
- Updated
Dec 17, 2025 - Rust
AGiXT is a dynamic AI Agent Automation Platform that seamlessly orchestrates instruction management and complex task execution across diverse AI providers. Combining adaptive memory, smart features, and a versatile plugin system, AGiXT delivers efficient and comprehensive AI solutions.
- Updated
Dec 16, 2025 - Python
LSP-AI is an open-source language server that serves as a backend for AI-powered functionality, designed to assist and empower software engineers, not replace them.
- Updated
Jan 7, 2025 - Rust
RamaLama is an open-source developer tool that simplifies the local serving of AI models from any source and facilitates their use for inference in production, all through the familiar language of containers.
- Updated
Dec 17, 2025 - Python
Maid is a cross-platform Flutter app for interfacing with GGUF / llama.cpp models locally, and with Ollama and OpenAI models remotely.
- Updated
Jul 28, 2025 - Dart
Improve this page
Add a description, image, and links to thellamacpp topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with thellamacpp topic, visit your repo's landing page and select "manage topics."