self-hosted-ai
Here are 23 public repositories matching this topic...
Language:All
Sort:Most stars
Perplexica is an AI-powered answering engine.
- Updated
Feb 13, 2026 - TypeScript
High-performance lightweight proxy and load balancer for LLM infrastructure. Intelligent routing, automatic failover and unified model discovery across local and remote inference backends.
- Updated
Feb 20, 2026 - Go
Local first speech AI engine for transcription, TTS, and voice workflows.
- Updated
Feb 20, 2026 - Rust
Small Language Model Inference, Fine-Tuning and Observability. No GPU, no labeled data needed.
- Updated
Feb 3, 2026 - Python
Free, open-source alternative to Weavy AI, Krea Nodes, Freepik Spaces & FloraFauna AI — node-based AI workflow builder for generative image & video pipelines
- Updated
Feb 19, 2026 - JavaScript
emotional AI Companions for personal relationships
- Updated
Jan 21, 2026 - Python
Run IBM Granite 4.0 locally on Raspberry Pi 5 with Ollama.This is a privacy-first AI. Your data never leaves your device because it runs 100% locally. There are no cloud uploads and no third-party tracking.
- Updated
Dec 30, 2025 - Shell
Recallium is a local, self-hosted universal AI memory system providing a persistent knowledge layer for developer tools (Copilot, Cursor, Claude Desktop). It eliminates "AI amnesia" by automatically capturing, clustering, and surfacing decisions and patterns across all projects. It uses the MCP for universal compatibility and ensures privacy
- Updated
Feb 11, 2026 - Batchfile
A private, local RAG (Retrieval-Augmented Generation) system using Flowise, Ollama, and open-source LLMs to chat with your documents securely and offline.
- Updated
Nov 21, 2024
AI SMS Auto-Responder for Android. Turn your Android device into an autonomous AI communication hub. A Python-based SMS auto-responder running natively on Termux, powered by LLMs (OpenRouter/Ollama) with a sleek Web & Terminal UI.
- Updated
Feb 15, 2026 - Python
Self-hosted AI chat interface with RAG, long-term memory, and admin controls. Works with TabbyAPI, Ollama, vLLM, and any OpenAI-compatible API.
- Updated
Feb 17, 2026 - Python
Production-ready guide for connecting OpenClaw to a Telegram Bot. Build a self-hosted Telegram AI Agent using OpenClaw Gateway, pairing, and streaming responses.
- Updated
Feb 18, 2026
Powers the local RAG pipeline in the BrainDrive Chat w/ Docs plugin.
- Updated
Nov 16, 2025 - Python
Production-ready test-time compute optimization framework for LLM inference. Implements Best-of-N, Sequential Revision, and Beam Search strategies. Validated with models up to 7B parameters.
- Updated
Jan 27, 2026 - Python
🚀 7 Ways to Run Any LLMs Locally - Simple Methods
- Updated
Jan 25, 2026
Deployment of a self-hosted LLM infrastructure using Ollama and Open WebUI on Linux, including custom model creation, API integration, and system-level troubleshooting.
- Updated
Feb 14, 2026
Kernel-first, lightweight OSS for self-hosted multi-turn AI chat in Docker, with transparent runtime logic, baseline performance data, and split-licensed extensibility.
- Updated
Feb 19, 2026 - TypeScript
LocalPrompt is an AI-powered tool designed to refine and optimize AI prompts, helping users run locally hosted AI models like Mistral-7B for privacy and efficiency. Ideal for developers seeking to run LLMs locally without external APIs.
- Updated
Feb 10, 2025 - Python
Web-Based Q&A Tool enables users to extract and query website content using FastAPI, FAISS, and a local TinyLlama-1.1B model—without external APIs. Built with React, it offers a minimal UI for seamless AI-driven search
- Updated
Mar 20, 2025 - Python
🌳 Open-source RAPTOR: Recursive Abstractive Processing for Tree-Organized Retrieval - Complete open-source implementation with 100% local LLMs (Granite Code 8B + mxbai-embed-large)
- Updated
Oct 18, 2025 - Python
Improve this page
Add a description, image, and links to theself-hosted-ai topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with theself-hosted-ai topic, visit your repo's landing page and select "manage topics."