llama-cpp
Here are 108 public repositories matching this topic...
Language:All
Sort:Most stars
Maid is a cross-platform Flutter app for interfacing with GGUF / llama.cpp models locally, and with Ollama and OpenAI models remotely.
- Updated
Mar 16, 2025 - Dart
Run AI models locally on your machine with node.js bindings for llama.cpp. Enforce a JSON schema on the model output on the generation level
- Updated
Feb 21, 2025 - TypeScript
LLama.cpp rust bindings
- Updated
Jun 27, 2024 - Rust
This repo is to showcase how you can run a model locally and offline, free of OpenAI dependencies.
- Updated
Jul 12, 2024 - Python
Local ML voice chat using high-end models.
- Updated
Mar 15, 2025 - C++
Making offline AI models accessible to all types of edge devices.
- Updated
Feb 12, 2024 - Dart
Review/Check GGUF files and estimate the memory usage and maximum tokens per second.
- Updated
Mar 14, 2025 - Go
LLaMA Server combines the power of LLaMA C++ with the beauty of Chatbot UI.
- Updated
Jun 10, 2023 - Python
InsightSolver: Colab notebooks for exploring and solving operational issues using deep learning, machine learning, and related models.
- Updated
Jun 12, 2024 - Jupyter Notebook
Improve this page
Add a description, image, and links to thellama-cpp topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with thellama-cpp topic, visit your repo's landing page and select "manage topics."