llm-evaluation
Here are 374 public repositories matching this topic...
Language:All
Sort:Most stars
The open source developer platform to build AI agents and models with confidence. Enhance your AI applications with end-to-end tracking, observability, and evaluations, all in one integrated platform.
- Updated
Nov 29, 2025 - Python
🪢 Open source LLM engineering platform: LLM Observability, metrics, evals, prompt management, playground, datasets. Integrates with OpenTelemetry, Langchain, OpenAI SDK, LiteLLM, and more. 🍊YC W23
- Updated
Nov 28, 2025 - TypeScript
Debug, evaluate, and monitor your LLM applications, RAG systems, and agentic workflows with comprehensive tracing, automated evaluations, and production-ready dashboards.
- Updated
Nov 28, 2025 - Python
The LLM Evaluation Framework
- Updated
Nov 28, 2025 - Python
Test your prompts, agents, and RAGs. AI Red teaming, pentesting, and vulnerability scanning for LLMs. Compare performance of GPT, Claude, Gemini, Llama, and more. Simple declarative configs with command line and CI/CD integration.
- Updated
Nov 29, 2025 - TypeScript
AI Observability & Evaluation
- Updated
Nov 25, 2025 - Jupyter Notebook
the LLM vulnerability scanner
- Updated
Nov 26, 2025 - Python
ReLE评测:中文AI大模型能力评测(持续更新):目前已囊括303个大模型,覆盖chatgpt、gpt-5、o4-mini、谷歌gemini-2.5、Claude4.5、智谱GLM-Z1、文心一言、qwen3-max、百川、讯飞星火、商汤senseChat、minimax等商用模型, 以及kimi-k2、ernie4.5、minimax-M1、DeepSeek-R1-0528、deepseek-v3.2、qwen3-2507、llama4、GLM4.5、gemma3、mistral等开源大模型。不仅提供排行榜,也提供规模超200万的大模型缺陷库!方便广大社区研究分析、改进大模型。
- Updated
Nov 28, 2025
🐢 Open-Source Evaluation & Testing library for LLM Agents
- Updated
Nov 18, 2025 - Python
🧊 Open source LLM observability platform. One line of code to monitor, evaluate, and experiment. YC W23 🍓
- Updated
Nov 27, 2025 - TypeScript
AutoRAG: An Open-Source Framework for Retrieval-Augmented Generation (RAG) Evaluation & Optimization with AutoML-Style Automation
- Updated
Nov 20, 2025 - Python
The LLM's practical guide: From the fundamentals to deploying advanced LLM and RAG apps to AWS using LLMOps best practices
- Updated
Mar 8, 2025 - Python
The open-source LLMOps platform: prompt playground, prompt management, LLM evaluation, and LLM observability all in one place.
- Updated
Nov 28, 2025 - Python
Evaluation and Tracking for LLM Experiments and AI Agents
- Updated
Nov 25, 2025 - Python
Laminar - open-source all-in-one platform for engineering AI products. Create data flywheel for your AI app. Traces, Evals, Datasets, Labels. YC S24.
- Updated
Nov 27, 2025 - TypeScript
Agentic LLM Vulnerability Scanner / AI red teaming kit 🧪
- Updated
Nov 27, 2025 - Python
Comprehensive resources on Generative AI, including a detailed roadmap, projects, use cases, interview preparation, and coding preparation.
- Updated
Nov 25, 2025 - Jupyter Notebook
Build, enrich, and transform datasets using AI models with no code
- Updated
Oct 23, 2025 - TypeScript
Prompty makes it easy to create, manage, debug, and evaluate LLM prompts for your AI applications. Prompty is an asset class and format for LLM prompts designed to enhance observability, understandability, and portability for developers.
- Updated
Nov 26, 2025 - Python
UQLM: Uncertainty Quantification for Language Models, is a Python package for UQ-based LLM hallucination detection
- Updated
Nov 21, 2025 - Python
Improve this page
Add a description, image, and links to thellm-evaluation topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with thellm-evaluation topic, visit your repo's landing page and select "manage topics."