llm-observability
Here are 29 public repositories matching this topic...
Language:All
Sort:Most stars
🪢 Open source LLM engineering platform: LLM Observability, metrics, evals, prompt management, playground, datasets. Integrates with OpenTelemetry, Langchain, OpenAI SDK, LiteLLM, and more. 🍊YC W23
- Updated
Nov 29, 2025 - TypeScript
Debug, evaluate, and monitor your LLM applications, RAG systems, and agentic workflows with comprehensive tracing, automated evaluations, and production-ready dashboards.
- Updated
Nov 28, 2025 - Python
Next-generation AI Agent Optimization Platform: Cozeloop addresses challenges in AI agent development by providing full-lifecycle management capabilities from development, debugging, and evaluation to monitoring.
- Updated
Nov 29, 2025 - Go
🧊 Open source LLM observability platform. One line of code to monitor, evaluate, and experiment. YC W23 🍓
- Updated
Nov 27, 2025 - TypeScript
Open Source TypeScript AI Agent Framework with built-in LLM Observability
- Updated
Nov 29, 2025 - TypeScript
The open-source LLMOps platform: prompt playground, prompt management, LLM evaluation, and LLM observability all in one place.
- Updated
Nov 28, 2025 - Python
Laminar - open-source all-in-one platform for engineering AI products. Create data flywheel for your AI app. Traces, Evals, Datasets, Labels. YC S24.
- Updated
Nov 27, 2025 - TypeScript
The open source post-building layer for agents. Our environment data and evals power agent post-training (RL, SFT) and monitoring.
- Updated
Nov 29, 2025 - Python
Build, Improve Performance, and Productionize your LLM Application with an Integrated Framework
- Updated
Nov 26, 2024 - TypeScript
React components for visualizing traces from AI agents
- Updated
Nov 14, 2025 - TypeScript
Modular, open source LLMOps stack that separates concerns: LiteLLM unifies LLM APIs, manages routing and cost controls, and ensures high-availability, while Langfuse focuses on detailed observability, prompt versioning, and performance evaluations.
- Updated
Feb 15, 2025
A powerful AI observability framework that provides comprehensive insights into agent interactions across platforms, enabling developers to monitor, analyze, and optimize AI-driven applications with minimal integration effort.
- Updated
May 14, 2025 - Python
A comprehensive solution for monitoring your AI models in production
- Updated
Nov 13, 2025 - Python
Open-source observability for your LLM application.
- Updated
Jan 2, 2025 - Python
🪢 Auto-generated Java Client for Langfuse API
- Updated
Aug 21, 2025 - Java
The reliability layer between your code and LLM providers.
- Updated
Jan 6, 2025 - Go
A Python package for tracking and analyzing LLM usage across different models and applications. It is primarily designed as a library for integration into development process of LLM-based agentic workflow tooling, providing robust tracking capabilities.
- Updated
Jul 7, 2025 - Python
AI agent platform for building multi-agent systems with orchestration, memory, RAG, workflows, and enterprise observability.
- Updated
Oct 27, 2025
Check if your AI sounds like your brand, stays safe, and behaves consistently. Works with your custom GPTs, hosted APIs, and local models. Get detailed reports in minutes, not days.
- Updated
Nov 12, 2025 - HTML
Improve this page
Add a description, image, and links to thellm-observability topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with thellm-observability topic, visit your repo's landing page and select "manage topics."