Skip to content
#

llm-observability

Here are 19 public repositories matching this topic...

RAG Eval Observability is a production-ready, open-source platform for building, evaluating, and monitoring Retrieval-Augmented Generation (RAG) systems. It pairs a ChatGPT-style UI with a robust backend for document ingestion, multiple retrieval strategies, offline evaluation, and real-time observability, along with backend CI/CD deployed on Azure

  • Updated Dec 17, 2025
  • Python

A practical and critical evaluation of Retrieval-Augmented Generation (RAG) systems on legal/insurance documents using RAGAS. This project analyzes metric failures, false negatives, retrieval pitfalls, and proposes a more realistic composite evaluation score.

  • Updated Feb 3, 2026
  • Python

Token cost is a design problem, not a billing problem. Most LLM cost overruns come from architectural waste, not model pricing. This tool is a token waste profiler that helps you understand where your tokens are going and which ones are useless.

  • Updated Jan 18, 2026
  • Python

An AI-powered multi-agent system that demonstrates clinical triage, OTC medication recommendations, and e-pharmacy integration for respiratory conditions. Built with modular agents that collaborate to provide safe, intelligent healthcare assistance.

  • Updated Dec 8, 2025
  • Python

Improve this page

Add a description, image, and links to the llm-observability topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the llm-observability topic, visit your repo's landing page and select "manage topics."

Learn more