diff --git a/README.md b/README.md index 300bf46..72c61a2 100644 --- a/README.md +++ b/README.md @@ -206,6 +206,7 @@ An awesome & curated list of the best LLMOps tools for developers. | [Humanloop](https://humanloop.com) | The LLM evals platform for enterprises, providing tools to develop, evaluate, and observe AI systems. | | | [Hypersigil](https://github.com/hypersigilhq/hypersigil) | Open-source prompt lifecycle management and gateway with a Web UI. | ![GitHub Badge](https://img.shields.io/github/stars/hypersigilhq/hypersigil.svg?style=flat-square) | | [Izlo](https://getizlo.com/) | Prompt management tools for teams. Store, improve, test, and deploy your prompts in one unified workspace. | | +| [kb-arena](https://github.com/xmpuspus/kb-arena) | Open-source RAG evaluation arena that benchmarks 7 retrieval strategies (naive vector, contextual, QnA pairs, knowledge graph, RAPTOR, PageIndex, hybrid) on your own docs. ELO-based blind A/B Strategy Arena, RAGAS metrics, auto-generated question sets across 5 difficulty tiers, multi-provider support (Anthropic, OpenAI, Ollama), and per-strategy cost tracking. | ![GitHub Badge](https://img.shields.io/github/stars/xmpuspus/kb-arena.svg?style=flat-square) | | [Keywords AI](https://keywordsai.co/) | A unified DevOps platform for AI software. Keywords AI makes it easy for developers to build LLM applications. | | | [MLflow](https://github.com/mlflow/mlflow/tree/master) | An open-source framework for the end-to-end machine learning lifecycle, helping developers track experiments, evaluate models/prompts, deploy models, and add observability with tracing. | ![GitHub Badge](https://img.shields.io/github/stars/mlflow/mlflow.svg?style=flat-square) | | [Laminar](https://github.com/lmnr-ai/lmnr) | Open-source all-in-one platform for engineering AI products. Traces, Evals, Datasets, Labels. | ![GitHub Badge](https://img.shields.io/github/stars/lmnr-ai/lmnr.svg?style=flat-square) |