LLM Observability Tools  
 
  	
	
        
  
	
	  
- 01.    Langfuse LLM Engineering Platform - 
Traces, evals, prompt management and metrics to debug and improve your LLM application.  
 
- 02.      Dynatrace LLM Observability
Monitor, optimize, and secure Generative AI applications, LLMs, and agentic workflows
 
 - 03.    Datadog LLM Observability - Develop, evaluate, and monitor LLM Applications with confidence
 
 - 04.     Lunary - An observability tool focused on retrieval-augmented generation (RAG) 
 
- 05.     Traceloop monitors what your model says, how fast it responds, and when things start to slip 
 
- 06.    LangSmith is a unified observability & evals platform where teams can debug, test, and monitor AI app performance — whether building with LangChain or not. 
   
 
- 07.     Portkey equips AI teams with everything they need to go to production - Gateway, Observability, Guardrails, Governance, and Prompt Management, all in one platform. 
 
- 08.     TruLens is a software tool that helps you to objectively measure the quality and effectiveness of your LLM-based applications using feedback functions. 
 
- 09.     Phoenix Arize - An Open-source LLM tracing, evaluation and Observability - Built on top of OpenTelemetry - is agnostic of vendor, framework, and language. 
 
- 10.    HoneyHive Modern AI Observability and Evaluation - to develop, evaluate, and observe AI agents 
 
- 11.    Helicone is an open source platform for monitoring, debugging, and improving LLM applications.