_CORE
AI & Agentic Systems Core Information Systems Cloud & Platform Engineering Data Platform & Integration Security & Compliance QA, Testing & Observability IoT, Automation & Robotics Mobile & Digital Banking & Finance Insurance Public Administration Defense & Security Healthcare Energy & Utilities Telco & Media Manufacturing Logistics & E-commerce Retail & Loyalty
References Technologies Blog Know-how Tools
About Collaboration Careers
CS EN
Let's talk

AI Observability — Monitoring LLMs in Production

08. 06. 2024 1 min read CORE SYSTEMSai
AI Observability — Monitoring LLMs in Production

You’ve deployed an LLM to production. How well is it performing? How much does it cost? Is it hallucinating more? You need AI observability.

What to Measure

  • Latency: TTFT, total generation time
  • Cost: Token usage per request/user/feature
  • Quality: User feedback, LLM-as-judge scores
  • Errors: API failures, rate limits, timeouts

Tooling

LangSmith: Tracing, evaluation. Langfuse: Open-source, self-hostable — our choice. Arize Phoenix: Evals and experiments.

Cost Management

  • Dashboard with real-time cost per feature
  • Alerting on cost anomalies
  • Prompt optimization reviews
  • Model routing — cheaper model where it suffices

AI Without Observability Is a Ticking Bomb

Implement tracing from day one. Langfuse for self-hosted, LangSmith for convenience.

ai observabilityllm monitoringmlopsproduction ai
Share:

CORE SYSTEMS

Stavíme core systémy a AI agenty, které drží provoz. 15 let zkušeností s enterprise IT.

Need help with implementation?

Our experts can help with design, implementation, and operations. From architecture to production.

Contact us