LLM-Tracing, Eval, Prompt-Management
Langfuse traced jeden LLM-Call der Plattform: Prompt-Input, Response-Output, Latency, Token-Counts, Cost in EUR/USD, User, Session, Feedback-Score. Prompt-Versioning + A/B-Testing. Evaluations gegen Golden-Sets. Alerts bei Anomalien (Cost-Spike, Latency-Drift). Integration über LiteLLM-Hooks + nativen Langfuse-SDKs in Cowork-AI, NemoClaw, OpenClaw.