As enterprises scale large language models (LLMs) into production, site reliability engineers (SREs) and platform operators face a new set of challenges. Traditional application metrics—CPU usage, request throughput, memory consumption—are no longer enough. With LLMs, reliability and efficacy are defined by entirely new dynamics—token-level performance, cache efficiency, and inference pipeline latency. This article explores how llm-d, an open source project co-developed with the leading AI vendors (Red Hat, Google, IBM, etc.) and integrated into Red Hat OpenShift AI 3.0, redefines observability for LLM workloads. New service level objectives (SLO) in the age of LLMs Users expect responsive applications, including those enhanced with AI, and enterprises require consistent performance to turn a pilot project into a profitable production application at scale. While SLOs in traditional microservice architectures are usually framed around request latency and error rate, the user experience for LLMs depends on more nuanced measures, including: - Time to first token (TTFT): Measures the delay before the initial token of a response is streamed to the user. A lower TTFT is needed to provide a responsive and immediate user experience, especially in interactive applications. - Time per output token (TPOT): Indicates the speed at which tokens are generated after the process begins. A consistent and low TPOT provides a smooth and efficient streaming of the complete response to the user. - Cache hit rate: Represents the proportion of requests that can use previously computed context stored in GPU memory. A high cache hit rate significantly reduces computational overhead and improves overall system throughput by avoiding redundant processing. - Prefill vs. decode latency: Distinguishes between the time taken for expensive prompt processing (prefill) and token-by-token decoding. Understanding this distinction helps to optimize resource allocation and identify bottlenecks in the different stages of request processing. - Goodput: The number of requests successfully served within their defined SLO budgets. Maximizing goodput is needed for the system to reliably meet its performance commitments. Together, these represent what users experience: responsiveness, throughput, and consistency. The challenge of managing these new SLOs Existing monitoring tools don’t know about tokens or caches. A traditional latency metric can tell you that a request took 400ms, but it can’t break down whether that was due to routing delays, cache misses, or GPU scheduling. In a distributed inference stack—where requests move through gateways, routers, schedulers, and GPU workers—observability blind spots are everywhere. Without token-aware metrics, operators can’t tell whether performance degradation is caused by routing imbalance, prefill-cache fragmentation, or overloaded GPUs. This makes it difficult to debug issues quickly in production, plan GPU capacity, or guarantee user-facing SLOs. What is llm-d? llm-d is a community-driven, Kubernetes-native project that disaggregates inference into composable services, including: - Endpoint picker (EPP): An semantic router that makes cache-aware and load-aware scheduling decisions. - Decode and prefill services: Separating heavy prompt ingestion (prefill) from sequential token generation (decode). Prefill can even run on CPU if GPU resources are constrained. - Key-value (KV) cache management: Centralized indexing to maximize cache reuse across workloads. On OpenShift AI, llm-d provides an operationally consistent way to run vLLM and other serving engines with advanced observability. How llm-d Helps solve the observability gap llm-d integrates deeply with Prometheus, Grafana, and OpenTelemetry, exposing both system-level and LLM-specific metrics: - Cache-aware metrics: Cache hit ratios, cache size utilization, and KV movements. - Token-level latency metrics: TTFT, TPOT, and end-to-end latency. - Routing transparency: Metrics and traces showing why requests were routed to specific pods. - Tracing across components: With OpenTelemetry, operators can follow a request through Inference Gateway (IGW) to Scheduler to vLLM workers. This transforms llm-d from a black box into an auditable, measurable inference stack. Example PromQL queries Here are some PromQL queries that illustrate how SREs can use llm-d metrics to monitor token-level in production as of llm-d version 0.2.0. 1. TTFT P95 histogram_quantile(0.95, sum(rate(llmd_inference_ttft_bucket[5m])) by (le))undefinedTracks the 95th percentile latency until the first token is emitted. 2. TPOT average rate(llmd_inference_output_tokens_sum[5m]) undefined/ rate(llmd_inference_requests_total[5m]) Measures how quickly tokens are generated once decoding starts. 3. Cache hit rate (sum(rate(llmd_kvcache_hits_total[5m])) by (pool)) undefined/ (sum(rate(llmd_kvcache_requests_total[5m])) by (pool)) Shows the ratio of cache hits to total cache lookups, per inference pool. 4. Routing efficiency (score distribution) histogram_quantile(0.90, sum(rate(llmd_router_routing_score_bucket[5m])) by (le)) These metrics can be scraped in OpenShift using llm-d by setting monitoring.podmonitor.enabled=true in the configuration, with a similar setting for the epp servicemonitor that manages the prefill-cache-based routing. This observes routing scores from the EPP to validate that cache-aware routing decisions prioritize efficiency. Example dashboard Having these metrics displayed in a single Grafana dashboard transforms LLM observability from guesswork into actionable insights. Instead of sifting through logs or raw counters, SREs and platform engineers can instantly see TTFT values, cache hit rates, and routing efficiency side by side, correlated across the entire llm-d stack on OpenShift AI. This visibility makes it possible to spot real-time performance regressions, validate that SLOs are being met, and more easilydiagnose whether latency is due to routing, caching, or GPU saturation—all from one dashboard. Conclusion The age of LLMs requires a new observability playbook. Traditional metrics fail to capture token-level performance and cache-aware efficiency. By combining vLLM’s engine-level metrics with llm-d’s routing and cache-aware insights, Red Hat OpenShift AI provides SREs and platform teams the tools they need to more effectively meet modern AI SLOs at production scale. With llm-d on OpenShift AI, users gain: - Transparency into distributed inference workloads - Confidence in meeting token-level SLOs - Lower GPU costs through cache efficiency - Actionable insights with Grafana dashboards powered by Prometheus and OpenTelemetry This is observability purpose-built for AI at scale, and it’s available today in the open source community and on the OpenShift AI 3.0 platform. Resource The adaptable enterprise: Why AI readiness is disruption readiness About the authors More like this Browse by channel Automation The latest on IT automation for tech, teams, and environments Artificial intelligence Updates on the platforms that free customers to run AI workloads anywhere Open hybrid cloud Explore how we build a more flexible future with hybrid cloud Security The latest on how we reduce risks across environments and technologies Edge computing Updates on the platforms that simplify operations at the edge Infrastructure The latest on the world’s leading enterprise Linux platform Applications Inside our solutions to the toughest application challenges Virtualization The future of enterprise virtualization for your workloads on-premise or across clouds