Skip to main content

Token Usage Dashboard

The Token Usage Dashboard is part of Virtana’s AI Factory Observability (AIFO).  It provides real-time visibility into the cost, performance, and reliability of Large Language Model (LLM) workloads across hosts, containers, and services.

A token represents a basic unit of processing, such as a word fragment or symbol, that an LLM handles during inference or training operations. Monitoring token usage helps you align performance, cost, and efficiency across your AI infrastructure.

token1.png

Overview

The Token Usage Dashboard in the Virtana Platform gives you a clear view of how your AI workloads use system resources. It helps you see where GPU power is being spent, so you can easily spot performance issues and make better use of your hardware. 

Virtana shows token usage in two simple ways:

Container-Level Tracking

  • The dashboard collects overall usage data from your containers (for example, NVIDIA NIM containers).

  • You can see how each container uses tokens and how that affects GPU performance.

  • This helps you find if some containers are using too much or too little GPU capacity.

    Example: If one container uses only 2% of GPU capacity while another uses 98%, you may have a “noisy neighbor” problem, where one process takes most of the GPU. You can then adjust workloads to balance performance.

Application-Level Tracking

  • Some applications send OpenTelemetry data that includes token usage and timing information.

  • Virtana combines this data to show how tokens are used across different applications, helping you relate it to performance.

    Example: The dashboard may show that an AI imaging app consumes many tokens but produces low GPU output. This can mean the app needs performance tuning or load redistribution.

Understanding NIM and Non-NIM Models

In Virtana AIFO, token data can be extracted differently based on the model type:

NIM Models (NVIDIA Inference Microservices)

NIM models are containerized inference services optimized for GPU workloads. Each NIM container operates as a self-contained LLM instance with built-in metrics collection capabilities. This allows Virtana to directly retrieve token usage metrics from the instance without additional configuration.

Non-NIM Models:

For non-NIM models, integration requires a LiteLLM proxy. The proxy emits OpenTelemetry traces containing token usage, latency, and request metadata on behalf of the underlying model. Virtana then ingests these traces to calculate token metrics and performance patterns.

Token Usage dashboard views

You can create a token usage view by selecting filters at the top of the dashboard. Start by choosing the deployments you want to analyze, then refine your view further with optional filters.

These filters include Model, Service, Host, and NIM Container, allowing you to narrow the scope for more precise insights. This flexible filtering helps you focus on specific aspects of your AI workloads for better monitoring and optimization.

token_dashboard2.png

Prerequisites

To get started using the GPU fleet analysis dashboard, ensure that you have the following requirements:

  • Virtana monitoring agents or API integrations must be deployed on all GPU-capable nodes for cloud and on-premises.

  • An active Virtana account with dashboard access permissions refer to Setting up access between IO and Global View

  • Network connectivity to Virtana Platform  from all GPU environments.

  • Cost Meter Card configured to track GPU usage and associated costs across environments.

  • (Optional) Integration with NVIDIA DCGM (NVIDIA Data Center GPU Manager) Exporter for enhanced Kubernetes support.

  • OpenTelemetry traces enabled in all LLMs or applications to send token data.

  • Create and securely store API tokens under Settings > API Token Management.

Configure cost meter cards

Create a cost meter card to capture GPU-related cost metrics that you can view in the GPU fleet analysis dashboard. Cost meter cards help define pricing categories, such as compute, GPU, and storage, as well as corresponding units, so you can track costs and view reports of GPU usage.

To get started, see Creating Cost Meter Cards

Open the dashboard

Be sure that your Virtana account has permissions for the Token usage dashboard. After opening the dashboard, set the filters for your token.

  1. Log in to the Virtana Global View and click Dashboards.

  2. Select Token Usage Dashboard. The dashboard opens.

  3. Create your view by choosing filters.

Download the dashboard

You can download the main dashboard page as a PDF document. This is useful when you want to create a filtered view of the dashboard to share as a report with your teams.

To save your dashboard locally, click Download > PDF document.

pdf_token.png

Viewing the Token Usage dashboard

When you open the dashboard, you immediately see a snapshot of the health of your selected token usage data. The Token Usage Dashboard visualizes token metrics alongside key performance and cost indicators, offering a comprehensive view of your AI workloads.

view_dashboard.png

Let’s explore each of these key metrics in detail:

Total Prompt Tokens 

The total number of input tokens processed by the models during the selected period. Helps estimate costs based on prompt length and design.

Total Completion Tokens 

The total number of output tokens generated by the models. Useful for correlating output size with user prompts.

Total Requests

The total number of API requests processed, including successful and failed ones.

Average Request Duration (s)

The average time taken for a request to complete. A key measure of responsiveness.

Throughput (req/s)

The number of requests handled per second. Indicates system scalability and efficiency.

Time to First Token (TTFT)

Measures how quickly a response starts streaming after a request is made — a key user experience metric.

Tokens consumption summary

Below the snapshot, you can explore detailed insights through the token consumption trend and model utilization views. These sections help you analyze how tokens are being used over time and how different models contribute to overall consumption.

consumption.png

Token Consumption Trend

Displays the number of Prompt and Completion tokens consumed over time. Use this chart to:

  • Identify spikes in usage.

  • Correlate token peaks with engineering deployments, feature rollouts, or testing activity.

  • Monitor optimization progress when prompt efficiency changes are implemented.

Utilization by Model

Shows the proportion of total tokens consumed by each model (for example, azure/gpt-5, gemini-2.5-pro, writer/palmyra).

Use this data to:

  • Identify high-cost models driving spend.

  • Evaluate cheaper alternatives without sacrificing performance.

  • Align model usage with business and budget goals.

Cost Summary

Below the snapshot, you can explore detailed insights into token costs. Use the token cost trend to understand how expenses evolve over time.

cost_summary_token.png

Cost views translate token usage directly into USD-based spend estimates.

Enables users to:

  • Reconcile token consumption with billing data.

  • Identify cost-heavy workloads and optimize prompts.

  • Apply rate limits or off-peak scheduling to minimize spend.

  • Select alternative models without sacrificing required performance.

Request Analysis

The Request Analysis view gives you a clear picture of how requests are distributed across different AI models and how those requests trend over time (see image above). Use this dashboard to quickly spot any unusual activity, such as sudden spikes in requests, which might mean a particular process or test is using up more resources than expected.

request_analysis_token.png

The Request Trend chart shows you how many requests succeed versus how many fail in each period. This helps you understand where resources might be getting wasted due to failed requests, so your team can take steps to lower unnecessary retries, save costs, and keep things running smoothly.

It’s a good idea to check this dashboard regularly. Reviewing wasted resource data helps you identify areas for improvement and track the impact of any changes you make to your system. This way, you can plan ongoing optimizations and see the return on investment (ROI) of your efforts.