<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Rock Darko</title>
    <description>The latest articles on DEV Community by Rock Darko (@rockdarko).</description>
    <link>https://dev.to/rockdarko</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3917732%2Ffd450f22-64e6-4efb-835b-909f0fd8bb8b.jpg</url>
      <title>DEV Community: Rock Darko</title>
      <link>https://dev.to/rockdarko</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/rockdarko"/>
    <language>en</language>
    <item>
      <title>A Grafana dashboard for Claude Code on Prometheus</title>
      <dc:creator>Rock Darko</dc:creator>
      <pubDate>Sat, 09 May 2026 01:31:41 +0000</pubDate>
      <link>https://dev.to/rockdarko/a-grafana-dashboard-for-claude-code-on-prometheus-2nbl</link>
      <guid>https://dev.to/rockdarko/a-grafana-dashboard-for-claude-code-on-prometheus-2nbl</guid>
      <description>&lt;p&gt;Claude Code emits OpenTelemetry metrics over OTLP. Anthropic publishes the metric names. So all that was missing, if you run a Prometheus-compatible backend, was a dashboard. Here is one.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fn2da9rpuuhptd95edxtv.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fn2da9rpuuhptd95edxtv.png" alt="Claude Code Metrics dashboard cover image" width="800" height="147"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Grafana Labs: &lt;a href="https://grafana.com/grafana/dashboards/25255-claude-code-metrics-prometheus/" rel="noopener noreferrer"&gt;dashboard 25255&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;Source: &lt;a href="https://github.com/rockdarko/claude-code-metrics-prometheus" rel="noopener noreferrer"&gt;github.com/rockdarko/claude-code-metrics-prometheus&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;License: MIT&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;It's a port, not original work. The dashboard concept and panel set come from &lt;a href="https://grafana.com/grafana/dashboards/25052-claude-code/" rel="noopener noreferrer"&gt;grafana.com 25052 by 1w2w3y&lt;/a&gt;, which targets Azure Application Insights via KQL. I rebuilt every panel in PromQL so the same view works against the OSS metrics stack. Credit upstream.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why a port was worth doing
&lt;/h2&gt;

&lt;p&gt;When my team started piloting Claude Code, the first thing I went looking for was a dashboard. 25052 was right there, well thought out, exactly the panels I wanted. But it's KQL on Application Insights, and our observability stack is Prometheus and Grafana. As far as I can tell, that's most teams.&lt;/p&gt;

&lt;p&gt;The metrics themselves are fine. Claude Code speaks OTLP, and OTLP works with everyone. The gap was just that nobody had wired the receiving end into a PromQL dashboard yet. So I did. Same panels, same intent, different query language. Compatible with Prometheus, VictoriaMetrics, Grafana Mimir, and Thanos.&lt;/p&gt;

&lt;h2&gt;
  
  
  The pipeline
&lt;/h2&gt;



&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Claude Code  →  OTLP  →  OTel Collector  →  /metrics  →  Prometheus  →  dashboard
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;If you already run a Collector and a Prometheus-flavored backend, this is three small additions: tell Claude Code where to send OTLP, add a Prometheus exporter to your Collector, add a scrape job. Full setup is in the README, including a minimal Collector config and a scrape file you can copy.&lt;/p&gt;

&lt;h2&gt;
  
  
  What you can actually see
&lt;/h2&gt;

&lt;p&gt;The dashboard has five sections.&lt;/p&gt;

&lt;p&gt;The Overview gives the at-a-glance KPIs: sessions, users, total cost, total tokens, commits, PRs, lines added and removed, active time, tokens by type.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fz3ktiaujyfg9c66lskh9.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fz3ktiaujyfg9c66lskh9.png" alt="Overview KPIs and leaderboards at the top of the dashboard" width="800" height="329"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Leaderboards answer "who is using this and on what." Top users by cost and tokens, top sessions by cost, cost broken down by model, edit decisions by language, sessions by terminal.&lt;/p&gt;

&lt;p&gt;Cost &amp;amp; Tokens is the time-series view: cost over time overall and per model, tokens over time by type and by model. Useful when somebody asks "are we still in budget" and you need an answer that isn't a guess.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fe0k7tdv4xie2i0tkapv2.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fe0k7tdv4xie2i0tkapv2.png" alt="Cost and token usage over time, broken down by model" width="800" height="229"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Activity &amp;amp; Productivity covers active time per hour, lines of code per hour, and tool decisions over time (accept, reject, other).&lt;/p&gt;

&lt;p&gt;Cost Breakdown is the part I find most interesting in practice: cost by query source, cost by effort, and cache hit ratio. The cache hit ratio in particular is worth watching. It's the difference between a sustainable bill and an alarming one.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F05u0a6flsyyj25a5l7gv.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F05u0a6flsyyj25a5l7gv.png" alt="Per-hour activity, cost decomposition, and cache hit ratio panels" width="800" height="348"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Three filter variables sit at the top of every panel: organization, user, model. Default time range is seven days.&lt;/p&gt;

&lt;h2&gt;
  
  
  Per-team, per-project, per-repo views
&lt;/h2&gt;

&lt;p&gt;Out of the box, the metric series carry organization, user, model, terminal, and a session id. There is no &lt;code&gt;repository&lt;/code&gt; or &lt;code&gt;project&lt;/code&gt; label, because Claude Code has no opinion about your team's taxonomy.&lt;/p&gt;

&lt;p&gt;What it does support is &lt;code&gt;OTEL_RESOURCE_ATTRIBUTES&lt;/code&gt;. Anything you put there becomes a label on every metric:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight shell"&gt;&lt;code&gt;&lt;span class="nb"&gt;export &lt;/span&gt;&lt;span class="nv"&gt;OTEL_RESOURCE_ATTRIBUTES&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="s2"&gt;"team=platform,project=billing-svc,cost_center=eng-123"&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Set it per shell, per direnv, per repo &lt;code&gt;.envrc&lt;/code&gt;, per team's onboarding script, whatever fits. Each Claude Code session inherits the value from its environment, the Collector forwards it on, and you can group and filter by it in Grafana like any other label.&lt;/p&gt;

&lt;p&gt;The dashboard's built-in template variables (organization, user, model) don't include custom labels yet, but extending it is small work: add a Grafana template variable with &lt;code&gt;label_values(claude_code_session_count_total, project)&lt;/code&gt; and reference &lt;code&gt;project=~"$project"&lt;/code&gt; in the panel queries. Or skip the editing and use Grafana's ad-hoc filters, which read the label cardinalities at query time and don't require changing the dashboard JSON.&lt;/p&gt;

&lt;h2&gt;
  
  
  Gotchas worth repeating
&lt;/h2&gt;

&lt;p&gt;A few things that bit me or that are easy to miss.&lt;/p&gt;

&lt;p&gt;Pin temporality to cumulative. Prometheus-family backends expect cumulative counters. The OpenTelemetry SDK currently defaults to cumulative, but defaults drift across SDK versions, and the failure mode is silent: wrong-looking rates, not an error. Set it explicitly in Claude Code's environment:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight shell"&gt;&lt;code&gt;&lt;span class="nb"&gt;export &lt;/span&gt;&lt;span class="nv"&gt;OTEL_EXPORTER_OTLP_METRICS_TEMPORALITY_PREFERENCE&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;cumulative
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;The PR counter only counts pull requests that Claude Code itself opened. If your team opens PRs manually after a Claude Code session, the dashboard will show zero. That's how the metric is defined upstream, not a bug in the dashboard.&lt;/p&gt;

&lt;p&gt;Cost is a client-side estimate. Claude Code computes it from token counts and known model prices. It tracks billing closely but won't match to the cent, particularly around price changes or cached-token billing.&lt;/p&gt;

&lt;p&gt;If &lt;code&gt;Sessions by Terminal&lt;/code&gt; is empty, set &lt;code&gt;resource_to_telemetry_conversion: enabled: true&lt;/code&gt; on your Collector's Prometheus exporter. Without it, attribute-derived labels don't make it through.&lt;/p&gt;

&lt;h2&gt;
  
  
  What's next
&lt;/h2&gt;

&lt;p&gt;A couple of additions I'd like to make: a token-spend rate panel against a configurable budget, and (once there's enough adoption to ground them in real failure modes) some sample alert rules.&lt;/p&gt;

&lt;p&gt;If you run it and find missing panels or buggy queries, open an issue. PRs are welcome, especially around custom labels people are adding via Collector processors. The repo is MIT and the dashboard is on Grafana Labs, ready to import.&lt;/p&gt;

</description>
      <category>grafana</category>
      <category>prometheus</category>
      <category>opentelemetry</category>
      <category>observability</category>
    </item>
  </channel>
</rss>
