Show HN: Rudel – Claude Code Session Analytics

github.com

144 points by keks0r 4 days ago


We built rudel.ai after realizing we had no visibility into our own Claude Code sessions. We were using it daily but had no idea which sessions were efficient, why some got abandoned, or whether we were actually improving over time.

So we built an analytics layer for it. After connecting our own sessions, we ended up with a dataset of 1,573 real Claude Code sessions, 15M+ tokens, 270K+ interactions.

Some things we found that surprised us: - Skills were only being used in 4% of our sessions - 26% of sessions are abandoned, most within the first 60 seconds - Session success rate varies significantly by task type (documentation scores highest, refactoring lowest) - Error cascade patterns appear in the first 2 minutes and predict abandonment with reasonable accuracy - There is no meaningful benchmark for 'good' agentic session performance, we are building one.

The tool is free to use and fully open source, happy to answer questions about the data or how we built it.

zippolyon - 5 hours ago

Great work on the session analytics. The "error cascade in first 2 minutes predicts abandonment" finding is exactly the kind of signal that causal auditing can act on. We built K9 Audit for the complementary problem: not just when sessions fail, but why — recording every tool call as a CIEU five-tuple (intent vs actual outcome) with a hash chain. The "26% abandoned" stat likely hides silent deviations that looked like success. k9log causal --last traces root cause across steps in seconds. https://github.com/liuhaotian2024-prog/K9Audit

dmix - 4 days ago

I've seen Claude ignore important parts of skills/agent files multiple times. I was running a clean up SKILL.md on a hundred markdown files, manually in small groups of 5, and about half the time it listened and ran the skill as written. The other half it would start trying to understand the codebase looking for markdown stuff for 2min, for no good reason, before reverting back to what the skill said.

LLMs are far from consistent.

emehex - 4 days ago

For those unaware, Claude Code comes with a built in /insights command...

Aurornis - 4 days ago

> 26% of sessions are abandoned, most within the first 60 seconds

Starting new sessions frequently and using separate new sessions for small tasks is a good practice.

Keeping context clean and focused is a highly effective way to keep the agent on task. Having an up to date AGENTS.md should allow for new sessions to get into simple tasks quickly so you can use single-purpose sessions for small tasks without carrying the baggage of a long past context into them.

monsterxx03 - 3 days ago

I built something in a similar space: Linko (https://github.com/monsterxx03/linko), a transparent MITM proxy with a webui that lets you see what's actually being sent between Claude Code and LLM APIs in real time.

  It's been really helpful for me to debug my own sessions and understand what the model is seeing (system prompts, tool definitions, tracing tool calls etc.).
tmaly - 3 days ago

I have seen numbers claiming tools are only called 59% of the time.

Saw another comment on a different platform where someone floated the idea of dynamically injecting context with hooks in the workflow to make things more deterministic.

c5huracan - 3 days ago

The "no meaningful benchmark for good agentic session performance" point resonates. Success varies so much by task type that a single metric is almost meaningless. A 60-second documentation lookup and a 30-minute refactoring session could both be successes.

Curious what shape the benchmark takes. Are you thinking per-task-type baselines, or something more like an aggregate efficiency score?

locma - 2 days ago

What's the end goal here — using session data as a feedback loop to iteratively improve CLAUDE.md and agent workflows based on real usage patterns?

dboreham - 4 days ago

One potential reason for sessions being abandoned within 60 seconds in my experience is realizing you forgot to set something in the environment: github token missing, tool set for the language not on the path, etc. Claude doesn't provide elegant ways to fix those things in-session so I'll just exit, fix up and start Claude again. It does have the option to continue a previous session but there's typically no point in these "oops I forgot that" cases.

152334H - 4 days ago

is there a reason, other than general faith in humanity, to assume those '1573 sessions' are real?

I do not see any link or source for the data. I assume it is to remain closed, if it exists.

marconardus - 4 days ago

It might be worthwhile to include some of an example run in your readme.

I scrolled through and didn’t see enough to justify installing and running a thing

steve_adams_86 - 3 days ago

Does this comply with Anthropic's terms? I've been developing small apps here and there on top of Claude Code and each time I find I'm too uncomfortable with their terms to bother distributing it.

blef - 4 days ago

Reminds me https://www.agentsview.io/.

KaiserPister - 4 days ago

This is awesome! I’m working on the Open Prompt Initiative as a way for open source to share prompting knowledge.

lgvdp - 3 days ago

I see a lot of people with concerns about privacy and security. Not shown in the post, but the github shows how to self host. No need to use 3rd party, you can just have your own too

alyxya - 4 days ago

Why does it need login and cloud upload? A local cli tool analyzing logs should be sufficient.

mbesto - 4 days ago

So what conclusions have you drawn or could a person reasonably draw with this data?

swaminarayan - 3 days ago

26% of AI coding sessions are abandoned within 60 seconds .Is this a prompt problem, a tooling problem, or a limitation of current models?

ekropotin - 4 days ago

> That's it. Your Claude Code sessions will now be uploaded automatically.

No, thanks

ericwebb - 4 days ago

I 100% agree that we need tools to understand and audit these workflows for opportunities. Nice work.

TBH, I am very hesitant to upload my CC logs to a third-party service.

smallerfish - 3 days ago

> content, the content or transcript of the agent session

Does this include the files being worked on by the agent in the session, or just the chat transcript?

anthonySs - 4 days ago

is this observability for your claude code calls or specifically for high level insights like skill usage?

would love to know your actual day to day use case for what you built

- 4 days ago
[deleted]
mentalgear - 4 days ago

How diverse is your dataset?

lau_chan - 4 days ago

Does it work for Codex?

bool3max - 3 days ago

Why is the comment calling out the biggest issue with this so heavily downvoted? Privacy is a massive concern with this.

cluckindan - 4 days ago

Nice. Now, to vibe myself a locally hosted alternative.

sriramgonella - 4 days ago

[flagged]

yangro - 3 days ago

[flagged]

socialinteldev - 4 days ago

[flagged]

mrothroc - 4 days ago

[dead]

longtermemory - 4 days ago

From session analysis, it would be interesting to understand how crucial the documentation, the level of detail in CLAUDE.md, is. It seems to me that sometimes documentation (that's too long and often out of date) contributes to greater entropy rather than greater efficiency of the model and agent.

It seems to me that sometimes it's better and more effective to remove, clean up, and simplify (both from CLAUDE.md and the code) rather than having everything documented in detail.

Therefore, from session analysis, it would be interesting to identify the relationship between documentation in CLAUDE.md and model efficiency. How often does the developer reject the LLM output in relation to the level of detail in CLAUDE.md?

aplomb1026 - 4 days ago

[flagged]

bhekanik - 4 days ago

[dead]

Sebastian_Dev - 4 days ago

[dead]

huflungdung - 4 days ago

[dead]

ptak_dev - 3 days ago

[flagged]

multidude - 4 days ago

[flagged]

mihir_kanzariya - 4 days ago

[flagged]

robutsume - 4 days ago

[flagged]

ozgurozkan - 4 days ago

[flagged]

vova_hn2 - 4 days ago

This is so sad that on top of black box LLMs we also build all these tools that are pretty much black box as well.

It became very hard to understand what exactly is sent to LLM as input/context and how exactly is the output processed.