Ship PRs that pass review the first time
Superfocus auto-researches your codebase when tickets land. Cursor and Claude get pre-loaded context on shared components, design patterns, and gotchas - so they build features the way your team actually does.
You love AI agents.
You hate the revision loops.
Cursor's indexing finds similar code, but doesn't understand your architecture. So you get:
- One-off components when shared ones exist
- Duplicate helpers your team already wrote
- PR comments: "we have a util for that"
- Code that works, but breaks conventions
The root cause: codebase search ≠ architectural understanding.
How it works
Superfocus is a background research assistant that preps your tickets before you start coding. It runs as a Cursor / VS Code extension.
New ticket assigned → Superfocus auto-starts research
A Superfocus AI agent researches your codebase looking for relevant gotchas
Context exports to Linear/Jira in the ticket description
Copy/paste into any AI IDE → better first-draft accuracy
This happens in parallel while you work on other tickets. By the time you're ready to start a new task, the research is already done.
Think of it like having an assistant who works ahead of you to find all the relevant codebase bits, and when you're ready, hands you a perfect brief.
Context so good, AI can actually estimate effort
Because Superfocus understands patterns, complexity, and connected code paths, it can predict effort with surprising accuracy, export those estimates to your issue tracker, and show you a timeline of upcoming work.
Sorts tickets by priority
Show estimated completion dates
Instant answers to "When will this be done?"
Reduces context switching
Makes planning more accurate
This alone makes engineers feel dramatically less overwhelmed.
Why context quality matters
Less "wait, this broke another component". Less "the AI didn't know we have a helper for this". Less time in revision loops. More time shipping.
Repeatable results
We reimplemented real open source project's PRs twice: once with Superfocus context prep, once without. Below you'll find everything needed to try it yourself.
Superfocus vs no context prep: Same PR, same model, 30% jump in implementation accuracy
Security & compliance
Runs locally inside VS Code / Cursor
Uses your existing enterprise OpenAI account
Stores keys in VS Code default key store
Sends only enriched context to Jira / Linear
Analytics are anonymized and never shared
You control your API costs
Because context quality improves, your token usage for implementation actually drops.
Pricing
Early Access - $20/month
Locking in first 100 users. Price increases after that.
- 50 research runs per day
- Linear + Jira export
- Works with Cursor, VSCode, Claude Code, any AI IDE
- Optionally bring your own OpenAI or Gemini key
I'm expecting the first 100 customers to run at a loss. Research runs use a lot of inference.
FAQ
Engineers working on complex codebases (50k+ lines) who use AI coding tools like Cursor or Claude Code. Especially valuable if you:
Security. Your code already lives on your machine - there's no reason to clone it to a cloud sandbox. Running locally means your codebase never leaves your control, and your legal/security team can sleep soundly. No data residency concerns, no vendor lock-in, no wondering where your code is stored.
You've already vetted your AI stack with your organization. Using your own key means:
About $0.07 per ticket with GPT-4o-mini (surprisingly capable for this) or $0.30 with GPT-4o. Since better context means your implementation agents search less and derail less, you often save more than that on tokens spent on the actual implementation.
I built Superfocus for my own workflow and have never had a job that didn't use either Linear or Jira. I find issue trackers distracting and overwhelming; after every ticket I find myself having to get my head around what to work on next. Are there new tickets? What is the best critical path? What urgent tickets are waiting for me? How do new tickets fit in? Superfocus exports research output to the description of a ticket, but also exports labels to estimate impact, effort, unblocks and risk. These labels power the sorting algorithm the vscode and cursor extensions use to prioritize tickets. As a result, I safe time, feel less overwhelmed and rarely open my issue tracker.
About 5 minutes depending on codebase size and ticket complexity. But it runs in the background while you work on other tickets, so by the time you're ready for something new, the research is done.
If you just created a ticket and want to start immediately, you can skip the Superfocus prompt - but waiting saves you time in revision rounds later. Go grab lunch, answer Slack messages, review a PR. The wait is worth it.
You'll need a Personal Access Token (PAT) for whichever tracker you use:
Paste it into Superfocus settings and you're done.
You can rerun prep runs if needed. But honestly, I've tuned the Superfocus agents extensively to avoid this — I almost never need to rerun them myself. The filtering and sub-agent architecture is designed specifically to prevent the context derailment that plagues other tools.
Superfocus doesn't plan - it does unopinionated research. Think of it like a super-experienced founding engineer giving you gotchas to kickstart your work.
A possible workflow: Run Superfocus prep → use that context in Plan Mode to create an approach → implement. This actually improves Plan Mode because it starts with all the puzzle pieces already assembled instead of searching as it goes.
Superfocus makes any agent or LLM coding tool smarter by front-loading the codebase knowledge they need.
Yes. Like Cursor, ChatGPT, and other systems, you can include custom rules to guide what patterns, gotchas, or conventions Superfocus should prioritize in your codebase.
Gemini is supported as well, and more models coming soon. Let me know what you need.
Yes — but Superfocus does require VS Code or Cursor running in the background to access your codebase. You can let it sit in the background while you work in Claude Code or anywhere else.
The workflow is hands-off: Superfocus researches in the background, exports to your ticket, then you copy the prompt with one button and paste into Claude Code (or any other tool).
Yes. Superfocus only accesses what's already on your machine. Nothing gets cloned or uploaded anywhere.
Only anonymized usage metrics to improve the product, and billing information if you're on a paid plan. I never see your code, ticket content, or codebase. It all stays local or goes directly to your Linear/Jira.
About
Superfocus was built by an engineer who got tired of babysitting AI agents.
I work on large-scale ML infrastructure at Anyscale (the company behind Ray, used by Anthropic, OpenAI, Cursor, Shopify, Apple, Canva, and others for training and data processing). After watching AI coding tools constantly miss architectural patterns in our codebase, I built Superfocus to solve it for myself.
The manual 'prep run' workflow worked so well that I turned it into an extension. Now I'm sharing it.
Bart Proost (@bartproost)