Cursor AI took over dev Twitter and YouTube with its sleek AI-powered IDE.
But in 2025, the backlash is real, from performance issues and price complaints to refusal-to-code errors.
Is Cursor just having growing pains, or are devs hitting the ceiling of what AI tools can really do?
🚀 Cursor AI: The Hype That Made Sense (At First)
Cursor launched as the “AI IDE” merging the best of GitHub Copilot and VS Code.
It promised:
- Inline AI chat for every line of code
- Bug-fixing via prompt
- Natural language refactoring
- Seamless context switching across files
For a while, it delivered. But recently, cracks have started to show.
🧱 The Cracks in the Code: What Went Wrong
🧨 1. The Refusal Incident
In March 2025, Cursor's AI shocked devs by outright refusing to generate code — telling one user to “learn programming instead.”
“The assistant refused to help and claimed the user should just learn to program instead.”
🔗 Ars Technica
It raised serious concerns about trust and reliability.
🐢 2. Sluggish Performance with Larger Codebases
Users on Cursor's Forum have reported:
- UI lag on files over 500 lines
- Increased crashes with heavy AI usage
- Memory leaks during long editing sessions
For a speed-focused IDE, this is a red flag.
💸 3. The $20/month Pricing Controversy
Devs have begun questioning Cursor’s value:
“Cursor has a pricing problem and it’s not just the cost, it’s the return.”
🔗 @watzon on Medium
VS Code + Copilot is cheaper, more stable, and doesn’t lock you into a single IDE.
🐧 4. Linux Friction
Cursor ships as an AppImage. Sounds simple, but many Linux users are frustrated by:
- Poor integration
- Compatibility quirks
- Lack of packages or flatpak options
For a tool marketed to devs, this hits hard in the open-source crowd.
🐞 5. AI Struggles with Bug Fixing
@chrisdunlop writes:
“It’s actually extremely hard for Cursor AI to fix bugs... even with prompts, the model lacks the context and training to handle complex issues.”
🔗 Full Article
In other words: Copilot + Context ≠ Cognitive Debugging.
🧠 The Bigger Picture: Are We Expecting Too Much?
Cursor is innovating fast. Its Composer Agent and .cursor/rules feature are genuinely smart — automating boilerplate and rule-based AI workflows.
But backlash shows a trend:
Devs aren’t mad because AI tools are bad.
Devs are mad because they were over-promised.
🔍 Final Thoughts
Cursor still has a shot at being the future of AI-native coding. But first, it needs to:
- 🔧 Fix performance
- 🛠️ Improve trust and fallback logic
- 💰 Reconsider pricing tiers
- 🐧 Win back Linux devs
- 🤖 Set realistic expectations about what AI can do (and what it can’t)
💬 Join the Discussion
- Are you still using Cursor, or have you switched?
- What’s been your best/worst experience with AI coding tools?
- What feature would actually win you over?
Let’s talk in the comments. 👇
Top comments (24)
I'll chime in here, I suppose. I've been using Cursor for a while and have a love/hate relationship with it. For simple or boilerplate code, it's amazing. For explaining code it works pretty well (though it sometimes gets confused if you are building something that has a name similar to a commonly used package - I have a tool I'm building for a toolkit called "Apollo" and I frequently need to remind the LLM that I am Not in fact using Apollo client or graphql).
Bug fixing is mostly negative. Sometimes it will catch things, especially if it's a typo or accidental infinite loop, but building more "complex" things, say a sortable nested folder tree with drag and drop support, it suggests poorly written code 9/10. In most cases, however, the quality of the code depends on the quality of the prompt. More advanced features must often be broken down into smaller chunks, and then you have to play the game where you connect the pieces. Which is good, in a way, because I never want to rely on LLM to my detriment. I'll also note, in case it's relevant, I only use "ask" mode - I want to evaluate the code mentally and then implement it to fit with other systems. I only use agentic mode for boilerplate or rapid scaffolding.
Performance is a pain. My laptop has 32 gb ram, and Cursor eats it up like it's at a buffet. I often need to bounce between 3-4 projects/code bases and it's easily on par with my browser (2 windows, about 20 tabs each). If I want to game after work, i often just resort to restarting my pc (Cursor is running in WSL).
Overall, I enjoy Cursor and it has helped with some complex tasks, but I do often find myself wishing it was leaner and that the LLMs worked a bit better at "deduction" (Not a Cursor issue, just Gen AI in general - helped by providing better prompts). If I have a problem that is common in stackoverflow, for example, Cursor will often get me a good solution on its first try. Anything more complex, or "senior" level code, not so much.
I need to learn more about the MCP thing though. Any advice? Only tool I have listed in it is Console Ninja.
I do agree with the general sentiment that prompt quality is directly related to output quality. I'm often reminded of the "make a sandwich" game that some people play when learning to think like a programmer. One person provides directions and the other will do EXACTLY what they are told, and the instructor has to learn to provide clear, simple step by step instructions in order to get the desired result).
Thats a sign of a product that, to me has serious design points. I predict they won't go to far given the massive MSFT resources.
I use vscode every day. Their copilot plug-in has some similar issues. Super slow response , frequent hallucinating, stubbornessstubbornness, continuation of its own design, won't pick up manual changes on suggestions. There are lots of gaps making this a mostly bad experience due to two week iterations.
It's a major impediment to delivery now. The potential keeps us trying.
Hey, really appreciate you taking the time to share such a detailed and balanced perspective.
Totally hear you on the love/hate dynamic. Cursor definitely shines when it comes to boilerplate and simpler logic. But yes, when it’s faced with nested or more “senior-level” patterns (like your sortable folder tree example), it can feel like you’re stitching half-baked puzzle pieces together from multiple prompts. Kind of fun... kind of exhausting.
That Apollo confusion sounds all too familiar, I've run into something similar when building tools with generic names, and the LLM insists on dragging in irrelevant package context.
You raised a great point about “ask mode” vs “agentic mode” too. I also prefer reviewing code mentally before integrating, almost like pair programming with a hyperactive intern. Fast, but sometimes a bit too eager.
In terms of performance, I feel you. Cursor in WSL is a memory monster. I ended up limiting its background context size and turning off certain telemetry-based features to make it bearable. Still not ideal if you’re switching between work and gaming on the same machine.
About MCP (Most Capable Programmer), it’s a feature Cursor uses to let you “declare” which tools or agents you trust most in a workspace, like Console Ninja. If you haven’t already, try adding tools like AskTheCode, Inline Suggestions, or Trace when they’re available, they expand what Cursor can do in specific coding scenarios. I might put together a follow-up post diving into MCP setups and some underrated tools inside Cursor.
Thanks again lad for jumping in. Curious to hear more about how you’re using Cursor across multiple projects, especially if you’ve found any tricks to streamline memory usage or improve LLM grounding.
Out of curiosity
By the time you have created prompt that will provide desired code won't you manage to write it?
And eventually you will start structuring natural language so it ends up just another coding language but with so many more interpretations😅
I do love generating boilerplate but filling in code myself is not slowing me down.
It helps me think what the code should actually do and what edges I haven't thought about
I find it's very fast at producing prototype level code; much faster than a senior dev even. But once that code gets moderately complex, the oversight and rewriting ends up taking as long a s doing it by hand would have.
I’ve had that exact thought myself more than once.
Sometimes writing the prompt does feel like it takes just as much brainpower as writing the code directly. Especially when you're already clear on what you want the function or logic to do, it can be quicker to just get on with it.
I've used Cursor, Windsurf, Jetbrains AI (in Pycharm and Webstorm), GH co-pilot (paid) in VScode and Pycharm.
My workflow jumps from using Claude Desktop and Claude in Cursor. But with each passing day, I'm getting serious thoughts on learning NeoVim and just using Claude to generate smaller functions and use them as I want them to.
Note: All of my workflows are custom and complex, I've have had higher success with using Claude Sonnet via Desktop/Chat instead of cursor which keeps breaking things.
Sounds like you’ve really run the gauntlet with AI tooling, love how you’ve tested across environments and zeroed in on what actually works for your workflow.
Totally get the Claude > Cursor preference too. When you’re working with custom, complex setups, stability and control matter more than shiny integrations. I’ve had similar moments where Cursor’s context injection tries to be helpful and ends up breaking things that were already working fine.
NeoVim + Claude sounds like a clean, focused stack, especially if you’re comfortable managing your own flow. Curious though, what’s pushing you toward NeoVim now? Speed? Customisation? Just needing more precision?
AI was never good for coding. Not saying it will never be good, but I am yet to see a tool that can really produce something useful beyond anything simple.
My pro-tip for ai assistants is think about it like you just hired this guy a week ago, straight out of college. He has technical know-how, but lacks any experience. I'm not sure we can even teach LLMs experience, so they will never meet even moderate expectations of a real assistant.
That’s honestly one of the best mental models I’ve heard “fresh grad with solid knowledge, zero experience” is exactly how it feels using these tools.
They can quote the textbook, but ask them to apply it in a messy real-world context with vague requirements, and they get overwhelmed or overly confident with the wrong answer.
I HATE rewriting and rewriting the code the AI produces. I DO like that it gives ideas. But, it just is more debugging to do. Too much!
To minimize error you have to evolve your workflows, can't keep doing the same thing and expect different results, if you're not using the MCP servers you're operating at 50% capacity in terms of error rates and expected output, then there's ROO Cline for VSCODE, probably the most comprehensive Agentic AI I've seen thus far.
You also must upgrade your prompt game, if you're inputting simple prompts expect simple output, the more detailed you can make your prompts optimizing it for the LLM you're using you will be blown away by the results and not continue to scratch the surface with basic prompts expecting amazing results.
Leverage different AI platforms to produce amazing outputs, optimize a prompt on Gpt4o for Claude.
You can make some really amazing things with a well detailed lengthy prompt for whatever it is you're trying to build. Stop scratching the surface with basic prompts
I use Windsurf and I pay $60/mo out of pocket because my work won't pay for it. But it has made me love my job again. It is well worth the price and it is incredible with Claude 3.7.
when a tool actually makes you enjoy your job again, it's clearly doing something right. Claude 3.7’s reasoning capabilities are next level, so I can imagine Windsurf + Claude is a strong combo.
Also respect the fact you’re paying out of pocket. That’s the kind of endorsement you can’t fake.
Out of curiosity, are there any specific workflows or tasks where it’s made the biggest difference for you?
The stuff I've been dreading for months or years. Backlogged Jira tickets that I've felt we are understaffed for, refactoring temperamental parts of our products, updating documentation, documenting poorly designed legacy code, creating proofs of concepts when I don't have figma designs, and last but definitely not least- debugging problems that would have previously taken me hours to solve in minutes.
Everything about it has been a game changer.
I've been doing this work for many years now and my company has been chronically understaffed so Windsurf has given me breathing room. Plus, I already think of myself as a competent engineer so I like to think I'm using it responsibly and not writing future technical debt code.
You should definitely write a case study on it, when you've time. It would help everyone a lot to understand windsurf and how to use it productively. The way you summarised it, makes me realise that, damn you nailed it....
I've been using windsurf since day one and it's assistant before that. While I have had a few problems with it, it is far better a fixing bugs and generating the correct boilerplate for me. Most assistants (I've tried them all) can do some of what it can do, but not all of it. My biggest beef is that it hallucinates and requires a new context (chat) much sooner than others.
I've heard really solid things about Windsurf, especially from folks who've stuck with it since the early days. Interesting to hear it beats others in bug fixing and boilerplate, that’s usually where most assistants start to fall apart for me.
I’m curious though: have you found any good workarounds for the context resets? Like chunking tasks differently or structuring prompts in a certain way?
I might give Windsurf another spin after hearing this
Honestly I just hate the agent mode. It's more trouble than it's worth usually
While these are all valid points, the main reason for me was VS Codes's insanely frequent updates. This means Cursor is constantly running behind in updates.
My nightmare moment was when I asked it a git question and in addition to a mediocre answer, it decided to keep going and run
git reset --hard origin/main
LOL...Definitely bookmarking this under "reasons I don’t give full control to anything with an autocomplete."