DEV Community

Cover image for Cursor AI Was Everyone’s Favourite AI IDE. Until Devs Turned on It
Abdul Basith
Abdul Basith

Posted on

37 2 1 2 2

Cursor AI Was Everyone’s Favourite AI IDE. Until Devs Turned on It

Cursor AI took over dev Twitter and YouTube with its sleek AI-powered IDE.

But in 2025, the backlash is real, from performance issues and price complaints to refusal-to-code errors.

Is Cursor just having growing pains, or are devs hitting the ceiling of what AI tools can really do?


🚀 Cursor AI: The Hype That Made Sense (At First)

Cursor launched as the “AI IDE” merging the best of GitHub Copilot and VS Code.

It promised:

  • Inline AI chat for every line of code
  • Bug-fixing via prompt
  • Natural language refactoring
  • Seamless context switching across files

For a while, it delivered. But recently, cracks have started to show.


🧱 The Cracks in the Code: What Went Wrong

🧨 1. The Refusal Incident

In March 2025, Cursor's AI shocked devs by outright refusing to generate code — telling one user to “learn programming instead.”

“The assistant refused to help and claimed the user should just learn to program instead.”

🔗 Ars Technica

It raised serious concerns about trust and reliability.


🐢 2. Sluggish Performance with Larger Codebases

Users on Cursor's Forum have reported:

  • UI lag on files over 500 lines
  • Increased crashes with heavy AI usage
  • Memory leaks during long editing sessions

For a speed-focused IDE, this is a red flag.


💸 3. The $20/month Pricing Controversy

Devs have begun questioning Cursor’s value:

“Cursor has a pricing problem and it’s not just the cost, it’s the return.”

🔗 @watzon on Medium

VS Code + Copilot is cheaper, more stable, and doesn’t lock you into a single IDE.


🐧 4. Linux Friction

Cursor ships as an AppImage. Sounds simple, but many Linux users are frustrated by:

  • Poor integration
  • Compatibility quirks
  • Lack of packages or flatpak options

For a tool marketed to devs, this hits hard in the open-source crowd.


🐞 5. AI Struggles with Bug Fixing

@chrisdunlop writes:

“It’s actually extremely hard for Cursor AI to fix bugs... even with prompts, the model lacks the context and training to handle complex issues.”

🔗 Full Article

In other words: Copilot + Context ≠ Cognitive Debugging.


🧠 The Bigger Picture: Are We Expecting Too Much?

Cursor is innovating fast. Its Composer Agent and .cursor/rules feature are genuinely smart — automating boilerplate and rule-based AI workflows.

But backlash shows a trend:

Devs aren’t mad because AI tools are bad.

Devs are mad because they were over-promised.


🔍 Final Thoughts

Cursor still has a shot at being the future of AI-native coding. But first, it needs to:

  • 🔧 Fix performance
  • 🛠️ Improve trust and fallback logic
  • 💰 Reconsider pricing tiers
  • 🐧 Win back Linux devs
  • 🤖 Set realistic expectations about what AI can do (and what it can’t)

💬 Join the Discussion

  • Are you still using Cursor, or have you switched?
  • What’s been your best/worst experience with AI coding tools?
  • What feature would actually win you over?

Let’s talk in the comments. 👇


📚 Sources

Top comments (24)

Collapse
 
xuroth profile image
Stephan Franz • Edited

I'll chime in here, I suppose. I've been using Cursor for a while and have a love/hate relationship with it. For simple or boilerplate code, it's amazing. For explaining code it works pretty well (though it sometimes gets confused if you are building something that has a name similar to a commonly used package - I have a tool I'm building for a toolkit called "Apollo" and I frequently need to remind the LLM that I am Not in fact using Apollo client or graphql).

Bug fixing is mostly negative. Sometimes it will catch things, especially if it's a typo or accidental infinite loop, but building more "complex" things, say a sortable nested folder tree with drag and drop support, it suggests poorly written code 9/10. In most cases, however, the quality of the code depends on the quality of the prompt. More advanced features must often be broken down into smaller chunks, and then you have to play the game where you connect the pieces. Which is good, in a way, because I never want to rely on LLM to my detriment. I'll also note, in case it's relevant, I only use "ask" mode - I want to evaluate the code mentally and then implement it to fit with other systems. I only use agentic mode for boilerplate or rapid scaffolding.

Performance is a pain. My laptop has 32 gb ram, and Cursor eats it up like it's at a buffet. I often need to bounce between 3-4 projects/code bases and it's easily on par with my browser (2 windows, about 20 tabs each). If I want to game after work, i often just resort to restarting my pc (Cursor is running in WSL).

Overall, I enjoy Cursor and it has helped with some complex tasks, but I do often find myself wishing it was leaner and that the LLMs worked a bit better at "deduction" (Not a Cursor issue, just Gen AI in general - helped by providing better prompts). If I have a problem that is common in stackoverflow, for example, Cursor will often get me a good solution on its first try. Anything more complex, or "senior" level code, not so much.

I need to learn more about the MCP thing though. Any advice? Only tool I have listed in it is Console Ninja.

I do agree with the general sentiment that prompt quality is directly related to output quality. I'm often reminded of the "make a sandwich" game that some people play when learning to think like a programmer. One person provides directions and the other will do EXACTLY what they are told, and the instructor has to learn to provide clear, simple step by step instructions in order to get the desired result).

Collapse
 
jwp profile image
JWP

Thats a sign of a product that, to me has serious design points. I predict they won't go to far given the massive MSFT resources.

I use vscode every day. Their copilot plug-in has some similar issues. Super slow response , frequent hallucinating, stubbornessstubbornness, continuation of its own design, won't pick up manual changes on suggestions. There are lots of gaps making this a mostly bad experience due to two week iterations.

It's a major impediment to delivery now. The potential keeps us trying.

Collapse
 
abdulbasithh profile image
Abdul Basith

Hey, really appreciate you taking the time to share such a detailed and balanced perspective.

Totally hear you on the love/hate dynamic. Cursor definitely shines when it comes to boilerplate and simpler logic. But yes, when it’s faced with nested or more “senior-level” patterns (like your sortable folder tree example), it can feel like you’re stitching half-baked puzzle pieces together from multiple prompts. Kind of fun... kind of exhausting.

That Apollo confusion sounds all too familiar, I've run into something similar when building tools with generic names, and the LLM insists on dragging in irrelevant package context.

You raised a great point about “ask mode” vs “agentic mode” too. I also prefer reviewing code mentally before integrating, almost like pair programming with a hyperactive intern. Fast, but sometimes a bit too eager.

In terms of performance, I feel you. Cursor in WSL is a memory monster. I ended up limiting its background context size and turning off certain telemetry-based features to make it bearable. Still not ideal if you’re switching between work and gaming on the same machine.

About MCP (Most Capable Programmer), it’s a feature Cursor uses to let you “declare” which tools or agents you trust most in a workspace, like Console Ninja. If you haven’t already, try adding tools like AskTheCode, Inline Suggestions, or Trace when they’re available, they expand what Cursor can do in specific coding scenarios. I might put together a follow-up post diving into MCP setups and some underrated tools inside Cursor.

Thanks again lad for jumping in. Curious to hear more about how you’re using Cursor across multiple projects, especially if you’ve found any tricks to streamline memory usage or improve LLM grounding.

Collapse
 
hugohub profile image
Karlis Melderis

Out of curiosity

By the time you have created prompt that will provide desired code won't you manage to write it?

And eventually you will start structuring natural language so it ends up just another coding language but with so many more interpretations😅

I do love generating boilerplate but filling in code myself is not slowing me down.
It helps me think what the code should actually do and what edges I haven't thought about

Collapse
 
tizio profile image
Tiz.io

I find it's very fast at producing prototype level code; much faster than a senior dev even. But once that code gets moderately complex, the oversight and rewriting ends up taking as long a s doing it by hand would have.

Collapse
 
abdulbasithh profile image
Abdul Basith

I’ve had that exact thought myself more than once.

Sometimes writing the prompt does feel like it takes just as much brainpower as writing the code directly. Especially when you're already clear on what you want the function or logic to do, it can be quicker to just get on with it.

Collapse
 
srbhr profile image
Saurabh Rai

I've used Cursor, Windsurf, Jetbrains AI (in Pycharm and Webstorm), GH co-pilot (paid) in VScode and Pycharm.
My workflow jumps from using Claude Desktop and Claude in Cursor. But with each passing day, I'm getting serious thoughts on learning NeoVim and just using Claude to generate smaller functions and use them as I want them to.

Note: All of my workflows are custom and complex, I've have had higher success with using Claude Sonnet via Desktop/Chat instead of cursor which keeps breaking things.

Collapse
 
abdulbasithh profile image
Abdul Basith

Sounds like you’ve really run the gauntlet with AI tooling, love how you’ve tested across environments and zeroed in on what actually works for your workflow.

Totally get the Claude > Cursor preference too. When you’re working with custom, complex setups, stability and control matter more than shiny integrations. I’ve had similar moments where Cursor’s context injection tries to be helpful and ends up breaking things that were already working fine.

NeoVim + Claude sounds like a clean, focused stack, especially if you’re comfortable managing your own flow. Curious though, what’s pushing you toward NeoVim now? Speed? Customisation? Just needing more precision?

Collapse
 
instalab profile image
Samuel Boczek

AI was never good for coding. Not saying it will never be good, but I am yet to see a tool that can really produce something useful beyond anything simple.

Collapse
 
skamansam profile image
Samuel

My pro-tip for ai assistants is think about it like you just hired this guy a week ago, straight out of college. He has technical know-how, but lacks any experience. I'm not sure we can even teach LLMs experience, so they will never meet even moderate expectations of a real assistant.

Collapse
 
abdulbasithh profile image
Abdul Basith

That’s honestly one of the best mental models I’ve heard “fresh grad with solid knowledge, zero experience” is exactly how it feels using these tools.

They can quote the textbook, but ask them to apply it in a messy real-world context with vague requirements, and they get overwhelmed or overly confident with the wrong answer.

Collapse
 
nicholasrjohnson profile image
Nicholas R Johnson

I HATE rewriting and rewriting the code the AI produces. I DO like that it gives ideas. But, it just is more debugging to do. Too much!

Collapse
 
captaindigitals profile image
CaptainDigitals

To minimize error you have to evolve your workflows, can't keep doing the same thing and expect different results, if you're not using the MCP servers you're operating at 50% capacity in terms of error rates and expected output, then there's ROO Cline for VSCODE, probably the most comprehensive Agentic AI I've seen thus far.

You also must upgrade your prompt game, if you're inputting simple prompts expect simple output, the more detailed you can make your prompts optimizing it for the LLM you're using you will be blown away by the results and not continue to scratch the surface with basic prompts expecting amazing results.

Leverage different AI platforms to produce amazing outputs, optimize a prompt on Gpt4o for Claude.

You can make some really amazing things with a well detailed lengthy prompt for whatever it is you're trying to build. Stop scratching the surface with basic prompts

Collapse
 
bobdotjs profile image
Bob Bass

I use Windsurf and I pay $60/mo out of pocket because my work won't pay for it. But it has made me love my job again. It is well worth the price and it is incredible with Claude 3.7.

Collapse
 
abdulbasithh profile image
Abdul Basith

when a tool actually makes you enjoy your job again, it's clearly doing something right. Claude 3.7’s reasoning capabilities are next level, so I can imagine Windsurf + Claude is a strong combo.

Also respect the fact you’re paying out of pocket. That’s the kind of endorsement you can’t fake.

Out of curiosity, are there any specific workflows or tasks where it’s made the biggest difference for you?

Collapse
 
bobdotjs profile image
Bob Bass

The stuff I've been dreading for months or years. Backlogged Jira tickets that I've felt we are understaffed for, refactoring temperamental parts of our products, updating documentation, documenting poorly designed legacy code, creating proofs of concepts when I don't have figma designs, and last but definitely not least- debugging problems that would have previously taken me hours to solve in minutes.

Everything about it has been a game changer.

I've been doing this work for many years now and my company has been chronically understaffed so Windsurf has given me breathing room. Plus, I already think of myself as a competent engineer so I like to think I'm using it responsibly and not writing future technical debt code.

Thread Thread
 
abdulbasithh profile image
Abdul Basith

You should definitely write a case study on it, when you've time. It would help everyone a lot to understand windsurf and how to use it productively. The way you summarised it, makes me realise that, damn you nailed it....

Collapse
 
skamansam profile image
Samuel

I've been using windsurf since day one and it's assistant before that. While I have had a few problems with it, it is far better a fixing bugs and generating the correct boilerplate for me. Most assistants (I've tried them all) can do some of what it can do, but not all of it. My biggest beef is that it hallucinates and requires a new context (chat) much sooner than others.

Collapse
 
abdulbasithh profile image
Abdul Basith

I've heard really solid things about Windsurf, especially from folks who've stuck with it since the early days. Interesting to hear it beats others in bug fixing and boilerplate, that’s usually where most assistants start to fall apart for me.

I’m curious though: have you found any good workarounds for the context resets? Like chunking tasks differently or structuring prompts in a certain way?

I might give Windsurf another spin after hearing this

Collapse
 
michelangelo17 profile image
Michelangelo Markus

Honestly I just hate the agent mode. It's more trouble than it's worth usually

Collapse
 
jesterly profile image
jesterly

While these are all valid points, the main reason for me was VS Codes's insanely frequent updates. This means Cursor is constantly running behind in updates.

Collapse
 
tizio profile image
Tiz.io • Edited

My nightmare moment was when I asked it a git question and in addition to a mediocre answer, it decided to keep going and run
git reset --hard origin/main

Collapse
 
abdulbasithh profile image
Abdul Basith

LOL...Definitely bookmarking this under "reasons I don’t give full control to anything with an autocomplete."