> Cursor Composer 1 was Qwen and this is Kimi. IDE is based on VSCode. The entire company is build on packaging open source and reselling it.
The question is, where's the outrage? Why are there no headlines "USA steals Chinese tech?" "All USA can do is make a cheap copy of Chinese SOTA models".
> So funny to see Twitter go wild saying "a 50 person team just beat Anthropic" blah blah.
Well, if it's an American company, then it's a noble underdog story. When Chinese do it, they are thieves leeching on the US tech investment.
It's a bit more than that. They have plenty of data to inform any finetunes they make. I don't know how much of a moat it will turn out to be in practice, but it's something. There's a reason every big provider made their own coding harness.
Cursor’s integration is much deeper than just plugging an LLM into VSCode
That said I have a feeling both VSCode and Claude code will catch up to their integration. But neither comes close yet (I say that as someone who mainly uses Claude Code).
We know Composer 2 is Kimi K2.5 from that tweet. Where is the evidence for Composer 1 being based on Qwen?
> So funny to see Twitter go wild saying "a 50 person team just beat Anthropic" blah blah.
In this case, it will be the other way round: Anthropic will see Cursor as a competitor AI lab using open weight models for Composor 2 (actually Kimi K2.5) which was allegedly distilled from Opus 4.6, and would be enough for Anthropic to cut off Cursor from using any of models.
It just means Kursor is sharing data with Chinese llm which enables them to improve their LLM by training on outputs and input of all data which cursor collects.
Looks like two Moonshot employees confirmed that it's not licensed before Moonshot made the decision to get out of the debate and delete their posts [0][1].
Cursor is mostly an IDE / coding-agent harness company.
So it probably makes sense for them not to train their own base model, but instead license something like Kimi and fine-tune it for their own harness and workflows.
Their moat looks pretty thin. A VSCode fork with an open-source LLM fork on top. In the fast-moving coding-agent market, it’s not obvious they keep their massive valuation forever.
"Just" Kimi K2.5 with RL—people really misunderstand how difficult it is to achieve these reults with RL. Cursor's research team is highly respected within the industry, and what they've done is quite impressive.
Before people go jumping to conclusions about model theft, it's worth considering the possibility that they did reach an agreement with Moonshot which their researchers were not aware of. That would certainly explain the deleted tweets. Until Moonshot makes an official statement, I'm not particularly concerned.
There are many reasons to make fun of Cursor. However, one of the things get right is their autocomplete model.
Are there any open models that come close? Why doesnt OAI or Anthropic dedicate some resources to blowing Cursor's model out of the water? Cursor's completion model is a sticking point for a lot of users.
They’re pretty upfront in their release post that they took an open source model and improved it with their own coding data. They mention “continued pretraining” (on top of the base model) and RL. Cursor never claimed to have done a full pretraining run.
More to the point, beating Opus 4.6 at coding and coming within striking distance of gpt-5.4 is impressive! The benchmarks outperform raw Kimi K2.5.
It’s particularly impressive given larger labs like Meta are struggling to catch up to OpenAI/Anthropic.
This is exactly what Cursor should be doing, within the obvious bounds of the law and such. Not everyone needs a pristine foundation model. What a waste of compute. Anthropic & OpenAI need product-level competition to knock them off their $25/Mtok horse.
The cursor investor pitch was we're training our own models to do coding. If your amazing model is just an RL repack, you need a new pitch to justify your 50bn valuation
What do people like about cursor? I've been using it for the past couple days, and I just don't see many positive things about it. It seems people like the autocomplete so I'll have to give that a try.
There's just too many "features" the ux ends up being all over the place. I thought having the browser inside of the editor would be great for design, but it's not that much better than just having your browser open along with your editor.
I noticed something strange with Cursor lately. When I am using Opus 4.6, sometimes it is giving ridiculously dumb answers as if they were actually using something like Qwen with a prompt to present itself as Opus. I have to close the session and start again hoping I'll get actual Opus.
I don't know - it works okay (yet to be tested whether it is actually smarter than Opus 4.6), but it is not bad at all. So far, it works quite fine (I'm not testing the "fast" version).
I'm annoyed that we still don't know for certain which base model they used for Cursor 1.
This feels really rude to me. I have no problem with them fine-tuning open weight models to create their own - they are getting great results, and Cursor's research term should be respected for that. But deliberately hiding the base model they use is disrespectful of the researchers who created that model.
168 comments
Ollama is also doing this.
There is so much money to be made repackaging open source these days.
So funny to see Twitter go wild saying "a 50 person team just beat Anthropic" blah blah.
> Cursor Composer 1 was Qwen and this is Kimi. IDE is based on VSCode. The entire company is build on packaging open source and reselling it.
The question is, where's the outrage? Why are there no headlines "USA steals Chinese tech?" "All USA can do is make a cheap copy of Chinese SOTA models".
> So funny to see Twitter go wild saying "a 50 person team just beat Anthropic" blah blah.
Well, if it's an American company, then it's a noble underdog story. When Chinese do it, they are thieves leeching on the US tech investment.
It's all so predictable, even the comments here.
> packaging open source and reselling it.
It's a bit more than that. They have plenty of data to inform any finetunes they make. I don't know how much of a moat it will turn out to be in practice, but it's something. There's a reason every big provider made their own coding harness.
That said I have a feeling both VSCode and Claude code will catch up to their integration. But neither comes close yet (I say that as someone who mainly uses Claude Code).
> Cursor Composer 1 was Qwen...
We know Composer 2 is Kimi K2.5 from that tweet. Where is the evidence for Composer 1 being based on Qwen?
> So funny to see Twitter go wild saying "a 50 person team just beat Anthropic" blah blah.
In this case, it will be the other way round: Anthropic will see Cursor as a competitor AI lab using open weight models for Composor 2 (actually Kimi K2.5) which was allegedly distilled from Opus 4.6, and would be enough for Anthropic to cut off Cursor from using any of models.
That's where it is going.
> There is so much money to be made repackaging open source these days
These days? Almost every tech offering in existence is 1000+ OSS dependencies gaffer taped together with a sprinkling of business logic.
Cursor isn't a shocking bit of software to pay for, its investment however...
It's a two way street.
[0] https://chainthink.cn/zh-CN/news/113784276696010804 - may have originally been https://x.com/apples_jimmy/status/2034920082602864990
[1] https://pbs.twimg.com/media/HD2Ky9jW4AAAe0Y?format=jpg&name=...
Their moat looks pretty thin. A VSCode fork with an open-source LLM fork on top. In the fast-moving coding-agent market, it’s not obvious they keep their massive valuation forever.
Before people go jumping to conclusions about model theft, it's worth considering the possibility that they did reach an agreement with Moonshot which their researchers were not aware of. That would certainly explain the deleted tweets. Until Moonshot makes an official statement, I'm not particularly concerned.
Cursor accesses Kimi-k2.5 via FireworksAI_HQ hosted RL and inference platform as part of an authorized commercial partnership.
https://x.com/Kimi_Moonshot/status/2035074972943831491
Cursor teams take:
Only ~1/4 of the compute spent on the final model came from the base, the rest is from our training. This is why evals are very different.
https://x.com/leerob/status/2035035355364081694
Are there any open models that come close? Why doesnt OAI or Anthropic dedicate some resources to blowing Cursor's model out of the water? Cursor's completion model is a sticking point for a lot of users.
More to the point, beating Opus 4.6 at coding and coming within striking distance of gpt-5.4 is impressive! The benchmarks outperform raw Kimi K2.5.
It’s particularly impressive given larger labs like Meta are struggling to catch up to OpenAI/Anthropic.
https://www.bloomberg.com/news/articles/2026-03-12/ai-coding...
They should have disclosed it though. If they didn't it's a bad look for sure.
Would this have been extensively fine tuned, beyond what Anthropic/OAI would do themselves?
I guess this is Cursor's own benchmark, so you can finetune on your own dataset and get better results on your own specific tasks I guess.
There's just too many "features" the ux ends up being all over the place. I thought having the browser inside of the editor would be great for design, but it's not that much better than just having your browser open along with your editor.
Currently I’ve not hit any of the limits despite using it quite rigorously, I wonder if this will change with a renewal?
This feels really rude to me. I have no problem with them fine-tuning open weight models to create their own - they are getting great results, and Cursor's research term should be respected for that. But deliberately hiding the base model they use is disrespectful of the researchers who created that model.