Is cursor using DeepSeek R1 671B? Or one of the distill versions? because the difference between one of the distill version and the original version is huge.
Cursor team, please look into this issue. Multiple people are complaining about the same. @danperks@deanrie
It feels like Cursor is using another model! Try it for yourself: use the same prompt in Cursor and on DS website. The difference is huge! We want to use R1’s full power. And an essential part of that power comes from longer thinking process.
Some model providers might use the 671B model but quantize it (different from distillation) to use less resources or produce faster responses, but it reduces quality.
Same! deepseek r1 is my new favorite model, full support for the 685B version would be great as I did notice some differences from the model exposed from Cursor to chat.deepseek.com, the quality seems better on the website.
If someone from Cursor could chime in to give some clarity on the matter it’d be awesome!
Not sure if not hallucination, but R1 told me he is instructed to “Always provide brief explanation”. That “brief” might cause less thinking and thus worse results. It is only a speculation, but I believe similar wording has negative effect on performance on Perplexity as well. R1 seems to take some user prompts on Perplexity very literally, eg my prompt for Sonnet which was designed so URLs will be outputted as raw URLs, with R1 it forces the model to respond only with list of URLs. So from this I would guess R1 might need more special system prompt tailored just for R1.