While the model performs quite well, the token generation speed appears to be somewhat slower compared to other models. According to the usage dashboard, it is currently being served via Fireworks.
Since Groq offers significantly faster speeds—around 200 to 300 tokens per second—it may be beneficial for Cursor to consider integrating Kimi K2 through Groq for improved performance.
Hey, thanks for the feedback. Since fireworks.ai is our partner, it was chosen as the provider for this model. Regarding the generation speed, I believe they will soon increase their capacity, and the generation speed will improve.
I toggled-on Kimi-K2 Instruct in
Models in Cursor Settings, however I don’t see it in the chat under models selection… I saw something on the forum about regions… Is it because I am in Canada? I don’t see any warning when I toggle on Kimi-K2 in settings models…
@condor I feel so dumb LOL… yes of course, scrolling! I didn’t realize because there was no scrolling bar. Happy to report scrolling works! Thanks!
Looking forward to be able to actually use it though… I tried it now, and had to wait 17 minutes for it to do my request of creating two readme.md files: the first one was completed after 15 minutes, and the second one was aborted (failed generating) at 17 minutes…