Hey everyone! Not sure if many people experienced some slowness like I did lately, but I wanted to understand if it’s normal for a model to be slow, and when are the peak hours are to try to avoid them (we all like our AI to be fast). So, I built this tool to track providers and models performance over time.
We’re all somewhat irritated when it takes forever to get a response for a simple request, and I hope that this could help you “ease” the frustration by knowing when it’s just obviously slow and not necessarily the fault of Cursor. I know this won’t fix the speed in the actual moment you need it, but at least you’ll know why it’s happening.
Thanks, glad you find it useful! It’s not tracked per specific region, but that would be a good point to consider. Maybe the people at Cursor could tell me different, but I don’t think there’s a way to do it directly from Cursor (without hacking it, per se), so instead I used direct access to the providers since Cursor also use them directly, so the results are probably more or less the same (but for sure not 100% accurate).
The timing is from start to finish, so latency plus entire completion of prompt output, and the prompt is random up to 512 tokens. The data is cached but since the amount of data stored to do all the calculation is required it is what takes a longer time if I am correct, but for sure there might be a way to cache the end result to speed it up.
Thank you for your feedback! I’ll see what I can do.
For Deepseek I am not using the Fireworks version, I know it is not as accurate, but I didn’t have access to Fireworks, but I am open to sponsoring to help cover any related costs.