What are your opinions on llama3 70b? Also groq team made something incredible i guess. Their response speed is unbelievable right now. If you do not get into queue, you will have full thousands of tokens output in 2 seconds.
What are your thoughts on this?
And also there are rumors about gpt5 coming in Apr 22
Yes we want Llama3 8b instruct on Groq w 800 tokens/s.
It’s very close to gpt-4 level. Not huge context, but Cursor covers this already.
That would be an INSANE upswing for Cursor. I catch myself sometimes just going there, instead of cursor, or web supported AI like Perplexity. So far I always got my answer.
Can you share where you got that from? Looking at the coding category on https://chat.lmsys.org/?leaderboard it seems pretty meh. Good for it’s size, especially since it’s open, but nowhere near gpt-4 from what I can see. Would love to be convinced otherwise
gpt5 propably will be something to do with agents etc. ofc it will be 5x better than gpt4 because of the data and training time. but i think llm plateau is close…
Groq just announced a set of SOTA fine-tuned Llama3 models for function calling. I would love to see these guys integrated into Cursor. x.com
Right now, a huge amount of my dev time is sitting waiting for changes to be applied to files after using the composer UI. If Groq could be used to speed up any part of this cycle, it could save me tons of time.