I don’t care if I have to attach my own API, really I don’t.
Claude 3.5 in Cursor ran me in circles for two days and I dumped to my last commit over 15 times. I read a solid review on the latest 0827 Gemini and thought “hell, what could possibly go worse?”
Solved the issue in under 3 hours with a major refactor and it didn’t try to “use client” on a server file even once.
The new Gemini needs to be part of Cursor because manually copying over git diffs and attaching files to Google Drive is the most necessary tedium I’ve experienced since Claude went batsh*t crazy a few weeks back.
Thanks! I’m going to try tomorrow. I’m wondering if it’s possible to use it in the long context chat? Usually adding “-200k” or something to the model makes it available in the long context chat. I wish there were no restrictions on the number of tokens. It’s been long asked of Cursor team to let us choose arbitrary token limits when using own API, but it’s been ignored so far. Hopefully at least 200k works in the long context chat. Or perhaps it’s possible to append any value like 500k or 2000k? If so I’ll take my words back, I’m gonna test this tomorrow.
Yeah it’s working for me in long context chat. But I don’t know what’s the maximum token for it gemini flash maximum is 500k token maybe it’s the same for gemini pro.
I’m glad this post is here! I tried Gemini 1.5 pro the latest version yesterday in the AI Studio and on Vertex AI and I think they’ve might some there! If I understand correctly, every time you have a chat, it adds tokens to the overall conversation. I have not gotten near 2 million tokens but with only building up to around 30,000 to 40,000, it just seemed to get better and better on every prompt. I’m excited to build it up to 2 million tokens and then see if it can work with that and create new, original code related to the entire project. The speed is insanely fast as well. I concur it would be amazing to have the 2 million token chat in Cursor. In the Ai Studio there is an option to add more tokens up to it seems the 2 million, and Vertex AI seems to allow only 8196 tokens output at a time.
Looks like Google might be finally pulling out the big AI guns with fine tuned gemini, gems (personally instructed Gemini’s) agent armies, and the whole bit. I’m too tired now, headed to bed! Looks like we have a fun weekend ahead of us!
EDIT: Make sure you create your API key after creating your paid account. I was using the same key I generated weeks ago. “Turning it off and on again” still works.
“The model gemini-1.5-pro-exp-0827 does not work with your current plan or API key”
I have a paid plan with like $300 in credits. What haven’t I turned on?
The context is 2mil in AI Studio but the API may be different. For example, there’s no rate limiter in the studio but the API cuts me off after 1-2 hours of usage in Cursor.
Not specifically, though after several days I have feedback.
Claude is still the best at coding new features for Next 13+, if you are using React. I find it curious that Gemini consistently codes for an outdated version of React. I wonder if this extends to other languages.
Gemini is vastly better at complex refactors, which isn’t surprising given it’s context window. It’s also far better at thinking through application structure on a grander scale than “I need a new view attached to this section of the app”
Gemini and Claude are about equally likely to get stuck in a logical loop and give me code that doesn’t actually change anything. Claude is slightly more likely to loop back and forth between mistakes if it’s context gets poisoned.
Claude almost causally deletes key features in order to “fix” a bug. It’s also vastly more likely to overcomplicate a refactor.
ChatGPT is the least likely to get caught in a logical loop, which makes it useful for breaking out of a funk when Claude and Gemini can’t find the root of an issue. It’s also much better at conversing around error logs and thinking through logical puzzles to get back a few steps in a process. Claude and Gemini don’t like to go backwards and ask if they’re even solving the right problem.
All in all, it’s incredibly useful to have all three available.