Add Support for Ollama natively

Integrate local Ollama model support in Cursor to enable AI code completion locally, without relying on cloud services.

Why ?

Some user could need to rely on an off-cloud model because they can’t always have access to internet, their internet is slow, or they work in sensitive environments. Additionally, some could be concerned about privacy

What is unique ?

While Cursor’s current AI completions rely on the cloud, integrating Ollama allow users to keep their data local, this could also reduce latency, as code completion is done on the user machine, or a server on their networks.

What would it contain ?

  • Integration with Ollama API: Embed Ollama API within cursor for the user to chose models they prefer.

  • Configuration and requirement: Allow the user to choose local or a home/distant server running Ollama.

Merge with:

It could be merged with current existing AI feature that are cloud based of Cursor, the user could easily switch if needed to the cloud service, it should be seen as an alternative when people can’t or don’t want to use Cloud-based models.

54 Likes

Yes, please!

4 Likes

i am with you

3 Likes

Also interested in this

2 Likes

yes please - also interested. Can also consider Open Interpreter: GitHub - OpenInterpreter/open-interpreter: A natural language interface for computers

This is a crucial feature, as some organizations prohibit sharing code with external AI tools. It will particularly benefit developers constrained by internal privacy policies, especially those working in government agencies and large financial corporations.

5 Likes

Yes, please!

2 Likes

it would be great

1 Like

Yes this is becoming the next level for AI enabled IDEs. Its clear that some of Cursors magic is in their own models, the vectorization etc. but at least to offload some of the remote API work when e.g. Claude is overwhelmed it would help to have access to local models.

1 Like

absolutely necessary. and I’m happy to keep paying for cursor. it’s not about the money. it’s about being able to code with a decent model with no wifi.
and it should work with composer too!

1 Like

Hey, unfortunately, Cursor’s servers do a lot of the heavy lifting when it comes to communicating directly with the LLM, so it’s unlikely there will be offline support in the near future!

Sorry to be the bearer of bad news here thought, I do sympathise that there are workflows with Cursor that could benefit from offline / local LLM support!

I guess would reveal a lot of the secret sauce of how the LLM calls are being formatted ?

Perhaps it could be integrated with the mention that it wouldn’t have refined capabilities we would have going through Cursor backend. I’m sure it would drive a lot of new users to the platform. Right now codebase embeddings and composer features give cursor an edge to tools such as Continue.dev but that might change in the near future…

might allow better custom model integration, then people can set up their proxy server with their deployed model and still use them with cursor, they might run R1 or future models on their own server with the distil size they want and use them with the cursor agent / composer?

1 Like

Hi,

Any chance you can adda feature where we set a public URL for the LLM API to use for requests on a per account/user basis?

I would pay more for a feature that lets you set your own URL, so I can reverse proxy to my own liteLLM or ollama instance. This way you can keep all the backend features the same, this would a a drop in replacement for the calls your backend already made to other LLM APIs.

1 Like

The issues is thats if we do that, the whole point of the existance of this integration is none existant now (the whole point of this integration would be privacy)

You can do that with cloudflare tunnel for free. Check this repo for ideas GitHub - danilofalcao/cursor-deepseek: A high-performance HTTP/2-enabled proxy server designed specifically to enable Cursor IDE's Composer to use DeepSeek's and OpenRouter's language models. This proxy translates OpenAI-compatible API requests to DeepSeek/OpenRouter API format, allowing Cursor's Composer and other OpenAI API-compatible tools to seamlessly work with these models. (tunnel, nginx settings)

How come this isn’t an option?

Sometimes i’m offline and i’m having to do the old copy and paste from ollama to my editor and vice-versa, really feel wrong :frowning:

This makes Cline a lot more attractive for people who work offline sometimes

It’s possible with Zed: Configuration - Zed