I’m currently using Cursor.sh to assist with my development work on a client project. My client has raised concerns about potential NDA violations and data protection, even though I’ve been using Privacy Mode to prevent data collection.
Specifically, my client wants assurance that enabling Privacy Mode ensures that no data is saved or stored in Cursor.sh’s database when scanning files on my local computer. I’ve reached out to the Cursor.sh team for documentation or evidence to verify this, but I was wondering if anyone here has dealt with similar concerns or has any insights.
Have any of you had to prove to your clients that Privacy Mode effectively prevents data storage? How did you go about getting this assurance? Any documentation, experiences, or advice would be greatly appreciated.
You can read a new page on privacy.
Click [ Full Policy ] toggle.
In addition, Business plan includes OpenAI’s zero-data retention policy.
OpenAI’s zero-data retention policy refers to a practice where OpenAI does not retain any personal data or conversation history provided by users through its services. This policy ensures that:
No Data Storage: Any data sent to OpenAI’s servers is not stored permanently. Once the interaction is completed, the data is deleted and not saved in any database or storage system.
Privacy Protection: This policy enhances user privacy by ensuring that their information is not retained or accessible for future use. It minimizes the risk of data breaches and unauthorized access.
Compliance: It helps in complying with various privacy laws and regulations that require minimizing data retention and protecting user privacy.
Security: By not retaining data, OpenAI reduces potential security vulnerabilities associated with data storage.
Overall, OpenAI’s zero-data retention policy is designed to protect user privacy and data security by ensuring that personal data is not stored or retained beyond the duration of the interaction.
Thanks @kinopee but I need a piece of proper evidence to prove that when I turn privacy mode on, Cursor won’t store any of my code to the DB, your privacy page alone doesn’t convince my client enough
This is a privacy hell. Under no circumstances, you are okay using cursor without getting your client’s okay. Not only they will pass the prompt to other servers, they will also use their servers to proxy and you cannot guarantee data residency will be honoured. This is the major limitation of Cursor. I doubt anyone in the EU and doing any kind of business with customer data can use this.
If you enable “Privacy Mode” in Cursor’s settings, none of your code will ever be stored by us or any third-party (except for OpenAI and Anthropic, which persist the prompts we send to them for 30 days for trust and safety. Business plan users’ data will not be retained at all by OpenAI or Anthropic.)
If you choose to keep Privacy Mode off, we may save prompts / collect telemetry data to improve the product. If you use autocomplete, Fireworks (our inference provider) may also collect prompts to improve inference speed.
Other notes
Even if you use your API key, your requests will still go through our backend! That’s where we do our final prompt building.
If you choose to index your codebase, Cursor will upload your codebase in small chunks to our server to compute embeddings, but all plaintext code ceases to exist after the life of the request. The embeddings and metadata about your codebase (hashes, file names) may be stored in our database, but none of your code is.
I don’t get the sense that the Cursor team are ‘bad guys’, they always seem to be open and transparent about what they are building, how they are doing it, and how their services work, but for their benefit, and that of users, I’d love to see a published checklist of business security ‘must haves’ to facilitate adoption and onboarding and save everyone’s time when having related discussions.
To be honest, I find even the big players aren’t very good with explaining their security practices ‘precisely’, I’ve already had too many discussions essentially speculating on what, exactly, is occurring, because the docs aren’t clear enough.
One of the questions I have is if there is a way to remove the index / embeddings on the server? I know there is the delete index option in the editor, but this does not delete the embedding on the server (the resync happens immediately).
Was also wondering what happens when I delete my account in the profile settings. Do the embeddings then get removed from the server?
Did this change, does anyone know they removed the clause
(except for OpenAI and Anthropic, which persist the prompts we send to them for 30 days for trust and safety. Business plan users’ data will not be revealed at all by OpenAI or Anthropic.)
Correct! Zero data retention with OpenAI and Anthropic is now available for all users that have privacy mode enabled, regardless of whether on the business plan or not.