Agent and Chat should accept video as input

Agent and Chat should accept video as input

That’s it, that’s the request.

Today I was doing some UI work and I wanted an animation effect that was hard to describe in words. I ended up doing a screen recording of a similar animation, putting that into Google AI Studio, asking it to generate code, and then giving that code to Cursor to integrate into my project.

I was honestly surprised Cursor didn’t accept video inputs.

This feature would be amazing for describing animations or complex user experience workflows to the models.

5 Likes

yea I second this

I third this.

I fourth this

I fifth this

+1 on this, GPT, Gemini, etc all allow videos on their models in their webUI

Been a cursor user for a long while but found this today would have been most helpful trying to explain some odd UI behavior. I tried using a GIF as it suggested but that was too choppy for the agent model to pick up on it.

+11 for this ; would be so much easier than painstakingly narrating multiple screenshot and also capturing stuff that a screenshot can’t quite get (a nuanced hover state, e.g. … my current fallback for when I can’t screencap something d/t limitations of a live UX interaction that precludes using a screenshot keyboard shortcut is to literally use my iphone to capture a video then grab specific frames). Would be a gamechanger if I could just loom and narrate what I am trying to describe, pass cursor the loom link and then it could use the loom transcript + loom frame markers (e.g. thumbs-up/reactions) + whatever video it wanted to digest. Or…sure, upload a short webp/mov.

(I can appreciate the challenge of trying to support someone uploading long video and expecting quick digestion of it… ¯\_(ツ)_/¯ )