Today I was doing some UI work and I wanted an animation effect that was hard to describe in words. I ended up doing a screen recording of a similar animation, putting that into Google AI Studio, asking it to generate code, and then giving that code to Cursor to integrate into my project.
I was honestly surprised Cursor didn’t accept video inputs.
This feature would be amazing for describing animations or complex user experience workflows to the models.
Been a cursor user for a long while but found this today would have been most helpful trying to explain some odd UI behavior. I tried using a GIF as it suggested but that was too choppy for the agent model to pick up on it.
+11 for this ; would be so much easier than painstakingly narrating multiple screenshot and also capturing stuff that a screenshot can’t quite get (a nuanced hover state, e.g. … my current fallback for when I can’t screencap something d/t limitations of a live UX interaction that precludes using a screenshot keyboard shortcut is to literally use my iphone to capture a video then grab specific frames). Would be a gamechanger if I could just loom and narrate what I am trying to describe, pass cursor the loom link and then it could use the loom transcript + loom frame markers (e.g. thumbs-up/reactions) + whatever video it wanted to digest. Or…sure, upload a short webp/mov.
(I can appreciate the challenge of trying to support someone uploading long video and expecting quick digestion of it… ¯\_(ツ)_/¯ )