The more advanced code models are here, the more the interaction with the cursor is changing. From getting a help when stuck and automating boilerplatte to merely overseeing what agents are doing.
This shift in paradigm would tremendously profit from adding more sophisticated ways in interacting with the cursor.
The first one is obviously speech-to-text, but there is a lot of software for that, so it’s not crucial.
What is crucial IMO, is adding more text-to-speech capabilities.
I’d love also to have possibility to have text-to-speech, where “agent” would speak about what he is doing currently.
Adding it now will also be beneficial for the future, as it will be more easy to implement it in another areas.
Additional points if something like xtts-2 would be used, so I can provide a VEGA (from Doom) voice to read me this.