Cursor 2.2: Multi-Agent Judging

A few issues with the process of multi-agent workflow.

  • sometimes they generate wildly different file names and structures
  • sometimes they generate the same filename
  • comparison is difficult for a human reviewer
  • presenting the result from one model to other models is difficult as the changes flow only back into the original folder
  • sometimes the models have different ideas, findings and knowledge, it is hard to manually steal their homework and share it between the models so they can copy and remix
  • sending a new query to all the agents at once after the initial query, is unstable and difficult
  • what happens to the forks after we close the chat session, are they going to use up space forever?