Dream Team is an innovative dual-purpose platform that serves as both a simulator for team-based software development processes and an agent environment for real-world product development.

The system simulates the entire development lifecycle, where all participants—clients, developers, and project managers—are represented as autonomous agents powered by large language models (LLMs). In this simulated environment, agents undergo training, interacting dynamically with both other agents and real developers, building critical skills in development that prepare them for real-world applications.
In the LLM-based sandbox, it’s possible to create virtual alter egos of real developers. This is achieved by analyzing their code repositories, digital footprints, and other data that capture their unique style and approach to development. These digital personas can assist in testing, support, and even accelerate development processes by anticipating the actions and methodologies of the developers they emulate.
At the core of Dream Team’s approach is the DevAgent-Zero methodology, which enables agents to continuously improve their capabilities by leveraging extensive knowledge bases and LLM insights, while also learning from both successful and failed scenarios. Simulation results demonstrate that agents in Dream Team can incrementally enhance their productivity and task success, making the platform a powerful tool for analyzing and building agent-developer hybrid teams.

In this way, Dream Team becomes a unique environment where trained agents not only learn but are prepared to transition to real-world development projects, applying their refined skills to boost both effectiveness and accuracy in project delivery.
Dream Team environment implementation
The Dream Team sandbox environment is built using the Phaser web development framework . The visual environment sprites, including agent avatars, as well as an environment map and collision map that we authored, are imported into Phaser.

We supplement the sandbox development framework with a server that makes the sandbox information available to generative agents and enables generative agents to move and influence the sandbox environment. The server maintains a JSON data structure that contains information about each agent in the sandbox world, including their current location, a description of their current action, and the sandbox object they are interacting with. At each sandbox time step, the sandbox server parses the JSON for any changes coming from the generative agents, moves the agents to their new positions, and updates the status of any sandbox objects that the agents are interacting with (e.g., changing the status of the dev env from “idle” to “regression tests ” if an agent’s action is “start test”). The sandbox server is also responsible for sending all agents and objects that are within a preset visual range for each agent to that agent’s memory, so the agent can react appropriately. The agent’s output action then updates the JSON, and the process loops for the next time step.

End users initialize a new agent with a brief natural language description, we split this semicolon-delimited list of characteristics up into a set of memories. These serve as the initial memories that determine the agent’s behavior. These memories are initial starting points: as the agents gain more experience in the sandbox world, and as more records saturate the memory stream, the agent’s summary and behavior will evolve.
The Relevance and Future of Agent Sandboxes
To understand the importance of creating agent sandboxes, it is worth turning to the thoughts of Dario Amodei, CEO of Anthropic, presented in the article “Machines of Loving Grace. How AI Could Transform the World for the Better.” He emphasizes that the risks and opportunities of powerful artificial intelligence can be enormous, and managing these risks effectively is the key to a positive future.
Agent sandboxes are controlled environments where AI systems can operate, analyze, and learn with minimal impact on the outside world. These environments help test and hone AI in a safe environment, mitigating risks and minimizing the possibility of errors and uncontrolled situations.
As Amodei notes, one of the key aspects of risk research is preventing unpredictable consequences of powerful AI that can cause significant harm. Agent sandboxes allow you to simulate a wide range of situations, avoiding scenarios where the consequences of AI actions can be catastrophic. In controlled environments, it is possible to quickly and reliably test hypotheses about how a system reacts in different circumstances, allowing one to identify potential vulnerabilities and dangerous patterns of behavior.
Moreover, agent sandboxes can demonstrate the “beautiful future with AI” that Amodei talks about. By testing systems in a safe environment, one can develop positive scenarios for their use, while simultaneously working to minimize undesirable consequences. Sandboxes help developers focus on developing “positive outcomes,” creating space for innovation that can be useful to society.
Analyzing Dario Amodei’s thoughts on how humans and AI can complement each other in production chains, it becomes clear that even in the conditions of almost complete automation, there will remain a critical share of work that requires human intervention. Amodei suggests that when AI performs 90% of tasks, the remaining 10% will become much more valuable and will lead to increased compensation for the people working on these tasks. This situation will create a new type of job, where people will play an important role in coordinating the actions of AI agents and giving them direction. In mixed teams consisting of humans and AI agents, it is the person who performs the remaining 10% of the work who will take the leading position. This person – be it a team leader or a development manager – will become the link responsible for the most important strategic decisions and the task of correctly distributing roles in the team. The main goal of the leader in such teams will be to adjust the interaction between AI and people, achieving harmonious and productive work. The phenomenon of the “fixed work fallacy” that
Amodei refers to shows that despite the automation of routine and technical processes, the amount of work available to people is not decreasing, but rather transforming. As technology advances and the share of automated processes increases, the demand for this 10% of human work will increase. Team leads and development managers who can effectively coordinate the work of AI agents will prove to be indispensable specialists, and their professional value in the labor market will increase. The importance of human leadership is also emphasized by the fact that human participation in real physical tasks will remain relevant in the coming decade. Despite advances in automation, it is people who will perform those actions that require flexibility, intuition, and skills that go beyond standard algorithms.
Links
- https://arxiv.org/abs/2405.02957
- https://arxiv.org/html/2405.10467v2
- https://arxiv.org/abs/2304.03442
- https://phaser.io
- https://darioamodei.com/machines-of-loving-grace
- https://github.com/modelcontextprotocol/servers
- https://modelcontextprotocol.io/quickstart

