Create your own trees of AI agents that work towards a common objective. Together, let's explore the potential of Agent Actors and inspire the LLM community to delve deeper into this exciting realm of possibilities!
Watch this 2-minute demo walkthrough — https://www.loom.com/share/8e60585f069c4a9f8ac9f01204b41704
Here's a sample generated execution plan:
- Time Weighted Long-Term Memory using
langchain.retrievers.TimeWeightedVectoreStoreRetriever
- Synthesized Working Memory: An agent draws insights from and synthesizes their relevant memories into a "working memory" of 1–12 items for use with zero-shot prompts.
- Implements the Plan-Do-Check-Adjust (PDCA) operational framework for continuous improvement.
- Automatic Planning and Distribution of Tasks to Agents Our
ParentAgent
class plans tasks for its children to do and distributes them to be completed in parallel. - Parallel Execution of Agents
ChildAgent
s work in parallel to Do and Check their results. Before running, they wait for all task dependencies to be resolved and inject that into their context. - Create your own trees of autonomous AI agents You can nest
ParentAgent
s underParentAgent
s, or comingle them withChildAgent
s. Use your own vector store, retriever, or embedding function with ourParentAgent
andChildAgent
classes. See how easy it is intest_system.py
.
- Terminating auto-GPTs that converges across successive runs
- Your own research and reporting teams of agents
- Simulation-driven organizational behaviour research
- Create a developer team of AutoGPTs that code for you together
- Proof of Concept, not production ready
- We've only tested used GPT3.5
Requires Python: ^3.10
Install through your choice of package manager:
poetry add git+https://github.com/shaman-ai/agent-actors.git
git clone https://github.com/shaman-ai/agent-actors
cd agent-actors
poetry install --with dev --with typing
from agent_actors import (
Agent, # subclass and replace with your own `run` method
ChildAgent, # Do and Check
ParentAgent, # Plan and Adjust
ConsolePrettyPrinter, # Helpful for printing JSON task outputs, pass as a handler to CallbackManager
)
Read test_system.py
!
- Modify
test_system.py
to your own needs - Run
poetry run pytest -s -k 'test_name_str_to_filter_by'
You can also run all tests with poetry run pytest
, but this may take a while to execute, and is likely to hit into API rate limits.
Check out this diagram to understand how the system works: https://beta.plectica.com/maps/W26XSGD28
- Improved Agent Prompts: Develop better prompts for the Plan, Do, Check, and Adjust chains
- Visualization Tooling: Develop an interface for exploring first, then composing, an execution tree of Agent Actors, allowing researchers to better understand and visualize the interaction between the supervisory agent and worker agents.
- Evaluation Data: Understanding how this performs in different contexts is key to developing a better AGI architecture.
- Unlock Talking to Agents: The dialogue functions are there, and we're looking for help on how we can "talk" to these agents from another, say, IPython, to get a look into their state.
- Unlock Inter-Agent Communication: What happens if agents can talk to each other, not just return results to their parents and write memories to the global store?
Our project is licensed under the GNU Lesser General Public License v3 (LGPLv3). This provides a balance between open-source values and flexibility for developers. Key points include:
- You can use this as a library in your closed-source software.
- Modifications to our library must be released under an LGPLv3 license, the easiest way to do this is to submit a PR to this project.
- Contributors benefit from patent protection.
For the full legal text, see the LICENSE file.
We extend our gratitude to the contributors of the Python packages langchain and ray, without which this wouldn't be possible. We extend our gratitude to the amazing researchers who wrote Generative Simulacra [TODO], ReAct [TODO], and Jeremy Howard [TODO] and FastAI, without which this wouldn't be possible. And to BabyAGI and AutoGPT for inspiring us.
Citation Please cite the repo if you use the data or code in this repo.
@misc{agentactors,
author = {Shaman AI},
title = {Agent Actors: Plan-Do-Check-Adjust with Parallelized LLM Agent Trees},
year = {2023},
publisher = {GitHub},
journal = {GitHub repository},
howpublished = {\url{https://github.com/shaman-ai/agent-actors}},
}