AgentCoord/README.md
2024-04-19 10:19:24 +08:00

66 lines
3.4 KiB
Markdown
Raw Permalink Blame History

This file contains ambiguous Unicode characters

This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.

# AgentCoord: Visually Exploring Coordination Strategy for LLM-based Multi-Agent Collaboration
<p align="center">
<a ><img src="static/icon.png" alt=" AgentCoord: Visually Exploring Coordination Strategy for LLM-based Multi-Agent Collaboration" width="200px"></a>
</p>
AgentCoord is an experimental open-source system to help general users design coordination strategies for multiple LLM-based agents.
[Research paper](https://arxiv.org/abs/2404.11943)
## System Usage
<a href="https://youtu.be/s56rHJx-eqY" target="_blank"><img src="static/videoPreview.png"
alt="System Usage Video" width="800" border="5" /></a>
## Installation
### Install with Docker (Recommended)
If you have installed [docker](https://www.docker.com/) and [docker-compose](https://docs.docker.com/compose/) on your machine, we recommend running AgentCoord in docker
Step 1: Clone the project:
```bash
git clone https://github.com/AgentCoord/AgentCoord.git
cd AgentCoord
```
Step 2: Config LLM (see [LLM configuration (use docker)](README.md#llm-configuration-use-docker)):
Step 3: Start the servers
```bash
docker-compose up
```
Step 4: Open http://localhost:8080/ to use the system.
### Install on your machine
If you want to install and run AgentCoord on your machine without using docker:
Step 1: Clone the project
```bash
git clone https://github.com/AgentCoord/AgentCoord.git
cd AgentCoord
```
Step 2: Config LLM (see [LLM configuration (install on your machine)](README.md#llm-configuration-install-on-your-machine)):
Step 3: Install required packages, then run the backend and frontend servers separately (see readme.md for [frontend](frontend/README.md) and [backend](backend/README.md#Installation)
Step 4: Open http://localhost:8080/ to use the system.
## Configuration
### LLM configuration (use docker)
You can set the configuration (i.e. API base, API key, Model name) for default LLM in ./docker-compose.yml. Currently, we only support OpenAIs LLMs as the default model. We recommend using gpt-4-turbo-preview as the default model (WARNING: the execution process of multiple agents may consume a significant number of tokens). You can switch to a fast mode that uses the Mistral 8×7B model with hardware acceleration by [Groq](https://groq.com/) for the first time in strategy generation to strike a balance of response quality and efficiency. To achieve this, you need to set the FAST_DESIGN_MODE field in the yaml file as True and fill the GROQ_API_KEY field with the api key of [Groq](https://wow.groq.com/).
### LLM configuration (install on your machine)
You can set the configuration in ./backend/config/config.yaml. See [LLM configuration (use docker)](#llm-configuration-use-docker) for configuration explanations.
### Agent configuration
Currently, we support config agents by [role-prompting](https://arxiv.org/abs/2305.14688). You can customize your agents by changing the role prompts in AgentRepo\agentBoard_v1.json. We plan to support more methods to customize agents (e.g., supporting RAG, or providing a unified wrapper for customized agents) in the future.
## More Papers & Projects for LLM-based Multi-Agent Collaboration
If youre interested in LLM-based multi-agent collaboration and want more papers & projects for reference, you may check out the [corpus](https://docs.google.com/spreadsheets/d/1HSl4AqIVXhUjZh0pRhz-brzfA7evh20Q/edit?usp=sharing&ouid=112400145401551512954&rtpof=true&sd=true) collected by us. Any contribution to the corpus is also welcome.