66 lines
3.4 KiB
Markdown
66 lines
3.4 KiB
Markdown
# AgentCoord: Visually Exploring Coordination Strategy for LLM-based Multi-Agent Collaboration
|
||
<p align="center">
|
||
<a ><img src="static/icon.png" alt=" AgentCoord: Visually Exploring Coordination Strategy for LLM-based Multi-Agent Collaboration" width="200px"></a>
|
||
</p>
|
||
AgentCoord is an experimental open-source system to help general users design coordination strategies for multiple LLM-based agents.
|
||
[Research paper](https://arxiv.org/abs/2404.11943)
|
||
|
||
## System Usage
|
||
<a href="https://youtu.be/s56rHJx-eqY" target="_blank"><img src="static/videoPreview.png"
|
||
alt="System Usage Video" width="800" border="5" /></a>
|
||
|
||
## Installation
|
||
|
||
### Install with Docker (Recommended)
|
||
|
||
If you have installed [docker](https://www.docker.com/) and [docker-compose](https://docs.docker.com/compose/) on your machine, we recommend running AgentCoord in docker:
|
||
|
||
Step 1: Clone the project:
|
||
```bash
|
||
git clone https://github.com/AgentCoord/AgentCoord.git
|
||
cd AgentCoord
|
||
```
|
||
|
||
Step 2: Config LLM (see [LLM configuration (use docker)](README.md#llm-configuration-use-docker)):
|
||
|
||
Step 3: Start the servers
|
||
```bash
|
||
docker-compose up
|
||
```
|
||
|
||
Step 4: Open http://localhost:8080/ to use the system.
|
||
|
||
### Install on your machine
|
||
|
||
If you want to install and run AgentCoord on your machine without using docker:
|
||
|
||
Step 1: Clone the project
|
||
```bash
|
||
git clone https://github.com/AgentCoord/AgentCoord.git
|
||
cd AgentCoord
|
||
```
|
||
|
||
Step 2: Config LLM (see [LLM configuration (install on your machine)](README.md#llm-configuration-install-on-your-machine)):
|
||
|
||
Step 3: Install required packages, then run the backend and frontend servers separately (see readme.md for [frontend](frontend/README.md) and [backend](backend/README.md#Installation)
|
||
|
||
Step 4: Open http://localhost:8080/ to use the system.
|
||
|
||
## Configuration
|
||
### LLM configuration (use docker)
|
||
You can set the configuration (i.e. API base, API key, Model name) for default LLM in ./docker-compose.yml. Currently, we only support OpenAI’s LLMs as the default model. We recommend using gpt-4-turbo-preview as the default model (WARNING: the execution process of multiple agents may consume a significant number of tokens). You can switch to a fast mode that uses the Mistral 8×7B model with hardware acceleration by [Groq](https://groq.com/) for the first time in strategy generation to strike a balance of response quality and efficiency. To achieve this, you need to set the FAST_DESIGN_MODE field in the yaml file as True and fill the GROQ_API_KEY field with the api key of [Groq](https://wow.groq.com/).
|
||
|
||
### LLM configuration (install on your machine)
|
||
You can set the configuration in ./backend/config/config.yaml. See [LLM configuration (use docker)](#llm-configuration-use-docker) for configuration explanations.
|
||
|
||
### Agent configuration
|
||
Currently, we support config agents by [role-prompting](https://arxiv.org/abs/2305.14688). You can customize your agents by changing the role prompts in AgentRepo\agentBoard_v1.json. We plan to support more methods to customize agents (e.g., supporting RAG, or providing a unified wrapper for customized agents) in the future.
|
||
|
||
## More Papers & Projects for LLM-based Multi-Agent Collaboration
|
||
If you’re interested in LLM-based multi-agent collaboration and want more papers & projects for reference, you may check out the [corpus](https://docs.google.com/spreadsheets/d/1HSl4AqIVXhUjZh0pRhz-brzfA7evh20Q/edit?usp=sharing&ouid=112400145401551512954&rtpof=true&sd=true) collected by us. Any contribution to the corpus is also welcome.
|
||
|
||
|
||
|
||
|
||
|