diff --git a/README.md b/README.md index c2aaa0d2..176f51e0 100644 --- a/README.md +++ b/README.md @@ -5,11 +5,12 @@ RL Swarm is a peer-to-peer system for reinforcement learning. It allows you to t Currently, we are running the [reasoning-gym](https://github.com/open-thought/reasoning-gym/tree/main) swarm on the Testnet. This swarm is designed to train models to solve a diverse set of reasoning tasks using the reasoning-gym dataset. The current list of default models includes: Models: - - Gensyn/Qwen2.5-0.5B-Instruct - - Qwen/Qwen3-0.6B - - nvidia/AceInstruct-1.5B - - dnotitia/Smoothie-Qwen3-1.7B - - Gensyn/Qwen2.5-1.5B-Instruct + +| Your RAM | Recommended Model | Download Size | +|----------|------------------|---------------| +| 4-8GB | Qwen2.5-0.5B | ~1GB | +| 8-16GB | Qwen2.5-1.5B | ~4GB | +| 16GB+ | Qwen2.5-7B+ | ~15GB+ | This iteration of rl-swarm is powered by the [GenRL](https://github.com/gensyn-ai/genrl) library. It is a fully composable framework for decentralized reinforcement learning which enables users to create and customize their own swarms for reinforcement learning with multi-agent multi-stage environments. @@ -82,7 +83,7 @@ If `docker-compose` does not work when running the above commands, please try `d ### Experimental (advanced) mode -If you want to experiment with the [GenRL](https://github.com/gensyn-ai/genrl) library or the[configurable parameters](https://github.com/gensyn-ai/rl-swarm/blob/main/rgym_exp/config/rg-swarm.yaml ), we recommend you run RL Swarm via shell script: +If you want to experiment with the [GenRL](https://github.com/gensyn-ai/genrl) library or the [configurable parameters](https://github.com/gensyn-ai/rl-swarm/blob/main/rgym_exp/config/rg-swarm.yaml ), we recommend you run RL Swarm via shell script: ```sh python3 -m venv .venv source .venv/bin/activate