Echo

The Distributed Reinforcement Learning Framework

circle-check
circle-check

As AI systems grow in scale and generality, RL has become a critical layer for alignment. However, most co-located RL frameworks face a deadlock: inference interrupts training, training delays inference.

Echo is Gradient’s distributed RL framework, running on everyday consumer devices.

Dual Swarm Architecture

Echo breaks the cycle by decoupling them into two independent swarms:

  • Inference on distributed devices (via Parallax)

  • Training on high-performance clusters

Echo Modes

We have open-sourced several early explorations with Echo, including:

  • A 7B math model beating Qwen2.5–32B

  • A 30B Sokoban model outperforming DeepSeek-R1 & GPT-OSS-120B

  • A 32B LoRA logic model ≥99% accuracy

See all models in our Hugging Face Collectionarrow-up-right.

Learn More

Read the blogarrow-up-right. Read the full research paperarrow-up-right.

Last updated