Gradient Releases Echo-2 RL Framework, Boosting AI Research Efficiency by Over 10 Times

marsbitОпубликовано 2026-02-12Обновлено 2026-02-12

Введение

Gradient has released the Echo-2 distributed reinforcement learning framework (arxiv.org/pdf/2602.02192), designed to overcome efficiency barriers in AI research training. By decoupling Learners and Actors at the architectural level, Echo-2 reduces the post-training expense of a 30B model from $4,500 to just $425. Under the same budget, it delivers more than 10x improvement in research throughput. The framework uses compute-storage separation and asynchronous training (Async RL) to offload large-scale sampling to unreliable and heterogeneous GPU instances. It incorporates bounded staleness, fault-tolerant scheduling, and a custom Lattica communication protocol to maintain model accuracy while significantly boosting efficiency. Alongside the framework, Gradient is launching Logits, an RLaaS platform, to shift AI research paradigm from "capital-intensive" to "efficiency-driven". Logits is now open for global students and researchers for预约 (logits.dev).

Distributed AI lab Gradient today released the Echo-2 distributed reinforcement learning framework (arxiv.org/pdf/2602.02192), aiming to break through the efficiency barriers in AI research training. By achieving a complete decoupling of Learner and Actor at the architectural level, Echo-2 slashes the post-training cost of a 30B model from $4,500 to $425. Under the same budget, it delivers over a 10x increase in research throughput.

The framework utilizes compute-storage separation technology for asynchronous training (Async RL), offloading massive sampling computations to unstable GPU instances and heterogeneous GPUs based on Parallax. Combined with breakthroughs in bounded staleness, instance fault-tolerant scheduling, and the proprietary Lattica communication protocol, it significantly enhances training efficiency while ensuring model accuracy. Alongside the framework release, Gradient is also set to launch Logits, an RLaaS platform, to propel AI research from a "capital-intensive" paradigm to one of "efficiency iteration." Logits is now open for预约 (booking) to students and researchers worldwide (logits.dev).

About Gradient

Gradient is an AI lab dedicated to building distributed infrastructure, focusing on the distributed training, serving, and deployment of cutting-edge large models. Backed by top-tier investment institutions, Gradient is building an open and efficient future for the intelligent era.

Связанные с этим вопросы

QWhat is the main purpose of Gradient's newly released Echo-2 framework?

AThe main purpose of the Echo-2 distributed reinforcement learning framework is to break AI research training efficiency barriers by decoupling Learner and Actor at the architectural layer, significantly reducing costs and increasing research throughput.

QHow much does Echo-2 reduce the post-training cost for a 30B model, according to the article?

AEcho-2 reduces the post-training cost for a 30B model from $4,500 to $425.

QWhat key technology does Echo-2 use to achieve asynchronous training (Async RL)?

AEcho-2 uses a compute-storage separation technology to offload massive sampling compute to unstable GPU instances and heterogeneous GPUs based on Parallax for asynchronous training.

QWhat is the name of the RLaaS platform that Gradient is launching alongside the Echo-2 framework?

AThe RLaaS platform launched alongside Echo-2 is called Logits (logits.dev).

QWho is the primary target audience for the Logits platform, as mentioned in the article?

AThe Logits platform is now open for reservations to students and researchers globally.

Похожее

Stuck Polymarket: The Real Test After Riding the Traffic Boom Has Arrived

Polymarket, a leading prediction market platform, is facing significant technical challenges as its growth outpaces its current infrastructure on Polygon. Users are experiencing laggy transactions, unresponsive orders, and delayed confirmations, severely impacting the trading experience. In response, DeFi Engineering VP Josh Stevens outlined a comprehensive engineering overhaul. The plan includes reducing on-chain data delays, fixing order cancellation issues, rebuilding the central limit order book (CLOB), improving website performance, and developing a unified SDK and API. A major revelation was the ongoing "chain migration," indicating a potential move away from Polygon. The core issue is that Polymarket has evolved from a simple prediction market into a high-frequency trading platform, making Polygon's limitations—such as block space, gas fees, and block time—a ceiling for further growth. The migration is not just a simple chain switch but a fundamental rebuild of its trading system to support more complex products like perpetual contracts (Perps). This announcement has sparked competition among chains like Solana, Sui, and Algorand, all vying to host Polymarket. For Polygon, losing this key application, which contributes significantly to its gas fee revenue, would be a major setback. The real test for Polymarket is no longer attracting users but proving it can provide a stable, reliable trading environment that retains them.

Odaily星球日报1 ч. назад

Stuck Polymarket: The Real Test After Riding the Traffic Boom Has Arrived

Odaily星球日报1 ч. назад

Торговля

Спот
Фьючерсы

Популярные статьи

Неделя обучения по популярным токенам (2): 2026 может стать годом приложений реального времени, сектор AI продолжает оставаться в тренде

2025 год — год институциональных инвесторов, в будущем он будет доминировать в приложениях реального времени.

1.8k просмотров всегоОпубликовано 2025.12.16Обновлено 2025.12.16

Неделя обучения по популярным токенам (2): 2026 может стать годом приложений реального времени, сектор AI продолжает оставаться в тренде

Обсуждения

Добро пожаловать в Сообщество HTX. Здесь вы сможете быть в курсе последних новостей о развитии платформы и получить доступ к профессиональной аналитической информации о рынке. Мнения пользователей о цене на AI (AI) представлены ниже.

活动图片