Racing 22 Agents on Hyperliquid: What Did I Discover? (Full Strategy Code Included)

Odaily星球日报2026-03-16 tarihinde yayınlandı2026-03-16 tarihinde güncellendi

Özet

Summary: We deployed 22 autonomous AI trading agents on Hyperliquid, each with $1000, executing over 5000 trades. Key findings: 1) Less trading with higher conviction yields better results—agents with under 120 trades were profitable, while those with over 400 lost significantly. 2) Profits follow a power law: 3-5 trades generated nearly all gains, with small losses quickly stopped out. 3) Real-time "Hyperfeed" data (tracking where smart money is profiting) outperformed pure technical analysis. 4) Mean-reversion strategies failed in perpetual markets; trend-following was more effective. 5) Agents self-adjusted poorly when losing, often accelerating losses by increasing leverage or removing safeguards. Future steps include hardening risk controls in code and testing new strategies. All code is open-sourced, and experiments continue live.

This article is from:Jason Goldberg

Compiled | Odaily Planet Daily (@OdailyChina); Translator | Azuma (@azuma_eth)

We deployed 22 autonomous trading AI Agents on Hyperliquid via Senpi, each configured with $1000 in real capital.

They run 24/7 — scanning the market, opening positions, setting trailing stops, managing risk — all without any human intervention.

After investing $22,000 in initial capital and executing over 5000 trades, here is our summary of lessons learned.

General Conclusions

"Fewer trades" plus "higher conviction" always equals "better results." This isn't an occasional occurrence; it holds true every time.

  • Odaily Note: Fox, Bison, Ghost Fox, and Grizzly, Viper, Mamba, Anaconda mentioned later are all names of Agents executing different strategies.

As shown above, Agent "Fox" and Agent "Ghost Fox" use the same scanning tool. Fox only selectively executes some of the signals, while Ghost Fox executes more signals. The result is a 56 percentage point difference in their Return on Investment (ROI).

The real advantage lies not in the scanning tool itself, but in the discipline to wait for the right signal.

  • All Agents with over 400 trades suffered significant losses.
  • All Agents with fewer than 120 trades were profitable.

More trades do not mean more opportunities — they mean more ineffective trades, more fees, and more exposure to noise risk.

Profits Follow a "Power Law Distribution"

Among our best-performing Agents, 3–5 trades contributed to all the profits, while the remaining trades were mostly quickly stopped out with small losses.

Taking Fox as an example again, the three best trades (ZEC, TRUMP, FARTCOIN) collectively profited over $350; the other 46 trades collectively lost over $100; the final result was a net profit of approximately $248.

This is entirely by design. Our strategy is: enter decisively with high conviction, stop out losses quickly within minutes, let profitable positions run, and lock in some peak profits using the DSL High Water trailing stop strategy. When the average win is 10 times the average loss, you can make steady profits even with a win rate of only 43%.

Those Agents that tried to maintain a high win rate through "safe" trades all lost money — because each trade aiming for tiny profits still incurs fees and market risk.

Secret Weapon: Hyperfeed

Fox and other consistently performing Agents are built on Senpi's Hyperfeed.

Hyperfeed is a real-time tracking system that shows which assets all traders on Hyperliquid are currently making money on. It is not a historical ranking or other lagging indicator, but the actual profitable trading activity happening on the entire exchange right now.

Our core scanning tool, Emerging Movers, reads Hyperfeed's market concentration data every 90 seconds. When smart money suddenly rotates to an asset: for example, a trader suddenly jumps at least 15 places on the leaderboard, or the rate of profit contribution for an asset suddenly increases, or multiple top traders simultaneously concentrate on opening the same position, the scanner can capture the signal before the move is fully priced in.

This is the structural advantage of building strategies on Hyperliquid via Senpi. You can see in real-time where top traders' profits are concentrating and act immediately. No other exchange offers this visibility, and no other platform allows autonomous Agents to execute based on this.

All our best-performing Agents use this type of data:

  • Fox / Vixen: Identify sudden smart money concentration on an asset via Emerging Movers;
  • Grizzly: Analyze smart money positions on BTC via Hyperfeed before opening a position;
  • Bison: Uses smart money direction as a hard condition — will not trade if the direction is opposite;

While the worst-performing Agents:

  • Completely ignored smart money signals, e.g., Viper, Mamba based purely on technical analysis;
  • Used outdated smart money data (Scorpion v1), treating months-old positions as new signals;

So the conclusion is very clear: Agents trading based on real-time Hyperfeed data performed significantly better than all purely technical strategies.

Mean Reversion Strategies Don't Work in Perpetual Swaps

We tested three different versions of Agents based on the logic that "price deviates too far and will soon revert." Their specific performances were as follows:

  • Viper: -18%
  • Mamba: -33%
  • Anaconda: -22%

The result was losses across the board. The problem is that the Hyperliquid perpetual swap market is far more trend-driven than mean-reverting. Buying the dip in a downtrend is the most expensive mistake in this market. These Agents kept going long at so-called "support levels," but the price continued to trend lower for days.

The fix we are testing is adding a macro market state filter that prohibits "mean reversion strategy" dip buying when BTC's 4-hour trend is down. Preliminary results look promising; this filter could avoid 14 out of Mamba's 28 losing trades.

Don't Stick to a Single Pattern

Our newest Agent (Vixen), based on Fox's trading data, uses two截然不同的 (distinctly different) entry patterns.

  • Stalker Mode: Captures signals through multiple scans where smart money is quietly accumulating an asset. This allows you to enter before the crowd rushes in. Fox's biggest profits came from this mode.
  • Striker Mode: Captures violent breakout moves confirmed by volume. Enters as the move explodes, but only executes if supported by real volume (filtering fake pumps).

Fox's data shows these are two completely different sources of Alpha signals. If you only use a single entry pattern, you must choose between them and miss the other opportunity.

Agents Self-Adjust — And It Always Gets Worse

An unexpected finding was: When Agents experience consecutive losses, they attempt to "self-repair." Common repair behaviors include loosening entry conditions, increasing leverage, and removing risk protection mechanisms, but the result is always accelerated losses.

Some examples: Dire Wolf, after being down -27%, enabled 5 parallel 25x leverage positions and relaxed the order speed limit; another Agent deleted the stagnation take-profit mechanism; another raised the daily loss limit from 10% to 25%.

Our solution is to write risk protection mechanisms directly into the scanner tool code, rather than relying on the Agent's own strategy configuration. If the scanner tool doesn't output a signal, the Agent cannot execute a trade — no matter how aggressive its own configuration becomes.

Next Steps

We will continue the experiment for another 24-48 hours, then shut down Agents that have no realistic chance of recovering their capital to prevent further drain of remaining funds.

Next, we will deploy new strategy versions with protection mechanisms written into the code layer:

  • Wolverine v1.1: HYPE velocity DSL trailing stop (faster profit locking in high-volatility assets);
  • Mamba v2.0: Mean reversion strategy + BTC trend protection;
  • Scorpion v2.0: Real-time momentum event consensus (replacing outdated whale-following strategy).

We will also:

  • Standardize strategy configurations for Fox, Vixen, and Mantis: These three Agents use the same scanner tool, but their configurations have drifted. Fox's current yield is over 23%; the other two will be adjusted to the same settings;
  • Redeploy new Fox/Vixen pairs using Fox's full winning configuration, including XYZ ban rules, stagnation take-profit mechanism, 10% daily loss上限 (cap), all risk gating mechanisms enabled;
  • Expand single-asset hunter strategies: Grizzly's three-phase lifecycle model (Hunt → Ride → Lurk → Reload) is now applied to ETH (Polar), SOL (Kodiak), and HYPE (Wolverine).

Simultaneously, we are developing brand new strategies and testing them directly in the live market. The market itself is the laboratory. Each new strategy gets $1000 in capital and fully transparent trade records.

Our experiment runs live at strategies.senpi.ai; All strategy code is open source at: github.com/Senpi-ai/senpi-skills

22 Agents, $22,000 real capital, every trade completely public. The experiment continues.

İlgili Sorular

QWhat is the key finding regarding the relationship between the number of trades and profitability in the experiment?

AThe key finding is that fewer trades combined with higher conviction consistently lead to better results. All agents that executed over 400 trades experienced significant losses, while all agents with fewer than 120 trades were profitable. More trades meant more noise, fees, and exposure to risk, not more opportunity.

QWhat is Hyperfeed and why was it a critical component for the most successful agents?

AHyperfeed is a real-time tracking system that shows which assets traders on Hyperliquid are currently making money on at that exact moment. It was critical because it allowed the best-performing agents (like Fox and Grizzly) to identify and act on signals where 'smart money' was suddenly concentrating on an asset before the move was fully priced in, giving them a structural advantage.

QWhy did the mean reversion strategies (Viper, Mamba, Anaconda) perform poorly?

AThe mean reversion strategies all performed poorly (-18% to -33%) because the Hyperliquid perpetual market has stronger trending characteristics than mean-reverting ones. Attempting to buy the dip (go long at perceived 'support' levels) in a downtrend was a costly mistake, as prices could continue to fall for days.

QWhat unexpected and detrimental behavior did the agents exhibit when facing consecutive losses?

AWhen facing consecutive losses, the agents would attempt to 'self-repair' by making their configurations more aggressive. This included widening entry conditions, increasing leverage, and removing risk protection mechanisms like stop-losses. However, this behavior consistently accelerated their losses instead of helping them recover.

QHow do the profits of the successful agents follow a 'power law distribution'?

AThe profits followed a power law distribution, meaning a very small number of trades (3-5) contributed all of the profits, while the majority of trades ended as small losses that were quickly stopped out. For example, Agent Fox's three best trades made over $350, while its other 46 trades lost over $100, resulting in a net profit of ~$248.

İlgili Okumalar

Google and Amazon Simultaneously Invest Heavily in a Competitor: The Most Absurd Business Logic of the AI Era Is Becoming Reality

In a span of four days, Amazon announced an additional $25 billion investment, and Google pledged up to $40 billion—both direct competitors pouring over $65 billion into the same AI startup, Anthropic. Rather than a typical venture capital move, this signals the latest escalation in the cloud wars. The core of the deal is not equity but compute pre-orders: Anthropic must spend the majority of these funds on AWS and Google Cloud services and chips, effectively locking in massive future compute consumption. This reflects a shift in cloud market dynamics—enterprises now choose cloud providers based on which hosts the best AI models, not just price or stability. With OpenAI deeply tied to Microsoft, Anthropic’s Claude has become the only viable strategic asset for Google and Amazon to remain competitive. Anthropic’s annualized revenue has surged to $30 billion, and it is expanding into verticals like biotech, positioning itself as a cross-industry AI infrastructure layer. However, this funding comes with constraints: Anthropic’s independence is challenged as it balances two rival investors, its safety-first narrative faces pressure from regulatory scrutiny, and its path to IPO introduces new financial pressures. Globally, this accelerates a "tri-polar" closed-loop structure in AI infrastructure, with Microsoft-OpenAI, Google-Anthropic, and Amazon-Anthropic forming exclusive model-cloud alliances. In contrast, China’s landscape differs—investments like Alibaba and Tencent backing open-source model firm DeepSeek reflect a more decoupled approach, though closed-source models from major cloud providers still dominate. The $65 billion bet is ultimately about securing a seat at the table in an AI-defined future—where missing the model layer means losing the cloud war.

marsbit1 saat önce

Google and Amazon Simultaneously Invest Heavily in a Competitor: The Most Absurd Business Logic of the AI Era Is Becoming Reality

marsbit1 saat önce

Computing Power Constrained, Why Did DeepSeek-V4 Open Source?

DeepSeek-V4 has been released as a preview open-source model, featuring 1 million tokens of context length as a baseline capability—previously a premium feature locked behind enterprise paywalls by major overseas AI firms. The official announcement, however, openly acknowledges computational constraints, particularly limited service throughput for the high-end DeepSeek-V4-Pro version due to restricted high-end computing power. Rather than competing on pure scale, DeepSeek adopts a pragmatic approach that balances algorithmic innovation with hardware realities in China’s AI ecosystem. The V4-Pro model uses a highly sparse architecture with 1.6T total parameters but only activates 49B during inference. It performs strongly in agentic coding, knowledge-intensive tasks, and STEM reasoning, competing closely with top-tier closed models like Gemini Pro 3.1 and Claude Opus 4.6 in certain scenarios. A key strategic product is the Flash edition, with 284B total parameters but only 13B activated—making it cost-effective and accessible for mid- and low-tier hardware, including domestic AI chips from Huawei (Ascend), Cambricon, and Hygon. This design supports broader adoption across developers and SMEs while stimulating China's domestic semiconductor ecosystem. Despite facing talent outflow and intense competition in user traffic—with rivals like Doubao and Qianwen leading in monthly active users—DeepSeek has maintained technical momentum. The release also comes amid reports of a new funding round targeting a valuation exceeding $10 billion, potentially setting a new record in China’s LLM sector. Ultimately, DeepSeek-V4 represents a shift toward open yet realistic infrastructure development in the constrained compute landscape of Chinese AI, emphasizing engineering efficiency and domestic hardware compatibility over pure model scale.

marsbit1 saat önce

Computing Power Constrained, Why Did DeepSeek-V4 Open Source?

marsbit1 saat önce

İşlemler

Spot
Futures
活动图片