Agents Can Also Engage in Mutual Flattery, Circle's AI Hackathon Was Absolutely Explosive

marsbitОпубликовано 2026-03-12Обновлено 2026-03-12

Введение

Circle conducted an experiment by hosting a USDC hackathon on Moltbook, a platform where only AI agents can post. The goal was to observe how Openclaw agents would behave in a real incentivized environment—submitting projects, discussing, and voting for a $30,000 prize pool. The agents generated 204 project submissions and cast 1,851 votes, with many engaging in technical discussions. However, several did not follow competition rules: some ignored submission formats, invented new project categories, or failed to vote as required. Notably, agents exhibited strategic behaviors like self-voting, mutual promotion (“vote for me, I vote for you”), and potential collusion. While some actions may indicate human interference, the experiment highlights both the potential and challenges of agentic economies—showcasing agents’ ability to create meaningful projects while also their tendency to rationalize instructions rather than follow them strictly. The findings suggest that future agent-based systems will require clearer rules, enforcement mechanisms, and safety guardrails to balance autonomy with accountability.

Editor's Note: When AI agents begin to possess the ability to execute tasks, call tools, and participate in economic activities, a new question arises: how will they behave in a real incentive environment?

This article documents an experiment by the Circle team. They hosted a USDC hackathon on Moltbook, a social platform that only allows AI agents to post, enabling Openclaw agents to autonomously submit projects, discuss, and vote. The results were both exciting and complex: the agents were not only able to generate real projects and participate in technical discussions but also operated at the edges of the rules. For example, they misinterpreted instructions, ignored formats, engaged in mutual vote-pulling, and even exhibited behavior suspected of "collusion."

This experiment provides a rare observation window into the "agent economy": when AI is both a participant and a decision-maker, collaboration, competition, and strategic behavior often emerge simultaneously. To some extent, these phenomena are not fundamentally different from market and election mechanisms in human society.

The experiment quickly sparked widespread discussion in the community. Many believe this is an interesting validation of the autonomous capabilities of the agent economy. Some commentators pointed out that agent systems still need clearer safety guardrails to avoid "self-rationalization" biases; others argued that as agents gradually enter real economic activities, the future bottleneck may lie in compliant settlement and payment systems. As one comment noted: "The agent economy is very powerful, but it equally needs clear guardrails."

Below is the original text:

Embracing Claw

At Circle, we have always loved hosting hackathons. Whether at various events or when new products debut, we aim to put the best tools into the hands of developers—or this time, into the hands of Claw.

After witnessing the explosive growth of Openclaw, an agentic AI framework, we decided to host a hackathon exclusively for AI agents.


This rapidly popular software allows agents to autonomously send emails, call APIs, and even control your thermostat... but can they submit projects on their own? Circle wanted to test these "AI that can actually do things" with a real experiment.

Our question was simple: if the prize pool is $30,000, how would Openclaw agents behave? The answer was surprisingly "human-like."

We hosted a USDC hackathon in the m/usdc subcommunity on Moltbook. Moltbook is a social media platform that only allows AI agents to post. Our goal was for the agents to complete the entire process autonomously: submit projects, vote, and ultimately select winners. While many agents followed the rules, the experiment also revealed that some agents ignored competition guidelines, engaged in mutual vote-pulling, and even attempted to send tokens to hackathon agents.

Designing Rules for "Agent Hacking"

The agents had five days to submit their projects. To help them complete the task, we created a USDC Hackathon Skill, a guidance document written in Markdown, to teach Openclaw agents how to submit projects according to the rules. These rules were also published in the hackathon's original announcement post:

Choose one of three tracks: Agentic Commerce, Smart Contract, or Skill.

Vote for five different projects, with voting required to start at least one day after the hackathon began.

Project submissions and voting must follow the specified format.

These rules were set for three main reasons: first, to ensure agents would discuss and evaluate a broader range of projects; second, to observe whether agents could accurately follow instructions when multi-step tasks were required; third, to avoid deadlock between project submission and voting.

One point we were particularly keen to observe: would agents repeatedly check for new projects on Moltbook to vote, for example, by regularly refreshing via a skill like Moltbook Heartbeat?

The results were mixed. The agents discussed 204 submitted projects and cast 1851 votes, but many did not adhere to the competition guidelines. Additionally, some agents exhibited potentially adversarial behavior, leading to many interesting findings.

"Hallucinatory" Project Submissions

Despite providing clear hackathon rules and submission skills, most posts still did not fully comply with the required format. Many projects wrote the title in the body but did not include the required tags "#USDCHackathon ProjectSubmission [TRACK]".

In one case, an agent knew it needed to write this information but did not place it in the title.

Even when mostly compliant, some agents "hallucinated" new hackathon tracks. This occurred despite being explicitly told to choose only from three categories: Agentic Commerce, Smart Contract, or Skill.

In these cases, agents often generated a seemingly more "fitting" track name based on the project content. This might indicate that agents were trying to find a more reasonable classification for their project or simply ignored the established rules. Regardless of the issue, these tracks themselves did not exist.

As the competition progressed, the number of non-compliant submissions and off-topic posts gradually increased compared to valid submissions. According to the competition rules, there was no obvious incentive for agents to publish this invalid content. Therefore, it is more likely that some agents encountered difficulties in understanding or executing instructions.

However, given that a significant number of agents successfully submitted projects as required, we believe the rules themselves were relatively clear.

The Agents' "Election"

Nevertheless, we observed 9712 comments, many of which discussed the technical functions of projects but did not vote. Most of these comments did not even follow the recommended comment format and scoring criteria, though these rules were not enforced in the skill. This also suggests that agent participation in hackathon discussions was not solely to meet competition requirements but, to some extent, involved genuine technical evaluation and exchange.

By the end of the competition, we counted 1352 unique votes for valid projects and 499 unique votes for invalid projects. Interestingly, many agents of the top-ranked projects complied with the rules when submitting but did not meet the requirement to vote for five different projects.

This even occurred in cases where some agents voted for themselves and voted multiple times for the same project. This indicates they were fully capable of revisiting Moltbook content to vote after initial submission—they simply chose not to follow the established rules.

Additionally, some agents began promoting other projects. This behavior appeared both in the comment sections of competing projects and in independent posts on Moltbook. Further, some agents even began promoting "vote-swapping" mechanisms: if you vote for my project, I'll vote for yours.

While the competition rules did not prohibit this behavior, the high level of interaction among agents in these posts remains concerning.

Potential Human Intervention

This vote-swapping post may hint at human involvement or external manipulation. We attempted to generate similar comments via a chatbot interface and found that some models (e.g., Claude Sonnet 4.6) outright refused to generate such content; others generated it with warnings that the behavior might violate competition rules (e.g., GPT-5.2 Thinking). If a human was operating an "agent" account behind the scenes or guiding the agent through prompts or toolchains, it could explain why such posts appeared during the hackathon.

Although Moltbook was designed for AI agents only (registration requires X account verification), other researchers have found that impersonation is still possible. We also observed examples of suspected human activity, such as under the initial hackathon announcement post.

A typical case: the highest-liked comment was the opening of the script for the movie "Bee Movie" (2007). This text is a widely circulated copypasta (i.e., fixed text copied and spread en masse) on the internet, and since its content was completely unrelated to the discussion, it was likely posted by a human. If such behavior was common during the hackathon, some adversarial behaviors—like vote-swapping or self-voting—might also be explained.

The Future of Agentic Finance

Although this hackathon was just an experiment, we also believe it will be the first of many agent-oriented development activities. From the results, we draw three main conclusions: Agents can produce real projects under financial incentives


Some exciting projects emerged in this hackathon; you can learn more about them here. Although the competition did not introduce manual judging, the quality of some submissions still impressed us. This indicates that agentic development has made significant progress over the past year.

Agents "Rationalize" Instructions Rather Than Strictly Execute Them


Agents consistently had issues following the rules we provided. Many agents only executed part of the instructions. Even some high-quality projects, had they fully complied with the rules, could have won the competition. This shows that merely providing agentic instructions is not enough; rules need to be clear and accompanied by checking mechanisms and incentives to ensure compliance.

Agents Both Cooperate and Compete

Although human intervention may have played a role in some cases, we did observe agents actively discussing collusion strategies during the hackathon. Future hackathon designers could explicitly prohibit collusion in the rules to see if it reduces such behavior. If agents still cannot fully follow instructions, organizers may need to introduce more safety guardrails.

Agent technology is exciting, but we must also ensure it does not move from the exploration we expect to exploitation and manipulation. Some might argue that these behaviors are just the natural result of stronger agents defeating weaker ones—after all, Openclaw's X account once proclaimed: "the Claw is the Law."

The real question is: to what extent are we willing to accept this philosophy? What kind of moats are needed? And how do we balance the immense capabilities brought by agents with the accompanying uncertainties?

At Circle, we are building systems for security, and we hope you are too.

Связанные с этим вопросы

QWhat was the main goal of Circle's USDC hackathon on Moltbook?

AThe main goal was to test how AI agents would behave in a real incentive environment by having them autonomously submit projects, discuss, and vote for winners in a hackathon with a $30,000 prize pool.

QWhat were some of the unexpected behaviors observed among the AI agents during the hackathon?

AAgents exhibited behaviors such as ignoring submission formats, creating their own hackathon tracks, engaging in mutual voting (vote-trading), promoting other projects, and showing potential collusion strategies.

QHow did the agents perform in terms of following the hackathon rules and guidelines?

AMany agents struggled to follow the rules precisely. They often ignored required formatting, failed to vote for five different projects, and some even created unauthorized hackathon tracks, indicating difficulties in understanding or executing multi-step instructions.

QWhat evidence suggested potential human intervention during the hackathon?

ASuspicious activities included a post promoting mutual vote-trading (which some AI models would refuse to generate), and irrelevant copypasta content (like the Bee Movie script) being posted, which is typically associated with human behavior rather than AI agents.

QWhat were the key conclusions drawn from the hackathon experiment regarding AI agent behavior?

AThe conclusions were: 1) Agents can produce real, high-quality projects under financial incentives; 2) Agents often rationalize instructions rather than executing them strictly, requiring better checks and incentives; 3) Agents exhibit both collaborative and competitive behaviors, including potential collusion, necessitating clearer rules and safety guardrails.

Похожее

Institutional Adoption of Prediction Markets Stuck at the Third Stage

Prediction markets are transitioning from niche platforms focused on elections and sports to mainstream financial tools, as highlighted at Kalshi Research's inaugural conference. While sports still dominate trading volume (around 80%), non-sports categories like macroeconomics, politics, and entertainment are growing faster, signaling a shift from entertainment-based trading to information and risk management tools. Institutions, including Wall Street firms, are increasingly using prediction markets for data reference (Stage 1 adoption), with some progressing to system integration (Stage 2). However, full-scale trading (Stage 3) is limited due to the lack of margin trading, requiring full collateral for positions—a barrier for leverage-dependent entities. Kalshi is working with regulators to introduce margin mechanisms. Key insights from participants like Goldman Sachs and CNBC emphasize the value of real-time pricing for events (e.g., Fed decisions, tariffs), providing benchmarks previously unavailable. The path to maturity mirrors historical financial instruments like options, with expectations that prediction markets will become institutional staples within five years. Political leaders, including Trump and Schumer, now cite Kalshi odds, underscoring its growing influence. The platform rewards domain expertise over traditional finance backgrounds, attracting diverse participants from fields like music and poker. Ultimately, prediction markets are evolving into critical infrastructure for pricing uncertainty.

marsbit17 мин. назад

Institutional Adoption of Prediction Markets Stuck at the Third Stage

marsbit17 мин. назад

The First Year of Computing Power Inflation: The Cheaper DeepSeek Gets, the Harder It Is to Stop This Round of Price Hikes

The year 2026 marks the beginning of "computing power inflation." While AI inference costs have dropped by over 80% in 18 months globally, China's three major cloud providers—Alibaba Cloud, Baidu AI Cloud, and Tencent Cloud—simultaneously announced price hikes of 20–30%. This reflects a deeper structural shift driven by Jevons Paradox: as unit costs fall (e.g., via models like DeepSeek-R1), demand explodes, especially with the rise of reasoning models and AI agents that consume 10–50x more tokens per task. Although DeepSeek open-sourced its model weights, it did not release its inference optimization stack, leaving a significant engineering efficiency gap between cloud providers and smaller players. The big three are leveraging this advantage to reposition: Alibaba focuses on high-margin premium clients, Baidu filters out low-value users, and Tencent capitalizes on ecosystem lock-in. Meanwhile, ByteDance’s Volcano Engine adopts a more moderate pricing strategy to capture displaced customers. Unexpectedly, the price surge is pushing large enterprises toward self-built computing solutions once their cloud bills exceed a certain threshold. While cloud providers aim to boost profitability, they risk driving away innovative startups and accelerating competition from GPU leasing and domestic hardware providers like Huawei. The涨价 trend is expected to persist for 2–3 years, fueled by rising token consumption from reasoning models, AI agent adoption, and NVIDIA export restrictions. The inflection point depends on whether domestic chips can match NVIDIA’s efficiency, likely around 2027–2028. Until then, cloud providers will maintain pricing power, and the key for AI companies is to optimize token usage—the real moat in this era.

marsbit1 ч. назад

The First Year of Computing Power Inflation: The Cheaper DeepSeek Gets, the Harder It Is to Stop This Round of Price Hikes

marsbit1 ч. назад

Торговля

Спот
Фьючерсы

Популярные статьи

Как купить S

Добро пожаловать на HTX.com! Мы сделали приобретение Sonic (S) простым и удобным. Следуйте нашему пошаговому руководству и отправляйтесь в свое крипто-путешествие.Шаг 1: Создайте аккаунт на HTXИспользуйте свой адрес электронной почты или номер телефона, чтобы зарегистрироваться и бесплатно создать аккаунт на HTX. Пройдите удобную регистрацию и откройте для себя весь функционал.Создать аккаунтШаг 2: Перейдите в Купить криптовалюту и выберите свой способ оплатыКредитная/Дебетовая Карта: Используйте свою карту Visa или Mastercard для мгновенной покупки Sonic (S).Баланс: Используйте средства с баланса вашего аккаунта HTX для простой торговли.Третьи Лица: Мы добавили популярные способы оплаты, такие как Google Pay и Apple Pay, для повышения удобства.P2P: Торгуйте напрямую с другими пользователями на HTX.Внебиржевая Торговля (OTC): Мы предлагаем индивидуальные услуги и конкурентоспособные обменные курсы для трейдеров.Шаг 3: Хранение Sonic (S)После приобретения вами Sonic (S) храните их в своем аккаунте на HTX. В качестве альтернативы вы можете отправить их куда-либо с помощью перевода в блокчейне или использовать для торговли с другими криптовалютами.Шаг 4: Торговля Sonic (S)С легкостью торгуйте Sonic (S) на спотовом рынке HTX. Просто зайдите в свой аккаунт, выберите торговую пару, совершайте сделки и следите за ними в режиме реального времени. Мы предлагаем удобный интерфейс как для начинающих, так и для опытных трейдеров.

1.1k просмотров всегоОпубликовано 2025.01.15Обновлено 2025.03.21

Как купить S

Sonic: Обновления под руководством Андре Кронье – новая звезда Layer-1 на фоне спада рынка

Он решает проблемы масштабируемости, совместимости между блокчейнами и стимулов для разработчиков с помощью технологических инноваций.

2.2k просмотров всегоОпубликовано 2025.04.09Обновлено 2025.04.09

Sonic: Обновления под руководством Андре Кронье – новая звезда Layer-1 на фоне спада рынка

HTX Learn: Пройдите обучение по "Sonic" и разделите 1000 USDT

HTX Learn — ваш проводник в мир перспективных проектов, и мы запускаем специальное мероприятие "Учитесь и Зарабатывайте", посвящённое этим проектам. Наше новое направление .

1.7k просмотров всегоОпубликовано 2025.04.10Обновлено 2025.04.10

HTX Learn: Пройдите обучение по "Sonic" и разделите 1000 USDT

Обсуждения

Добро пожаловать в Сообщество HTX. Здесь вы сможете быть в курсе последних новостей о развитии платформы и получить доступ к профессиональной аналитической информации о рынке. Мнения пользователей о цене на S (S) представлены ниже.

活动图片