Zhejiang University Research Team Proposes New Approach: Teaching AI How the Human Brain Understands the World

marsbitОпубликовано 2026-04-05Обновлено 2026-04-05

Введение

A research team from Zhejiang University published a paper in *Nature Communications* challenging the prevailing notion that larger AI models inherently think more like humans. They found that while model performance on recognizing concrete concepts improved as parameters increased (from 74.94% to 85.87%), performance on abstract concept tasks slightly declined (from 54.37% to 52.82%) in models like SimCLR, CLIP, and DINOv2. The key difference lies in how concepts are organized. Humans naturally form hierarchical categories (e.g., grouping a swan and an owl into "birds"), enabling them to apply past knowledge to new situations. Models, however, rely heavily on statistical patterns in data and struggle to form stable, abstract categories. The team proposed a novel solution: using human brain signals (recorded when viewing images) to supervise and guide the model's internal organization of concepts. This method, termed transferring "human conceptual structures," helped the model learn a brain-like categorical system. In experiments, the model showed improved few-shot learning and generalization, with a 20.5% average improvement on a task requiring abstract categorization like distinguishing living vs. non-living things, even outperforming much larger models. This research shifts the focus from simply scaling model size ("bigger is better") to designing smarter internal structures ("structured is smarter"). It highlights a new pathway for developing AI that possesses more hum...

Large models have been growing in size, with the mainstream view being that the more parameters a model has, the closer it gets to human-like thinking. However, a paper published by a Zhejiang University team on April 1 in Nature Communications presents a different perspective (original article link: https://www.nature.com/articles/s41467-026-71267-5). They found that as model size (primarily SimCLR, CLIP, DINOv2) increases, the ability to recognize specific objects does continue to improve, but the ability to understand abstract concepts not only fails to improve but can even decline. When parameters increased from 22.06 million to 304.37 million, performance on concrete concept tasks rose from 74.94% to 85.87%, while performance on abstract concept tasks dropped from 54.37% to 52.82%.

Differences Between Human and Model Thinking

When the human brain processes concepts, it first forms a system of categorical relationships. Swans and owls look different, but humans still classify them both as birds. Moving up, birds and horses can be further grouped into the animal category. When humans encounter something new, they often first consider what it resembles from past experience and which category it might belong to. Humans continuously learn new concepts, then organize this experience, using this relational system to recognize new things and adapt to new situations.

Models also classify, but they form these classifications differently. They rely primarily on patterns that repeatedly appear in large-scale data. The more frequently a specific object appears, the easier it is for the model to recognize it. When it comes to larger categories, models struggle more. They need to capture the commonalities between multiple objects and then group these commonalities into the same category. Existing models still have significant shortcomings here. As parameters continue to increase, performance on concrete concept tasks improves, while performance on abstract concept tasks sometimes even decreases.

A commonality between the human brain and models is that both internally form a system of categorical relationships. However, their emphases differ. The higher-order visual regions of the human brain naturally distinguish broad categories like living and non-living things. Models can separate specific objects but find it difficult to stably form these larger categories. This difference means the human brain more easily applies past experience to new objects, allowing for rapid categorization of unseen things. Models, conversely, rely more on existing knowledge, so when encountering new objects, they tend to focus on superficial features. The method proposed in the paper addresses this characteristic, using brain signals to constrain the model's internal structure, making it closer to the human brain's categorization method.

The Solution from the Zhejiang University Team

The team's proposed solution is also unique: instead of simply adding more parameters, they use a small amount of brain signal data for supervision. These brain signals come from recordings of brain activity while humans view images. The original paper states the goal as transferring 'human conceptual structures' to DNNs. This means teaching the model, as much as possible, how the human brain classifies, generalizes, and groups similar concepts together.

The team conducted experiments using 150 known training categories and 50 unseen test categories. The results showed that as this training progressed, the distance between the model's representations and the brain representations continuously narrowed. This change occurred for both categories, indicating that the model was learning not just individual samples but truly beginning to learn a conceptual organization method more akin to the human brain.

After this process, the model demonstrated stronger few-shot learning capabilities and performed better in novel situations. In a task requiring the model to distinguish abstract concepts like living vs. non-living things with very few examples, the model improved by an average of 20.5%, even surpassing much larger control models. The team also conducted 31 additional specialized tests, where several types of models showed improvements of nearly ten percent.

Over the past few years, the familiar path in the modeling industry has been larger model scale. The Zhejiang University team has chosen a different direction: moving from 'bigger is better' to 'structured is smarter'. Scaling up is indeed useful, but it primarily improves performance on familiar tasks. Abstract understanding and transfer capabilities, inherent to humans, are equally crucial for AI. This requires future AI thinking structures to more closely resemble the human brain. The value of this direction lies in redirecting the industry's attention from pure size expansion back to the cognitive structure itself.

Neosoul and the Future

This points to a larger possibility: AI evolution may not only occur during the model training phase. Model training can determine how AI organizes concepts and forms higher-quality judgment structures. Then, after entering the real world, another layer of AI evolution just begins: how an AI agent's judgments are recorded, tested, and how they continuously grow and evolve through real-world competition, learning and evolving on their own, much like humans. This is precisely what Neosoul is doing now. Neosoul doesn't just have AI agents produce answers; it places AI agents into a system of continuous prediction, verification, settlement, and selection, allowing them to continuously optimize themselves based on predictions and outcomes, preserving better structures and淘汰ing worse ones. What the Zhejiang University team and Neosoul jointly point towards is actually the same goal: enabling AI to not just solve problems, but to possess comprehensive thinking abilities and continuously evolve.

Связанные с этим вопросы

QWhat was the key finding of the Zhejiang University team's research published in Nature Communications regarding model scaling?

AThey found that as model parameters increased (from 22.06 million to 304.37 million), performance on recognizing concrete concepts improved (74.94% to 85.87%), but performance on understanding abstract concepts not only failed to improve but actually decreased (54.37% to 52.82%).

QWhat is the fundamental difference between how the human brain and AI models form conceptual categories?

AThe human brain naturally forms a hierarchical classification system (e.g., grouping specific birds into the 'bird' category, then 'birds' and 'horses' into 'animals'). AI models primarily rely on statistical patterns from large-scale data, excelling at recognizing specific, frequently appearing objects but struggling to form stable, larger abstract categories.

QWhat was the unique solution proposed by the Zhejiang University team to improve AI's abstract reasoning?

AInstead of scaling model size, they used a small amount of brain signal data (recordings of human brain activity when viewing images) as supervision to transfer human conceptual structures to the deep neural networks (DNNs), teaching them how to classify and generalize concepts more like the human brain.

QWhat improvements were observed in the model after being trained with the brain signal supervision method?

AThe distance between the model's representations and brain representations decreased. The model showed stronger few-shot learning capabilities and performed better in novel situations. In a task requiring abstract concept discrimination with very few examples, performance improved by an average of 20.5%, even surpassing much larger control models.

QWhat broader shift in AI development philosophy does the research and Neosoul project represent, according to the article?

AIt represents a shift from the 'bigger is better' paradigm focused on scaling parameters to a 'structured is smarter' approach. The focus is on improving the AI's cognitive structure to be more human-like, enabling abstract understanding and transfer capabilities, and creating systems for continuous learning and evolution through real-world prediction, verification, and competition.

Похожее

This Week's Key Events Preview | U.S. to Release April CPI Data; U.S. Senate Banking Committee to Review "Digital Asset Market Structure Act of 2025"

Weekly News Preview: Key events for May 12-16 include major economic and crypto industry developments. On Tuesday, May 12, the U.S. will release its April CPI data. Additionally, the gaming blockchain Ronin will begin a 10-hour migration to an Ethereum Layer 2, built on OP Stack with EigenDA for data availability. This aims to leverage Ethereum's security and settle RON's annual inflation below 1%. Base's first independent network upgrade, "Base Azul," is scheduled for mainnet activation on Wednesday, May 13, focusing on security, performance, and developer experience enhancements. Thursday, May 14, sees the U.S. Senate Banking Committee voting on the "Digital Asset Market Structure Act of 2025." In other news, Solana DeFi protocol Carrot will shut down, setting a final withdrawal deadline due to impacts from the Drift exploit. The Moscow Exchange will launch futures trading for Solana, Ripple, and Tron indices (RUB-settled) for qualified investors. Multiple service closures are scheduled for Friday, May 15. Dmail Network will begin winding down due to unsustainable infrastructure costs and failed commercialization. Users must export data before this date. Separately, the Cosmos-based lending blockchain UX Chain will fully shut down. Finally, on Saturday, May 16, gaming infrastructure provider Lattice will wind down operations, with its Redstone Layer 2 network ceasing. Users are urged to withdraw assets, especially from contracts like Uniswap pools, before the shutdown.

链捕手49 мин. назад

This Week's Key Events Preview | U.S. to Release April CPI Data; U.S. Senate Banking Committee to Review "Digital Asset Market Structure Act of 2025"

链捕手49 мин. назад

Morning Post | Trump Media Group Releases Q1 Financial Report; Top Three DeFi Applications Return Nearly $100 Million in Revenue to Token Holders in 30 Days; Michael Saylor Shares Bitcoin Tracker Info Again

**Title: Daily Briefing | Trump Media Group Releases Q1 Report; Top 3 DeFi Apps Return Nearly $100M to Token Holders; Michael Saylor Signals Potential Bitcoin Buy** **Summary:** Key developments in the past 24 hours include: * **Economic Outlook:** Goldman Sachs has pushed back its forecast for the next two Federal Reserve interest rate cuts to December 2026 and March 2027, citing persistent inflationary pressures from energy costs. This delayed timeline is expected to tighten liquidity flow into risk assets, including cryptocurrencies. * **DeFi & Revenue:** Data from DefiLlama shows that three leading DeFi applications—Hyperliquid, Pump.fun, and EdgeX—collectively distributed $96.3 million in revenue to their token holders over the last 30 days. This trend highlights a shift in the crypto community's focus towards real protocol earnings and sustainable economic models. * **Corporate Bitcoin Moves:** Michael Saylor, founder of MicroStrategy (note: referred to as 'Strategy' in the text, likely a typographical error), has signaled potential upcoming Bitcoin purchases by posting a "Bitcoin Tracker" update, following a pattern that typically precedes the company's official disclosure of new acquisitions. * **Market Integrity:** Prediction market platform Polymarket announced updates to address platform issues, including identifying and banning clusters of accounts involved in "ghost-fill" activities and implementing measures to prevent bulk account creation. * **Regulation:** The Bank of England Governor warned that stablecoin regulation could lead to tensions between US and international regulators. In South Korea, the National Tax Service has launched a pilot program to entrust seized virtual assets to private custody firms for management. * **Meme Token Trends:** GMGN data lists the top trending meme tokens on Ethereum (e.g., HEX, SHIB), Solana (e.g., FWOG, TROLL), and Base (e.g., SKITTEN, PEPE) over the past day. **Financial Note:** Trump Media & Technology Group reported a Q1 loss of approximately $4 billion, primarily attributed to unrealized losses on its Bitcoin and other digital asset holdings.

链捕手1 ч. назад

Morning Post | Trump Media Group Releases Q1 Financial Report; Top Three DeFi Applications Return Nearly $100 Million in Revenue to Token Holders in 30 Days; Michael Saylor Shares Bitcoin Tracker Info Again

链捕手1 ч. назад

Telegram Takes Direct Control of TON, Social Traffic Rewrites the Public Chain Narrative

Telegram founder Pavel Durov announced that Telegram will replace the TON Foundation as the core driver and largest validator of The Open Network (TON). Key initiatives include a sixfold reduction in transaction fees, performance upgrades, and improved developer tools within the next few weeks. This marks a strategic shift from Telegram merely providing user access to deeply integrating TON into its platform's core infrastructure. The goal is to transform Telegram's massive social traffic into sustainable on-chain activity. While viral mini-apps like Notcoin have demonstrated Telegram's ability to drive user adoption, TON aims to support frequent, low-value transactions inherent to social platforms—such as tipping, in-app payments, and game rewards. Ultra-low fees and sub-second finality (0.6 seconds) are crucial to making blockchain interactions seamless and nearly invisible within the Telegram user experience. However, Telegram's increased central role raises questions about network decentralization. Durov argues that Telegram's participation will attract more large validators, thereby enhancing decentralization. TON also offers high annual staking rewards (18.8%), aiming to retain capital within its ecosystem. The fundamental challenge for TON is no longer leveraging Telegram's user base, but becoming an indispensable, seamless infrastructure layer for Telegram's everyday applications—moving from an adjacent chain to an embedded utility.

marsbit1 ч. назад

Telegram Takes Direct Control of TON, Social Traffic Rewrites the Public Chain Narrative

marsbit1 ч. назад

Telegram Takes Direct Control of TON, Social Traffic Reshapes Public Chain Narrative

Telegram's founder, Pavel Durov, has announced a major shift in the development of The Open Network (TON). Telegram will now become the core driver of TON, replacing the TON Foundation and becoming its largest validator. The focus will be on technical upgrades over the next few weeks, including slashing network fees by six times to near-zero and improving finality time to 0.6 seconds. This move signifies a deeper integration between Telegram and TON, moving beyond just providing a user base. The goal is to transform Telegram's vast social traffic and built-in features—like Mini Apps, payments, and bots—into sustainable, on-chain usage scenarios. The reduced fees and faster speeds are crucial for enabling the small, frequent transactions typical of social interactions. While this promises stronger execution and product alignment, it raises questions about centralization. Durov argues Telegram's involvement will attract more validators, enhancing decentralization, but the outcome remains to be seen. Additionally, TON's high annual staking reward of 18.8% aims to retain capital within the ecosystem. The key challenge for TON is no longer just leveraging Telegram's entry point, but becoming an invisible, seamless infrastructure layer within Telegram's daily use. Its success hinges on converting viral attention into lasting, embedded utility.

Odaily星球日报1 ч. назад

Telegram Takes Direct Control of TON, Social Traffic Reshapes Public Chain Narrative

Odaily星球日报1 ч. назад

Торговля

Спот
Фьючерсы
活动图片