I Dropped Out of High School, Learned with AI, and Made a Comeback as an OpenAI Researcher

深潮Publicado em 2025-12-17Última atualização em 2025-12-17

Resumo

The article tells the story of Gabriel Petersson, who dropped out of high school in Sweden and eventually became a research scientist at OpenAI working on the Sora video project. He achieved this through a self-directed, AI-powered learning method he calls "recursive knowledge filling." Instead of following a traditional "bottom-up" educational path, he starts with a concrete project and uses AI to deeply understand each component through relentless questioning. He treats AI not as a tool to generate answers, but as an infinitely patient tutor. For example, to learn about diffusion models, he began by asking an AI for the core concepts, then had it generate code. He then interrogated every part of that code, asking "why" and "how" until he built an intuitive understanding from the top down. This method allows him to rapidly acquire the essence of a subject in days rather than years. The author contrasts this with how most people use AI, which often leads to a decline in their own critical thinking and skills, as evidenced by research. The key difference is mindset: using AI as a "co-pilot for thinking" rather than an "answer generator." The article concludes with a five-step framework for applying this method to learn any subject and suggests that this approach could lead to a future of more "one-person companies" where individuals use AI to master multiple disciplines. The core advice is to never stop at the first answer—to keep asking questions.

Author: Jin Guanghao

A while ago, I attended an AI gathering in Shanghai.

The event itself covered a lot about AI implementation.

But what impressed me the most was a learning method shared by an investor veteran.

He said this method saved him and changed his criteria for evaluating people in investments.

What exactly is it? It's learning to "ask questions."

When you're interested in a question, go talk to DeepSeek, keep talking until it can't answer anymore.

This technique of "infinite questioning" was quite shocking to hear at the time, but after the event, I put it aside.

I didn't try it, nor did I think about it again.

Until recently, when I came across the story of Gabriel Petersson dropping out of school, using AI to learn, and joining OpenAI.

It suddenly dawned on me what that veteran meant by "asking to the end" in this AI era.

Gabriel's Interview Podcast | Image: YouTube

01 "High School Dropout" Makes a Comeback as OpenAI Researcher

Gabriel is from Sweden and dropped out of high school.

Gabriel's Social Media Profile | Image: X

He used to think he was too dumb to ever work in AI.

The turning point came a few years ago.

His cousin started a startup in Stockholm, working on an e-commerce product recommendation system, and asked him to help.

So Gabriel went. With little technical background and no savings, he even slept on the couch in the company's common lounge for a whole year during the early startup phase.

But he learned a lot that year. Not in school, but forced by the pressure of real problems: programming, sales, system integration.

Later, to optimize his learning efficiency, he switched to being a contractor, allowing him more flexibility in choosing projects, specifically seeking out the best engineers to collaborate with and actively seeking feedback.

When applying for a U.S. visa, he faced an awkward problem: this type of visa requires proof of "extraordinary ability" in the field, typically needing academic publications, paper citations, etc.

How could a high school dropout have these?

Gabriel came up with a solution: he compiled his high-quality technical posts from programmer communities as a substitute for "academic contributions." This proposal was actually approved by immigration.

After arriving in San Francisco, he continued to use ChatGPT to self-study math and machine learning.

Now he is a research scientist at OpenAI, working on building the Sora video model.

At this point, you must be curious, how did he do it?

Gabriel's Viewpoint | Image: X

02 Recursive Knowledge Filling: A Counterintuitive Learning Method

The answer is "infinite questioning"—find a specific problem and solve it thoroughly with AI.

Gabriel's learning method is contrary to most people's intuition.

The traditional learning path is "bottom-up": build the foundation first, then learn applications. For example, to learn machine learning, you first need linear algebra, probability, calculus, then statistical learning, then deep learning, and finally you can touch actual projects. This process might take years.

His method is "top-down": start directly with a specific project, solve problems as they arise, and fill knowledge gaps when discovered.

He said in the podcast that this method was hard to promote before because you needed an omniscient teacher to always tell you "what to learn next."

But now, ChatGPT is that teacher.

Gabriel's Viewpoint | Image: X

How exactly? He gave an example: how to learn diffusion models.

Step one, start with the macro concept. He asks ChatGPT: "I want to learn about video models, what's the core concept?" AI tells him: autoencoders.

Step two, code first. He asks ChatGPT to directly write a piece of diffusion model code. At first, many parts are incomprehensible, but that's okay, just get the code running first. Once it runs, you have a basis for debugging.

Step three, the core: recursive questioning. He asks questions about every module in the code.

Drill down layer by layer like this until the underlying logic is thoroughly understood. Then return to the previous layer and ask about the next module.

He calls this process "recursive knowledge filling."

Recursive Knowledge Filling | Image: nanobaba2

This is much faster than studying step-by-step for six years; it might take just three days to build basic intuition.

If you're familiar with the Socratic method, you'll find this is essentially the same idea: approaching the essence of things through layered questioning, where every answer is the starting point for the next question.

Only now, he uses AI as the person being questioned, and since AI is nearly omniscient, it continuously expresses the essence of things in an easy-to-understand way to the questioner.

In fact, by adopting this approach, Gabriel performs "knowledge extraction" from AI, learning the essence of things.

03 Most of Us Using AI Are Actually Getting Dumber

After listening to the podcast, Gabriel's story made me wonder:

Why can he learn so well using AI, while many people feel they are regressing after using AI?

This isn't just my subjective feeling.

A 2025 Microsoft Research paper shows [1] that when people frequently use generative AI, their use of critical thinking significantly decreases.

In other words, we outsource our thinking to AI, and then our own thinking ability atrophies.

Skills follow the "use it or lose it" rule: when we use AI to write code, our ability to code by hand and brain quietly deteriorates.

The working style of "vibe coding" with AI seems efficient, but in the long run, the programmer's own skills decline.

You throw requirements to AI, it spits out a bunch of code, you run it, feel great. But if you turn off AI and try to write the core logic by hand, many find their minds blank.

A more extreme case comes from the medical field. A medical paper points out [2] that doctors' colonoscopy detection skills decreased by 6% three months after introducing AI assistance.

This number seems small, but think: this is real clinical diagnostic ability, concerning patients' health and lives.

So the question is: the same tool, why do some get stronger and some weaker?

The difference lies in what you treat AI as.

If you treat AI as a tool to do work for you, letting it write code, write articles, make decisions for you, then your ability will indeed degrade. Because you skip the thinking process and only get the result. Results can be copied and pasted, but thinking ability doesn't grow out of thin air.

But if you treat AI as a coach or mentor, using it to test your understanding, question your blind spots, force yourself to clarify vague concepts: then you are actually using AI to accelerate your learning cycle.

The core of Gabriel's method is not "let AI learn for me" but "let AI learn with me." He is always the one actively asking questions; AI just provides feedback and material. Every "why" is asked by himself, every layer of understanding is dug by himself.

This reminds me of an old saying: Give a man a fish, and you feed him for a day; teach a man to fish, and you feed him for a lifetime.

Recursive Knowledge Filling | Image: nanobaba2

04 Some Practical Insights

At this point, some might ask: I'm not in AI research, nor a programmer, what use is this method to me?

I think Gabriel's methodology can be abstracted into a more universal five-step framework. Everyone can use AI to learn any unfamiliar field.

1. Start from actual problems, not from chapter one of a textbook.

Whatever you want to learn, just start doing it, and fill in the gaps when you get stuck.

Knowledge learned this way has context and purpose, much more effective than memorizing concepts in isolation.

Gabriel's Viewpoint | Image: X

2. Treat AI as an infinitely patient tutor.

You can ask it any stupid question, have it explain the same concept in different ways, ask it to "explain like I'm five."

It won't laugh at you, nor will it get impatient.

3. Actively question until you build intuition. Don't settle for superficial understanding.

Can you restate a concept in your own words? Give an example not mentioned in the original text?

Explain it to a layperson? If not, keep asking.

4. There's a trap to be wary of: AI can also hallucinate.

During recursive questioning, if the AI explains a underlying concept wrong, you might go further down the wrong path.

So it's advised to cross-validate with multiple AIs at key nodes, ensuring the foundation of your questioning is solid.

5. Record your questioning process.

This forms reusable knowledge assets: next time you encounter a similar problem, you have a complete thought process to review.

In traditional观念, the value of tools lies in reducing resistance and improving efficiency.

But learning is恰恰相反: moderate resistance, necessary friction, is反而 the prerequisite for learning. If everything is too smooth, the brain enters power-saving mode and remembers nothing.

Gabriel's recursive questioning本质上 creates friction.

He constantly asks why, constantly pushes himself to the edge of not understanding, then slowly fills the gaps.

This process is uncomfortable, but it's this discomfort that truly embeds knowledge into long-term memory.

05 Future Career Trends

In this era, the monopoly of academic qualifications is being broken, but the cognitive threshold is invisibly rising.

Most people only use AI as an "answer generator," while a very few, like Gabriel, use AI as a "thinking exerciser."

Actually, similar uses have already appeared in different fields.

For example, on Jike, I see many parents using nanobanana to tutor their children. But they don't let AI give the answer directly; instead, they have AI generate解题 steps, showing the thinking process step by step, then analyze the logic of each step with the child.

This way, the child learns not the answer, but the method of solving the problem.

Prompt "Solve the given integral and write the complete solution on a whiteboard" | Image: nanobaba2

Others use features like Listenhub or NotebookLM to convert long articles or papers into podcast form, with two AI voices discussing, explaining, and questioning. Some think this is lazy, but others find that listening to the conversation and then going back to the original text actually improves comprehension efficiency.

Because the conversation naturally raises questions, forcing you to think: Do I really understand this point?

Gabriel's Interview Podcast to Podcast | Image: notebooklm

This points to a future career trend: specialized and versatile.

Before, to create a product, you needed to know front-end, back-end, design, operations, marketing. Now, you can, like Gabriel, use the "recursive gap-filling" method to quickly master 80% of the knowledge in your weak areas.

You were originally a programmer; by using AI to补全 design and business logic, you can become a product manager.

You were originally a good content creator; through AI, you can quickly补全 coding skills and become an independent developer.

Based on this trend, it can be inferred: "Perhaps, in the future, more 'one-person company' forms will appear."

06 Reclaim Your Initiative

Now, thinking back to that investor veteran's words, I understand what he really meant.

"Keep asking until it can't answer."

This phrase is a great mindset in the AI era.

If we are satisfied with the first answer AI gives, we are quietly degenerating.

But if we can, through questioning, force AI to explain the logic thoroughly, then internalize it into our own intuition: then AI truly becomes our external tool, not us becoming appendages of AI.

Don't let ChatGPT think for you; let it think with you.

Gabriel went from a dropout sleeping on a couch to an OpenAI researcher.

There's no secret in between, just tens of thousands of questions.

In this era full of anxiety about being replaced by AI, the most practical weapon might be:

Don't stop at the first answer, keep asking.

Perguntas relacionadas

QWhat is the core learning method Gabriel Petersson used to become an OpenAI researcher without a high school diploma?

AGabriel Petersson used a 'recursive knowledge filling' method, which involves starting with a practical project, using AI to ask infinite questions, and drilling down into each concept until he fully understood it, rather than following a traditional bottom-up learning path.

QHow does the 'recursive knowledge filling' method work in practice when learning a new concept like diffusion models?

AHe starts by asking AI for the core concepts, then gets code examples to run, and recursively questions every module in the code until he understands the underlying logic, filling knowledge gaps layer by layer.

QWhat negative effect can frequent use of generative AI have on human cognitive abilities, according to the article?

AFrequent use of generative AI can lead to a decline in critical thinking skills, as people outsource their thinking to AI, causing their own abilities to atrophy due to disuse.

QWhat is the key difference between using AI as a tool for outsourcing work versus using it as a learning coach?

AUsing AI as a tool for outsourcing work leads to skill degradation because it bypasses the thinking process, while using it as a coach involves active questioning and probing, which accelerates learning and deepens understanding.

QWhat future career trend does the article suggest is emerging due to AI-enabled learning methods like Gabriel's?

AThe article suggests a trend towards 'one-person companies,' where individuals can use AI to quickly master multiple domains, such as a programmer learning design and business logic to become a product manager or a content creator becoming an independent developer.

Leituras Relacionadas

Google and Amazon Simultaneously Invest Heavily in a Competitor: The Most Absurd Business Logic of the AI Era Is Becoming Reality

In a span of four days, Amazon announced an additional $25 billion investment, and Google pledged up to $40 billion—both direct competitors pouring over $65 billion into the same AI startup, Anthropic. Rather than a typical venture capital move, this signals the latest escalation in the cloud wars. The core of the deal is not equity but compute pre-orders: Anthropic must spend the majority of these funds on AWS and Google Cloud services and chips, effectively locking in massive future compute consumption. This reflects a shift in cloud market dynamics—enterprises now choose cloud providers based on which hosts the best AI models, not just price or stability. With OpenAI deeply tied to Microsoft, Anthropic’s Claude has become the only viable strategic asset for Google and Amazon to remain competitive. Anthropic’s annualized revenue has surged to $30 billion, and it is expanding into verticals like biotech, positioning itself as a cross-industry AI infrastructure layer. However, this funding comes with constraints: Anthropic’s independence is challenged as it balances two rival investors, its safety-first narrative faces pressure from regulatory scrutiny, and its path to IPO introduces new financial pressures. Globally, this accelerates a "tri-polar" closed-loop structure in AI infrastructure, with Microsoft-OpenAI, Google-Anthropic, and Amazon-Anthropic forming exclusive model-cloud alliances. In contrast, China’s landscape differs—investments like Alibaba and Tencent backing open-source model firm DeepSeek reflect a more decoupled approach, though closed-source models from major cloud providers still dominate. The $65 billion bet is ultimately about securing a seat at the table in an AI-defined future—where missing the model layer means losing the cloud war.

marsbitHá 1h

Google and Amazon Simultaneously Invest Heavily in a Competitor: The Most Absurd Business Logic of the AI Era Is Becoming Reality

marsbitHá 1h

Computing Power Constrained, Why Did DeepSeek-V4 Open Source?

DeepSeek-V4 has been released as a preview open-source model, featuring 1 million tokens of context length as a baseline capability—previously a premium feature locked behind enterprise paywalls by major overseas AI firms. The official announcement, however, openly acknowledges computational constraints, particularly limited service throughput for the high-end DeepSeek-V4-Pro version due to restricted high-end computing power. Rather than competing on pure scale, DeepSeek adopts a pragmatic approach that balances algorithmic innovation with hardware realities in China’s AI ecosystem. The V4-Pro model uses a highly sparse architecture with 1.6T total parameters but only activates 49B during inference. It performs strongly in agentic coding, knowledge-intensive tasks, and STEM reasoning, competing closely with top-tier closed models like Gemini Pro 3.1 and Claude Opus 4.6 in certain scenarios. A key strategic product is the Flash edition, with 284B total parameters but only 13B activated—making it cost-effective and accessible for mid- and low-tier hardware, including domestic AI chips from Huawei (Ascend), Cambricon, and Hygon. This design supports broader adoption across developers and SMEs while stimulating China's domestic semiconductor ecosystem. Despite facing talent outflow and intense competition in user traffic—with rivals like Doubao and Qianwen leading in monthly active users—DeepSeek has maintained technical momentum. The release also comes amid reports of a new funding round targeting a valuation exceeding $10 billion, potentially setting a new record in China’s LLM sector. Ultimately, DeepSeek-V4 represents a shift toward open yet realistic infrastructure development in the constrained compute landscape of Chinese AI, emphasizing engineering efficiency and domestic hardware compatibility over pure model scale.

marsbitHá 2h

Computing Power Constrained, Why Did DeepSeek-V4 Open Source?

marsbitHá 2h

Trading

Spot
Futuros

Artigos em Destaque

O que é WL

I. Introdução ao ProjetoWorldLand é uma L2 ou side chain do Ethereum, concebida como uma solução de baixo para cima para melhorar o ecossistema Ethereum.II. Informação sobre o Token1) Informação BásicaNome do token: WL (WorldLand)III. Links RelacionadosWebsite:https://worldland.foundation/Exploradores:https://bscscan.com/address/0x8aaB31fbc69C92fa53f600910Cf0f215531F8239Redes Sociais:https://x.com/WorldLand_space Nota: A introdução ao projeto provém dos materiais publicados ou fornecidos pela equipa oficial do projeto, que é apenas para referência e não constitui aconselhamento de investimento. A HTX não se responsabiliza por quaisquer perdas diretas ou indiretas resultantes.

212 Visualizações TotaisPublicado em {updateTime}Atualizado em 2026.03.28

O que é WL

Como comprar WL

Bem-vindo à HTX.com!Tornámos a compra de WorldLand (WL) simples e conveniente.Segue o nosso guia passo a passo para iniciar a tua jornada no mundo das criptos.Passo 1: cria a tua conta HTXUtiliza o teu e-mail ou número de telefone para te inscreveres numa conta gratuita na HTX.Desfruta de um processo de inscrição sem complicações e desbloqueia todas as funcionalidades.Obter a minha contaPasso 2: vai para Comprar Cripto e escolhe o teu método de pagamentoCartão de crédito/débito: usa o teu visa ou mastercard para comprar WorldLand (WL) instantaneamente.Saldo: usa os fundos da tua conta HTX para transacionar sem problemas.Terceiros: adicionamos métodos de pagamento populares, como Google Pay e Apple Pay, para aumentar a conveniência.P2P: transaciona diretamente com outros utilizadores na HTX.Mercado de balcão (OTC): oferecemos serviços personalizados e taxas de câmbio competitivas para os traders.Passo 3: armazena teu WorldLand (WL)Depois de comprar o teu WorldLand (WL), armazena-o na tua conta HTX.Alternativamente, podes enviá-lo para outro lugar através de transferência blockchain ou usá-lo para transacionar outras criptomoedas.Passo 4: transaciona WorldLand (WL)Transaciona facilmente WorldLand (WL) no mercado à vista da HTX.Acede simplesmente à tua conta, seleciona o teu par de trading, executa as tuas transações e monitoriza em tempo real.Oferecemos uma experiência de fácil utilização tanto para principiantes como para traders experientes.

238 Visualizações TotaisPublicado em {updateTime}Atualizado em 2026.03.28

Como comprar WL

Como comprar BASED

Bem-vindo à HTX.com!Tornámos a compra de Based (BASED) simples e conveniente.Segue o nosso guia passo a passo para iniciar a tua jornada no mundo das criptos.Passo 1: cria a tua conta HTXUtiliza o teu e-mail ou número de telefone para te inscreveres numa conta gratuita na HTX.Desfruta de um processo de inscrição sem complicações e desbloqueia todas as funcionalidades.Obter a minha contaPasso 2: vai para Comprar Cripto e escolhe o teu método de pagamentoCartão de crédito/débito: usa o teu visa ou mastercard para comprar Based (BASED) instantaneamente.Saldo: usa os fundos da tua conta HTX para transacionar sem problemas.Terceiros: adicionamos métodos de pagamento populares, como Google Pay e Apple Pay, para aumentar a conveniência.P2P: transaciona diretamente com outros utilizadores na HTX.Mercado de balcão (OTC): oferecemos serviços personalizados e taxas de câmbio competitivas para os traders.Passo 3: armazena teu Based (BASED)Depois de comprar o teu Based (BASED), armazena-o na tua conta HTX.Alternativamente, podes enviá-lo para outro lugar através de transferência blockchain ou usá-lo para transacionar outras criptomoedas.Passo 4: transaciona Based (BASED)Transaciona facilmente Based (BASED) no mercado à vista da HTX.Acede simplesmente à tua conta, seleciona o teu par de trading, executa as tuas transações e monitoriza em tempo real.Oferecemos uma experiência de fácil utilização tanto para principiantes como para traders experientes.

217 Visualizações TotaisPublicado em {updateTime}Atualizado em 2026.03.30

Como comprar BASED

Discussões

Bem-vindo à Comunidade HTX. Aqui, pode manter-se informado sobre os mais recentes desenvolvimentos da plataforma e obter acesso a análises profissionais de mercado. As opiniões dos utilizadores sobre o preço de A (A) são apresentadas abaixo.

活动图片