Demis Hassabis: Agents, AGI & The Next Big Scientific Breakthrough
Demis Hassabis, CEO of Google DeepMind and 2024 Nobel laureate in Chemistry, sits down with YC's Garry Tan for a wide-ranging conversation about the path to AGI. He identifies continual learning, long-term reasoning, and memory as the remaining unsolved problems, estimates AGI around 2030, and explains how ideas from AlphaGo are resurging in modern foundation models. The discussion covers the power of distillation for small models, the early state of AI agents, the 'Einstein test' for genuine AI creativity, and why founders should combine AI with deep tech domains like materials science and drug discovery.
Google DeepMind CEO、2024 年诺贝尔化学奖得主 Demis Hassabis 与 YC 的 Garry Tan 进行了一场关于 AGI 之路的深度对话。他指出持续学习、长期推理和记忆是剩余未解决的关键问题,预计 AGI 将在 2030 年左右到来,并解释 AlphaGo 的思想(包括蒙特卡洛树搜索)如何在现代基础模型中复兴。讨论涵盖蒸馏技术对小模型的赋能(尚未发现理论极限)、AI 智能体处于早期但快速进步的阶段、模型能解 IMO 金牌却做错基础数学的「锯齿状智能」问题,以及真正创造力所需的是什么(发明围棋,而不只是下出第 37 手)。对于创始人,Hassabis 给出了具体建议:将 AI 与材料科学、药物发现等深科技领域结合,物理世界的复杂性本身就能构成抵御下一个基础模型更新的护城河。他还分享了 AlphaFold 式突破的三要素公式——巨大的组合搜索空间、清晰的目标函数、足够的数据——以及判断 AI 能否做出真正原创发现的「爱因斯坦测试」。
What's Still Missing for AGI
Demis estimates about 50/50 odds that one or two big ideas are still needed beyond current scaling — with continual learning, long-term reasoning, and memory being the key unsolved problems.
AGI 还缺什么
Demis 估计大约有五成概率,在当前扩展范式之外还需要一两个大突破——持续学习、长期推理和记忆是仍未解决的关键问题。
Memory Is More Than a Bigger Context Window
Even with million-token context windows, naively storing everything is brute force. The real challenge is efficiently retrieving what's relevant for the current decision, much like how the hippocampus consolidates memories during sleep.
记忆不只是更大的上下文窗口
即使有了百万 token 的上下文窗口,简单粗暴地存储一切仍是蛮力做法。真正的挑战是高效检索与当前决策相关的信息,就像海马体在睡眠中巩固记忆一样。
AlphaGo Ideas Are Coming Back at Scale
Chain-of-thought reasoning and thinking modes in today's models are descendants of AlphaGo's techniques. DeepMind is revisiting ideas like Monte Carlo tree search at massive scale for the next wave of advances.
AlphaGo 的思想正在大规模回归
当今模型的思维链推理和思考模式是 AlphaGo 技术的后裔。DeepMind 正在大规模重新审视蒙特卡洛树搜索等思路,以推动下一波突破。
The Jagged Intelligence Problem
Models can solve IMO gold-medal problems yet still make elementary math errors — what Demis calls 'jagged intelligence.' This is an introspection gap in how models monitor their own thought process.
锯齿状智能问题
模型能解决国际数学奥林匹克的金牌题,却仍会犯基础数学错误——Demis 称之为「锯齿状智能」。这是模型对自身思维过程缺乏内省能力所致。
The Sweet Spot for Startups: AI + Deep Tech
The most defensible startup opportunities lie at the intersection of AI and deep technology areas involving the physical world — materials, medicine, and other hard sciences — where the next foundation model update won't render you obsolete.
创业最佳切入点:AI + 深科技
最具防御性的创业机会在 AI 与深科技(涉及物理世界——材料、医学和其他硬科学)的交叉点,下一个基础模型的更新不会让你被淘汰。
The AlphaFold Breakthrough Pattern
Three ingredients for an AI breakthrough: a massive combinatorial search space where brute force fails, a clear objective function for hill climbing, and enough data or simulators to generate in-distribution synthetic data.
AlphaFold 突破模式
AI 突破的三要素:暴力搜索无法解决的巨大组合搜索空间、可以用于爬坡的清晰目标函数、以及足够的数据或模拟器来生成分布内的合成数据。
The Einstein Test for AI Creativity
Train a system with knowledge up to 1901 and see if it can produce what Einstein did in 1905, including special relativity. That would signal AI can make genuinely novel discoveries beyond pattern matching.
AI 创造力的「爱因斯坦测试」
用 1901 年以前的知识训练一个系统,看它能否得出爱因斯坦 1905 年的发现(包括狭义相对论)。这将标志着 AI 能够超越模式匹配,做出真正原创的发现。
"Continual learning, long-term reasoning, some aspects of memory — these are still unsolved. I think all of these are going to be required for AGI."
"持续学习、长期推理、记忆的某些方面——这些问题仍未解决。我认为实现 AGI 需要所有这些。"
— Demis Hassabis"It's not enough to come up with move 37. That's pretty cool, very useful, but can it invent Go? That's what I want — a system that can invent Go."
"仅仅想出第 37 手还不够。那很酷、很有用,但它能发明围棋吗?我想要的是一个能发明围棋的系统。"
— Demis Hassabis"Step one was solve intelligence, build AGI, and then step two was use it to solve everything else — what I call root node problems in science."
"第一步是解决智能问题,构建 AGI;第二步是用它来解决其他所有问题——我称之为科学中的「根节点问题」。"
— Demis Hassabis"Nothing that's really long-lasting and worthwhile is easy. And so, I've always been drawn to deep technologies."
"真正持久、有价值的东西没有容易的。所以,我一直被深科技所吸引。"
— Demis Hassabis"I would have worked on AI no matter what happened. I just decided from a very young age it was the thing that could be the most consequential thing I could think of."
"无论发生什么,我都会从事 AI 工作。我从很小的时候就认定,这是我能想到的最具影响力的事情。"
— Demis Hassabis"There is a compound out there that would solve this disease if one could find it. As long as the laws of physics allow it, the only question is how do you find it in an efficient way."
"存在一种化合物可以治愈这种疾病,只要能找到它。只要物理定律允许,唯一的问题就是如何高效地找到它。"
— Demis Hassabis"Can you train a system with the knowledge cutoff of 1901 and then will it come up with what Einstein did in 1905, including special relativity? Once that is, then I think we're on the verge of these systems being able to invent something truly novel."
"你能否用 1901 年的知识截止点训练一个系统,然后看它能否得出爱因斯坦 1905 年的发现,包括狭义相对论?一旦可以做到,我们就站在这些系统能够发明全新事物的门槛上了。"
— Demis Hassabis"Going after hard problems and deep problems is no more difficult in some ways than going after a shallower, simpler, more superficial problem. They're just differently difficult."
"追求困难的、深层的问题,在某些方面并不比追求浅显、简单、表面的问题更难。它们只是难法不同而已。"
— Demis HassabisAGI Timeline: ~2030, With Possible Missing Pieces
Demis still targets ~2030 for AGI. Current techniques will be part of the final architecture, but one or two major innovations may still be needed. Founders building deep tech should plan for AGI arriving mid-journey.
AGI 时间线:约 2030 年,可能仍有缺失环节
Demis 仍将 AGI 目标定在约 2030 年。当前技术将是最终架构的一部分,但可能还需要一两项重大创新。做深度科技的创业者应预设 AGI 会在项目中途到来。
Small Models Are the Practical Workhorse
Distillation lets flash models reach 90-95% of frontier performance at a fraction of the cost and latency. No theoretical limit to distillation has been found yet.
小模型才是实用的主力
蒸馏技术让 flash 模型以极低的成本和延迟达到前沿模型 90-95% 的性能。目前尚未发现蒸馏的理论极限。
Agents Need Continual Learning to Level Up
Today's agents are useful for parts of tasks but can't adapt to new contexts. Cracking continual learning is the key missing piece for fire-and-forget agents.
智能体需要持续学习才能真正升级
当今的智能体在子任务上很有用,但无法适应新上下文。突破持续学习是实现「发射后不管」式智能体的关键缺失环节。
Combine AI with Physical-World Expertise
The most defensible startups will combine frontier AI with deep expertise in areas involving atoms — materials science, drug discovery, medicine. Real-world complexity creates a moat.
将 AI 与物理世界的专业知识结合
最具防御性的创业公司会将前沿 AI 与涉及原子的领域的深厚专业知识结合——材料科学、药物发现、医学。现实世界的复杂性本身构成了护城河。
Pursue Passion, Not Consensus
When Hassabis started DeepMind in 2010, investors and academics told him AI was a dead end. His advice: work on what you're genuinely passionate about, because conviction carries you through when no one else believes.
追求热爱,而非共识
当 Hassabis 2010 年创立 DeepMind 时,投资者和学者告诉他 AI 是一条死路。他的建议:从事你真正热爱的事情,因为信念能带你度过无人相信的那些年。
The Three-Ingredient Formula for AI Breakthroughs
A problem is ripe for an AlphaFold-style breakthrough when it has: (1) massive combinatorial search space, (2) clear objective function, (3) enough data or simulation capacity. Drug discovery fits perfectly.
AI 突破的三要素公式
当一个问题具备以下条件时,就适合实现 AlphaFold 式的突破:(1) 巨大的组合搜索空间,(2) 清晰的目标函数,(3) 足够的数据或模拟能力。药物发现完美符合。
Design for a World Where AGI Appears Mid-Journey
True deep tech takes about 10 years. If AGI arrives around 2030, founders must build things that AGI can amplify rather than obsolete.
为 AGI 半途出现的世界做设计
真正的深科技大约需要 10 年。如果 AGI 在 2030 年左右到来,创始人必须构建 AGI 能放大而非淘汰的东西。
AGI
The pursuit of artificial general intelligence and what's still missing
通用人工智能:对通用人工智能的追求以及仍然缺失的环节
Continual Learning
The ability of AI systems to learn new information without forgetting old knowledge
持续学习:AI 系统在不遗忘旧知识的情况下学习新信息的能力
AlphaFold
DeepMind's breakthrough in protein structure prediction that won the Nobel Prize
AlphaFold:DeepMind 在蛋白质结构预测方面的突破,获得了诺贝尔奖
Memory Systems
How AI systems store, retrieve, and consolidate information beyond context windows
记忆系统:AI 系统如何在上下文窗口之外存储、检索和巩固信息
Distillation
Techniques to compress frontier model capabilities into smaller, faster models
蒸馏技术:将前沿模型能力压缩到更小、更快模型中的技术
AI Agents
Systems that can accomplish goals autonomously, make active decisions and plans
AI 智能体:能够自主完成目标、做出主动决策和规划的系统
Scientific Discovery
Using AI to accelerate breakthroughs in science, from virtual cells to drug discovery
科学发现:利用 AI 加速科学突破,从虚拟细胞到药物发现
Deep Tech Startups
Founders combining AI with hard science domains for defensible businesses
深科技创业:创始人将 AI 与硬科学领域结合,建立有护城河的业务
Multimodal AI
Models built from the ground up to understand text, images, audio, and video together
多模态 AI:从一开始就构建为同时理解文本、图像、音频和视频的模型
AlphaGo
The techniques behind AlphaGo are resurging in modern foundation models
AlphaGo:AlphaGo 背后的技术正在现代基础模型中复兴
Intro
Brief highlights from the conversation teasing the key themes: continual learning, agents, and AGI.
开场:对话精华预告,引出持续学习、智能体和 AGI 等核心主题。
From Chess Prodigy to DeepMind
Garry introduces Demis's extraordinary career journey — chess prodigy, video game designer at 17, PhD in cognitive neuroscience, co-founding DeepMind in 2010, and winning the 2024 Nobel Prize.
从国际象棋神童到 DeepMind:Garry 介绍 Demis 非凡的职业历程——国际象棋神童、17 岁设计电子游戏、认知神经科学博士、2010 年联合创立 DeepMind,以及获得 2024 年诺贝尔奖。
What's Missing Before AGI
Demis explains that current techniques (pre-training, RLHF, chain-of-thought) will be part of the final AGI architecture, but continual learning, long-term reasoning, and memory remain unsolved. He puts 50/50 odds on needing one or two more big ideas.
AGI 之前还缺什么:Demis 解释当前技术(预训练、RLHF、思维链)将是最终 AGI 架构的一部分,但持续学习、长期推理和记忆仍未解决。他认为需要一两个大突破的概率约五五开。
Why Memory Is Still Unsolved
Drawing from his PhD research on the hippocampus, Demis explains how the brain consolidates memories during sleep through experience replay — and why simply expanding context windows is brute force, not real memory.
为什么记忆问题仍未解决:结合他关于海马体的博士研究,Demis 解释大脑如何在睡眠中通过经验回放巩固记忆——以及为什么单纯扩展上下文窗口只是蛮力,不是真正的记忆。
How AlphaGo Shaped Gemini
DeepMind's heritage in reinforcement learning and search (AlphaGo, AlphaZero) directly informs today's thinking modes and chain-of-thought reasoning. Ideas like Monte Carlo tree search are being revisited at massive scale.
AlphaGo 如何影响了 Gemini:DeepMind 在强化学习和搜索方面的积累(AlphaGo、AlphaZero)直接影响了今天的思考模式和思维链推理。蒙特卡洛树搜索等思路正在大规模重新应用。
Why Smaller Models Are Getting So Powerful
Distillation lets Google pack frontier capabilities into flash models at 1/10th the cost with 90-95% of the quality. No theoretical limit to distillation has been found yet. This matters for serving billions of users across Google's products.
为什么小模型越来越强:蒸馏技术让 Google 将前沿能力打包到 flash 模型中,成本仅十分之一,质量达 90-95%。尚未发现蒸馏的理论极限。这对服务 Google 数十亿用户的产品至关重要。
The 1000x Engineer
Engineers using AI tools are achieving 500-1000x the output of engineers from just six months ago. Fast small models enable rapid iteration that more than compensates for the 5-10% quality gap versus frontier models.
1000 倍工程师:使用 AI 工具的工程师正在实现六个月前工程师 500-1000 倍的产出。快速的小模型实现了高速迭代,足以弥补与前沿模型 5-10% 的质量差距。
Continual Learning and the Future of Agents
Current agents are cobbling together capabilities with 'duct tape.' True continual learning is needed for agents that can adapt to new situations without requiring full retraining.
持续学习与智能体的未来:当前的智能体是在用「胶带」拼凑能力。需要真正的持续学习才能让智能体适应新情况而无需完全重新训练。
Why AI Still Fails at Basic Reasoning
Models can solve IMO gold-medal math problems yet make elementary errors — Demis calls this 'jagged intelligence.' The root cause is a lack of introspection and self-monitoring in the reasoning process.
为什么 AI 仍在基础推理上失败:模型能解决国际数学奥林匹克金牌题却犯基础错误——Demis 称之为「锯齿状智能」。根本原因是推理过程中缺乏内省和自我监控。
Are Agents Overhyped or Just Getting Started?
Demis believes agents are just getting going. The technology is finally reaching the point where it's not just a toy demo but actually adds real value to productivity and efficiency.
智能体是炒作还是刚刚开始?:Demis 认为智能体才刚刚起步。技术终于到了不再是玩具演示、而是真正为生产力和效率增值的阶段。
Can AI Become Truly Creative?
AlphaGo's move 37 was creative within the rules of Go, but can AI invent Go itself? Demis wants to see systems that can create entirely new paradigms, not just optimize within existing ones.
AI 能真正具有创造力吗?:AlphaGo 的第 37 手在围棋规则内是创造性的,但 AI 能发明围棋本身吗?Demis 希望看到能创造全新范式的系统,而不只是在现有范式中优化。
Open Models, Gemma, and Local AI
Gemma 4 models demonstrate DeepMind's distillation prowess. Local models are critical for privacy, security, and edge deployment — especially for personal devices and home robotics.
开源模型、Gemma 与本地 AI:Gemma 4 模型展示了 DeepMind 的蒸馏实力。本地模型对隐私、安全和边缘部署至关重要——尤其是个人设备和家庭机器人。
Why Gemini Was Built Multimodal
Building multimodal from day one was harder initially but now pays off in world modeling, Gemini Robotics, and real-world assistants that understand physical environments.
为什么 Gemini 从一开始就做多模态:从第一天起就走多模态路线初期更难,但如今在世界模型构建、Gemini Robotics 和理解物理环境的现实世界助手中正在获得回报。
What Happens When Inference Gets Cheap?
As inference costs plummet, the bottleneck shifts to having the right ideas and asking the right questions. Creativity and taste become the scarce resources.
推理成本趋近于零时会怎样?:随着推理成本急剧下降,瓶颈转移到拥有正确的想法和提出正确的问题上。创造力和品味成为稀缺资源。
From AlphaFold to the Virtual Cell
DeepMind's mission: solve intelligence first, then use it to solve 'root node' problems. AlphaFold is the prototype — 3M+ researchers use it. Next frontiers include virtual cells (~10 years away), material science, and climate modeling.
从 AlphaFold 到虚拟细胞:DeepMind 的使命:先解决智能,再用它解决「根节点」问题。AlphaFold 是范本——超过 300 万研究人员在使用。下一个前沿包括虚拟细胞(约 10 年后)、材料科学和气候建模。
AI as the Ultimate Tool for Science
AI is like a Promethean capability — transformative but requiring careful stewardship. The power to solve disease and discover new materials comes with responsibility for preventing misuse.
AI 作为终极科学工具:AI 就像普罗米修斯之火——具有变革性但需要谨慎管理。解决疾病和发现新材料的力量伴随着防止滥用的责任。
Advice for Founders
Combine AI with deep tech domains where real-world complexity creates a moat. The sweet spot is interdisciplinary teams at the intersection of machine learning and hard sciences. Pursue what you're genuinely passionate about.
给创始人的建议:将 AI 与现实世界复杂性构成护城河的深科技领域结合。最佳切入点是机器学习与硬科学交叉领域的跨学科团队。追求你真正热爱的事物。
The AlphaFold Breakthrough Pattern
Three ingredients: massive combinatorial search space, clear objective function, and enough data/simulators. Drug discovery fits this pattern — there exists a compound to cure a disease, the question is finding it efficiently.
AlphaFold 的突破模式:三要素:巨大的组合搜索空间、清晰的目标函数、以及足够的数据/模拟器。药物发现完美符合——存在治愈疾病的化合物,问题是如何高效找到它。
Can AI Make Real Scientific Discoveries?
The 'Einstein test': train a system with knowledge up to 1901 and see if it produces Einstein's 1905 breakthroughs, including special relativity. Once that's possible, AI can invent truly novel things.
AI 能做出真正的科学发现吗?:「爱因斯坦测试」:用 1901 年以前的知识训练一个系统,看它能否得出爱因斯坦 1905 年的突破(包括狭义相对论)。一旦可以做到,AI 就能发明真正全新的东西。
What to Build Before AGI Arrives
With AGI potentially arriving around 2030, founders should build things AGI can amplify rather than replace. The future architecture will be general models orchestrating specialized tools, not one monolithic brain.
AGI 到来之前应该构建什么:AGI 可能在 2030 年左右到来,创始人应构建 AGI 能放大而非替代的东西。未来架构将是通用模型调度专业工具,而非一个巨型大脑。
[0:00]continual learning, long-term reasoning, [music] uh some aspects of memory, these are still unsolved. I think all of these are going to be required for AGI. 持续学习、长期推理、[music] 嗯,记忆的某些方面,这些都还是未解的问题。我认为实现 AGI 需要所有这些能力。
[0:09]Depending on what your AGI timeline is, you know, mine's like 2030 or something like this, then [music] if you start off on a deep tech journey today, you have to just consider AGI appearing in the middle of that journey. It's not bad, necessarily, but you have to take that 取决于你的 AGI 时间线是什么,你知道,我的是 2030 年左右这样的。[music] 如果你今天开始一段深科技的创业之旅,你必须考虑到 AGI 会在旅途中出现。这不一定是什么坏事,但你必须把这个因素
[0:24]into account. [music] You have to have an active system that can actively solve problems for you to get to AGI. So, agents are that path, and I think we're Demis Hassabis has had one of the most 考虑进去。[music] 你必须有一个能主动为你解决问题的系统,才能达到 AGI。所以,Agent 就是那条路径,而我认为我们是……Demis Hassabis 拥有科技界最
[0:43]unusual careers in tech. He was a chess prodigy as a kid, then designed his first hit video game, Theme Park, at 17. 不寻常的职业之一。他小时候是个国际象棋神童,然后在 17 岁时设计了他的第一款大热电子游戏——Theme Park。
[0:53]He then went back to school, got a PhD in cognitive neuroscience, published foundational work on how memory and imagination work in the brain, and then in 2010 co-founded DeepMind with one mission, solve intelligence. 之后他重返校园,拿到了认知神经科学的博士学位,发表了关于大脑中记忆和想象力如何运作的基础性研究成果,然后在 2010 年联合创立了 DeepMind,只有一个使命——解决智能问题。
[1:08]And I think they've done it. Since then, his lab has gone on to do things most people thought were decades away. 而我认为他们做到了。从那以后,他的实验室完成了一些大多数人认为还要几十年才能实现的事情。
[1:17]AlphaGo beat a world champion at Go. AlphaFold cracked protein structure prediction, a 50-year grand challenge in biology, and they gave it away for free to every scientist on Earth. AlphaGo 击败了世界围棋冠军。AlphaFold 攻克了蛋白质结构预测——这是生物学领域一个 50 年来的重大挑战——而且他们把它免费提供给了地球上每一位科学家。
[1:29]That work won him the Nobel Prize in chemistry last year. Today, Demis leads Google DeepMind, where he's building Gemini and pushing toward the same goal he set when he was a teenager, artificial general intelligence. please 这项工作让他去年获得了诺贝尔化学奖。如今,Demis 领导着 Google DeepMind,在那里他正在打造 Gemini,并朝着他从少年时代就设定的同一个目标前进——通用人工智能。
[1:53]So, you've been thinking about AGI longer than almost anyone. When you look at the current paradigm, large-scale pre-training, RLHF, chain of thought, how much of the final architecture for AGI do you think we already have, and what's fundamentally missing right now? 那么,你思考 AGI 的时间比几乎任何人都要长。当你审视当前的范式——大规模预训练、RLHF、思维链——你认为 AGI 的最终架构中,我们已经拥有了多少?目前还缺少什么根本性的东西?
[2:08]Well, first of all, thank thanks Garry for that great introduction, and it's great to be here. Thanks for for welcoming here. It's amazing space, actually. I'll have to come back here often. Very inspiring that you all get to work in in in this space. So, the question is I think the components that you just mentioned, I'm pretty sure will 嗯,首先,谢谢 Garry 的精彩介绍,很高兴来到这里。谢谢你邀请我来。这个空间真的很棒。我得经常来这里。你们都能在这个空间工作,真的很令人振奋。所以,你的问题是——我认为你刚才提到的那些组件,我很确定会
[2:25]be part of the final architecture for AGI. So, I think they've come such a long way now, and we've proven out so many things about what they can do. 成为 AGI 最终架构的一部分。所以,我认为它们已经走了很长一段路了,我们已经验证了它们能做到的很多事情。
[2:35]I can't see a world in which we'll sort of realize in a couple of years this was a dead end. That doesn't make sense to me. But, there still might be one or two things missing on top of of of of what you've you know, what we already know works. So, continual learning, long-term reasoning, 我无法想象一个世界,几年后我们会突然发现这是一条死胡同。这对我来说说不通。但是,在我们已经知道有效的东西之上,可能仍然缺少一两个东西。所以,持续学习、长期推理,
[2:52]uh some aspects of memory, these are still unsolved. And how to get the systems to be more consistent across the board. I think all of these are going to be required for AGI. Now, it might be that the existing techniques can just scale up to that 嗯,记忆的某些方面,这些都还是未解的问题。还有如何让系统在各方面更加一致。我认为实现 AGI 需要所有这些能力。现在,也许是现有的技术只需要通过一些创新、一些渐进式的创新就能扩展到那个程度
[3:08]with some innovation and some incremental innovation. But, it could be that there's still one or two big ideas left that need to be cracked. I don't think it's more than one or two if they are out there. And I think you know, my betting is about 50/50 if that's the case. So, of course 但也许仍然有一两个大想法需要被攻克。如果真有的话,我认为不会超过一两个。而且你知道,我的判断是大概五五开吧。所以,当然
[3:25]at DeepMind at Google DeepMind we work on both those things. I guess that's a really working with a bunch of identic systems the wildest thing to me is to what degree it's the same weights over and over. So, this idea of continual learning is so interesting because like 在 DeepMind,在 Google DeepMind,我们同时在研究这两个方向。我猜这真的是……和一堆相同的系统一起工作,最让我惊讶的是,在多大程度上它是一遍又一遍的同一套权重。所以持续学习这个概念太有意思了,因为
[3:41]you know, right now we're sort of cobbling it together with duct tape, you know, these dream cycles at night and things like that. 你知道,目前我们基本上是在用胶带把它粘在一起,你知道,比如晚上做那种梦境循环之类的东西。
[3:47]>> Yeah. It's pretty cool the dream cycles, and we we used to think about this with consolidation with episodic memory. It's actually that's what I studied for my PhD is how the hippocampus works and integrates, you know, new knowledge gracefully into the existing knowledge base. So, the brain does that amazingly >> 没错。梦境循环确实很酷,我们以前会想到这个,和情节记忆的巩固有关。实际上,那正是我博士期间研究的内容——海马体是如何运作和整合新知识到现有知识库中的。大脑在这方面做得非常出色
[4:03]well. It it it does it through you know, during sleep, especially things like REM sleep, replaying back episodes that that are important so that you can learn from it. In fact, our very first Atari program, DQN, one of the ways it was able to master Atari games was by doing 非常好。它通过……你知道,在睡眠期间,尤其是类似 REM 睡眠的阶段,重放那些重要的经历片段,这样你就能从中学习。事实上,我们最早的 Atari 程序 DQN,它能够掌握 Atari 游戏的方式之一就是通过做
[4:19]experience replay. So, we sort of borrowed that from from neuroscience and replayed successful trajectories many times, you know, that's way back in 2013 now in the in the dark ages of AI. 经验回放。所以我们算是从神经科学那里借鉴了这个想法,然后多次重放成功的轨迹。你知道,那是 2013 年的事了,现在回看,算是 AI 的黑暗时代了。
[4:31]It was a really important thing, and and I agree with you, we're kind of using duct tape right now. So, like shove it all in the context window. 那确实是一个非常重要的突破,而且我同意你的看法,我们现在基本上是在用胶带粘。就像,把所有东西都塞进上下文窗口里。
[4:38]This but this seems a bit unsatisfying, right? And actually, even though we're working on machines, not biological brains, and so you potentially you could have, you know, millions or tens of millions size context window or memory, and it can be perfect, there's still a 但这似乎有点……不够令人满意,对吧?实际上,即使我们研究的是机器而不是生物大脑,所以你理论上可能有百万甚至千万大小的上下文窗口或记忆,而且它可以做到完美无缺,但仍然有一个
[4:54]cost to looking it up and finding the right thing that that's actually relevant for the specific decision you've got to make right now. And that's non-trivial that cost, even if you can potentially store it all. I think there's actually a lot of room for 查找和找到与你当前需要做出的具体决策真正相关的内容的成本。而且这个成本是不可忽视的,即使你理论上可以存储所有东西。我认为在
[5:10]innovation in in areas like memory. Yeah. I mean, the wild thing is that it feels like a million token context window is actually bigger than I mean, it's plenty big, honestly. 记忆等领域还有很大的创新空间。是的。我的意思是,让人惊讶的是,一百万 token 的上下文窗口实际上感觉比……我的意思是,说实话,它已经足够大了。
[5:18]You can do so >> Well, it's it's it's plenty big for for for most things that it should be useful. I mean, if you think about the context window is sort of equivalent to working memory. You know, humans have we have like a few digits, you know, it's like a dozen digits maybe, you know, 你可以做很多…… >> 嗯,它对于大多数需要它发挥作用的事情来说确实足够大了。我是说,如果你把上下文窗口想象成相当于工作记忆,你知道,人类有……我们有几个数字的容量,大概十几个数字吧,你知道,
[5:34]average of seven. We got million or you know, 10 million context windows, but the problem is is that we're trying to store everything in that. You know, things that aren't in not important, things that are wrong. It's pretty brute force currently, and that doesn't seem right. And then the problem is if you're 平均大概七个。而我们有一百万甚至一千万的上下文窗口,但问题是我们试图把所有东西都存储在里面。你知道,不重要的东西、错误的东西,全部都塞进去。目前这种方式相当粗暴,这感觉不太对。然后问题来了,如果你
[5:49]now trying to try and process live video, and you're just going to naively record all the tokens, then actually million tokens isn't that much. It's only like 20 minutes. So, actually you need more if you want something that's going to understand your you know, your 现在要处理实时视频,而你只是简单地记录所有 token,那一百万 token 其实也没多少。大概只有 20 分钟左右。所以实际上你需要更多,如果你想要一个能理解你的……你知道,你的
[6:05]what's going on in your life over maybe a month or two. DeepMind has uh historically leaned into reinforcement learning and search, AlphaGo, AlphaZero, and MuZero. How much of that philosophy is actually embedded in how you're building Gemini today. Is 过去一两个月生活中发生了什么。DeepMind 历史上一直很重视强化学习和搜索,AlphaGo、AlphaZero 和 MuZero。这种理念在多大程度上体现在你们今天构建 Gemini 的方式中?
[6:21]RL still underrated? Yeah, I think potentially it is. It sort of goes in in waves. You know, we've worked on agents since the beginning of DeepMind. In fact, we also that's what we said we were working on. So, all of the Atari work and AlphaGo and most specifically, 强化学习仍然被低估了吗?是的,我认为可能确实被低估了。它算是起起伏伏的。你知道,我们从 DeepMind 创立之初就在研究 Agent。事实上,我们当初也说了我们就是在做 Agent。所以所有的 Atari 工作和 AlphaGo,以及最具体的来说,
[6:37]they're agent systems, and what we meant by that is systems that are able to, you know, accomplish goals on their own and make active decisions and and make plans. And so, of course we were doing it in the domain of games to make it tractable, and then doing increasingly 它们都是 Agent 系统,我们的意思是那些能够自主完成目标的系统,能够做出主动决策,制定计划。当然,我们是在游戏领域做这件事,为了让问题可控,然后做越来越
[6:54]complex games, things like StarCraft after AlphaGo, AlphaStar. So, we basically did all the games that were out there. 复杂的游戏,比如 AlphaGo 之后的 StarCraft——AlphaStar。所以我们基本上攻克了市面上所有的游戏。
[7:02]And then of course, the question is can you generalize those models to be world models or models of language, not just models of simple games or even complex games. And that's what the last few years has been about. But really, you can think of a lot of the things we're doing today, all the leading models with 然后当然,问题是——你能不能把这些模型泛化成世界模型或语言模型,而不仅仅是简单游戏甚至复杂游戏的模型。这就是过去几年的事情了。但实际上,你可以把我们今天做的很多事情,所有领先的模型
[7:19]thinking modes and chain of thought reasoning as aspects of what was sort of pioneered with AlphaGo coming back now. 带有思考模式和思维链推理的,都可以看作是 AlphaGo 当年开创的那些理念的回归。
[7:26]And I actually think there's a lot of work we did back then that is relevant today, and we're sort of relooking at some of those old ideas at scale today in a more general way, including things like Monte Carlo tree search and other other ways of doing 实际上我认为我们当年做的很多工作在今天仍然相关,我们正在以更通用的方式大规模重新审视一些旧的想法,包括蒙特卡洛树搜索和其他一些增强
[7:41]augmenting the RL on top of the the reinforcement learning we're ready to do today. And I think a lot of those ideas both from AlphaGo and AlphaZero are really really relevant to to where we are with today's foundation models. And I think a lot of that is 强化学习的方法。我认为 AlphaGo 和 AlphaZero 中的很多想法与我们现在的基础模型所处阶段真的非常相关。而且我认为其中很多东西
[7:56]what we're going to see of the advances the next few years. One question I would have like obviously today you need bigger and bigger models to be smarter and smarter, but then we're also seeing distillation working, and then smaller models can be like quite a bit faster. I think you know, you guys have incredible 就是我们在未来几年将会看到的进步。我有一个问题,显然今天你需要越来越大的模型才能越来越聪明,但我们也在看到蒸馏技术的成功,更小的模型可以变得相当快。我认为你们有非常厉害的
[8:13]flash models that are like nine like you're finding that they're 95% as good as the frontier and at like 1/10 the price. Is that right? I think that's one of our core strengths is I mean, you have to build the biggest models to to to have the frontier capabilities. But, flash 模型,大概有 95% 的前沿模型能力,但价格只有十分之一。是这样吗?我认为这是我们核心优势之一。我的意思是,你必须构建最大的模型才能拥有前沿能力。但是,
[8:30]I think one of our biggest strengths has been distilling and packing that power into smaller and smaller models very quickly. 我认为我们最大的优势之一就是非常快速地将这些能力蒸馏到越来越小的模型中。
[8:37]Obviously, we we you know, we invented the kind of distillation process and people like Jeff and Oriol and and others, and we're still world experts in that. And we also have a huge need to do it because we've got to serve the 显然,你知道,我们发明了这种蒸馏方法,Jeff 和 Oriol 等人做的,我们在这一领域仍然是世界顶尖的。而且我们也有巨大的需求来做这件事,因为我们要服务
[8:52]biggest probably AI surfaces there are. Obviously, there's search with AI overviews and AI mode, then there's Gemini app, and now increasingly every single product at Google has, you know, Maps and YouTube and so on has some aspect of Gemini or Gemini-related 可能是最大的 AI 应用场景。显然,搜索方面有 AI overviews 和 AI mode,然后是 Gemini 应用,现在 Google 的每一个产品,你知道,Maps、YouTube 等等都融入了 Gemini 或与 Gemini 相关的
[9:08]technology in it. And so, that's billions of users, a dozen more than a dozen billion user products, and they have to be served extremely fast, extremely efficiently, and cheaply, and with low latency. So, that that gives us a really important incentive to to make these flash and 技术。所以那是数十亿用户,十多个拥有十亿用户的产品,它们需要被极其快速、高效、低成本地提供服务,而且要低延迟。所以这给了我们一个非常重要的动力来让这些 flash 和
[9:25]even smaller models, flash light models, extremely efficient. And hopefully that ends up then being really useful for many of the workloads that all of you use for. I'm curious about how much smarter these smaller models can actually be. Like are there limits to the distillation process? Like could a 甚至更小的模型、flash light 模型变得极其高效。希望这最终能对你们所有人的工作负载真的有用。我很好奇这些小模型到底能变得多聪明。蒸馏过程有极限吗?比如一个
[9:42]50B or 400B model be as smart as like a mythos for today? Yeah, I don't I don't see any I don't think we've got to any kind of or at least none of us know yet if we've got to any kind of informational limit. I mean, maybe at some point that will be the case where there's just an information density that 50B 或 400B 的模型能和今天的巨型模型一样聪明吗?是的,我不认为我们已经达到了任何类型的……或者至少我们中没有人知道我们是否已经达到了任何类型的信息极限。我的意思是,也许在某个时候会是这种情况,存在一个信息密度的上限,我们
[9:59]can't we can't get beyond. But I think for now there's the assumption we make is that, you know, a year later after one of our leading, you know, pro models or frontier models goes out, half a year later, a year later you'll have them in 无法突破。但我认为就目前而言,我们的假设是,你知道,在我们发布一个领先的 pro 模型或前沿模型之后,半年到一年,你就会在那些
[10:14]the the really tiny, almost edge models. And you also see some of that goodness in our Gemma models, which hopefully you're all enjoying our Gemma 4 models, which I think are really amazing power for their sizes. So, again, that uses a lot of this these distillation 非常小的、几乎是边缘端的模型中拥有同样的能力。你也可以在我们的 Gemma 模型中看到一些这样的优势,希望你们都在享受我们的 Gemma 4 模型,我认为它们在各自的尺寸下真的很厉害。所以,这同样使用了很多这种蒸馏
[10:29]techniques and and the idea of how to make things really efficient in these very small models. So, I don't really see any limit yet in terms of like some kind of theoretical limit. I think we're still pretty far off of that. That's amazing. I mean, that is really good. 技术,以及如何让这些非常小的模型变得极其高效的想法。所以我目前还看不到任何理论上的极限。我认为我们离那还很远。这太厉害了。我的意思是,这真的很棒。
[10:41]Yes. Uh you know, one of the weirder things that we're seeing right now is like engineers can do like 500 to 1,000 times the amount of work that they were doing like 6 months ago, I guess. I mean, the people in this room, there are people who are doing about like 1,000 x the 是的。嗯,你知道,我们现在看到的一个比较奇怪的现象是,工程师大概能做到 6 个月前 500 到 1000 倍的工作量。我是说,这个房间里的人,有些人正在做到大约 1000 倍的
[10:56]work that like I Stevie Yegge talks about this. It's like 1,000 x the work that a Google engineer from the 2000s was doing. I think it's very exciting. I mean, I think the small models have many uses. One is obviously cost, but the speed can allow, you know, if you think about coding even or other things, you 工作量,就像 Stevie Yegge 谈到的那样。是 2000 年代 Google 工程师工作量的 1000 倍。我认为这非常令人兴奋。我的意思是,我认为小模型有很多用途。一方面显然是成本,但速度可以带来,你知道,如果你想想编程或其他事情,你
[11:13]can iterate a lot faster also, especially if there's if you're collaborating with the system. I think there's a there's a a lot of need for having fast systems that maybe are not quite frontier, like you said, like 95% 90%, but that's 可以更快地迭代,特别是如果你在和系统协作的话。我认为对于快速系统有大量需求,这些系统可能不是最前沿的,就像你说的,大概 95%、90%,但这
[11:28]plenty good enough and actually gain back more than the 10% on the the iteration speed. So, and then the other big thing I think is running these things on the edge again for efficiency reasons, but also for privacy and security reasons, too. If you think about different devices that you might 完全足够好了,而且实际上通过迭代速度可以赢回超过那 10% 的差距。另外我认为另一件大事是在边缘端运行这些模型,同样是为了效率,但也为了隐私和安全。如果你想想你可能
[11:45]run these systems on that that, you know, process very personal information, can also think about robotics as well. 运行这些系统的各种设备,这些设备处理非常个人化的信息,也可以想想机器人领域。
[11:51]You know, robots in your house. I think you're going to want very efficient, very powerful local models, which maybe are orchestrated, you know, with some bigger models, frontier models that are in the in the cloud, but you only delegate to that in certain 你知道,家里的机器人。我认为你会需要非常高效、非常强大的本地模型,也许它们会和云端中一些更大的前沿模型进行协作编排,但你只在特定
[12:07]circumstances. And perhaps you, you know, you process all of the audio visual feed, let's say, locally and that stays local. I could imagine that would be a very good sort of end state. Y Combinator School is back. We're hand-selecting the most promising 情况下才会委托给云端。而且也许你,你知道,你在本地处理所有的音视频流,那些数据都留在本地。我可以想象那会是一个非常好的最终形态。Y Combinator School 回来了。我们正在精心挑选全球最有潜力的
[12:22]builders in the world and flying them out to San Francisco for July 25th and 26th to discuss the cutting edge of tech. Apply now for a spot. Okay, back to the video. Going back to context and memory, models currently stateless, but, you know, continue like what would the Builder,飞到旧金山参加 7 月 25 日和 26 日的活动,讨论最前沿的科技。现在就申请名额吧。好了,回到视频。回到上下文和记忆的话题,模型目前是无状态的,但是,你知道,比如
[12:39]developer experience even be like for someone who's using a continual learning model? Like, you know, any idea like how you'd steer it? I think it's really interesting. I think that's one of not having continual learning currently is one of the things holding back agents from doing full uh tasks. You know, I 对于使用持续学习模型的开发者来说,体验会是什么样的?比如,你知道,你有没有想过怎么驾驭它?我觉得这真的很有趣。我认为目前缺乏持续学习是阻碍 Agent 完成完整任务的瓶颈之一。你知道,我
[12:56]think they're really useful for aspects of tasks right now and you can patch them together and do some really cool things, but they don't adapt well with the context that you're in. And I think that's the missing piece for them being really kind of fire and forget and 认为它们目前对任务的某些方面确实很有用,你可以把它们拼接起来做一些很酷的东西,但它们不太能适应你所处的具体上下文。我认为这是它们真正实现那种「发射后不用管」能力的缺失拼图,
[13:11]they'll figure it out themselves. You know, I think they need to be able to learn about the specific context that you're going to put them in. So, I think we have to crack that to get full general intelligence. Where are we on reasoning? So, models can do really 它们自己就能搞定一切。你知道,我认为它们需要能够学习你将它们放入的具体上下文。所以我认为我们必须攻克这个问题才能实现真正的通用智能。推理方面我们处于什么阶段?所以,模型现在可以做到非常
[13:28]impressive chain of thought now, but they still fail on things a smart undergrad wouldn't. What specifically needs to change and what progress do you expect in reasoning? There's a lot of innovation left in in think the thinking paradigms, I would say. Again, I think we're fairly we're doing fairly 令人印象深刻的思维链,但它们仍然会在一个聪明的大学生不会出错的地方犯错。推理方面具体需要改变什么?你预期推理方面会有什么进展?在思考范式方面还有很多创新空间,我想这样说。同样,我认为我们目前做的还相当……我们做的还比较
[13:44]simplistic things, fairly brute force. One could imagine I think there's a lot of scope, for example, in monitoring the chain of thought, maybe interjecting midway through a thought process. I often get the impression with our systems and and 简单化,相当粗暴。可以想象,我认为有很大空间,比如,监控思维链,也许在思考过程中途介入。我经常对我们的系统和
[13:59]our competitor systems that they're almost overthinking. They're almost getting into sort of loops of things. 竞争对手的系统有一种感觉,它们几乎在过度思考。它们几乎陷入了某种循环之中。
[14:05]Like, one thing I sometimes like to do is is play chess against Gemini. And, you know, it's the all the leading foundation models are pretty poor at games, which is quite interesting. It's very cool to kind of look at the thinking traces cuz obviously these are can be a 比如,我有时候喜欢做的一件事就是和 Gemini 下国际象棋。你知道,所有领先的基础模型在游戏方面都相当差,这其实挺有趣的。看思维链的痕迹非常酷,因为这些是可以被
[14:20]well-understood, you know, I can tell quite quickly if it's going off on a tangent and it's very sort of provable what the what the the thinking is doing, whether it's useful or not. And so, what we see is that, you know, sometimes it will it will it will consider a move, it 很好理解的,你知道,我可以很快判断它是否在走偏,而且它的思考过程是否有效是可以被验证的。所以我们看到的是,你知道,有时候它会考虑一步棋,
[14:36]will realize it's a blunder, but it can't find anything better, so it kind of goes back to that move and does it anyway. So, it, you know, you just shouldn't be seeing that happening in a in a very precise reasoning system. So, there's just sort of huge gaps I think still, but it may 它会意识到这是一步臭棋,但它找不到更好的选择,所以它又回到那步棋然后还是走了那步。所以,你知道,在一个非常精确的推理系统中你不应该看到这种情况发生。所以我认为仍然存在相当大的差距,但可能
[14:52]only be one or two tweaks that are required to fix those kind of gaps, just to be clear. But I think that's pretty pretty obvious they are there. And that's why you get this kind of jagged intelligence. You know, on the one hand, it can solve gold medal problems in IMO, which is super hard, but on the other 只需要一两个调整就能修复这类差距,澄清一下。但我认为这些差距的存在是相当明显的。这就是为什么你会看到这种参差不齐的智能。你知道,一方面,它能解决 IMO 的金牌问题,这是超级难的,但另一方面
[15:08]hand, as we've all seen, it can still make basic elementary maths errors if you pose the question in a certain way, right? So, or elementary reasoning errors. So, there's just something to me about the almost an introspection about its own thought process that I feel like 另一方面,正如我们都看到的,它仍然会在某些情况下犯基础的小学数学错误,对吧?或者说基础的推理错误。所以我觉得,关于它对自己思维过程的一种内省能力,好像还缺点什么。
[15:24]there's there's something maybe missing there. Agents are really big. Some would say they're hyped. I personally think they're just getting started. It's [laughter] totally insane. What does DeepMind's internal research tell you about where agent capabilities actually are right now versus, you know, sort of the hype out there? I think we all I 好像确实还缺了点什么。Agents 现在真的很火。有人可能会说这是炒作。我个人认为才刚刚开始。[laughter] 这太疯狂了。DeepMind 内部的研究告诉你,Agent 的实际能力现在到底在什么水平?跟外面那些炒作相比怎么样?我觉得我们——我
[15:40]agree with you. I think we're just at the beginning. You have to have an active system that can actively solve problems for you to get to AGI. That was always clear to us. So, agents are that path. And I think we're just getting going. I think all of us are getting used to how do we best work and you're 同意你的看法。我觉得我们才刚起步。你必须有一个主动的系统,能够主动替你解决问题,才能达到 AGI。这一点我们一直很清楚。所以 Agent 就是那条路。我觉得我们才刚刚开始发力。我觉得我们所有人都在摸索——怎么最好地和它协作,而你
[15:56]leading the way in a lot of this in your own personal experiments. I'm sure many of you are doing that. I think how do you incorporate it into your workflow in a way that isn't just sort of a nice to have, but actually starting to do fundamental things. My impression is at the moment we're all experimenting 在很多方面都是走在前面的,在你自己的个人实验中。我相信你们很多人也在做类似的事情。问题是怎么把它融入你的工作流,不是那种锦上添花的东西,而是真正能做一些根本性的事情。我的感觉是,目前我们都在实验
[16:12]we're experimenting a lot of things, but we're only in the maybe the last couple of months starting to find the really valuable places. And the technology is probably only getting good enough for that to be the case, right? Where that it's not a kind of toy, nice demonstration, but actually really adding value to your to your to your 我们在尝试很多东西,但可能最近几个月才开始找到真正有价值的场景。而且技术可能也只是刚刚好到那个程度,对吧?不再是那种玩具式的、炫酷的演示,而是真正能为你的时间、你的效率带来价值的。
[16:29]time and efficiency. I'd often wonder I see a lot of people working on like setting off, you know, dozens of agents for like 40 hours, but I'm not sure I've seen the output that yet of that quite justify that level of input 我经常在想,我看到很多人让几十个 Agent 跑四十个小时,但我不确定我看到的产出已经能证明那个投入是值得的。
[16:45]going in. But I think it will come. So, I still think we're in the experimentation phase. We haven't seen a AAA game that tops the app store charts that was sort of live coded yet, right? 但我觉得这一天会来的。所以我们仍然处在实验阶段。我们还没有看到一个 AAA 级游戏登上应用商店排行榜,是完全靠实时编码做出来的,对吧?
[16:56]I've seen and I've programmed and I'm sure many we've all done little nice demonstrations and it's like amazing I can do a prototype of theme park in half an hour now, which took me 6 months back when I was 17. It's kind of mind-blowing. And I and I wish I I got this feeling if I spent the whole summer 我见过,我也写过代码,我相信在座很多人——我们都做过一些漂亮的小演示,确实很惊艳,现在我半小时就能做一个 Theme Park 的原型,而当年我 17 岁的时候花了 6 个月。这确实很震撼。而且我觉得——我有一种感觉,如果我整个夏天
[17:12]working on it, you could make something really incredible, but it still needs craft and, you know, human sort of soul into it and taste. I think that's that's something that can that's you have to make sure you still bring that to to whatever is your building. And I think 都投入进去,可能真的能做出一些了不起的东西,但它仍然需要工艺,需要人的灵魂和品味融入其中。我觉得这一点——你必须确保你在构建任何东西的时候仍然注入这些。
[17:27]it still shows like it's not quite there yet because why haven't we seen a kid making a hit game that's that sells 10 million copies, right? That should be possible given the effort that's gone in. So, something's still somehow missing. Maybe it's to do with the process or maybe it's to do with the 而且我觉得,现在还是能看出差距的,因为为什么我们还没有看到一个小孩做出一款销量一千万的爆款游戏呢?考虑到大家投入的精力,这应该是可能的。所以某些东西还是缺失了。也许是流程的问题,也许是
[17:43]tools. I'm not quite sure. You all will probably know better than me cuz I'm sure you're all experimenting on that, but I haven't seen the result yet which I would expect once this is really delivering that full value. Which I think will come in the next 6 to 12 months. Some of it is like how much of it will be autonomous versus I mean, I 工具的问题。我不太确定。在座的各位可能比我更清楚,因为你们肯定都在做这些实验,但我还没有看到那个结果——那种我期望中的、真正实现完整价值的结果。不过我觉得接下来 6 到 12 个月内会出现。还有一部分问题是,有多少会是自主完成的,对比——我
[18:00]don't think we'd see autonomous first. We would actually probably see people in this room operating at 1,000 x and then >> That's what you should see first. And then many of you, you know, they'll be like games companies or, you know, other types of companies that have built some 觉得我们不会先看到全自主。我们实际上可能先看到的是这个房间里的人效率提升到 1000 倍,然后 >> 那才是你应该先看到的。然后你们中的很多人,你们会做出游戏公司或者其他类型的公司,用这些工具做出一些
[18:16]kind of best-selling app, best-selling game using these tools. That's what you should see first. And then more of that will get automated. I mean, some of it is like there's a human in there and then the human doesn't want to say that the the the agents did it yet. I think 畅销应用、畅销游戏。那才是你应该先看到的。然后越来越多的部分会被自动化。我的意思是,这里面有人的参与,而那个人可能还不太愿意说这是 Agent 做的。我觉得
[18:32]part of it might be though that um this we want to discuss like creativity. What I often say about that is like if we look at the things we've done like AlphaGo. So, obviously very famously, you'll know about the move 37 in game two. And for me, I was waiting for a 部分原因可能是——我们想聊聊创造力这个话题。我经常举的例子是我们做过的 AlphaGo。大家都知道,非常著名的第二局的第 37 手。对我来说,我在等一个
[18:47]moment like that to start the science projects like AlphaFold. We started AlphaFold like the day we got back from Seoul, which is 10 years ago now. So, I'm going to Korea after this to celebrate the 10-year anniversary of AlphaGo. But it's not enough to come up with move 37. Like, that's pretty cool, 那样的时刻来启动科学项目,比如 AlphaFold。我们从首尔回来那天就启动了 AlphaFold,到现在已经 10 年了。所以我接下来要去韩国庆祝 AlphaGo 的十周年。但仅仅走出第 37 手是不够的。确实很酷,
[19:04]very useful, but can it invent Go? That's what I I want a system that can invent Go if you give it a high-level description, you know, like a game you can learn the rules of in 5 minutes, but it takes many lifetimes to master. It's 也很有用,但它能发明围棋吗?这才是我想要的——一个能发明围棋的系统,如果你给它一个高层次的描述,比如:一个你五分钟就能学会规则,但需要很多辈子才能精通的游戏。它
[19:19]beautiful aesthetically, but you can play it in a few hours in an afternoon. So, you know, maybe you could imagine that would be the high-level description I would give and then I'd want the the return, the thing I get back is Go. 在美学上很美,但你一个下午几个小时就能玩。所以,你可以想象我会给出这样一个高层次描述,然后我期望得到的回报是——围棋。
[19:33]Right? And clearly today's systems I think can't do that. So, the question is why? 对吧?但显然今天的系统还做不到。那问题是为什么?
[19:40]And I think there's something still missing there. Well, someone in this room might might make it. >> Then the answer would be there's nothing missing. It just was the way we were using the systems. And that might actually be the answer. It might be that our today's systems are capable of that with a brilliant enough creative person 我觉得还缺了点什么。不过,这个房间里可能有人会做出来。>> 那答案就是什么都不缺,只是我们使用系统的方式不对。这可能就是答案。也许今天的系统其实已经有这个能力了,只要你有一个足够有创造力的天才
[19:55]using it and providing that impetus that the soul of the project and being able to probably being O'Fei enough with the tools to like almost be at one with the tools. I could imagine that would be happening if you experimented with the tools all day and 在使用它,给它注入项目的灵魂,并且对工具有足够的熟悉度,几乎达到人机合一的境界。我可以想象这种事情会发生,如果你整天
[20:11]all night like probably many of you are doing that and you combine that with proper deep creativity. Um something you know more incredible could be done. 整夜地实验这些工具——可能你们中很多人就是这样——然后你把这种经验和真正的深层创造力结合起来,可能就会做出更不可思议的东西。
[20:19]Switching gears to open source. I mean or open open and open weights. I mean the recent release of Gemma you're making highly capable open and accessible ones that can actually run locally. What do you think that means for you will AI be something that is in 换个话题,聊聊开源。我是说开源和开放权重。最近发布的 Gemma,你们做出了非常强大、开放、可本地运行的模型。你觉得这意味着什么?AI 会不会变成
[20:35]the hands of the users instead of primarily in the cloud and does that change who gets to you know build with these models? We're huge proponents of in general of open source and open science and you mentioned AlphaFold at the beginning, you know, we put that all out there for free and all of our 掌握在用户手中的东西,而不是主要在云端?这会不会改变谁能用这些模型来构建东西?我们一直大力支持开源和开放科学,你一开始提到了 AlphaFold,我们把它全部免费开放了,我们所有的
[20:51]science work even still today we publish in you know the big journals. We wanted to create world-leading models for their their sizes. Right and so that's what we hopefully we've done with Gemma and we're you know very committed to that path and hopefully you all experiment and build and and enjoy using Gemma. I 科学工作至今仍在顶级期刊上发表。我们想打造同级别中世界领先的模型。所以我们希望 Gemma 做到了这一点,而且我们会持续投入这条路,希望大家都能去实验、构建、享受使用 Gemma。
[21:08]think it's been like 40 million downloads now and uh just in you know two and a half weeks. So we're really excited about that and I also think it's important for there to be Western stacks on open source, you know obviously a lot of the Chinese models are excellent and and they're currently well well leading 我想现在已经大概 4000 万次下载了,仅仅两周半的时间。所以我们非常兴奋。另外我觉得很重要的一点是,在开源领域要有西方的技术栈,你知道,显然很多中国模型都非常优秀,而且目前在开源方面确实领先
[21:24]in open source and we think Gemma's very competitive for its sizes in in all those respects. And for us I mean there is a question of resources, talent and compute like nobody has enough spare compute to just make two you know uh frontier models at maximum size, 我们认为 Gemma 在它的尺寸级别上各方面都非常有竞争力。对我们来说,确实存在资源、人才和算力的问题,没有人有足够的闲置算力去同时做两个——你知道,最大规模的
[21:41]right with different attributes. So that's pretty difficult. But also what for now what we've we've decided is that our edge models, the things we want to use for Android and glasses and robotics, um it's best that they're open models because they're vulnerable anyway 前沿模型,对吧?带有不同属性的。这确实很难。但目前我们决定的策略是,我们的边缘模型——用于 Android、眼镜和机器人的那些——最好是开放模型,因为反正它们一旦部署到设备上就容易泄露。
[21:57]on the once you put them out on the surfaces. So they might as well be actually fully open. Right? So we've sort of made a decision to kind of unify that at the at the kind of we call it nano size level. So that actually works for us 所以还不如干脆完全开放。对吧?所以我们做了个决定,在我们称之为 nano 级别的层面统一处理。所以这对我们来说实际上
[22:12]uh strategically as well. Um and you know we hope as many people as possible build on it and of course we'll be building on that, too. Earlier uh before we came on I got to show you a demo of my version of Samantha from Her, which is >> Yes. uh harrowing for me to try to demo something to you. Yeah, very good. Um 在战略上也是合理的。而且我们希望尽可能多的人基于它来构建,当然我们自己也会在上面继续建设。刚才在我们上台之前,我给你展示了一个我的 Samantha 的演示——就是《Her》里的那个。>> 是的。嗯,在你面前演示东西对我来说还挺紧张的。是的,效果很好。
[22:29]and it worked, which is amazing. Gemini was built multimodal and I spent a lot of time with a bunch of the models and Mhm. I mean the depth of the context and the tool use with speech directly to model like there's nothing like bar none like the best one actually. 而且居然成功了,太棒了。Gemini 从一开始就是多模态构建的,我花了很多时间跟各种模型打交道。嗯。我的意思是,那个上下文深度,加上直接语音到模型的工具调用——没有任何其他模型能比,毫无悬念,就是最好的。
[22:44]>> Yeah. Yeah, I think I think that's the sort of still a slightly underappreciated aspect of of of the Gemini series is we we started it being multimodal from the start. That made it a little bit more difficult actually to begin with cuz then just focusing on text for example. But I we believe we're >> 对。对,我觉得 Gemini 系列仍然有一个被稍微低估的方面,就是我们从一开始就把它做成了多模态的。这在一开始确实增加了不少难度,因为不能只专注于文本。但我们相信
[23:00]going to gain from that in the long run. And I think we're seeing that now for things like world model building so stuff like Genie that we build on top of Gemini. I think it's going to be really important for things like robotics. So this is why Gemini robotics which many 长期来看这会带来回报。我觉得我们现在正在看到这一点,比如世界模型构建——像我们在 Gemini 之上构建的 Genie 那样的东西。我觉得这对机器人领域会非常重要。这就是为什么 Gemini Robotics——你们很多人
[23:15]of you probably played around with. I think it's going to be built on multimodal foundation models, the robotics models and we think we have a sort of competitive advantage with with Gemini being so strong at multimodal. 可能都玩过了。我认为机器人模型会建立在多模态基础模型之上,而 Gemini 在多模态方面如此强大,我们认为这是一个竞争优势。
[23:27]We're using it increasingly in things like Waymo but also if you imagine devices and assistants that digital assistants that come with you into the real world, you know, maybe on your phone or glasses or some other device. It needs to understand the physical world around you 我们在 Waymo 这样的项目上越来越多地使用它,另外如果你想象一下那些跟着你进入现实世界的设备和数字助手——可能在你的手机、眼镜或其他设备上——它需要理解你周围的物理世界
[23:44]and intuitive physics and and the and the physical context you're in and that's what our systems are extremely good at and I think you found that's why you've enjoyed using it in your setup. 以及直觉物理和你所处的物理环境,这正是我们的系统非常擅长的,我觉得你也在使用中体会到了这一点。
[23:54]We're planning to continue on that and I think we're far and away the strongest models on on those types of problems. So the cost of inferences are dropping fast. What becomes possible when inference is essentially free and how does that change what your team is actually 我们会继续在这方面投入,我认为在这类问题上我们的模型遥遥领先。推理成本在快速下降。当推理本质上免费的时候,什么会成为可能?这会怎样改变你的团队实际
[24:09]optimizing for? Yeah, I'm not sure inference will ever be essentially free. I mean there's sort of Jevons paradox and other things about like I think we'll just end up using all of us will end up using whatever we can get our hands on and you could imagine 优化的方向?嗯,我不确定推理会不会真的变成免费的。有杰文斯悖论之类的——我觉得我们所有人最终都会把手头所有能用的算力都用了。你可以想象
[24:26]millions of agents, swarms of agents working together on things. So that's one way to use the inference or you could imagine single agents or smaller groups of agents thinking for in multiple directions and then ensembling that. So we're experimenting with all these things. Probably many of you are. All of 数百万个 Agent,成群的 Agent 协同工作。这是使用算力的一种方式;或者你也可以想象单个 Agent 或较小规模的 Agent 组合,往多个方向思考,然后做集成。我们在实验所有这些东西,你们很多人可能也是。
[24:43]that will use up any inference I think that's available. I mean one day maybe it can be almost cost zero, certainly the energy if we solve fusion or you know superconductors or you know optimal batteries or some set of those things which I think we will do with material 所有这些会消耗掉任何可用的推理算力。我的意思是,也许有一天成本可以接近零——当然,如果我们解决了核聚变,或者超导体,或者最优电池,或者其中一些技术,我觉得我们用材料
[24:58]science. Energy costs will be essentially zero but they'll still be the physical creation of the chips and other things. There'll some there'll be some bottleneck um at least for the next few decades I think. And so if that's the case, there'll still be rationing on 科学能做到。能源成本将趋近于零,但芯片的物理制造等方面仍然存在成本。至少未来几十年还是会有某种瓶颈。所以如果真的是这样,推理端仍然需要
[25:14]the inference side. You still have to use it I think efficiently. Yeah. Well, luckily the smaller models are getting smaller and smarter, which is fantastic. 配给。你仍然需要高效地使用它。是的。幸运的是,小模型正在变得越来越小、越来越聪明,这太棒了。
[25:21]Uh we got a lot of bio and biotech founders in the audience. I can see a few. AlphaFold 3 took us beyond proteins to a broad spectrum of biomolecules. Uh how close are we to modeling full cellular systems or is that still a fundamentally harder problem in a class 我们观众里有很多生物和生物科技的创始人。我能看到几位。AlphaFold 3 把我们从蛋白质带到了更广泛的生物分子。我们离建模完整的细胞系统有多近?还是说那仍然是一个本质上更难的问题,属于完全不同的
[25:37]of its own? Well, I Isomorphic Labs which we spun out from from from from DeepMind after we did AlphaFold 2. 级别?嗯,Isomorphic Labs 是我们在完成 AlphaFold 2 之后从 DeepMind 分拆出来的。
[25:45]Um it's it's which is going amazingly well. It's it's it's trying to build out uh not just AlphaFold. It's just one piece of the drug discovery process uh as many you know, but we're trying to do the the adjacent biochemistry and chemistry to design the right compounds 它的发展非常好。它不只是 AlphaFold。AlphaFold 只是药物发现流程中的一个环节,你们很多人都知道,但我们在做的是相关的生物化学和化学工作,来设计具有
[26:00]with the right properties and so on. We'll have some big announcements for you know very soon to talk about on on that front. I think that's going really well. Eventually you want a whole virtual cell. So I've talked about this in many of my science talks about a full working simulation of a cell that you 正确属性的化合物等等。我们很快会有一些重大公布,来聊聊这个方向的进展。我觉得一切进展得很顺利。最终你想要的是一个完整的虚拟细胞。我在很多科学演讲中都谈过这个——一个可以完整运行的细胞模拟,你可以
[26:16]can perturb and then the you know the the outputs of that would be close enough to experimental that it's useful. 对其进行扰动,然后输出的结果要足够接近实验数据,才有实用价值。
[26:23]Right? You could skip out a lot of the the search steps and generate lots of synthetic data to train other models that then would predict things about you know real cells. And um I think we're about 10 years away probably from something like a virtual cell, like a 对吧?你可以省掉很多搜索步骤,生成大量合成数据来训练其他模型,然后这些模型就能预测真实细胞的情况。我觉得我们大概离虚拟细胞还有 10 年左右,一个
[26:38]full virtual cell. You know, we're starting out this is we're working on the DeepMind side, science side on a you know virtual nucleus. Cell nucleus first cuz relatively self-contained. The trick with all of these things is can you pick uh a slice of the complexity, you know, eventually you want to want to model a 完整的虚拟细胞。你知道,我们正在起步——在 DeepMind 的科学这边,我们首先在做的是虚拟细胞核。因为细胞核相对来说比较自包含。这些事情的关键在于,你能不能从复杂性中切出一块来,你知道,最终你想要建模一个
[26:54]human body, but can you model it down to the right level of detail and what slice can you uh take out of it that will be self-contained enough you can kind of model and approximate the inputs and outputs into that self-contained system and then just focus on the 人体,但你能不能把它建模到合适的精度级别?你能切出哪一块,使得它足够自包含——你可以建模并近似它的输入输出——然后专注于那个
[27:10]self-contained system. So a nucleus is quite interesting from that perspective. Um then the other issue is just there's not enough data yet. So you need data uh and I talked to various you know top scientists about who work on electron microscopes and other imaging things. If 自包含的系统。从那个角度来看,细胞核很有意思。另一个问题是,目前数据还不够。你需要数据——我跟很多顶级科学家聊过,他们做电子显微镜和其他成像技术。如果
[27:27]we could image a live cell without killing the cell, that would be um game-changing obviously cuz then you could convert it into a vision problem which we would know how to solve. Right? 我们能在不杀死细胞的情况下对活细胞成像,那将是颠覆性的,因为那样你就可以把它转化成一个视觉问题,而这是我们擅长的。对吧?
[27:37]And but at the moment there are at least I'm not aware of any techniques that can give you a kind of you know nanometer resolution uh but without destroying but it in you know in a live dynamic cell. So you can see all the interactions. Right? You can take static images at that resolution 但目前,至少据我所知,还没有什么技术能在纳米级分辨率下——在不破坏细胞的情况下——观察一个活的、动态的细胞,看到所有的相互作用。你在那个分辨率下可以拍静态图像
[27:53]obviously um really detailed now and that's quite exciting. But it's not enough uh to turn it just into just into a complex vision problem. 当然,现在已经可以拍得非常精细了,这确实很令人兴奋。但这还不够,不足以把它转化成一个复杂的视觉问题。
[28:03]So that's one way it could be solved. So it could be a hardware driven data driven solution or it could be that we build better uh learn simulators of uh these dynamical systems. So that's that's the more modeling way of solving it. Uh you've been looking at all kinds 所以这是解决方案的一种可能路径。可以是硬件驱动的数据解决方案,也可以是我们构建更好的可学习的动态系统模拟器。那是更偏建模的解决方式。你还研究了很多其他
[28:18]of science and not just bio. Uh there's material science, drug discovery, climate modeling, mathematics. If you had a rank which scientific domain will transform the most dramatically the next 5 years, what's in your list? 科学领域,不只是生物。材料科学、药物发现、气候建模、数学。如果要你排名,未来 5 年哪个科学领域会发生最戏剧性的变革,你的名单是什么?
[28:30]>> all so exciting and that's why I mean that that for me has been my main passion and always the reason why I've worked on AI for my whole career for 30 plus years now is to use AI as the ultimate tool. I always thought AI would be the ultimate tool for science and to invite such advanced scientific >> 全都太令人兴奋了,这就是为什么——我的意思是,那一直是我最大的热情所在,也是我整个职业生涯从事 AI 工作的原因——三十多年了——就是把 AI 当作终极工具。我一直认为 AI 会是科学的终极工具,用来推动如此先进的科学
[28:47]understanding, scientific discovery and things like medicine and just our understanding of the universe around us. 理解、科学发现,以及医学等领域,还有我们对周围宇宙的理解。
[28:53]So actually when you mentioned our original way we used to articulate our mission statement, which is still the way we think about it is there was two steps to it. One was step one was solve intelligence, I build AGI and then step two was use it to solve everything else. 实际上,当我们提到我们最初表述使命的方式——这也是我们至今仍在用的方式——分两步。第一步是解决智能,构建 AGI;第二步是用它来解决其他所有事情。
[29:06]We had to change that a bit over time cuz people were like, "Do you really mean solve everything else?" And we did mean that and I think people are sort of understanding what that means today. But specifically I was meaning solve other what I call root node problems in science. So areas of science that would unlock whole new branches or avenues of 后来我们不得不稍微调整了一下说法,因为人们会说,'你真的要解决所有其他事情吗?'我们确实是这个意思,而且我觉得人们现在开始理解那意味着什么了。具体来说,我的意思是解决科学中我称之为'根节点问题'的东西——就是那些能解锁全新分支或
[29:23]discovery. And AlphaFold is the prototypical example of what we want to do. So over 3 million researchers around the world, pretty much every biology researcher in the world uses AlphaFold now. And I was told by some of my you know farmer executive friends that you 发现路径的科学领域。AlphaFold 就是我们想做的事情的典型例子。全世界有超过 300 万研究人员在使用 AlphaFold,几乎地球上每一个生物学研究者都在用它。我一些制药行业高管朋友告诉我,
[29:38]know almost every drug discovered from now on will have used AlphaFold at some point in its in the drug discovery process. So that's something we're very proud of and it's the sort of impact that we hope to have with with AI. But, I do think it's just the beginning. I I 从现在开始,几乎每一种新药在药物发现过程中的某个环节都会用到 AlphaFold。这是我们非常自豪的事情,也是我们希望通过 AI 产生的那种影响力。但我觉得,这只是开始。
[29:54]don't really see any area of science or engineering that this won't be able to help be helpful with. And the ones you mentioned, I think we're almost like an AlphaFold one moment. So, it's we've got very promising results, but it's not quite solved the the grand challenge yet in that domain. But, I think we're going 我真的不觉得有任何科学或工程领域是 AI 帮不上忙的。你提到的那些领域,我觉得我们差不多都处在一个 AlphaFold 1 的阶段——我们已经有了很有希望的结果,但还没有完全解决那个领域的终极挑战。不过我认为我们会
[30:10]to have a lot to talk about in the next couple of years on all those areas you mentioned, materials, which I I think is very exciting, all the way to mathematics. In in science, I mean, it feels Promethean. It's like here is this capability and I think so. I mean, of course, along with that include 未来几年我们在你提到的所有领域都有很多可以聊的——材料科学,我觉得非常令人兴奋,一直到数学。在科学领域,我的感觉是它有种普罗米修斯式的意味,就像我们拥有了这样一种能力。当然,我的意思是,伴随着这种能力而来的还有
[30:26]including what the the the parable of Prometheus, we have to also be careful with how we use that and what we use it for and also the misuse that can happen with those same tools. A lot of people in this room are trying to build companies applying AI to science. For them, what's the difference between a 包括普罗米修斯的寓言所警示的那样,我们必须谨慎地使用这些能力,注意用途,以及这些工具可能被滥用的风险。在座的很多人都在尝试创办将AI应用于科学的公司。对他们来说,一个
[30:41]startup that actually advances the frontier in your view versus one that's just wrapping an API around a foundation model and calling it AI for science? 真正推动前沿的创业公司,和那种只是把基础模型的API包一层就号称做AI for Science的公司,在你看来区别在哪里?
[30:49]Well, look, I think there's one of the things I would recommend I'm trying to think about and I think you mentioned this to me before. What would I do today myself if I was sitting in your place and Y Combinator, you know, looking at things? One thing you have to do is obviously intercept where the AI tech is going. So, that's one hard part of it. 嗯,我觉得有一件事我会推荐——我也在想这个问题,我记得你之前也跟我提过。如果我今天坐在你的位置上,在Y Combinator看项目,我会怎么做?首先你必须要能判断AI技术的发展方向,这本身就是一个难点。
[31:06]But, I do think there's huge scope for combining where AI is going with some other deep technology area. I just think that that's sweet spot is is whether it's materials or medicine or other really hard areas of science. 但我确实认为,将AI的发展方向与某个其他深层技术领域结合起来,空间是巨大的。我觉得最佳切入点就在那里——不管是材料、医学,还是其他真正有难度的科学领域。
[31:21]I think that those kinds of interdisciplinary teams, especially if it involves the world of atoms as well, there's not going to be a shortcut to that at least in the foreseeable future. 我认为这种跨学科的团队,尤其是涉及原子世界的那些,至少在可预见的未来不会有什么捷径。
[31:31]Those areas that are pretty safe from just getting swamped by whatever the next update is to the foundation models. 这些领域相当安全,不会轻易被基础模型的下一次更新所淹没。
[31:38]So, I think if you're looking for things like that, that's one of the more defensible areas, I would say. And I've always loved deep tech, so I'm kind of biased towards deep tech things. I think nothing that's really long-lasting and worthwhile is easy. And so, I'm always 所以如果你在寻找这样的方向,我认为这是最具防御性的领域之一。我一直都很喜欢深科技,所以我对深科技是有些偏好的。我觉得真正持久、有价值的东西没有一个是容易的。所以我一直
[31:54]been drawn to to deep technologies. Obviously, AI was like that back in 2010 when we started out, right? It was it was thought to just we know we know it doesn't work kind of thing is what I was told by investors and even in academia, it was considered to be a very niche 被深科技所吸引。显然,AI在2010年我们刚起步的时候也是这样,对吧?当时人们觉得,我们知道这东西不行——投资者就是这么跟我说的。甚至在学术界,它也被认为是一个非常小众的
[32:10]subject that we sort of tried in the '90s and we know doesn't work. But, if you, you know, if you have belief and conviction in your idea why it's different this time or what special combination from your background that you had, ideally you're expert in both those areas, both the machine learning 领域,就是那种我们在90年代已经尝试过了、知道行不通的东西。但是,如果你对你的想法有信念和信心——为什么这次不一样,或者你背景中有什么特别的组合——最理想的情况是你在两个领域都是专家,既是机器学习
[32:26]and the other area you're applying it to, or you can create a founding team with that expertise, I think there's huge impact to be made there and huge value to be built there. That's a really important message. I mean, even I mean, it's it's easy to forget. Like basically, once you've done it, you've done it. But, before you've done it, 也是你要应用的另一个领域,或者你能组建一个具备这种专业知识的创始团队,我认为那里可以产生巨大的影响、创造巨大的价值。这是一个非常重要的信息。我的意思是,甚至……说起来很容易忘记——基本上,一旦你做成了,你就做成了。但在你做成之前,
[32:42]people are arrayed against you. Oh, sure. That mean, no one believes in it, which is why I think you've got to you've also got to work in things that you're genuinely passionate about. Like for me, I would have worked on AI no matter what happened. I just decided from a very young age it was the thing 所有人都在反对你。哦,当然。意思就是,没有人相信你。所以我认为你必须做自己真正有热情的事情。对我来说,无论发生什么,我都会做AI。我很早就认定了这是
[32:58]that could be the most consequential thing I could think of. It's turned out that way, but it might not have. Maybe we would have been 50 years too early. And it was also the most interesting thing I could think of working on. And so, I would have still be working on AI today even if we were still, you know, in a 我能想到的最具深远影响的事情。结果确实如此,但当初也可能并非如此。也许我们早了50年。而且这也是我能想到的最有趣的工作方向。所以即使今天我们仍然,你知道,待在某个
[33:14]little garage somewhere and it still wasn't quite working. I would have still been trying to find maybe I'd have been back in academia or something, but I would have found some way of of continuing to work on it. So, I mean, AlphaFold was like an example of a spike that you pursued and it worked. You know, what makes a scientific domain 小车库里面,AI还是不太行,我依然会继续尝试——也许我会回到学术界什么的,但我会找到某种方式继续做下去。所以说,AlphaFold就是你坚持追求并取得成功的一个典型案例。你觉得什么样的科学领域
[33:30]ripe for an AlphaFold style breakthrough? And is there a pattern, a certain objective function, like The way I I'm I should write this up at some point when I have 5 minutes spare, but the lesson I've learned from all the Alpha projects we've done, specifically AlphaGo and AlphaFold, is 已经成熟到可以迎来AlphaFold式的突破?是否存在某种模式、某种特定的目标函数?比如……我应该在有空的时候把这个写下来,但我从我们做过的所有Alpha项目中学到的经验,特别是AlphaGo和AlphaFold,就是
[33:47]I think the techniques we have and the problems I look like to look for are great in if the if the situation can be described as massive combinatorial search space. The more massive, the better in some ways. So, no brute force or special case algorithm will will solve it. And that's true of Go moves 我认为我们现有的技术和我喜欢寻找的问题,最大的特点就是:如果一个问题可以被描述为一个巨大的组合搜索空间,那就非常适合。某种程度上越大越好,这样暴力搜索或者特制算法都解决不了。围棋的落子就是如此
[34:04]and of, you know, different configurations of proteins, far more than the atoms in the universe, both of those. And then, you have a clear objective function. So, you know, you can think of it as minimizing the free energy in the proteins or, you know, the winning the game of Go. So, you need to be able to you need to specify your 蛋白质的不同构型也是如此,两者的可能性都比宇宙中的原子还多。然后,你需要有一个清晰的目标函数。比如,你可以把它理解为最小化蛋白质的自由能,或者赢得围棋比赛。所以你需要能够明确地定义你的
[34:20]objective function clearly so you can hill climb. And then, enough data and or simulator that can generate you lots of in-distribution synthetic data. If those things are true, then I think with today's methods, 目标函数,这样你才能不断攀登优化。然后,你还需要足够的数据,或者能够生成大量分布内合成数据的模拟器。如果这些条件都满足,那么我认为用今天的方法,
[34:37]you can go a long way into tackling and finding the kind of needle in the haystack that you need to for the solution that you're trying to look for. 你就可以在很大程度上解决并找到你所需要的那个大海捞针般的解决方案。
[34:44]And I think just drug discovery, by the way, in the same way. Right? There is a compound out there that would solve this disease if one could find it. If one could only find it. Right? And that wouldn't have any side effects and so on. And as long as the laws of physics allows it, then the only question is how 顺便说一下,药物发现也是同样的道理,对吧?一定存在某种化合物能治愈某种疾病——只要能找到它。只要能找到就好了。而且没有副作用等等。只要物理定律允许,那么唯一的问题就是如何
[35:01]do you find it in an efficient way, in a tractable way? I think we showed for the first time, actually, with AlphaGo that these systems could find those kinds of needles in the haystack. In that case, you know, the perfect Go move. I guess to get a little meta, I mean, we've we're talking about humans using these 以高效的方式、可行的方式找到它。我认为我们实际上第一次证明了——通过AlphaGo——这些系统能够找到那种大海捞针般的答案。在那个例子中,就是完美的围棋落子。如果往更高一层想,我们一直在谈论人类利用这些
[35:17]methods to create AlphaFold, but then there's a meta level, which is humans using AI to explore the space of possible hypotheses. How close are we to AI systems that can do genuine scientific reasoning, not just pattern matching on data? 方法来创造AlphaFold,但还有更高的一层,那就是人类利用AI来探索可能的假说空间。我们离AI系统能够进行真正的科学推理——而不仅仅是在数据上做模式匹配——还有多远?
[35:32]>> we're close. Um, we're working on these general systems like that like we we have this system called co-scientist and we have other algorithms like AlphaFold that can go a little bit beyond what the basic Gemini will do. And obviously, all the frontier >> 我们已经很接近了。嗯,我们正在研发这类通用系统——比如我们有一个叫co-scientist的系统,还有像AlphaFold这样的算法,能比基础的Gemini做得更多。显然,所有前沿
[35:47]labs are experimenting in this way. I've yet to seen anything so far and we we all tinker with the same things, you know, some math problems that are a little bit harder than IMO and so on. I haven't seen anything yet um that is a true genuine, you know, massive discovery. That's my personal 实验室都在朝这个方向尝试。但到目前为止我还没见过——我们都在琢磨同样的问题,比如比国际数学奥林匹克稍难的数学题之类的——我还没有见过任何真正的、重大的发现。这是我个人的
[36:04]opinion. I think it's coming. I think it may be related to this earlier this thing we discussed about creativity and and actually going on beyond the bounds of what's known. So, clearly that's just not pattern matching at that point cuz there is no pattern to match 看法。我认为这即将到来。这也许跟我们之前讨论过的创造力有关——真正超越已知边界的创造。显然,到了那个阶段就不仅仅是模式匹配了,因为根本没有模式可以匹配。
[36:20]to. And it's a bit more than extrapolation. It's some kind of analogical reasoning. And I don't think these systems have that or at least we're not using them in the in the right way to do that. So, the way I often say that in science is, can it come up with a hypothesis that's really interesting, 它也不只是外推。它是某种类比推理。我认为目前的系统还不具备这种能力,或者至少我们还没有以正确的方式使用它们来实现这一点。所以我经常用科学领域的说法来描述这个标准:它能不能提出一个真正有趣的假说,
[36:36]not just solve one? When I say just, we're not talking about just like solving the Riemann hypothesis or something. This would be obviously amazing. Well, one of the Millennium Prize problems and maybe we're a couple of years out from doing that. 而不仅仅是解决一个已有的问题?我说'仅仅'的时候,我们说的可不是解决黎曼猜想之类的——那当然也很了不起——但那是千禧年大奖难题之一,也许我们离做到那个还有几年时间。
[36:47]But, I'd like to solve P equals NP. That's that's my favorite one. But, can you but even harder than that would be to come up with a new set of of Millennium Prize problems that were regarded by top mathematicians to be as, you know, deep and meaningful and worthy 不过我个人想解决的是P等于NP的问题,那是我的最爱。但比那更难的是——你能不能提出一组新的千禧年大奖难题,让顶级数学家们认为它们同样深刻、同样有意义、同样值得
[37:03]of lifetime of study and effort to solve. Right? I think that's another level harder. And we don't have, you know, I still don't think we know how to do that. I don't think it's it's magical, though. I do think these systems will be eventually able to do 用一生去研究和解决的难题。对吧?我觉得这是更高一个层次的难度。我们目前还不知道怎么做。但我也不认为这有什么魔法般的神秘,我确实相信这些系统最终会能做到。
[37:19]that. Maybe we're missing one or two things. And then, the way we would test that is, you know, sometimes call it my Einstein test, which is, you know, can you train a system with the knowledge of cutoff of 1901 and then will it come up with, you know, what Einstein did in 1905, including special relativity, you 也许我们还缺一两样东西。然后我们会怎么测试呢——有时候我叫它'爱因斯坦测试'——就是你能不能用一个知识截止到1901年的系统来训练,然后看它能不能像爱因斯坦在1905年那样提出包括狭义相对论在内的那些发现,你知道的
[37:36]know, his annus mirabilis. Can it can it do that? Right? And then, I think we could run that test. Maybe we should just run that test and keep seeing if that's possible. Once that is, then I think we're on the verge of these systems being able to invent something new, truly novel. So, last 他的奇迹年。它能不能做到?对吧?我觉得我们可以跑一下这个测试。也许我们就应该跑一下,不断检验这是否可行。一旦做到了,我认为我们就到了这些系统能够发明全新的、真正新颖的东西的边缘。那么,最后一个
[37:53]last question. For the people who are deeply technical in this room who want to work on something, you know, even close to the scale that what you have created with you know, it's one of the largest AI efforts in the world and you've been a pioneer for all these years. So, for that, I think everyone in 最后一个问题。在座有很多技术功底很深的人,他们想做些事情,也许接近你所创造的规模——你知道,那是世界上最大的AI项目之一,而且你这么多年来一直是先驱。所以,我想在座的每一个人
[38:08]this room thanks you and the folks at DeepMind very, very deeply from the bottom of our hearts. Thank you. What's the thing that you know now about building at the frontier that you wish you known at 25? 都在发自内心地感谢你和DeepMind的团队。谢谢你。关于在前沿领域构建东西,你现在知道了什么是你希望在25岁时就知道的?
[38:20]I think we covered some of it in terms of actually you you work out that going after hard problems and deep problems um, is no more difficult in some ways than than going after a shallower, simpler, more superficial problem. 我觉得我们已经聊了其中一些。实际上你会发现,去攻克那些困难的、深入的问题,在某些方面并不比去做那些浅层的、简单的、表面的东西更难。
[38:33]They're they're they're just differently difficult. There's different things that are hard about each of those things. But, I think given life's very short and you you know, you only have so much time and energy, you might as well put your life force into something that will really make a difference if you hadn't done it, if you 它们只是难的方式不同。每种事情都有各自困难的地方。但我觉得,人生苦短,你知道,你的时间和精力都是有限的,不如把你的生命投入到一个真正有意义的事情上——如果你没去做的话,如果你
[38:50]hadn't been there to push it. So, I would just think of it through that lens. And then, the other thing is if you're if you are and then we talked about deep tech and I love interdisciplinary work and I think that's going to be even more prevalent in the next few years in combinations of fields and 没有去推动它,世界就会因此不同。我会用这个视角来看待。另一件事就是——我们之前谈到深科技——我非常喜欢跨学科的工作,而且我认为在未来几年,不同领域的交叉融合会变得更加普遍,
[39:07]finding the the the the connections between those fields. And it's going to be even easier to do that with AI. And then, the only other thing I would say is if, you know, if you have your depending on what your AGI timeline is, you know, mine's like 2030 or something like this, then if you start off on a 去发现这些领域之间的联系。而且有了AI,做这件事会变得更容易。然后我想说的最后一点是,你知道,取决于你对AGI时间线的判断——我的大概是在2030年左右——那么如果你今天开始一段
[39:23]deep tech journey today, usually that you're talking about a 10-year journey for for true deep tech, in my opinion. So, then now you have to just consider AGI appearing in the middle of that journey. So, what does that mean? It doesn't it's not bad, 深科技之旅,真正的深科技在我看来通常需要十年的旅程。那么你现在就必须考虑到AGI可能在这个旅程的中间出现。这意味着什么?这不一定是坏事,
[39:38]necessarily, but you You to take that into account. Right? To Will it be able to leverage it? What will the AGI system do with it? And it goes a little bit back to what you said earlier about AlphaFold and general AI systems. So, one thing I can think see happening is Gemini, Claude, or one of these general 但你必须把这个因素考虑进去。对吧?它能不能利用AGI?AGI系统会怎么处理你正在做的事?这有点像你之前说的AlphaFold和通用AI系统的关系。我能预见的一件事是,Gemini、Claude或者这些通用
[39:55]systems making use of AlphaFold like specialized systems as tools. I don't think we're going to have it just in one giant brain cuz it will have too much regression in if I put all the proteins into, you know, Gemini, that wouldn't make sense. We don't need Gemini to do 系统会把AlphaFold这样的专业化系统当作工具来使用。我不认为我们会把所有东西都塞进一个巨型大脑里,因为那样会有太多性能回归——如果我把所有蛋白质的东西都放进Gemini里,那没什么意义。我们不需要Gemini来做
[40:11]protein folding. Going back to your information efficiency, it will definitely affect its language skills or something like that, right? In a bad way. So, much better I think is to have really good general-purpose tool usage models that will then maybe they could even train those specific tools, but they would be in a 蛋白质折叠。回到你说的信息效率问题,这肯定会反过来影响它的语言能力之类的能力,而且是负面的。所以好得多的方式是拥有真正优秀的通用工具使用模型——它们甚至可以训练那些专用工具,但这些工具会是独立的
[40:27]separate system. So, I think that's kind of interesting to think through the implications of that and then what you might build today. Also, physical things, too, like what kinds of factories would you build? What sorts of you know, finance systems, and so on. 系统。所以我觉得思考一下这种架构的影响很有意思,然后再想想你今天应该构建什么。还有物理层面的东西也是,比如你会建什么样的工厂?什么样的金融系统?等等。
[40:42]So, I just think you need to really take that seriously and and and on the one hand is like and imagine what that world would look like and then build something that would be useful if that comes in halfway through. 所以我认为你真的需要认真对待这件事。一方面要想象那个世界会是什么样子,然后构建一些即使AGI在中途出现也依然有用的东西。
[40:53]Demis Hassabis everyone. 以上就是Demis Hassabis。