AI world models need to understand cause and effect | 人工智能世界模型需要理解因果关系 - FT中文网
登录×
电子邮件/用户名
密码
记住我
请输入邮箱和密码进行绑定操作:
请输入手机号码,通过短信验证(目前仅支持中国大陆地区的手机号):
请您阅读我们的用户注册协议隐私权保护政策,点击下方按钮即视为您接受。
FT英语电台

AI world models need to understand cause and effect
人工智能世界模型需要理解因果关系

They should be able to map how reality works, not just how it looks
他们应该能够描绘现实的运作机制,而不仅仅是它的表象 。

The writer is chair of SpaceTime Labs

Human intellect rests on three pillars: seeing (observing the world), doing (intervening in it) and imagining (simulating what might happen under different choices). Right now, artificial intelligence inhabits only one of these pillars.

Expanding existing frontier AI models will not address this problem. The breakthrough that set off today’s frenzy was the transformer architecture, developed at Google and scaled up into large language models trained on much of the public internet and used to write text and code. Then came agents that stitch these models together into automated workflows. Now the focus is on “world models”, which try to capture the physical environment from vast streams of video and other inputs.

World models are an important evolution from LLMs. This so-called spatial intelligence is being used to develop technology that can enable driverless cars and robotic factory workers. The trouble is that systems built in this way do not really understand the world they record. Instead, they mimic it one 3D object at a time. They conflate coincidence with cause. They can act without being able to explain why, optimise without grasping what happens if conditions change and hallucinate with great confidence. In domains such as healthcare, energy grids or, worse, autonomous weapons, the repercussions of this may not just be embarrassing but lethal.

Decades ago, Alan Turing argued that a truly intelligent machine should “learn from experience”. It should not passively observe but act. It should learn from the consequences of its actions and ask “what if?” Training a machine to do this will require something new — a “causal world model” that acts as an internal map of how a slice of reality works, not just how it looks.

Over the past 20 years, a small but determined group of scientists has been building a mathematical language of cause and effect, and with it a solid theoretical foundation for such models. The work, popularised in Judea Pearl’s The Book of Why, explains how to distinguish correlation from causation, formalise interventions and generate counterfactuals — in other words, the worlds that might have been.

Current AI models focus on correlations between variables. This works well in predictive situations where pattern recognition can be used. But causal world models are needed if we are to address the problems that matter most this century. Planning climate adaptation scenarios in megacities like São Paulo, where I live, requires asking “what if” questions about extreme events that have not yet occurred (and may not).

Genuine scientific discovery is not possible without models that can generalise, follow the causal rules of a system and generate realistic scenarios, and go beyond simply extrapolating and automating existing processes. Take complex biological networks, for example. How can we discover the novel bioproducts needed to accelerate the energy transition or solve complex diseases? Designing drought-resilient crops is not a matter of finding patterns in past yields; it requires an understanding of the ways in which the soil microbiome, plant genetics, water, nutrients, pests, diseases and weather interact — and an understanding of what drives what, when and where.

Emerging markets, which are both vulnerable and full of challenges that provide useful experimental data, should be at the forefront of this. They are ideally-suited innovation test beds, partners and co-developers.

The world faces a choice. It can continue racing to build hyperscale infrastructure to support existing AI models, or it can opt to focus some of that attention towards developing models that grasp how the world really works and how it can be deliberately changed and controlled for the better.

Developing causal models may have other unexpected upsides. The brute-force approach of testing trillions of possible correlations and weighting them by trial and error will consume data, energy, emissions and money. But causal models should be parsimonious by design. Training and inference can be orders of magnitude more efficient because the machine would not be blindly searching; it would be probing along meaningful lines of causality under the constraints of the laws of physics that govern the real world.

From São Paulo to Nairobi to Mumbai, the costs of delay are counted in failed harvests and avoidable emissions. Without a revolution in how machines reason about cause and effect, the current AI boom risks ending in disappointment.

版权声明:本文版权归FT中文网所有,未经允许任何单位或个人不得转载,复制或以任何其他方式使用本文全部或部分,侵权必究。

囤积行为加剧伊朗战争引发的经济损害

随着霍尔木兹海峡的对峙进入第三个月,全球各国政府都在艰难应对同一个难题:如何防止囤积者加剧从汽油到注射器等各类产品的短缺。

FT社评:伊朗战争让各国央行进退两难

如果各国央行过早通过加息来遏制通胀压力,可能令本已受创的经济雪上加霜;如但果按兵不动、观望冲突的进展,又可能贻误时机。

反弹的通胀与不耐烦的特朗普:凯文•沃什面临双重压力

美国参议院本周有望批准这位56岁的金融家接替杰伊•鲍威尔出任美联储主席。

伊朗战争推高燃气价格,印度工人纷纷逃离城市生活

伊朗战争推高了烹饪燃料价格,迫使印度许多务工人员返乡回村。

能源、军火与粮食:特朗普对伊战争日益沉重的代价

这场冲突正波及整个美国经济,造成了数千亿美元的产出损失。

肺纤维化生物科技公司Avalyn Pharma申请首次公开募股(IPO)

一家生物技术公司正开发可吸入剂型的已获批肺纤维化口服药,计划赴公开市场融资以支持其后期研发。
2天前
设置字号×
最小
较小
默认
较大
最大
分享×