Transcript

Transcript of YouTube Video: How AI Is Unlocking the Secrets of Nature and the Universe | Demis Hassabis | TED

Transcript of YouTube Video: How AI Is Unlocking the Secrets of Nature and the Universe | Demis Hassabis | TED

The following is a summary and article by AI based on a transcript of the video "How AI Is Unlocking the Secrets of Nature and the Universe | Demis Hassabis | TED". Due to the limitations of AI, please be careful to distinguish the correctness of the content.

Article By AIVideo Transcript
00:04

Chris Anderson: Demis, so good to have you here.

00:06

Demis Hassabis: It's fantastic to be here, thanks, Chris.

00:09

Now, you told Time Magazine,

00:11

"I want to understand the big questions,

00:13

the really big ones that you normally go into philosophy or physics

00:16

if you're interested in them.

00:18

I thought building AI

00:21

would be the fastest route to answer some of those questions."

00:25

Why did you think that?

00:27

DH: (Laughs)

00:28

Well, I guess when I was a kid,

00:30

my favorite subject was physics,

00:32

and I was interested in all the big questions,

00:35

fundamental nature of reality,

00:37

what is consciousness,

00:38

you know, all the big ones.

00:40

And usually you go into physics, if you're interested in that.

00:43

But I read a lot of the great physicists,

00:45

some of my all-time scientific heroes like Feynman and so on.

00:48

And I realized, in the last, sort of 20, 30 years,

00:50

we haven't made much progress

00:52

in understanding some of these fundamental laws.

00:54

So I thought, why not build the ultimate tool to help us,

00:59

which is artificial intelligence.

01:01

And at the same time,

01:03

we could also maybe better understand ourselves

01:05

and the brain better, by doing that too.

01:07

So not only was it incredible tool,

01:08

it was also useful for some of the big questions itself.

01:12

CA: Super interesting.

01:13

So obviously AI can do so many things,

01:16

but I think for this conversation,

01:17

I'd love to focus in on this theme of what it might do

01:21

to unlock the really big questions, the giant scientific breakthroughs,

01:25

because it's been such a theme driving you and your company.

01:29

DH: So I mean, one of the big things AI can do,

01:31

and I've always thought about,

01:33

is we're getting, you know, even back 20, 30 years ago,

01:36

the beginning of the internet era and computer era,

01:39

the amount of data that was being produced

01:43

and also scientific data,

01:44

just too much for the human mind to comprehend in many cases.

01:48

And I think one of the uses of AI is to find patterns and insights

01:52

in huge amounts of data and then surface that

01:54

to the human scientists to make sense of

01:57

and make new hypotheses and conjectures.

01:59

So it seems to me very compatible with the scientific method.

02:03

CA: Right.

02:04

But game play has played a huge role in your own journey

02:07

in figuring this thing out.

02:09

Who is this young lad on the left there?

02:12

Who is that?

02:13

DH: So that was me, I think I must have been about around nine years old.

02:17

I'm captaining the England Under 11 team,

02:20

and we're playing in a Four Nations tournament,

02:23

that's why we're all in red.

02:24

I think we're playing France, Scotland and Wales, I think it was.

02:27

CA: That is so weird, because that happened to me too.

02:32

In my dreams.

02:33

(Laughter)

02:34

And it wasn't just chess,

02:38

you loved all kinds of games.

02:39

DH: I loved all kinds of games, yeah.

02:41

CA: And when you launched DeepMind,

02:43

pretty quickly, you started having it tackle game play.

02:47

Why?

02:48

DH: Well, look, I mean, games actually got me into AI in the first place

02:51

because while we were doing things like,

02:54

we used to go on training camps with the England team and so on.

02:57

And actually back then,

02:58

I guess it was in the mid '80s,

03:01

we would use the very early chess computers,

03:03

if you remember them, to train against,

03:06

as well as playing against each other.

03:08

And they were big lumps of plastic,

03:10

you know, physical boards that you used to,

03:12

some of you remember, used to actually press the squares down

03:15

and there were LED lights, came on.

03:17

And I remember actually, not just thinking about the chess,

03:19

I was actually just fascinated by the fact that this lump of plastic,

03:23

someone had programmed it to be smart

03:26

and actually play chess to a really high standard.

03:29

And I was just amazed by that.

03:31

And that got me thinking about thinking.

03:33

And how does the brain come up with these thought processes,

03:37

these ideas,

03:38

and then maybe how we could mimic that with computers.

03:42

So yeah, it's been a whole theme for my whole life, really.

03:46

CA: But you raised all this money to launch DeepMind,

03:49

and pretty soon you were using it to do, for example, this.

03:55

I mean, this is an odd use of it.

03:57

What was going on here?

03:58

DH: Well, we started off with games at the beginning of DeepMind.

04:01

This was back in 2010, so this is from about 10 years ago,

04:04

it was our first big breakthrough.

04:05

Because we started off with classic Atari games from the 1970s,

04:09

the simplest kind of computer games there are out there.

04:12

And one of the reasons we used games is they're very convenient

04:15

to test out your ideas and your algorithms.

04:19

They're really fast to test.

04:21

And also, as your systems get more powerful,

04:24

you can choose harder and harder games.

04:26

And this was actually the first time ever that our machine surprised us,

04:30

the first of many times,

04:32

which, it figured out in this game called Breakout,

04:34

that you could send the ball round the back of the wall,

04:37

and actually, it would be much safer way to knock out all the tiles of the wall.

04:40

It's a classic Atari game there.

04:42

And that was our first real aha moment.

04:44

CA: So this thing was not programmed to have any strategy.

04:47

It was just told, try and figure out a way of winning.

04:51

You just move the bat at the bottom and see if you can find a way of winning.

04:54

DH: It was a real revolution at the time.

04:56

So this was in 2012, 2013

04:58

where we coined these terms "deep reinforcement learning."

05:01

And the key thing about them is that those systems were learning

05:04

directly from the pixels, the raw pixels on the screen,

05:07

but they weren't being told anything else.

05:09

So they were being told, maximize the score,

05:11

here are the pixels on the screen,

05:13

30,000 pixels.

05:15

The system has to make sense on its own from first principles

05:18

what’s going on, what it’s controlling,

05:20

how to get points.

05:21

And that's the other nice thing about using games to begin with.

05:24

They have clear objectives, to win, to get scores.

05:27

So you can kind of measure very easily that your systems are improving.

05:30

CA: But there was a direct line from that to this moment

05:33

a few years later,

05:35

where country of South Korea and many other parts of Asia

05:39

and in fact the world went crazy over -- over what?

05:42

DH: Yeah, so this was the pinnacle of -- this is in 2016 --

05:46

the pinnacle of our games-playing work,

05:48

where, so we'd done Atari,

05:50

we'd done some more complicated games.

05:52

And then we reached the pinnacle, which was the game of Go,

05:56

which is what they play in Asia instead of chess,

05:59

but it's actually more complex than chess.

06:01

And the actual brute force algorithms

06:05

that were used to kind of crack chess were not possible with Go

06:10

because it's a much more pattern-based game,

06:12

much more intuitive game.

06:14

So even though Deep Blue beat Garry Kasparov in the '90s,

06:17

it took another 20 years for our program, AlphaGo,

06:21

to beat the world champion at Go.

06:23

And we always thought,

06:24

myself and the people working on this project for many years,

06:27

if you could build a system that could beat the world champion at Go,

06:31

it would have had to have done something very interesting.

06:34

And in this case, what we did with AlphaGo,

06:36

is it basically learned for itself,

06:38

by playing millions and millions of games against itself,

06:40

ideas about Go, the right strategies.

06:42

And in fact invented its own new strategies

06:45

that the Go world had never seen before,

06:47

even though we've played Go for more than,

06:49

you know, 2,000 years,

06:51

it's the oldest board game in existence.

06:54

So, you know, it was pretty astounding.

06:56

Not only did it win the match,

06:57

it also came up with brand new strategies.

07:01

CA: And you continued this with a new strategy

07:03

of not even really teaching it anything about Go,

07:05

but just setting up systems

07:07

that just from first principles would play

07:10

so that they could teach themselves from scratch, Go or chess.

07:15

Talk about AlphaZero and the amazing thing that happened in chess then.

07:21

DH: So following this, we started with AlphaGo

07:24

by giving it all of the human games that are being played on the internet.

07:28

So it started that as a basic starting point for its knowledge.

07:32

And then we wanted to see what would happen if we started from scratch,

07:35

from literally random play.

07:37

So this is what AlphaZero was.

07:39

That's why it's the zero in the name,

07:40

because it started with zero prior knowledge

07:44

And the reason we did that is because then we would build a system

07:47

that was more general.

07:48

So AlphaGo could only play Go,

07:50

but AlphaZero could play any two-player game,

07:53

and it did it by playing initially randomly

07:57

and then slowly, incrementally improving.

07:59

Well, not very slowly, actually, within the course of 24 hours,

08:02

going from random to better than world-champion level.

08:06

CA: And so this is so amazing to me.

08:08

So I'm more familiar with chess than with Go.

08:10

And for decades,

08:11

thousands and thousands of AI experts worked on building

08:15

incredible chess computers.

08:16

Eventually, they got better than humans.

08:18

You had a moment a few years ago,

08:21

where in nine hours,

08:23

AlphaZero taught itself to play chess better than any of those systems ever did.

08:30

Talk about that.

08:32

DH: It was a pretty incredible moment, actually.

08:34

So we set it going on chess.

08:38

And as you said, there's this rich history of chess and AI

08:40

where there are these expert systems that have been programmed

08:43

with these chess ideas, chess algorithms.

08:46

And you have this amazing, you know,

08:48

I remember this day very clearly, where you sort of sit down with the system

08:52

starting off random, you know, in the morning,

08:55

you go for a cup of coffee, you come back.

08:57

I can still just about beat it by lunchtime, maybe just about.

09:00

And then you let it go for another four hours.

09:02

And by dinner,

09:03

it's the greatest chess-playing entity that's ever existed.

09:06

And, you know, it's quite amazing,

09:08

like, looking at that live on something that you know well,

09:11

you know, like chess, and you're expert in

09:13

and actually just seeing that in front of your eyes.

09:16

And then you extrapolate to what it could then do in science or something else,

09:20

which of course, games were only a means to an end.

09:23

They were never the end in themselves.

09:25

They were just the training ground for our ideas

09:27

and to make quick progress in a matter of, you know,

09:30

less than five years actually went from Atari to Go.

09:34

CA: I mean, this is why people are in awe of AI

09:37

and also kind of terrified by it.

09:40

I mean, it's not just incremental improvement.

09:42

The fact that in a few hours you can achieve

09:45

what millions of humans over centuries have not been able to achieve.

09:50

That gives you pause for thought.

09:53

DH: It does, I mean, it's a hugely powerful technology.

09:56

It's going to be incredibly transformative.

09:58

And we have to be very thoughtful about how we use that capability.

10:02

CA: So talk about this use of it because this is again,

10:04

this is another extension of the work you've done,

10:08

where now you're turning it to something incredibly useful for the world.

10:12

What are all the letters on the left, and what’s on the right?

10:15

DH: This was always my aim with AI from a kid,

10:19

which is to use it to accelerate scientific discovery.

10:23

And actually, ever since doing my undergrad at Cambridge,

10:26

I had this problem in mind one day for AI,

10:28

it's called the protein-folding problem.

10:30

And it's kind of like a 50-year grand challenge in biology.

10:33

And it's very simple to explain.

10:35

Proteins are essential to life.

10:38

They're the building blocks of life.

10:39

Everything in your body depends on proteins.

10:41

A protein is sort of described by its amino acid sequence,

10:47

which you can think of as roughly the genetic sequence

10:49

describing the protein, so that are the letters.

10:52

CA: And each of those letters represents in itself a complex molecule?

10:55

DH: That's right, each of those letters is an amino acid.

10:58

And you can think of them as a kind of string of beads

11:00

there at the bottom, left, right?

11:02

But in nature, in your body or in an animal,

11:06

this string, a sequence,

11:07

turns into this beautiful shape on the right.

11:10

That's the protein.

11:11

Those letters describe that shape.

11:14

And that's what it looks like in nature.

11:16

And the important thing about that 3D structure is

11:19

the 3D structure of the protein goes a long way to telling you

11:22

what its function is in the body, what it does.

11:24

And so the protein-folding problem is:

11:26

Can you directly predict the 3D structure just from the amino acid sequence?

11:31

So literally if you give the machine, the AI system,

11:34

the letters on the left,

11:35

can it produce the 3D structure on the right?

11:38

And that's what AlphaFold does, our program does.

11:40

CA: It's not calculating it from the letters,

11:42

it's looking at patterns of other folded proteins that are known about

11:47

and somehow learning from those patterns

11:50

that this may be the way to do this?

11:52

DH: So when we started this project, actually straight after AlphaGo,

11:55

I thought we were ready.

11:56

Once we'd cracked Go,

11:57

I felt we were finally ready after, you know,

12:00

almost 20 years of working on this stuff

12:02

to actually tackle some scientific problems,

12:05

including protein folding.

12:06

And what we start with is painstakingly,

12:09

over the last 40-plus years,

12:11

experimental biologists have pieced together

12:14

around 150,000 protein structures

12:17

using very complicated, you know, X-ray crystallography techniques

12:21

and other complicated experimental techniques.

12:24

And the rule of thumb is

12:26

that it takes one PhD student their whole PhD,

12:29

so four or five years, to uncover one structure.

12:33

But there are 200 million proteins known to nature.

12:36

So you could just, you know, take forever to do that.

12:39

And so we managed to actually fold, using AlphaFold, in one year,

12:43

all those 200 million proteins known to science.

12:46

So that's a billion years of PhD time saved.

12:49

(Applause)

12:52

CA: So it's amazing to me just how reliably it works.

12:55

I mean, this shows, you know,

12:58

here's the model and you do the experiment.

13:00

And sure enough, the protein turns out the same way.

13:03

Times 200 million.

13:04

DH: And the more deeply you go into proteins,

13:07

you just start appreciating how exquisite they are.

13:09

I mean, look at how beautiful these proteins are.

13:12

And each of these things do a special function in nature.

13:14

And they're almost like works of art.

13:16

And it's still astounds me today that AlphaFold can predict,

13:19

the green is the ground truth, and the blue is the prediction,

13:22

how well it can predict, is to within the width of an atom on average,

13:26

is how accurate the prediction is,

13:28

which is what is needed for biologists to use it,

13:31

and for drug design and for disease understanding,

13:34

which is what AlphaFold unlocks.

13:36

CA: You made a surprising decision,

13:38

which was to give away the actual results of your 200 million proteins.

13:44

DH: We open-sourced AlphaFold and gave everything away

13:47

on a huge database with our wonderful colleagues,

13:50

the European Bioinformatics Institute.

13:51

(Applause)

13:55

CA: I mean, you're part of Google.

13:57

Was there a phone call saying, "Uh, Demis, what did you just do?"

14:01

DH: You know, I'm lucky we have very supportive,

14:04

Google's really supportive of science

14:06

and understand the benefits this can bring to the world.

14:10

And, you know, the argument here

14:12

was that we could only ever have even scratched the surface

14:15

of the potential of what we could do with this.

14:17

This, you know, maybe like a millionth

14:19

of what the scientific community is doing with it.

14:22

There's over a million and a half biologists around the world

14:25

have used AlphaFold and its predictions.

14:27

We think that's almost every biologist in the world

14:29

is making use of this now, every pharma company.

14:32

So we'll never know probably what the full impact of it all is.

14:35

CA: But you're continuing this work in a new company

14:37

that's spinning out of Google called Isomorph.

14:40

DH: Isomorphic.

14:41

CA: Isomorphic.

14:43

Give us just a sense of the vision there.

14:45

What's the vision?

14:47

DH: AlphaFold is a sort of fundamental biology tool.

14:50

Like, what are these 3D structures,

14:52

and then what might they do in nature?

14:55

And then if you, you know,

14:57

the reason I thought about this and was so excited about this,

15:00

is that this is the beginnings of understanding disease

15:03

and also maybe helpful for designing drugs.

15:06

So if you know the shape of the protein,

15:09

and then you can kind of figure out

15:11

which part of the surface of the protein

15:13

you're going to target with your drug compound.

15:16

And Isomorphic is extending this work we did in AlphaFold

15:19

into the chemistry space,

15:21

where we can design chemical compounds

15:24

that will bind exactly to the right spot on the protein

15:27

and also, importantly, to nothing else in the body.

15:30

So it doesn't have any side effects and it's not toxic and so on.

15:34

And we're building many other AI models,

15:37

sort of sister models to AlphaFold

15:39

to help predict,

15:41

make predictions in chemistry space.

15:43

CA: So we can expect to see

15:44

some pretty dramatic health medicine breakthroughs

15:48

in the coming few years.

15:49

DH: I think we'll be able to get down drug discovery

15:51

from years to maybe months.

15:54

CA: OK. Demis, I'd like to change direction a bit.

15:58

Our mutual friend, Liv Boeree, gave a talk last year at TEDAI

16:02

that she called the “Moloch Trap.”

16:04

The Moloch Trap is a situation

16:06

where organizations,

16:09

companies in a competitive situation can be driven to do things

16:14

that no individual running those companies would by themselves do.

16:19

I was really struck by this talk,

16:21

and it's felt, as a sort of layperson observer,

16:25

that the Moloch Trap has been shockingly in effect in the last couple of years.

16:30

So here you are with DeepMind,

16:32

sort of pursuing these amazing medical breakthroughs

16:35

and scientific breakthroughs,

16:37

and then suddenly, kind of out of left field,

16:41

OpenAI with Microsoft releases ChatGPT.

16:46

And the world goes crazy and suddenly goes, “Holy crap, AI is ...”

16:50

you know, everyone can use it.

16:54

And there’s a sort of, it felt like the Moloch Trap in action.

16:58

I think Microsoft CEO Satya Nadella actually said,

17:03

"Google is the 800-pound gorilla in the search space.

17:08

We wanted to make Google dance."

17:12

How ...?

17:14

And it did, Google did dance.

17:16

There was a dramatic response.

17:18

Your role was changed,

17:20

you took over the whole Google AI effort.

17:24

Products were rushed out.

17:27

You know, Gemini, some part amazing, part embarrassing.

17:30

I’m not going to ask you about Gemini because you’ve addressed it elsewhere.

17:33

But it feels like this was the Moloch Trap happening,

17:37

that you and others were pushed to do stuff

17:40

that you wouldn't have done without this sort of catalyzing competitive thing.

17:45

Meta did something similar as well.

17:47

They rushed out an open-source version of AI,

17:50

which is arguably a reckless act in itself.

17:55

This seems terrifying to me.

17:57

Is it terrifying?

17:59

DH: Look, it's a complicated topic, of course.

18:01

And, first of all, I mean, there are many things to say about it.

18:05

First of all, we were working on many large language models.

18:10

And in fact, obviously, Google research actually invented Transformers,

18:13

as you know,

18:14

which was the architecture that allowed all this to be possible,

18:17

five, six years ago.

18:19

And so we had many large models internally.

18:21

The thing was, I think what the ChatGPT moment did that changed was,

18:25

and fair play to them to do that, was they demonstrated,

18:28

I think somewhat surprisingly to themselves as well,

18:31

that the public were ready to,

18:34

you know, the general public were ready to embrace these systems

18:37

and actually find value in these systems.

18:39

Impressive though they are, I guess, when we're working on these systems,

18:43

mostly you're focusing on the flaws and the things they don't do

18:46

and hallucinations and things you're all familiar with now.

18:49

We're thinking, you know,

18:50

would anyone really find that useful given that it does this and that?

18:54

And we would want them to improve those things first,

18:56

before putting them out.

18:58

But interestingly, it turned out that even with those flaws,

19:01

many tens of millions of people still find them very useful.

19:04

And so that was an interesting update on maybe the convergence of products

19:09

and the science that actually,

19:13

all of these amazing things we've been doing in the lab, so to speak,

19:16

are actually ready for prime time for general use,

19:19

beyond the rarefied world of science.

19:21

And I think that's pretty exciting in many ways.

19:24

CA: So at the moment, we've got this exciting array of products

19:27

which we're all enjoying.

19:29

And, you know, all this generative AI stuff is amazing.

19:31

But let's roll the clock forward a bit.

19:34

Microsoft and OpenAI are reported to be building

19:38

or investing like 100 billion dollars

19:40

into an absolute monster database supercomputer

19:45

that can offer compute at orders of magnitude

19:49

more than anything we have today.

19:52

It takes like five gigawatts of energy to drive this, it's estimated.

19:56

That's the energy of New York City to drive a data center.

20:00

So we're pumping all this energy into this giant, vast brain.

20:04

Google, I presume is going to match this type of investment, right?

20:09

DH: Well, I mean, we don't talk about our specific numbers,

20:11

but you know, I think we're investing more than that over time.

20:15

So, and that's one of the reasons

20:17

we teamed up with Google back in 2014,

20:19

is kind of we knew that in order to get to AGI,

20:23

we would need a lot of compute.

20:24

And that's what's transpired.

20:26

And Google, you know, had and still has the most computers.

20:30

CA: So Earth is building these giant computers

20:33

that are going to basically, these giant brains,

20:35

that are going to power so much of the future economy.

20:38

And it's all by companies that are in competition with each other.

20:42

How will we avoid the situation where someone is getting a lead,

20:47

someone else has got 100 billion dollars invested in their thing.

20:52

Isn't someone going to go, "Wait a sec.

20:54

If we used reinforcement learning here

20:57

to maybe have the AI tweak its own code

21:00

and rewrite itself and make it so [powerful],

21:03

we might be able to catch up in nine hours over the weekend

21:06

with what they're doing.

21:07

Roll the dice, dammit, we have no choice.

21:09

Otherwise we're going to lose a fortune for our shareholders."

21:12

How are we going to avoid that?

21:14

DH: Yeah, well, we must avoid that, of course, clearly.

21:16

And my view is that as we get closer to AGI,

21:20

we need to collaborate more.

21:22

And the good news is that most of the scientists involved in these labs

21:27

know each other very well.

21:29

And we talk to each other a lot at conferences and other things.

21:32

And this technology is still relatively nascent.

21:35

So probably it's OK what's happening at the moment.

21:37

But as we get closer to AGI, I think as a society,

21:42

we need to start thinking about the types of architectures that get built.

21:46

So I'm very optimistic, of course,

21:48

that's why I spent my whole life working on AI and working towards AGI.

21:53

But I suspect there are many ways to build the architecture safely, robustly,

22:00

reliably and in an understandable way.

22:03

And I think there are almost certainly going to be ways of building architectures

22:07

that are unsafe or risky in some form.

22:09

So I see a sort of,

22:11

a kind of bottleneck that we have to get humanity through,

22:14

which is building safe architectures as the first types of AGI systems.

22:20

And then after that, we can have a sort of,

22:23

a flourishing of many different types of systems

22:26

that are perhaps sharded off those safe architectures

22:29

that ideally have some mathematical guarantees

22:33

or at least some practical guarantees around what they do.

22:36

CA: Do governments have an essential role here

22:38

to define what a level playing field looks like

22:40

and what is absolutely taboo?

22:42

DH: Yeah, I think it's not just about --

22:44

actually I think government and civil society

22:46

and academia and all parts of society have a critical role to play here

22:49

to shape, along with industry labs,

22:52

what that should look like as we get closer to AGI

22:55

and the cooperation needed and the collaboration needed,

22:58

to prevent that kind of runaway race dynamic happening.

23:01

CA: OK, well, it sounds like you remain optimistic.

23:04

What's this image here?

23:05

DH: That's one of my favorite images, actually.

23:07

I call it, like, the tree of all knowledge.

23:09

So, you know, we've been talking a lot about science,

23:12

and a lot of science can be boiled down to

23:15

if you imagine all the knowledge that exists in the world

23:18

as a tree of knowledge,

23:19

and then maybe what we know today as a civilization is some, you know,

23:24

small subset of that.

23:26

And I see AI as this tool that allows us,

23:29

as scientists, to explore, potentially, the entire tree one day.

23:33

And we have this idea of root node problems

23:36

that, like AlphaFold, the protein-folding problem,

23:38

where if you could crack them,

23:40

it unlocks an entire new branch of discovery or new research.

23:45

And that's what we try and focus on at DeepMind

23:47

and Google DeepMind to crack those.

23:50

And if we get this right, then I think we could be, you know,

23:53

in this incredible new era of radical abundance,

23:56

curing all diseases,

23:58

spreading consciousness to the stars.

24:01

You know, maximum human flourishing.

24:03

CA: We're out of time,

24:04

but what's the last example of like, in your dreams,

24:06

this dream question that you think there is a shot

24:09

that in your lifetime AI might take us to?

24:12

DH: I mean, once AGI is built,

24:14

what I'd like to use it for is to try and use it to understand

24:18

the fundamental nature of reality.

24:20

So do experiments at the Planck scale.

24:23

You know, the smallest possible scale, theoretical scale,

24:26

which is almost like the resolution of reality.

24:29

CA: You know, I was brought up religious.

24:31

And in the Bible, there’s a story about the tree of knowledge

24:34

that doesn't work out very well.

24:36

(Laughter)

24:37

Is there any scenario

24:41

where we discover knowledge that the universe says,

24:46

"Humans, you may not know that."

24:49

DH: Potentially.

24:51

I mean, there might be some unknowable things.

24:53

But I think scientific method is the greatest sort of invention

24:58

humans have ever come up with.

24:59

You know, the enlightenment and scientific discovery.

25:03

That's what's built this incredible modern civilization around us

25:06

and all the tools that we use.

25:08

So I think it's the best technique we have

25:11

for understanding the enormity of the universe around us.

25:15

CA: Well, Demis, you've already changed the world.

25:18

I think probably everyone here will be cheering you on

25:21

in your efforts to ensure that we continue to accelerate

25:24

in the right direction.

25:25

DH: Thank you.

25:26

CA: Demis Hassabis.

25:28

(Applause)