The following is a summary and article by AI based on a transcript of the video "How AI Is Unlocking the Secrets of Nature and the Universe | Demis Hassabis | TED". Due to the limitations of AI, please be careful to distinguish the correctness of the content.
00:04 | Chris Anderson: Demis, so good to have you here. |
---|---|
00:06 | Demis Hassabis: It's fantastic to be here, thanks, Chris. |
00:09 | Now, you told Time Magazine, |
00:11 | "I want to understand the big questions, |
00:13 | the really big ones that you normally go into philosophy or physics |
00:16 | if you're interested in them. |
00:18 | I thought building AI |
00:21 | would be the fastest route to answer some of those questions." |
00:25 | Why did you think that? |
00:27 | DH: (Laughs) |
00:28 | Well, I guess when I was a kid, |
00:30 | my favorite subject was physics, |
00:32 | and I was interested in all the big questions, |
00:35 | fundamental nature of reality, |
00:37 | what is consciousness, |
00:38 | you know, all the big ones. |
00:40 | And usually you go into physics, if you're interested in that. |
00:43 | But I read a lot of the great physicists, |
00:45 | some of my all-time scientific heroes like Feynman and so on. |
00:48 | And I realized, in the last, sort of 20, 30 years, |
00:50 | we haven't made much progress |
00:52 | in understanding some of these fundamental laws. |
00:54 | So I thought, why not build the ultimate tool to help us, |
00:59 | which is artificial intelligence. |
01:01 | And at the same time, |
01:03 | we could also maybe better understand ourselves |
01:05 | and the brain better, by doing that too. |
01:07 | So not only was it incredible tool, |
01:08 | it was also useful for some of the big questions itself. |
01:12 | CA: Super interesting. |
01:13 | So obviously AI can do so many things, |
01:16 | but I think for this conversation, |
01:17 | I'd love to focus in on this theme of what it might do |
01:21 | to unlock the really big questions, the giant scientific breakthroughs, |
01:25 | because it's been such a theme driving you and your company. |
01:29 | DH: So I mean, one of the big things AI can do, |
01:31 | and I've always thought about, |
01:33 | is we're getting, you know, even back 20, 30 years ago, |
01:36 | the beginning of the internet era and computer era, |
01:39 | the amount of data that was being produced |
01:43 | and also scientific data, |
01:44 | just too much for the human mind to comprehend in many cases. |
01:48 | And I think one of the uses of AI is to find patterns and insights |
01:52 | in huge amounts of data and then surface that |
01:54 | to the human scientists to make sense of |
01:57 | and make new hypotheses and conjectures. |
01:59 | So it seems to me very compatible with the scientific method. |
02:03 | CA: Right. |
02:04 | But game play has played a huge role in your own journey |
02:07 | in figuring this thing out. |
02:09 | Who is this young lad on the left there? |
02:12 | Who is that? |
02:13 | DH: So that was me, I think I must have been about around nine years old. |
02:17 | I'm captaining the England Under 11 team, |
02:20 | and we're playing in a Four Nations tournament, |
02:23 | that's why we're all in red. |
02:24 | I think we're playing France, Scotland and Wales, I think it was. |
02:27 | CA: That is so weird, because that happened to me too. |
02:32 | In my dreams. |
02:33 | (Laughter) |
02:34 | And it wasn't just chess, |
02:38 | you loved all kinds of games. |
02:39 | DH: I loved all kinds of games, yeah. |
02:41 | CA: And when you launched DeepMind, |
02:43 | pretty quickly, you started having it tackle game play. |
02:47 | Why? |
02:48 | DH: Well, look, I mean, games actually got me into AI in the first place |
02:51 | because while we were doing things like, |
02:54 | we used to go on training camps with the England team and so on. |
02:57 | And actually back then, |
02:58 | I guess it was in the mid '80s, |
03:01 | we would use the very early chess computers, |
03:03 | if you remember them, to train against, |
03:06 | as well as playing against each other. |
03:08 | And they were big lumps of plastic, |
03:10 | you know, physical boards that you used to, |
03:12 | some of you remember, used to actually press the squares down |
03:15 | and there were LED lights, came on. |
03:17 | And I remember actually, not just thinking about the chess, |
03:19 | I was actually just fascinated by the fact that this lump of plastic, |
03:23 | someone had programmed it to be smart |
03:26 | and actually play chess to a really high standard. |
03:29 | And I was just amazed by that. |
03:31 | And that got me thinking about thinking. |
03:33 | And how does the brain come up with these thought processes, |
03:37 | these ideas, |
03:38 | and then maybe how we could mimic that with computers. |
03:42 | So yeah, it's been a whole theme for my whole life, really. |
03:46 | CA: But you raised all this money to launch DeepMind, |
03:49 | and pretty soon you were using it to do, for example, this. |
03:55 | I mean, this is an odd use of it. |
03:57 | What was going on here? |
03:58 | DH: Well, we started off with games at the beginning of DeepMind. |
04:01 | This was back in 2010, so this is from about 10 years ago, |
04:04 | it was our first big breakthrough. |
04:05 | Because we started off with classic Atari games from the 1970s, |
04:09 | the simplest kind of computer games there are out there. |
04:12 | And one of the reasons we used games is they're very convenient |
04:15 | to test out your ideas and your algorithms. |
04:19 | They're really fast to test. |
04:21 | And also, as your systems get more powerful, |
04:24 | you can choose harder and harder games. |
04:26 | And this was actually the first time ever that our machine surprised us, |
04:30 | the first of many times, |
04:32 | which, it figured out in this game called Breakout, |
04:34 | that you could send the ball round the back of the wall, |
04:37 | and actually, it would be much safer way to knock out all the tiles of the wall. |
04:40 | It's a classic Atari game there. |
04:42 | And that was our first real aha moment. |
04:44 | CA: So this thing was not programmed to have any strategy. |
04:47 | It was just told, try and figure out a way of winning. |
04:51 | You just move the bat at the bottom and see if you can find a way of winning. |
04:54 | DH: It was a real revolution at the time. |
04:56 | So this was in 2012, 2013 |
04:58 | where we coined these terms "deep reinforcement learning." |
05:01 | And the key thing about them is that those systems were learning |
05:04 | directly from the pixels, the raw pixels on the screen, |
05:07 | but they weren't being told anything else. |
05:09 | So they were being told, maximize the score, |
05:11 | here are the pixels on the screen, |
05:13 | 30,000 pixels. |
05:15 | The system has to make sense on its own from first principles |
05:18 | what’s going on, what it’s controlling, |
05:20 | how to get points. |
05:21 | And that's the other nice thing about using games to begin with. |
05:24 | They have clear objectives, to win, to get scores. |
05:27 | So you can kind of measure very easily that your systems are improving. |
05:30 | CA: But there was a direct line from that to this moment |
05:33 | a few years later, |
05:35 | where country of South Korea and many other parts of Asia |
05:39 | and in fact the world went crazy over -- over what? |
05:42 | DH: Yeah, so this was the pinnacle of -- this is in 2016 -- |
05:46 | the pinnacle of our games-playing work, |
05:48 | where, so we'd done Atari, |
05:50 | we'd done some more complicated games. |
05:52 | And then we reached the pinnacle, which was the game of Go, |
05:56 | which is what they play in Asia instead of chess, |
05:59 | but it's actually more complex than chess. |
06:01 | And the actual brute force algorithms |
06:05 | that were used to kind of crack chess were not possible with Go |
06:10 | because it's a much more pattern-based game, |
06:12 | much more intuitive game. |
06:14 | So even though Deep Blue beat Garry Kasparov in the '90s, |
06:17 | it took another 20 years for our program, AlphaGo, |
06:21 | to beat the world champion at Go. |
06:23 | And we always thought, |
06:24 | myself and the people working on this project for many years, |
06:27 | if you could build a system that could beat the world champion at Go, |
06:31 | it would have had to have done something very interesting. |
06:34 | And in this case, what we did with AlphaGo, |
06:36 | is it basically learned for itself, |
06:38 | by playing millions and millions of games against itself, |
06:40 | ideas about Go, the right strategies. |
06:42 | And in fact invented its own new strategies |
06:45 | that the Go world had never seen before, |
06:47 | even though we've played Go for more than, |
06:49 | you know, 2,000 years, |
06:51 | it's the oldest board game in existence. |
06:54 | So, you know, it was pretty astounding. |
06:56 | Not only did it win the match, |
06:57 | it also came up with brand new strategies. |
07:01 | CA: And you continued this with a new strategy |
07:03 | of not even really teaching it anything about Go, |
07:05 | but just setting up systems |
07:07 | that just from first principles would play |
07:10 | so that they could teach themselves from scratch, Go or chess. |
07:15 | Talk about AlphaZero and the amazing thing that happened in chess then. |
07:21 | DH: So following this, we started with AlphaGo |
07:24 | by giving it all of the human games that are being played on the internet. |
07:28 | So it started that as a basic starting point for its knowledge. |
07:32 | And then we wanted to see what would happen if we started from scratch, |
07:35 | from literally random play. |
07:37 | So this is what AlphaZero was. |
07:39 | That's why it's the zero in the name, |
07:40 | because it started with zero prior knowledge |
07:44 | And the reason we did that is because then we would build a system |
07:47 | that was more general. |
07:48 | So AlphaGo could only play Go, |
07:50 | but AlphaZero could play any two-player game, |
07:53 | and it did it by playing initially randomly |
07:57 | and then slowly, incrementally improving. |
07:59 | Well, not very slowly, actually, within the course of 24 hours, |
08:02 | going from random to better than world-champion level. |
08:06 | CA: And so this is so amazing to me. |
08:08 | So I'm more familiar with chess than with Go. |
08:10 | And for decades, |
08:11 | thousands and thousands of AI experts worked on building |
08:15 | incredible chess computers. |
08:16 | Eventually, they got better than humans. |
08:18 | You had a moment a few years ago, |
08:21 | where in nine hours, |
08:23 | AlphaZero taught itself to play chess better than any of those systems ever did. |
08:30 | Talk about that. |
08:32 | DH: It was a pretty incredible moment, actually. |
08:34 | So we set it going on chess. |
08:38 | And as you said, there's this rich history of chess and AI |
08:40 | where there are these expert systems that have been programmed |
08:43 | with these chess ideas, chess algorithms. |
08:46 | And you have this amazing, you know, |
08:48 | I remember this day very clearly, where you sort of sit down with the system |
08:52 | starting off random, you know, in the morning, |
08:55 | you go for a cup of coffee, you come back. |
08:57 | I can still just about beat it by lunchtime, maybe just about. |
09:00 | And then you let it go for another four hours. |
09:02 | And by dinner, |
09:03 | it's the greatest chess-playing entity that's ever existed. |
09:06 | And, you know, it's quite amazing, |
09:08 | like, looking at that live on something that you know well, |
09:11 | you know, like chess, and you're expert in |
09:13 | and actually just seeing that in front of your eyes. |
09:16 | And then you extrapolate to what it could then do in science or something else, |
09:20 | which of course, games were only a means to an end. |
09:23 | They were never the end in themselves. |
09:25 | They were just the training ground for our ideas |
09:27 | and to make quick progress in a matter of, you know, |
09:30 | less than five years actually went from Atari to Go. |
09:34 | CA: I mean, this is why people are in awe of AI |
09:37 | and also kind of terrified by it. |
09:40 | I mean, it's not just incremental improvement. |
09:42 | The fact that in a few hours you can achieve |
09:45 | what millions of humans over centuries have not been able to achieve. |
09:50 | That gives you pause for thought. |
09:53 | DH: It does, I mean, it's a hugely powerful technology. |
09:56 | It's going to be incredibly transformative. |
09:58 | And we have to be very thoughtful about how we use that capability. |
10:02 | CA: So talk about this use of it because this is again, |
10:04 | this is another extension of the work you've done, |
10:08 | where now you're turning it to something incredibly useful for the world. |
10:12 | What are all the letters on the left, and what’s on the right? |
10:15 | DH: This was always my aim with AI from a kid, |
10:19 | which is to use it to accelerate scientific discovery. |
10:23 | And actually, ever since doing my undergrad at Cambridge, |
10:26 | I had this problem in mind one day for AI, |
10:28 | it's called the protein-folding problem. |
10:30 | And it's kind of like a 50-year grand challenge in biology. |
10:33 | And it's very simple to explain. |
10:35 | Proteins are essential to life. |
10:38 | They're the building blocks of life. |
10:39 | Everything in your body depends on proteins. |
10:41 | A protein is sort of described by its amino acid sequence, |
10:47 | which you can think of as roughly the genetic sequence |
10:49 | describing the protein, so that are the letters. |
10:52 | CA: And each of those letters represents in itself a complex molecule? |
10:55 | DH: That's right, each of those letters is an amino acid. |
10:58 | And you can think of them as a kind of string of beads |
11:00 | there at the bottom, left, right? |
11:02 | But in nature, in your body or in an animal, |
11:06 | this string, a sequence, |
11:07 | turns into this beautiful shape on the right. |
11:10 | That's the protein. |
11:11 | Those letters describe that shape. |
11:14 | And that's what it looks like in nature. |
11:16 | And the important thing about that 3D structure is |
11:19 | the 3D structure of the protein goes a long way to telling you |
11:22 | what its function is in the body, what it does. |
11:24 | And so the protein-folding problem is: |
11:26 | Can you directly predict the 3D structure just from the amino acid sequence? |
11:31 | So literally if you give the machine, the AI system, |
11:34 | the letters on the left, |
11:35 | can it produce the 3D structure on the right? |
11:38 | And that's what AlphaFold does, our program does. |
11:40 | CA: It's not calculating it from the letters, |
11:42 | it's looking at patterns of other folded proteins that are known about |
11:47 | and somehow learning from those patterns |
11:50 | that this may be the way to do this? |
11:52 | DH: So when we started this project, actually straight after AlphaGo, |
11:55 | I thought we were ready. |
11:56 | Once we'd cracked Go, |
11:57 | I felt we were finally ready after, you know, |
12:00 | almost 20 years of working on this stuff |
12:02 | to actually tackle some scientific problems, |
12:05 | including protein folding. |
12:06 | And what we start with is painstakingly, |
12:09 | over the last 40-plus years, |
12:11 | experimental biologists have pieced together |
12:14 | around 150,000 protein structures |
12:17 | using very complicated, you know, X-ray crystallography techniques |
12:21 | and other complicated experimental techniques. |
12:24 | And the rule of thumb is |
12:26 | that it takes one PhD student their whole PhD, |
12:29 | so four or five years, to uncover one structure. |
12:33 | But there are 200 million proteins known to nature. |
12:36 | So you could just, you know, take forever to do that. |
12:39 | And so we managed to actually fold, using AlphaFold, in one year, |
12:43 | all those 200 million proteins known to science. |
12:46 | So that's a billion years of PhD time saved. |
12:49 | (Applause) |
12:52 | CA: So it's amazing to me just how reliably it works. |
12:55 | I mean, this shows, you know, |
12:58 | here's the model and you do the experiment. |
13:00 | And sure enough, the protein turns out the same way. |
13:03 | Times 200 million. |
13:04 | DH: And the more deeply you go into proteins, |
13:07 | you just start appreciating how exquisite they are. |
13:09 | I mean, look at how beautiful these proteins are. |
13:12 | And each of these things do a special function in nature. |
13:14 | And they're almost like works of art. |
13:16 | And it's still astounds me today that AlphaFold can predict, |
13:19 | the green is the ground truth, and the blue is the prediction, |
13:22 | how well it can predict, is to within the width of an atom on average, |
13:26 | is how accurate the prediction is, |
13:28 | which is what is needed for biologists to use it, |
13:31 | and for drug design and for disease understanding, |
13:34 | which is what AlphaFold unlocks. |
13:36 | CA: You made a surprising decision, |
13:38 | which was to give away the actual results of your 200 million proteins. |
13:44 | DH: We open-sourced AlphaFold and gave everything away |
13:47 | on a huge database with our wonderful colleagues, |
13:50 | the European Bioinformatics Institute. |
13:51 | (Applause) |
13:55 | CA: I mean, you're part of Google. |
13:57 | Was there a phone call saying, "Uh, Demis, what did you just do?" |
14:01 | DH: You know, I'm lucky we have very supportive, |
14:04 | Google's really supportive of science |
14:06 | and understand the benefits this can bring to the world. |
14:10 | And, you know, the argument here |
14:12 | was that we could only ever have even scratched the surface |
14:15 | of the potential of what we could do with this. |
14:17 | This, you know, maybe like a millionth |
14:19 | of what the scientific community is doing with it. |
14:22 | There's over a million and a half biologists around the world |
14:25 | have used AlphaFold and its predictions. |
14:27 | We think that's almost every biologist in the world |
14:29 | is making use of this now, every pharma company. |
14:32 | So we'll never know probably what the full impact of it all is. |
14:35 | CA: But you're continuing this work in a new company |
14:37 | that's spinning out of Google called Isomorph. |
14:40 | DH: Isomorphic. |
14:41 | CA: Isomorphic. |
14:43 | Give us just a sense of the vision there. |
14:45 | What's the vision? |
14:47 | DH: AlphaFold is a sort of fundamental biology tool. |
14:50 | Like, what are these 3D structures, |
14:52 | and then what might they do in nature? |
14:55 | And then if you, you know, |
14:57 | the reason I thought about this and was so excited about this, |
15:00 | is that this is the beginnings of understanding disease |
15:03 | and also maybe helpful for designing drugs. |
15:06 | So if you know the shape of the protein, |
15:09 | and then you can kind of figure out |
15:11 | which part of the surface of the protein |
15:13 | you're going to target with your drug compound. |
15:16 | And Isomorphic is extending this work we did in AlphaFold |
15:19 | into the chemistry space, |
15:21 | where we can design chemical compounds |
15:24 | that will bind exactly to the right spot on the protein |
15:27 | and also, importantly, to nothing else in the body. |
15:30 | So it doesn't have any side effects and it's not toxic and so on. |
15:34 | And we're building many other AI models, |
15:37 | sort of sister models to AlphaFold |
15:39 | to help predict, |
15:41 | make predictions in chemistry space. |
15:43 | CA: So we can expect to see |
15:44 | some pretty dramatic health medicine breakthroughs |
15:48 | in the coming few years. |
15:49 | DH: I think we'll be able to get down drug discovery |
15:51 | from years to maybe months. |
15:54 | CA: OK. Demis, I'd like to change direction a bit. |
15:58 | Our mutual friend, Liv Boeree, gave a talk last year at TEDAI |
16:02 | that she called the “Moloch Trap.” |
16:04 | The Moloch Trap is a situation |
16:06 | where organizations, |
16:09 | companies in a competitive situation can be driven to do things |
16:14 | that no individual running those companies would by themselves do. |
16:19 | I was really struck by this talk, |
16:21 | and it's felt, as a sort of layperson observer, |
16:25 | that the Moloch Trap has been shockingly in effect in the last couple of years. |
16:30 | So here you are with DeepMind, |
16:32 | sort of pursuing these amazing medical breakthroughs |
16:35 | and scientific breakthroughs, |
16:37 | and then suddenly, kind of out of left field, |
16:41 | OpenAI with Microsoft releases ChatGPT. |
16:46 | And the world goes crazy and suddenly goes, “Holy crap, AI is ...” |
16:50 | you know, everyone can use it. |
16:54 | And there’s a sort of, it felt like the Moloch Trap in action. |
16:58 | I think Microsoft CEO Satya Nadella actually said, |
17:03 | "Google is the 800-pound gorilla in the search space. |
17:08 | We wanted to make Google dance." |
17:12 | How ...? |
17:14 | And it did, Google did dance. |
17:16 | There was a dramatic response. |
17:18 | Your role was changed, |
17:20 | you took over the whole Google AI effort. |
17:24 | Products were rushed out. |
17:27 | You know, Gemini, some part amazing, part embarrassing. |
17:30 | I’m not going to ask you about Gemini because you’ve addressed it elsewhere. |
17:33 | But it feels like this was the Moloch Trap happening, |
17:37 | that you and others were pushed to do stuff |
17:40 | that you wouldn't have done without this sort of catalyzing competitive thing. |
17:45 | Meta did something similar as well. |
17:47 | They rushed out an open-source version of AI, |
17:50 | which is arguably a reckless act in itself. |
17:55 | This seems terrifying to me. |
17:57 | Is it terrifying? |
17:59 | DH: Look, it's a complicated topic, of course. |
18:01 | And, first of all, I mean, there are many things to say about it. |
18:05 | First of all, we were working on many large language models. |
18:10 | And in fact, obviously, Google research actually invented Transformers, |
18:13 | as you know, |
18:14 | which was the architecture that allowed all this to be possible, |
18:17 | five, six years ago. |
18:19 | And so we had many large models internally. |
18:21 | The thing was, I think what the ChatGPT moment did that changed was, |
18:25 | and fair play to them to do that, was they demonstrated, |
18:28 | I think somewhat surprisingly to themselves as well, |
18:31 | that the public were ready to, |
18:34 | you know, the general public were ready to embrace these systems |
18:37 | and actually find value in these systems. |
18:39 | Impressive though they are, I guess, when we're working on these systems, |
18:43 | mostly you're focusing on the flaws and the things they don't do |
18:46 | and hallucinations and things you're all familiar with now. |
18:49 | We're thinking, you know, |
18:50 | would anyone really find that useful given that it does this and that? |
18:54 | And we would want them to improve those things first, |
18:56 | before putting them out. |
18:58 | But interestingly, it turned out that even with those flaws, |
19:01 | many tens of millions of people still find them very useful. |
19:04 | And so that was an interesting update on maybe the convergence of products |
19:09 | and the science that actually, |
19:13 | all of these amazing things we've been doing in the lab, so to speak, |
19:16 | are actually ready for prime time for general use, |
19:19 | beyond the rarefied world of science. |
19:21 | And I think that's pretty exciting in many ways. |
19:24 | CA: So at the moment, we've got this exciting array of products |
19:27 | which we're all enjoying. |
19:29 | And, you know, all this generative AI stuff is amazing. |
19:31 | But let's roll the clock forward a bit. |
19:34 | Microsoft and OpenAI are reported to be building |
19:38 | or investing like 100 billion dollars |
19:40 | into an absolute monster database supercomputer |
19:45 | that can offer compute at orders of magnitude |
19:49 | more than anything we have today. |
19:52 | It takes like five gigawatts of energy to drive this, it's estimated. |
19:56 | That's the energy of New York City to drive a data center. |
20:00 | So we're pumping all this energy into this giant, vast brain. |
20:04 | Google, I presume is going to match this type of investment, right? |
20:09 | DH: Well, I mean, we don't talk about our specific numbers, |
20:11 | but you know, I think we're investing more than that over time. |
20:15 | So, and that's one of the reasons |
20:17 | we teamed up with Google back in 2014, |
20:19 | is kind of we knew that in order to get to AGI, |
20:23 | we would need a lot of compute. |
20:24 | And that's what's transpired. |
20:26 | And Google, you know, had and still has the most computers. |
20:30 | CA: So Earth is building these giant computers |
20:33 | that are going to basically, these giant brains, |
20:35 | that are going to power so much of the future economy. |
20:38 | And it's all by companies that are in competition with each other. |
20:42 | How will we avoid the situation where someone is getting a lead, |
20:47 | someone else has got 100 billion dollars invested in their thing. |
20:52 | Isn't someone going to go, "Wait a sec. |
20:54 | If we used reinforcement learning here |
20:57 | to maybe have the AI tweak its own code |
21:00 | and rewrite itself and make it so [powerful], |
21:03 | we might be able to catch up in nine hours over the weekend |
21:06 | with what they're doing. |
21:07 | Roll the dice, dammit, we have no choice. |
21:09 | Otherwise we're going to lose a fortune for our shareholders." |
21:12 | How are we going to avoid that? |
21:14 | DH: Yeah, well, we must avoid that, of course, clearly. |
21:16 | And my view is that as we get closer to AGI, |
21:20 | we need to collaborate more. |
21:22 | And the good news is that most of the scientists involved in these labs |
21:27 | know each other very well. |
21:29 | And we talk to each other a lot at conferences and other things. |
21:32 | And this technology is still relatively nascent. |
21:35 | So probably it's OK what's happening at the moment. |
21:37 | But as we get closer to AGI, I think as a society, |
21:42 | we need to start thinking about the types of architectures that get built. |
21:46 | So I'm very optimistic, of course, |
21:48 | that's why I spent my whole life working on AI and working towards AGI. |
21:53 | But I suspect there are many ways to build the architecture safely, robustly, |
22:00 | reliably and in an understandable way. |
22:03 | And I think there are almost certainly going to be ways of building architectures |
22:07 | that are unsafe or risky in some form. |
22:09 | So I see a sort of, |
22:11 | a kind of bottleneck that we have to get humanity through, |
22:14 | which is building safe architectures as the first types of AGI systems. |
22:20 | And then after that, we can have a sort of, |
22:23 | a flourishing of many different types of systems |
22:26 | that are perhaps sharded off those safe architectures |
22:29 | that ideally have some mathematical guarantees |
22:33 | or at least some practical guarantees around what they do. |
22:36 | CA: Do governments have an essential role here |
22:38 | to define what a level playing field looks like |
22:40 | and what is absolutely taboo? |
22:42 | DH: Yeah, I think it's not just about -- |
22:44 | actually I think government and civil society |
22:46 | and academia and all parts of society have a critical role to play here |
22:49 | to shape, along with industry labs, |
22:52 | what that should look like as we get closer to AGI |
22:55 | and the cooperation needed and the collaboration needed, |
22:58 | to prevent that kind of runaway race dynamic happening. |
23:01 | CA: OK, well, it sounds like you remain optimistic. |
23:04 | What's this image here? |
23:05 | DH: That's one of my favorite images, actually. |
23:07 | I call it, like, the tree of all knowledge. |
23:09 | So, you know, we've been talking a lot about science, |
23:12 | and a lot of science can be boiled down to |
23:15 | if you imagine all the knowledge that exists in the world |
23:18 | as a tree of knowledge, |
23:19 | and then maybe what we know today as a civilization is some, you know, |
23:24 | small subset of that. |
23:26 | And I see AI as this tool that allows us, |
23:29 | as scientists, to explore, potentially, the entire tree one day. |
23:33 | And we have this idea of root node problems |
23:36 | that, like AlphaFold, the protein-folding problem, |
23:38 | where if you could crack them, |
23:40 | it unlocks an entire new branch of discovery or new research. |
23:45 | And that's what we try and focus on at DeepMind |
23:47 | and Google DeepMind to crack those. |
23:50 | And if we get this right, then I think we could be, you know, |
23:53 | in this incredible new era of radical abundance, |
23:56 | curing all diseases, |
23:58 | spreading consciousness to the stars. |
24:01 | You know, maximum human flourishing. |
24:03 | CA: We're out of time, |
24:04 | but what's the last example of like, in your dreams, |
24:06 | this dream question that you think there is a shot |
24:09 | that in your lifetime AI might take us to? |
24:12 | DH: I mean, once AGI is built, |
24:14 | what I'd like to use it for is to try and use it to understand |
24:18 | the fundamental nature of reality. |
24:20 | So do experiments at the Planck scale. |
24:23 | You know, the smallest possible scale, theoretical scale, |
24:26 | which is almost like the resolution of reality. |
24:29 | CA: You know, I was brought up religious. |
24:31 | And in the Bible, there’s a story about the tree of knowledge |
24:34 | that doesn't work out very well. |
24:36 | (Laughter) |
24:37 | Is there any scenario |
24:41 | where we discover knowledge that the universe says, |
24:46 | "Humans, you may not know that." |
24:49 | DH: Potentially. |
24:51 | I mean, there might be some unknowable things. |
24:53 | But I think scientific method is the greatest sort of invention |
24:58 | humans have ever come up with. |
24:59 | You know, the enlightenment and scientific discovery. |
25:03 | That's what's built this incredible modern civilization around us |
25:06 | and all the tools that we use. |
25:08 | So I think it's the best technique we have |
25:11 | for understanding the enormity of the universe around us. |
25:15 | CA: Well, Demis, you've already changed the world. |
25:18 | I think probably everyone here will be cheering you on |
25:21 | in your efforts to ensure that we continue to accelerate |
25:24 | in the right direction. |
25:25 | DH: Thank you. |
25:26 | CA: Demis Hassabis. |
25:28 | (Applause) |