Lex Fridman Podcast - #371 – Max Tegmark: The Case for Halting AI Development
Episode Date: April 13, 2023Max Tegmark is a physicist and AI researcher at MIT, co-founder of the Future of Life Institute, and author of Life 3.0: Being Human in the Age of Artificial Intelligence. Please support this podcast ...by checking out our sponsors: - Notion: https://notion.com - InsideTracker: https://insidetracker.com/lex to get 20% off - Indeed: https://indeed.com/lex to get $75 credit EPISODE LINKS: Max's Twitter: https://twitter.com/tegmark Max's Website: https://space.mit.edu/home/tegmark Pause Giant AI Experiments (open letter): https://futureoflife.org/open-letter/pause-giant-ai-experiments Future of Life Institute: https://futureoflife.org Books and resources mentioned: 1. Life 3.0 (book): https://amzn.to/3UB9rXB 2. Meditations on Moloch (essay): https://slatestarcodex.com/2014/07/30/meditations-on-moloch 3. Nuclear winter paper: https://nature.com/articles/s43016-022-00573-0 PODCAST INFO: Podcast website: https://lexfridman.com/podcast Apple Podcasts: https://apple.co/2lwqZIr Spotify: https://spoti.fi/2nEwCF8 RSS: https://lexfridman.com/feed/podcast/ YouTube Full Episodes: https://youtube.com/lexfridman YouTube Clips: https://youtube.com/lexclips SUPPORT & CONNECT: - Check out the sponsors above, it's the best way to support this podcast - Support on Patreon: https://www.patreon.com/lexfridman - Twitter: https://twitter.com/lexfridman - Instagram: https://www.instagram.com/lexfridman - LinkedIn: https://www.linkedin.com/in/lexfridman - Facebook: https://www.facebook.com/lexfridman - Medium: https://medium.com/@lexfridman OUTLINE: Here's the timestamps for the episode. On some podcast players you should be able to click the timestamp to jump to that time. (00:00) - Introduction (07:34) - Intelligent alien civilizations (19:58) - Life 3.0 and superintelligent AI (31:25) - Open letter to pause Giant AI Experiments (56:32) - Maintaining control (1:25:22) - Regulation (1:36:12) - Job automation (1:45:27) - Elon Musk (2:07:09) - Open source (2:13:39) - How AI may kill all humans (2:24:10) - Consciousness (2:33:32) - Nuclear winter (2:44:00) - Questions for AGI
Transcript
Discussion (0)
The following is a conversation with Max Tecmark, his third time in the podcast.
In fact, his first appearance was episode number one of this very podcast.
He is a physicist and artificial intelligence researcher at MIT,
co-founder of Future Left Institute, and author of Life 3.0,
being human in the age of artificial intelligence.
Most recently, he's a key figure in spearheading
the open letter calling for a six-month pause
on giant AI experiments like training GPT-4.
The letter reads,
we're calling for a pause on training of models larger
than GPT-4 for six months.
This does not imply a pause or ban on all AI research
and development, or the use
of systems that have already been placed in the market. Our call is specific, and addresses
a very small pool of actors who possesses this capability. The letter has been signed by
over 50,000 individuals, including 1800 CEOs and over 1500 professors. Signatories include Joshua Bengeo, Stuart Russell, Elon Musk, Steve Wozniak, Yvonneau
Harari, Andrew Yang, and many others.
This is a defining moment in the history of human civilization, where the balance of power
between human and AI begins to shift.
And Max's mind and his voice is one of the most valuable and powerful in a time like this.
His support, his wisdom, his friendship has been a gift on forever, deeply grateful for.
And now a quick few second mention of each sponsor. Check them out in the description.
It's the best way to support this podcast. We got notion for project
and team collaboration inside Tracker for biological data and indeed for hiring. Choose wise
them my friends. Also, speaking of hiring, if you want to work with our amazing team,
we're always hiring whether it's through indeed or otherwise, could electsfreedman.com slash
hiring. And now onto the full batteries, as always, no ads in the middle.
I try to make this interesting, but if you must skip them, please still check out our
sponsors.
I enjoy this stuff.
Maybe you will too.
This show is brought to you by Notion.
A spoken endlessly about how amazing Notion is, how everybody, all the cool kids are recommending
it for just basic note-taking, but there's so, so much more.
It's the collaborative aspect of it, the project management aspect of it, the wikis, the
document sharing, all of that, all in a simple, powerful, beautifully designed solution.
What can I say?
On top of this, there's the notion AI tool.
This is the best integration of large language models into a productivity note-taking tool.
There are so many amazing features.
It's just endless.
Go to the website.
You can generate entire presentations and reports based on a to-do list.
You can summarize stuff.
You can short stuff.
You can generate tables based on the description.
You can write a summary.
You can expand the text. You can change the style of the text. You can fix spelling and grammar. You can write a summary, you can expand the text, you can change the style, the text,
you can fix spelling, grammar, you can translate, you can use simple language, more complicated
language, change the tone of the voice, make it shorter, longer, like I said, everything
is just so easy to play around with and all of it, no matter what you're doing, will challenge
you to think how you write, it will challenge you to expand the style of your writing
It will save you a lot of time of course, but I just think it makes you a better thinker and
productive being in this world and I think that's such a
great integration of AI into the the productivity
Workflow to me. It's not enough for a large language model to be effective at answering questions
and having good dialogue.
You have to really integrate it into the workflow.
And notion, better than anybody else I've seen,
has done that.
So if that's interesting to you,
notion AI helps you work faster, right?
Better than think bigger, doing tasks
that normally take you hours and just minutes,
try notion AI for free when you go to Notion.com slash Lex.
That's all lowercase Notion.com slash Lex to try the power of Notion AI today.
This shows also brought to you by Inside Tracker, a service I use to track biological data.
It's really good to do that kind of thing regularly to look at all the different markers
in your body and to understand what could be made better, the less talented, the diet changes.
It's kind of obvious that decisions about your life should be made based on the data that comes from your body.
Not some kind of population study, although those are good.
Not some spiritual guru, although those are good, not some novel, whether it's Harry Potter,
Dusty Aski, although those are sometimes good. Not your relative who says I heard a guy say that a
guy does this thing that is very broad sounds he's sounding, although sometimes it turns out to be
pretty effective. Overall, the best decisions about your life should be based
on the things that come from your own body. Inside tracker uses algorithms to analyze your
blood data, DNA data, data, fitness tracker, all that kind of stuff to give you recommendations.
You should be doing it. You should be doing it regularly. So it's not just a one time
thing, but regularly over time you see what changes lead to improvements in the various
markers that come from your
body.
Get special savings for a limited time when you go to inside tracker.com slash Lex.
This show is also brought to you by Indeed, a hiring website.
I think the most important thing in life, not to quote, quote in the barbarian because
I'll be very inappropriate to quote at this moment and
It's not actually accurate at all
As a reflection was important in life. It's only has comedic value
What I really want to say about what's important in life is the people you surround yourself with and we spent so much of our time
in the workplace
Seeking solutions to very difficult problems together,
passionately pursuing ambitious goals,
sometimes impossible goals.
That is the source of meaning,
a sort of a fapiness for people.
And I think part of that happiness
comes from the collaboration with other human beings,
the sort of professional depth of connection
that you have with other human beings,
of being together through the grind,
and surviving, and accomplishing the goal,
or failing in a big epic way,
knowing that you have tried together.
And so, doing that with the right team,
I think is one of the most important things in life,
so you should surround yourself with the right team.
If you're looking to join a team,
you should be very selective about that.
Or if you're looking to hire a team,
you should be very selective about that
and use the best tools of the job.
I've used indeed many, many times throughout my life
for the teams I've led.
Don't overspend on hiring.
Visit indeed.com slash Lex.
To start hiring now, that's indeed.com slash Lex terms and conditions apply.
This is Alex Friedman podcast.
To support it, please check out our sponsors in the description.
And now, dear friends, here's Max, tag mark. You were the first ever guest on this podcast episode number one.
So first of all, Max, I just have to say, thank you for giving me a chance.
Thank you for starting this journey.
It's been an incredible journey.
Just thank you for sitting down with me and just acting like I'm somebody who matters,
that I'm somebody who's interested in talk to.
And thank you for doing it.
I meant a lot.
All right.
Thanks to you for putting your heart and soul into this.
I know when you delve into controversial topics,
it's inevitable to hit by what Hamlet talks about the slings and arrows and stuff. And I really
admire this. It's in an era where YouTube videos are too long, and now it has to be like a 20-minute
TikTok, 22-second TikTok clip. It's just so refreshing to see you going exactly against all of the
advice and doing these really, really long form things and the people appreciate it.
Reality is nuanced.
And thanks for sharing it that way.
So let me ask you again, the first question I've ever asked on this podcast, episode
number one, talking to you, do you think there's intelligent life out there in the universe?
Let's revisit that question.
Do you have any updates? What's your view when you look out to the stars?
So when we look out to the stars, if you define our universe the way most astrophysicists do,
not as all of space, but the spherical region of space, that we can see with our telescopes
from which light has the time to reach us since our big bag. I'm in the minority. I'm estimate that we are the only life in this spherical volume
that has invented internet radio's gone our level of tech. And if that's true, then
it puts a lot of responsibility on us to not mess this one up because if that's true, then it puts a lot of responsibility on us
to not mess this one up, because if it's true,
it means that life is quite rare.
And we are stewards of this one spark of advanced consciousness,
which if we nurture it, then help it grow.
It eventually life can spread from here
out into much of our universe.
And we can have this just amazing future,
worse if we instead are reckless with the technology we build and just snuff it out
due to the stupidity or in fighting, then maybe the rest of cosmic history
in our universe was just going to be a playframe of the benches.
But I do think that we are actually
very likely to get visited by alien intelligence quite soon. But I think we are going to be building
that alien intelligence. So we're going to give birth to an intelligent alien civilization.
Unlike anything that human, the evolution here on Earth was able to create
in terms of the path, the biological path it took.
Yeah, and it's going to be much more alien than a cat or even the most exotic animal on
the planet right now, because it will not have been created through the usual Darwinian competition,
where it necessarily cares about self-preservation, the freight of death, any of those things.
The space of alien minds that you can build is just so much faster than what evolution
will give you.
And with that also comes a great responsibility for us to make sure that the kind of minds we create are the kind of minds that
it's good to create minds that will share our values and be good for humanity and life and also
don't create minds that don't suffer.
Do you try to visualize the full space of alien minds that AI could be? Do you try to consider all the different kinds of intelligence?
So, generalizing what humans are able to do to the full spectrum of what intelligent creatures entities could do?
I try, but I would say I fail. I mean, it's very difficult for a human mind to really grapple with something so completely
alien.
Even for us, right?
If we just try to imagine how would it feel if we were completely indifferent towards death
or individuality, even if you just imagine that for example,
you could just copy my knowledge of how to speak Swedish.
Boom, now you can speak Swedish.
And you could copy any of my cool experiences
and then you could delete the ones you didn't like in your own life.
Just like that.
It would already change quite a lot about how you feel as a human being, right?
You probably spend less effort studying things
if you just copy them and you might be less afraid of death
because if the plane you're on starts to crash,
you just be like, oh, shucks,
I'm gonna, I'd have him back in my brain up for four hours.
So I'm gonna lose all this wonderful experiences
of this flight.
We might also start feeling more compassionate, maybe
with other people, if we can so readily share each other's
experiences in our knowledge and feel more like a hive mind.
It's very hard, though.
I really feel very humble about this to grapple with it.
How it might actually feel. The one thing which is so obvious though, I think it's just really worth
reflecting on is because the mind space of possible intelligence is so different for ours,
it's very dangerous if we assume they're going to be like us or anything like us.
Well, there's the entirety of human written history has been through poetry, through novels, been trying to
describe through philosophy, trying to describe the human condition and what's entailed in it, like Jessica
said, fear of death and all those kinds of things, what is love and all of that changes. If you have a different kind of intelligence.
All of it. The entirety, all those poems that are trying to sneak up to what the hell it means
to be human, all of that changes. How AI concerns and existential crises that AI experiences, how that clashes with the human existential
crisis, the human condition.
Yeah.
It's hard to fathom, hard to predict.
It's hard, but it's fascinating to think about it also.
Even in the best case scenario where we don't lose control over the ever more powerful AI
that we're building to other humans whose goals we think are horrible
and where we don't lose control to the machines. AI provides the things we want.
Even then, you get into the questions, do you touch here?
Maybe it's the struggle that it's actually hard to do things,
it's part of the things that gives us meaning as well. So, for example, I found it so shocking that this new Microsoft GPT-4 commercial that
they put together has this woman talking about showing this demo of how she's going to
give a graduation speech to her beloved daughter, and she asks GPT-4 to write it. It was freaking 200 words
or so. If I realized that my parents couldn't be bothered struggling a little bit to write
200 words and outsource that to their computer, I would feel really offended actually.
And so I wonder if eliminating too much of this struggle from our existence,
If I'm eliminating too much of this struggle from our existence, do you think that would also take away a little bit of what means to be human?
Yeah.
I was, we can't even predict.
I had somebody mentioned to me that they use, they started using Chad G.P.TT with a 3.5 and not 4.0 to write what they really feel to a person
and they have a temper issue.
And they're basically trying to get ChaggPT to rewrite it in a nicer way, to get the point
across but rewrite it in a nicer way.
So we're even removing the inner asshole from our communication. So I don't, you know,
there's some positive aspects of that, but mostly it's just the transformation of how humans communicate.
And it's scary because so much of our society is based on this glue of communication. And if that we're not using AI as the medium of communication
that does the language for us, so much of the emotion that's laden in human communication,
so much of the intent that's going to be handled by outsourced AI, how does that change everything?
How does that change the internal state of how we feel about other human beings? What makes us lonely, what makes us excited?
What makes us afraid, how we fall in love, all that kind of stuff.
Yeah.
For me personally, I have to confess the challenge is one of the things really makes my life feel meaningful.
If I go hike and mountain with my wife,
I don't want to just press a button and be at the top.
I want the struggle and come up there sweaty
and feel like, wow, we did this in the same way.
I want to constantly work on myself
to become a better person.
If I say something in anger that I regret,
I want to go back and really work on myself
rather than just telling AI from now on, always filter what I regret. I want to go back and really work on myself rather than just tell an AI
from now on, always filter what I write, so I don't have to work on myself, because then
I'm not growing.
Yeah, but then again, it could be like with chess. An AI once it's significantly obviously
supersedes the performance of humans. It will live in its own world and provide me be a flourishing civilization for humans but we humans will continue hiking mountains and playing our games
even though ai is so much smarter so much stronger so much superior in every single way just like with chess
yeah so that that i mean that's one possible hopefully trajectory here is that humans will
continue to human. And AI will just be a kind of a medium that enables the human experience to flourish. Yeah.
I would phrase that as rebranding ourselves from homo sapiens,
the homo sentience, you know, right now, it's sapiens, the ability to be intelligence, we've even put it in our species name. So we're
branding ourselves as the smartest, yeah, information processing
We're branding ourselves as the smartest information processing entity on the planet. That's clearly going to change if AI continues ahead.
So maybe we should focus on the experience instead, the subjective experience that we have
with homo-sentience.
And say, that's what really valuable.
The love, the connection, the other things.
Get off our high horses and get rid of this hubris that only we can do integrals.
So consciousness, the subjective experience
is a fundamental value to what it means to be human.
Make that the priority.
That feels like a hopeful direction to me, but that also requires more
compassion, not just towards other humans, because they happen to be the smartest
on the planet, but also towards all our other fellow creatures on this planet.
And I personally feel right now we're treating a lot of farm animals horribly,
for example, and the excuse we're using is, oh, they're not as smart as us.
But if we get it, we're not that smart in the grand scheme of things either in the post AI epoch,
then surely we should value the subjective experience of a cow also.
Well, allow me to briefly look at the book, which at this point is becoming more and more visionary that you've written, I guess, over five years ago, Life 3.0. So first of all, 3.0.
What's 1.0?
What's 2.0?
What's 3.0?
And how's that vision sort of evolve, the vision and the book evolve to today?
Life 1.0 is really dumb, like bacteria, and that it can't actually learn anything at all during the lifetime.
The learning just comes from this genetic process from one generation to the next.
Life 2.0 is us and other animals which have brains which can learn during their lifetime a great deal. And so, and you know, you were born without being able
to speak English.
And at some point you decided, hey, I want to upgrade my software.
Let's install an English speaking module.
So you did.
And the live 3.0 does not exist yet.
It can replace not only its software the way we can, but also its hardware.
And that's where we're heading towards at high speed. We're already maybe 2.1 because we can
put in an artificial knee, a pacemaker, etc., etc. And if Newerlink, you know, other companies succeed,
et cetera, et cetera. And if Newerlink, you know, their company succeed, will be life, 2.2, et cetera.
But the companies trying to build age, UI, or trying to make is of course, full 3.0.
And you can put that intelligence in something that also has no biological basis whatsoever.
So less constraints and more capabilities, just like the leap from one point
out to two point out, there is nevertheless you speaking so harshly about bacteria. So disrespectfully
about bacteria, there is still the same kind of magic there that permeates life two point out
and and three point out. It seems like maybe the thing that's truly powerful about life, intelligence, and consciousness was already there in one point out.
Is it possible?
I think we should be humble and not be so quick to make everything binary and say either it's there or it's not clearly there's a great spectrum. And there is even controversy about whether some unicellular organisms like amoebas can
maybe learn a little bit, you know, after all.
So, apologies if I offended any bacteria.
It wasn't my intent.
It was more that I wanted to talk up how cool it is to actually have a brain where you
can learn dramatically within your lifetime.
Typical human.
And the higher up you get from 1.0 to 2.0 to 3.0,
the more you become the captain of your own ship,
the master of your own destiny,
and the less you become a slave to whatever evolution
gave you, right?
By upgrading your software, we can be so different
from previous generations and even from our parents,
much more so than even a bacterium, you know,
no offense to them. And if you can also swap out your hardware, take any physical form you want,
of course, it's really the sky's the limit. Yeah, so the, it accelerates the rate at which you
can perform the computation, the computation that determines your destiny. Yeah, and I think it's
worth commenting a bit
on what you means in this context also.
If you swap things out a lot, right?
This is controversial, but my current understanding
is that life is best thought of not as a bag of meat
or even a bag of elementary particles, but rather it is in as a system which can process information and retain its own complexity,
even though nature is always trying to mess it up.
So it's all about information processing.
And that makes it a lot like something like a wave in the ocean, which is not its water
molecules, right?
The water molecules bob up and down, but the wave moves forward.
It's an information pattern.
In the same way, you, Lex, you're not the same atoms as during the first time you did with
me.
You swapped out most of them, but it's still you.
The information pattern is still there. If you could swap out your arms and
you can still have this kind of continuity, it becomes more sophisticated sort of way before and time where the information lives on. I lost both my parents since I last podcast and it actually gives me a lot of solace that
this way of thinking about them.
They haven't entirely died because a lot of mommy and daddy's, sorry, I'm getting emotional
here, but a lot of their values and ideas and even
jokes and so on, they haven't gone away.
Some of them live on, I can carry on some of them and they also live on a lot of other
and a lot of other people.
So in this sense, even with Life 2.0, we can, to some extent, already transcend our physical bodies and our death.
And particularly if you can share your own information, your own ideas with many others,
like you do in your podcast, then that's the closest immortality we can get with our bio-bodies.
You carry a little bit of them in you. Yeah, that's the closest immortality we can get with our bio bodies.
You carry a little bit of them in you.
Yeah.
Yeah.
Do you miss them?
You miss your mom and dad?
Of course.
Of course.
What did you learn about life from them if it can take a bit of a tangent?
I don't have so many things for starters, my fascination for math and the physical mysteries of our university.
You got a lot of that for my dad, but I think my obsession for really big questions and consciousness
and so on that actually came mostly from my mom. And what I got from both of them, which is a very core part of really who I am, I think
is this feeling comfortable with not buying into what everybody else is saying.
When I think it's right, they both very much did their own thing and sometimes they got
flagged for it and did it anyway.
That's why you've always been an inspiration to me that you're at the top of your field and you still
You're still willing to
To tackle the big questions in your own way. You're one of the one of the people that represents
MIT best to me. He's always been an inspiration of that
So it's good to hear that you got that from your mom dad. Yeah, you're too kind, but yeah, I mean
So it's good to hear that you got that from your mom dad. Yeah, you're too kind, but yeah, I mean
The real the good reason to do science is
Because you're really curious you want to figure out the truth if you think
This is how it is and everyone else says no, no, that's bullshit and it's that way, you know
You stick with what you think is true. And even if everybody else keeps thinking,
it's bullshit, there's a certain...
Yeah.
I always root for the underdog when I watch movies.
And my dad once, I, one time, for example, when I wrote one of my craziest papers ever,
I'm talking about our universe ultimately being mathematical, which we're not going to get into today.
I got this email from a quite famous professor saying, this is not only bullshit, but it's
going to ruin your career.
You should stop doing this kind of stuff.
I sent it to my dad.
Do you know what he said?
He replied with a quote from Dante.
Seguil to Corso, elacidir la gente, follow your own path, and let the people talk.
Go that.
This is the kind of thing.
You're he's dead, but that attitude is not.
How did losing them as a man, as a human being, change you?
How did you expand your thinking about the world?
How did it expand?
You're thinking about this thing we're talking about, which is humans creating
another living sentient, perhaps being.
I think it mainly do two things.
One of them just going through all their stuff, they had passed away and so on, just drove
home to be how important it is to ask yourselves,
why are we doing this, things we do?
Because it's inevitable that you look at some things
they spent an enormous time on and you ask the,
at hindsight, would they really have spent so much time on this?
Or would they have done something that was more meaningful?
So I've been looking more in my life now and asking,
you know, why am I doing
what I'm doing? And I feel it should either be something I really enjoyed doing or it
should be something that I find really, really meaningful because it helps humanity. Um, though. If it's not enough, there's two categories.
Maybe I should spend less time on it, you know.
The other thing is dealing with death up in personal like this.
It's actually made me less afraid of the, um,
even less afraid of other people telling me that I'm an idiot, you know,
which happens regularly.
And I live my life, do my thing, you know.
And it's made it a little bit easier for me to focus on what I feel is really important.
What about the fear of your own death?
Has it made it more real that this is...
That this is something that happens?
Yeah, it's made it extremely real.
And I'm next to the next in line in our family now, right?
It's me and my brother, my younger brother.
But they both handled it with such dignity.
That was a true inspiration also.
They never complained about things. And you know, when you're old and your body start falling apart
It's more and more to complain about they looked at what could they still do that was meaningful and they focused on that rather than wasting time
Like talking about or even thinking much about things they were disappointed in. I think anyone can make themselves depressed if they start their morning by making a list of grievances,
whereas if you start your day on a little meditation, things are grateful for it.
You basically choose to be a happy person. Because you only have a finite
number of days to spend them. Make a count. Be grateful. Yeah. Well, you do happen to be working on a thing
which seems to have potentially some of the greatest impact on human civilization of anything
humans have ever created, which is artificial intelligence. This is on the both detail technical level and in a high philosophical level you work on. So
you've mentioned to me that there's an open letter
that you're working on. It's actually
going live in a few hours.
I've been having late nights and early mornings. It's been very exciting actually. I in short, I
Have you seen don't look up
the film? Yes, yes. I don't want to be the movie spoiler for anyone watching this who hasn't seen it.
But if you're watching this, you haven't seen it. Watch it. Because we are actually acting out. It's life-imitating art. Humanity is doing exactly that right now,
except it's an asteroid that we are building ourselves. Almost nobody is talking about it.
People are squabbling across the planet about all sorts of things which seem very minor compared
to the asteroid that's about to hit us, right? Most politicians don't even have their
radar, this on the radar, they think maybe in 100 years or whatever.
Right now, we're at a fork on the road.
This is the most important fork.
The humanity has reached and it's over 100,000 years on this planet.
We're building effectively a new species.
It's smarter than us.
It doesn't look so much like a species yet because it's mostly not embodied in robots, but that's a technicality which will soon be changed. And this arrival
of artificial general intelligence, they can do all our jobs as well as us, and probably
shortly thereafter superintelligence which greatly exceeds our cognitive abilities, it's
going to either be the best thing ever to happen to humanity or the worst.
I'm really quite confident that there is not that much middle ground there, but it would
be fundamentally transformative to human civilization, utterly and totally.
Again, we'd be branded ourselves as homo sapiens because it seemed like the basic thing where
the king of the castle on this planet, where the smart ones, if we can control everything
else, this could very easily change.
We're certainly not going to be the smartest on the planet very long, unless AI progress
just halts.
And we can talk more about why I think that's true because it's it's controversial and then we can also talk about
Reasons you might think it's gonna be the best thing ever and the reason you think it's gonna be the end of humanity
Which is of course super controversial, but
what I think we can anyone who's working on
advanced AI But what I think we can, anyone who's working on Advanced AI can agree on is it's much like
the film.
Don't look up in that it's just really comical how little serious public debate there
is about it, given how huge it is.
So what we're talking about is a development of currently things like GPT-4 and the signs
it's showing of rapid improvement that may in the near term lead to development of super
intelligent AGI, AI, general AI systems and what kind of impact that has on society. Exactly. When that thing is achieved general human level intelligence and then beyond that general
super human level intelligence.
There's a lot of questions to explore here.
So one, you mentioned HALT.
Is that the content of the letter to suggest that maybe we should pause the development of these systems?
Exactly. So this is very controversial.
From when we talked the first time we talked about, I was involved in starting the future life institutes.
And we worked very hard on 2014, 2015 was the mainstream AI safety.
The idea that even could be risks and that you could do things about them.
Before then, a lot of people thought it was just really cookie to even talk about it.
And a lot of AI researchers felt worried that this was too flaky and could be bad for
funding and that the people who talked about it were just not, didn't understand AI.
I'm very, very happy with how that's gone and that now, you know,
completely mainstream. You go on any AI conference and people talk about AI safety and it's a
tert-nerdy technical field full of equations and similar and blah, blah. As it should be,
but there is this other thing which has been quite taboo up until now,
calling for slowdown. So, what we've constantly been saying, including myself, I've been
biting my tongue a lot, you know, is that, you know, we don't need to slow down AI development.
We just need to win this race, the wisdom race between the growing power of the AI
and the growing wisdom with which we manage it. And rather than trying to slow down AI,
let's just try to accelerate the wisdom. Do all this technical work to figure out how you can
actually ensure that your powerful AI is going to do what you wanted to do and have society adapt
also with incentives and regulations so that
these things could put the good use. Sadly, that didn't pan out. The progress on technical
AI and capabilities has gone a lot faster than many people thought back when we started this in 2014. It turned out to be easier to build
real advanced AI than we thought. On the other side, it's gone much slower than we hoped with getting
policy makers and others to actually put in place incentives in place to make
Steer this in the in the good directions. We can know maybe we should unpack and talk a little bit about each so yeah
Why did it go faster than we then a lot of people thought them?
In hindsight, it's exactly like building
flying machines
People spent a lot of time wondering about how how the birds fly, you know, and that turned
out to be really hard. Have you seen the TED talk with a flying bird? Like a flying robotic
bird? Yeah, it flies around the audience, but it took a hundred years longer to figure
out how to do that than for the Wright Brothers to build the first airplane because it turned
out there was a much easier way to fly. And evolution picked a more complicated one because
it had its hands tied. It could only
build a fly M-machine that could assemble itself, which the Wright brothers didn't care about.
It can only build a machine that used only the most common atoms in the periodic table.
Wright brothers didn't care about that. They could use steel, iron atoms, and it had to be
able to repair itself, and it also had to be incredibly
fuel efficient. A lot of birds use less than half the fuel of a remote control plane
flying the same distance. If we humans throw a little more fuel in a roof, there you go,
100 years earlier. That's exactly what's happening now with these large language models.
The brain is incredibly complicated.
Many people made the mistake you're thinking we have to figure out how the brain does human
level AI first before we could build in a machine.
That was completely wrong.
You can take an incredibly simple computational system called a transformer network and just
train it to do something incredibly dumb.
Just read a gigantic amount of text
and try to predict the next word.
And it turns out, if you just throw a ton of compute
at that and a ton of data,
it gets to be frighteningly good, like GPT-4,
which I've been playing with so much since it came out, right?
And there's still some debate about whether that can get you all the way to full human level
or not.
But yeah, we can come back to the details of that and how you might get the human level
AI even if large language models don't.
Can you briefly, if it's just a small tangent comment on your feelings about GPT-4?
So just that you're impressed by this rate of progress, but where is it?
Can GPT-4 reason?
What are the intuitions?
What are human interpretable words you can assign to the capabilities of GPT-4 that makes
you so damn impressed with it?
I'm both very excited about it and terrified.
It's interesting mixture of emotions.
All the best things in life include those two somehow.
Yeah, I can absolutely reason.
Anyone who hasn't played with it, I highly recommend doing that before dising it.
It can do quite remarkable reasoning. I've had to do a lot of things which I
really like. I couldn't do that myself that well, even. And obviously, there's a dramatically
faster than we do, too, when you watch a type. And it's doing that while servicing a massive
number of other humans at the same time. At the same time, it cannot reason as well as a human can on some tasks,
just because it's obviously a limitation from its architecture.
I mean, we have in our heads what in GeekSpeak is called a recurrent neural network.
There are loops, information you can go from this neuron to this neuron to this neuron,
and then back to this one, you can like,
ruminate on something for a while, you can self-reflect a lot.
These large language models that are,
they cannot, like GPT-4, it's a so-called transformer,
where it's just like a one-way street of information,
basically, in GeekSpeak, it's called a feed-forward
neural network.
And it's only so deep, so it can only do logic
that's that many steps and that deep,
and it's not.
And you can create problems which will fail to solve for that
reason.
But the fact that it can do so amazing things with this incredible,
simple architecture already is quite stunning.
And what we see in my lab at MIT when we look inside
large language models to try to figure out how they're doing it, which that's the key core focus
of our research. It's called mechanistic interpretability in GeekSpeak. You have this machine,
it does something smart, you try to reverse version engineers, see how does it do it?
I think you've also got artificial neuroscience. Exactly.
The neuroscience is to do with actual brains, but here you have the advantage that you
don't have to worry about measurement errors.
You can see what every neuron is doing all the time.
And a recurrent thing we see again and again, there's been a number of beautiful papers quite
recently by a lot of researchers,
some of them here in this area, is where when they figure out how something is done, you
can say, oh man, that's such a dumb way of doing it.
And you immediately see how it can be improved.
For example, there was a beautiful paper recently where they figured out how a large language
model stores certain facts, like Eiffel Tower is in Paris, and they figured out how a large language model stores certain facts like
Eiffel Tower is in Paris.
And they figured out exactly how it's stored and the proof that they understood it was they
could edit it.
They changed some of the synapses in it.
And then they asked it, where is the Eiffel Tower?
And it said, it's in Rome.
And then they asked you, how do you get there?
Oh, how do you get there from Germany?
Oh, you take this train,
the Roma Termini train station and this and that. And what might you see if you're in front of it?
Oh, you might see the Colosseum. So they had edited it. So they literally moved it to Rome.
But the way that it's storing this information, it's incredibly dumb for any fellow nerds listening to this.
There was a big matrix, and roughly speaking, there are certain rowing column vectors, which
encode these things, and they correspond very hand-wavily to principle components.
And it would be much more efficient for sparse matrix to store in the database.
And everything so far, we've figured out how these things do.
Our ways you can see they can easily be improved.
And the fact that this particular architecture
has some roadblocks built into it is in no way
in a prevent crafty researchers from quickly finding work
around and making other kinds of architectures.
So to go all the way.
So it's in short, it's turned out to be a lot easier
to build close to human intelligence than we thought,
and that means our runway is a species.
Get our shit together, it has shortened.
And it seems like the scary thing about the effectiveness of large language models,
some Sam Altman, everything in the conversation with,
and he really showed that the leap from
GPT-3 to GPT-4 has to do with just a bunch of hacks.
A bunch of little explorations about with a smart researchers doing a few little
fixes here and there. It's not some fundamental leap and transformation in the architecture.
And more data and more compute. And more data and compute. But he said the big
leaps has to do with not the data and the compute, but just learning this new discipline, just
like you said.
So researchers are going to look at these architectures and there might be big leaps where you realize,
wait, why are we doing this in this dumb way?
And all of a sudden, this model is 10x smarter.
And that can happen on any one day, on any one Tuesday or one day afternoon.
And then all of a sudden, you have a system that's 10x smarter.
It seems like it's such a new discipline.
It's such a new, like we understand so little about why this thing works so damn well,
that the linear improvement of compute or exponential, but the steady improvement of compute,
steady improvement of the data may not be the thing that even leads to the next leap.
It could be a surprise little hack that improves everything.
There were a lot of little leaps here and there because because so much of this is out in the open
also.
So many smart people are looking at this and trying to figure out little leaps here and
there and it becomes this sort of collective race where if people, a lot of people feel
if I don't take the leap someone else with and it is actually very crucial for the other
part of it.
Why do we want to slow this down? So again, what this open letter is calling for is just pausing all training
of systems that are more powerful than GPT-4 for six months. Give a chance for the labs to
coordinate a bit on safety and for society to adapt, give the right incentives
to the labs because you know, you've interviewed a lot of these people who lead these labs and you know,
just as well as I do, they're good people, they're idealistic people, they're doing this first and
foremost because they believe that AI has a huge potential to help humanity. And but at the same time, they are trapped
in this horrible race to the bottom.
Have you read Meditations on Malok
by Scott Alexander?
Yes.
Yeah, it's a beautiful essay on this poem by Ginsberg
where he interprets it as being about this monster.
It's this game theory monster that pits people into against each other and this race at the bottom where everybody ultimately loses it.
The evil thing about this monster is even though everybody sees it and understands they still can't get out of the race.
They still can't get out of the race, right?
Most a good fraction of all the bad things that we humans do are caused by malloc and I like
Scott Alexander's
naming of the monster so we can we humans can think of it as an if a thing
If you look at why do we have overfishing?
Why do we have more generally the tragedy of the comments? Why is it that
to live or a, I don't know if you had her on your podcast? Yeah, she's become a friend.
Great. She made this awesome point recently that beauty filters that a lot of female
influencers feel pressure to use or exactly malloc in action again. First nobody was using them and people saw them just the way they were.
And then some of them started using it and becoming ever more plastic, fantastic.
And then the other ones that weren't using it started to realize that
if they want to just keep their their market share, they have to start using it too.
And then you're in the situation where they're all using it. And none of them has any more market share or less than before. So nobody gained anything. Everybody lost.
And they have to keep becoming ever more plastic, fantastic also.
But nobody can go back to the old way because it's just too costly, right?
Mollock is everywhere.
And Mollock is not a new arrival on the scene, either.
We humans have developed a lot of collaboration mechanisms to help us fight back against Mollock
through various kinds of constructive collaboration, the Soviet Union and the United States did sign
the number of arms control treaties against Malak who is trying to stoke them into unnecessarily
risky nuclear arms races, etc.
And this is exactly what's happening on the AI front.
This time, it's a little bit geopolitics, but it's mostly money where there's just so much commercial pressure
You know if you take any of these
leaders of the top tech companies
If they just say you know this too risky. I want to pause for six months. They're gonna get a lot of pressure from shareholders and others
I like well, you know if you pause, but those guys don't pause, we're
if you don't want to get our lunch eaten, and shareholders even have a power to replace
the executives in the worst case, right? So we did this open letter because we want to
help these idealistic tech executives to do what their heart tells them by providing enough
public pressure on the whole sector, just pause, so they can all pause in a coordinated fashion.
And I think without the public pressure, none of them can do it alone, push back against their
shareholders, no matter how good-hearted they are. Smaller is a really powerful foe.
So the idea is to, for the major developers of AI systems like this, so we're talking about
Microsoft, Google, Meta, and anyone else. Well, OpenAI is very close with Microsoft Word. And there are plenty of smaller players.
For example, Anthropa is very impressive.
There's conjecture.
There's many, many players that don't want to make a long list.
So we leave anyone out.
And for that reason, it's so important that some coordination happens.
That there's external pressure on all of them, saying you all need
to pause. Because then the researchers in, these organizations, the leaders, they want to slow down a
bit, they can say they're shareholders, you know, everybody's slowing down because of this pressure,
and it's the right thing to do. Have you seen in history their examples was possible to pop the mall?
Absolutely. Even like human cloning for example, you could make so much money on human cloning.
Why aren't we doing it? Because biologists thought hard about this and felt like this is way too risky.
We've they got together all in the 70s in the Silamar and decided even
to stop a lot more stuff also just editing the human germline.
Gene editing that goes in to our offspring and decided, let's not do this
because it's too unpredictable what it's going to lead to.
We could lose control over what happens to our species.
So they paused.
There was a ton of money to be made there.
So it's very doable, but you just need a public awareness of what the risks are.
The broader community coming in and saying, hey, let's slow down.
And another common pushback I get today is we can't stop in the West because China.
And in China and the outer living, they also get told we can't slow down because the West,
because both sides think they're the good guy.
But look at human cloning,
the China forge ahead at human cloning. The China forage ahead with human cloning. There's
been exactly one human cloning that's actually been done that I know of. It was done by Chinese
guy. Do you know where he is now? In jail. And you know, who put him there? Who? Chinese
government? Not because Westerners said China, I look, this is starting to... No, the Chinese government put them there because they also felt they like control.
The Chinese government, if anything, maybe they are even more concerned about having
control than Western governments, have no incentive of just losing control over where
everything is going.
And you can also see the Ernie bot that was released by I believe by due recently.
They got a lot of pushback from the government
and had to rain it in in a big way.
I think once this basic message comes out
that this isn't an arms race, it's a suicide race
where everybody loses if anybody's AI
goes out of control, it really changes the whole dynamic.
And I'll say this again, because this is this very basic point I think a lot of people get wrong. Because a lot of people
dismiss the whole idea that AI can really get very superhuman, because they think there's something
really magical about intelligence such that it can only exist in you in mind
You know because they believe that I think it's kind of kind of get to just more or less
GPT-4 plus plus and then that's it
They don't see it as a super as a suicide race
They think whoever gets that first they're gonna control the world. They're gonna win
That's not how it's gonna be and we can talk again about
the scientific arguments from why it's not going to stop there.
But the way it's going to be is if anybody completely loses control and you don't care,
if someone manages to take over the world who really doesn't share your goals, you probably
don't really even care very much about what nationality they have. You're not going to like it. It's much worse than today. If you live in
an orwellian dystopia, what do you care who created it? If someone, if it goes farther and
we just lose control even to the machines so that it's not us versus them, it's us versus it.
What do you care who created this this unaligned entity which has goals
different from you, and we get marginalized, we get made obsolete, we get
replaced. That's why what I mean when I say it's a suicide race. It's kind of like we're rushing towards this cliff.
But the closer the cliff we get, the more scenic the views are and the more money there is there.
So we keep going.
But we have to also stop at some point, right?
Well, we're ahead.
And it's. It's a suicide race, which cannot be won, but the way that really benefit from it is to
continue developing awesome AI a little bit slower.
So we make it safe, make sure it does the things that humans want, and create a condition
where everybody wins. The technology is shown as that, you
know, geopolitics and politics in general is not a zero sum game at all.
So there is some rate of development that will lead us as a human species to lose control
of this thing. And the hope you have is that there is some lower level of development,
which will not, which will not allow us to lose control.
This is an interesting thought you have about losing control.
So if you have somebody, if you are somebody like son of a proctor or a Sam Altman at the head of a company like this,
you're saying if they develop an AGI, they too will lose control of it.
So no one person can maintain control. No good-but-and-visuals can maintain control.
If it's created very very soon and is a big black box that we don't understand like the large language models
Yeah, then I'm very confident they're gonna lose control
But this isn't just me saying you know Sam Altman and Demis the Saba is have both said
They themselves acknowledge that thing. themselves acknowledge that there's really great
risks with this and they want to slow down once they feel it gets scary. But it's clear
that they're stuck and then again, Malaq is forcing them to go a little faster than they're
comfortable with because of pressure from just commercial pressures, right? To get a
bit optimistic here, of course, this is a problem that can be ultimately solved.
To win this wisdom race, it's clear that what we hope that was going to happen hasn't
happened.
The capability progress has gone faster than a lot of people thought, and the progress
in the public sphere of policy making and so on has gone slower than we thought, even the technical AI safety has gone slower. A lot of the technical safety research
was kind of banking on that large language models and other poorly understood systems couldn't
get us all the way, but you had to build more of a kind of intelligence that you could understand.
Maybe it could prove itself safe, you know, things like this. And I'm quite confident that this can
be done so we can reap all the benefits, but we cannot do it as quickly as this out of
control express train we are now is going to get the AGI. That's why we need a little
more time, I feel. Is there something you've said with like Sam Alman talked about, which is while we're in
the pre-AGI stage to release often and as transparently as possible, to learn a lot.
So as opposed to being extremely cautious, release a lot.
Don't invest in a closed development where you focus on AI safety while it's
somewhat dumb."
Released as often as possible. And as you start to see signs of human
level intelligence or superhuman level intelligence, then you put a halt on it.
Well, what a lot of safety researchers have been saying for many years is that
most dangerous things you can do, within AI, is first of all teach it to write code. Well, what a lot of safety researchers have been saying for many years is that the most
dangerous things you can do within AI is, first of all, teach it to write code, because
that's the first step towards recursive self-improvement, which you can take it from AGI to much higher
levels.
Okay, oops, we've done that.
And another thing, high risk, is connected to the internet, let it go to websites, download stuff on its own,
talk to people, oops, we've done that already.
You know, Elias Yukowsky, you said you interviewed him
recently, right?
Yes, he had this tweet recently, which
I've got, gave me one of the best laughs in a while,
where he was like, hey, people used to make fun of me
and say, you're so stupid, Elias, because you're saying,
you're saying you're saying
you have to worry. Obviously, developers, once they get to really strong AI, first thing you're going to do is never connect it to the internet, keep it in a box where you
can really study it. So he had written it in the meme form, so it was like then.
in the meme forms was like then. Yeah.
And then that.
And then now, let's make a chat bot.
Yeah.
And the third thing is to do it Russell.
Yeah.
You know, amazing AI research.
He has argued for a while that we should never teach AI anything about humans. Above all, we should
never let it learn about human psychology and how you manipulate humans. That's the most
dangerous kind of knowledge you can give it. Yeah, you can teach it all it needs to know
about how to cure cancer and stuff like that. But don't let it read Daniel Connemon's book
about cognitive biases and all that and then
Oops, LOL, you know, let's invent social media
I'll recommend our algorithms with do exactly that they they look get so good at knowing us and pressing our buttons
That we're we're starting to create a world now where we just have ever more hatred because they figured
out that these algorithms, not for out of evil, but just to make money on advertising
that the best way to get more engagement, the euphemism, get people glued to their little
rectangles, or it is just to make them pissed off.
Well, that's really interesting that a large AI system that's doing a recommender system
kind of task on social media is basically
just studying human beings because it's a bunch of us rats giving it signal, non-stop
signal.
It'll show a thing and we give signal and whether we spread that thing, we like that thing,
that thing increases our engagement, gets us to return to the platform, and it has that
on the scale of hundreds of millions of people constantly. So it's just learning and learning and learning. And presumably if the
the number of parameters in neural networks that's doing the learning, and more, and to end the
learning is the more it's able is just basically encode how to manipulate human behavior, how to
control humans at scale. Exactly. And that is not something I think is in humanity's interest.
Yes.
Right now, it's mainly letting some humans manipulate
other humans for profit and power, which already
caused a lot of damage.
And eventually, that's a sort of skill
that can make AI's persuade humans to let them escape whatever
safety precautions we have.
But, you know, there was a really nice article in the New York Times recently by Yval Noah
Harari and two co-authors, including Justin Harris from the Social Dilemma.
We have this phrase in there, I love. Humanity's first contact with advanced AI was social media, and we lost that one.
We now live in a country where there's much more hate in the world where there's much
more hate, in fact.
In our democracy, there we're having this conversation and people can't even agree
on who won the last election, you know.
And we humans often point fingers at other humans and say it's their fault, but it's really malloc and these AI algorithms.
We got the algorithms and then malloc
pitted the social media companies against each other. So nobody could have a less creepy algorithm because then they would lose out on the other company. Is there anywhere to win that battle back just if we just linger on this one battle that
we've lost in terms of social media? Is it possible to redesign social media? This very
medium in which we use as a civilization to communicate with each other to have these
kinds of conversation, to have discourse to try to figure out how to solve the biggest problems in the world, whether that's nuclear war or the development
of AGI.
Is it possible to do social media correctly?
I think it's not only possible, but it's necessary who are we kidding that we're going
to be able to solve all these other challenges if we can't even have a conversation with
each other.
It's constructive.
The whole idea, the key idea of democracy is that you get a bunch of people together,
and they have a real conversation.
The one you try to foster on this podcast, where you respectfully listen to people you disagree
with.
And you realize, actually, you know, there are some things, actually, we, some common ground
we have, and it's, it's, yeah, we both agree, let's not have a nuclear wars, let's not
do that, etc.
We're kidding ourselves, thinking we can face off the second contact with ever more powerful AI that's happening now with these
large language models, if we can't even have a functional conversation in the public space.
That's why I started the Improve the News project,
Improve the News.org.
But I am an optimist fundamentally in,
and there is a lot of intrinsic goodness in people
and that what makes the difference
between someone doing good things for humanity and bad things is not
some sort of fairy tale thing that this person was born with evil gene and this one is not born with
a good gene. No, I think it's whether we put, whether people find themselves in situations that
bring out the best in them or that bring out the worst in them. And I feel we're building an internet
and a society that brings out the worst. This doesn't have to be that way. No, it does not.
The possibility to create incentives and also create incentives that make money. They both make
money and bring out the best in people. I mean, in the long term, it's not a good investment for anyone.
You know, to have a nuclear war, for example.
And, you know, is it a good investment for humanity?
If we just ultimately replace all humans by machines and then we're so obsolete that
eventually the, they're no humans left?
Well, it depends against how you do the math.
But I would say by any reasonable economic standard, if you look at the future
income of humans and there aren't any, that's not a good investment, moreover, why can't
we have a little bit of pride in our species, dammit?
Why should we just build another species that gets rid of us?
If we were Neanderthals, would we really consider the smart move if we had really advanced
biotech to build homo sapiens?
You might say, hey Max, yeah, let's build these homo sapiens.
They're going to be smarter than us.
Maybe they can help us defend this better against predators and help fix our partaves,
make them nicer, and we'll control them undoubtedly.
You know, so then they build a couple, a little baby girl, a little baby boy, they either,
and then you have some wise old and the Andrathal elder who's like,
I'm scared that we're opening a Pandora's box here and that we're gonna get outsmarted by these
and super Neanderthal intelligences
and they're not being in Neanderthals left.
But then you have a bunch of others in the cave,
you're such a Luddite scaremonger, of course,
they're gonna wanna keep us around
because we are their creators and the smarter,
I think the smarter they get, the nicer they're gonna get,
they're gonna leave us, they're gonna want us around and it's gonna be fine.
And besides, look at these babies, they're so cute.
Clearly, they're totally harmless.
Those babies are exactly GPT-4.
It's not, I want to be clear, it's not GPT-4 that's terrifying. It's the GPT-4 is a baby technology.
Microsoft even had a paper recently out
with a title, something like Sparkles of AGI.
Well, they were basically saying this is baby AI,
like these little deandreth old babies.
And it's gonna grow up.
There's gonna be other systems from the same company, from other companies.
They'll be way more powerful, but they're going to take all the things, ideas from these
babies.
And before we know it, we're going to be like those last the underthoughts were pretty
disappointed when they realized that they were getting replaced.
Well, this interesting point you make, which is the programming is, it's, it's
entirely possible. The GPT-4 is already the kind of system that can change
everything by writing programs.
Like three, it's, yeah, it's because it's life 2.0.
I, the systems I'm afraid of are going to look nothing like a large language
model and they're not. I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of, I'm afraid of under the hood, they're like the minimum viable intelligence. They do everything in a dumbest way that still works sort of.
Yeah.
And so they are life 3.0 except when they replace their software, it's a lot faster than us too. So when we don't think on how one logical step, every nanoseconder, or the way they do,
and we can't also just suddenly scale up our hardware massively in the cloud, so limited,
right? So they are, they are also life,
consumed, become a little bit more like life,
people know in that if they need more hardware,
hey, just rent it in the cloud, you know, how do you pay for it?
Well, all the services you provide.
And what we haven't seen yet, which could change a lot, is an entire software system.
So right now programming is done sort of in bits and pieces as an assistant tool to
humans.
But I do a lot of programming and with the kind of stuff that GBT4 is able to do, I mean,
is replacing a lot when I'm able to do.
But I, you still need a human in the loop to kind of manage the design of things, manage
like what are the prompts that generate the kind of stuff to do some basic adjustment
of the code, to do some debugging.
But if it's possible to add on top of GPT-4 kind of feedback loop of
self-debugging, improving the code, and then you launch that system on to the wild on the internet because everything is connected and have it do things
Have it interact with humans and then get that feedback. Now you have this giant ecosystem of humans. It's one of the things that Elon Musk recently sort of tweeted
as a case why everyone needs to pay $7 or whatever for Twitter.
To make sure they're real.
They're making sure they're real. We're now going to be living in a world where the bots are getting
smarter and smarter and smarter to a degree where you can't tell the difference between a human and a bot.
That's right.
And now you can have bots outnumber humans by a 1 million to 1, which is why he's making
a case why you have to pay to prove your human, which is one of the only mechanisms to
prove, which is depressing.
And I feel we have to remember, as individuals, we should, from time to time,
ask ourselves, why are we doing what we're doing?
All right, and as a species, we need to do that too.
So if we're building, as you say, machines
that are outnumbering us and more and more outsmarting us
and replacing us on the job market,
not just for the dangerous and boring tasks,
but also for writing dangerous and boring tasks, but also
for writing poems and doing art and things that a lot of people find really meaningful.
I got to ask ourselves, why?
Why are we doing this?
We are, the answer is malloc is tricking us into doing it.
And it's such a clever trick that even though we see the trick, we still have no choice
but to fall for it, right?
Also the thing you said about you using
co-pilot AI tools to program faster how many time what factor faster would you say you code now?
It's like twice as faster. I don't really
Because it's a new tool. Yeah, it's I don't know, because it's a new tool. Yeah.
It's, I don't know if speed is significantly improved, but it feels like I'm a year away from
being 5 to 10 times faster.
So if that's typical for programmers, then you're already seeing another kind of self,
of course, of self-improvement, right?
Because previously, one, like a major generation of improvement of the codes
would happen on the human R&D time scale.
And now, if that's five times shorter, then it's going to take five times less time
than otherwise would to develop the next level of these tools and so on.
So these are the, this is exactly the sort of beginning of an intelligence explosion.
There can be humans in the loop a lot in the early stages and then eventually humans are
needed less and less and the machines can more kind of go along. But what you weren't, we said
there is just an exact example of these sort of things. Another thing which, which I was kind of
lying on my psychiatrist imagining,
I'm on a psychiatrist's couch here saying,
what are my fears that people would do with AI systems?
Another, so I mentioned three that I had fears
about many years ago that they would do,
namely, teach at the code,
connect it to the internet and teach it to manipulate humans.
A fourth one is building an API, where code can control
this super powerful thing, right?
That's very unfortunate, because one thing
that systems like GPT-4 have going for them
is that they are an oracle in the sense
that they just answer questions.
There is no robot connected to GPT-4.
GPT-4 can't go and do stock trading based on its thinking.
It is not an agent.
An intelligent agent is something that takes in information from the world, processes
it, to figure out what action to take based on its goals that it has, and then does
something back on the world. But once you have an API, for example,
TPD4, nothing stops Joe Schmo and a lot of other people
from building real agents, which just keep making calls
somewhere in some inner loops somewhere
to these powerful Oracle systems, which makes them
themselves much more powerful.
That's another kind of unfortunate development, which I think we would have been better off
delaying.
I don't want to pick on any particular companies.
I think they're all under a lot of pressure to make money.
And again, the reason we're calling for this pause is to give them all covered to do what they
know is the right thing.
Slow down a little bit at this point.
But everything we've talked about, I hope, we'll make it clear to people watching this,
why these sort of human level tools can cause a gradual acceleration.
You keep using yesterday's technology to build tomorrow's technology. And when you do that over and over again, you naturally get an explosion.
You know, that's the definition of an explosion in science, right? Like, if you have two people,
they fall in love. Now we have four people and then they can make more babies and now you have eight people
and then then you have 16, 32, 64, etc.
We call that an population explosion, whereas just that if it's instead free neutrons in
a nuclear reaction, if each one can make more than one, then you get an exponential growth
in that.
We call it a nuclear explosion.
All explosions are like that. An an intelligence explosion is just exactly the same principle.
That some quantum, some amount of intelligence can make more intelligence than that. And
then repeat. You always get the exponentials. What's your intuition? Why does you mention
there's some technical reasons why it doesn't stop at a certain point? What's your intuition?
And do you have any intuition
why it might stop?
It's obviously going to stop when it bumps up against the laws of physics.
There are some things you just can't do no matter how smart you are.
All right.
Allegedly.
And because we don't have the full laws of physics.
Yeah.
Yeah.
Right.
Seth Lloyd wrote a really cool paper on the physical limits on computation.
For example, if you make it put too much energy into it it and then the finite space will turn into a black hole.
You can't move information around fast in the speed of light, stuff like that, but it's
hard to store way more than a modest number of bits per atom, et cetera.
But, you know, those limits are just astronomically above 30 orders of magnitude above where we are now.
So bigger, different, bigger jump in intelligence
than if you go from an aunt to a human.
I think, of course, what we want to do
is have a controlled thing.
A nuclear reactor, you put moderators in to make sure exactly it doesn't blow up out of control. Of course, what we want to do is have a controlled thing.
A nuclear reactor, you put moderators in
to make sure exactly it doesn't blow up out of control, right?
When we do experiments with biology and cells and so on,
we also try to make sure it doesn't get out of control.
We can do this with AI too.
The thing is, we haven't succeeded yet.
And malloc is exactly doing the opposite.
Just fueling, just egging everybody on, faster, faster,
faster, or the other company is going to catch up with you,
or the other country is going to catch up with you.
We do this.
We have to want this stuff.
We have to, and I don't believe in this,
just asking people to look into their hearts
and do the right thing. It's easier for others to say that. But if you're in a situation where
your company is going to get screwed, by other companies, they're not stopping. You're putting
people in a very hard situation. The right thing to do is change the whole incentive structure
instead.
And this is not an old, maybe I should say one more thing
about this, because Mollock has been around as humanity's,
number one or number two enemy since the beginning
of civilization.
And we came up with some really cool countermeasures.
Like, first of all, already over 100,000 years ago,
evolution realized that it was very unhelpful
that people kept killing each other all the time.
So it genetically gave us compassion
and made it so that if you get too drunk dudes
getting into a pointless bar fight,
they might give each other black eyes, but they have a lot of inhibition towards cause towards just killing each other
That's a and similarly if you find a baby lying on the street when you go out for your morning jog tomorrow
You're gonna stop and pick it up, right?
Even though it may be a make you late for your next podcast
it up, right? Even though it may be make you late for your next podcast. So evolution gave us these genes that make our own egoistic incentives more aligned with what's good for the greater group
or part of, right? And then as we got a bit more sophisticated and developed language,
we invented gossip, which is also a fantastic anti-moloc, right?
Because now it really discourages liars, moochers, cheaters, because their own incentive now is not to do this, because word quickly gets around, and then suddenly people are going to invite them to their dinners anymore or trust them. And then when we got still more sophisticated and bigger societies, you know, invented the legal system, where even
strangers who didn't could rely on gossip and things like this would treat each other,
would have an incentive. Now, those guys in the bar fight, even if someone is so drunk that he
actually wants to kill the other guy, he also has a little thought in the back of his head
that, you know, do I really want to spend the next 10 years eating like really crappy food
in a small room? I'm just going to chill out, you know. So, and we similarly have tried
to give these incentives to our corporations by having regulation and also some oversight
so that their incentives are aligned with the greater good.
We tried really hard.
And the big problem that we're failing now
is not that we haven't tried before,
but it's just that the tech is growing much,
it's developing much faster than the regulators
been able to keep up, right?
So regulators, it's kind of comical.
The European Union right now is doing this AI act, right?
Which, and in the beginning, they had a little opt-out
exception that GPT-4 would be completely excluded
from regulation.
Brilliant idea.
What's the logic behind that?
Some lobbyists pushed successfully for this. so we were actually quite involved with the future life institute
Mark Brakel
Rista Uck Anthony gear and others, you know, we're quite involved with
Talking to very educating various people involved in this process about
These general purpose AI models coming and pointing out that they would become the laughing stock if they didn't put it in
So it the friends started pushing they didn't put it in.
So the French started pushing for it, it got put in to the draft and it looked like it was good.
Then there was a huge counter push from lobbyists. There were more lobbyists and Brussels from tech companies and from oil companies, for example. And it looked like it might maybe get taken out again.
And it looked like it might maybe get taken out again. And now GPT-4 happened.
And I think it's going to stay in.
But this just shows, you know, malloc can be defeated.
But the challenge we're facing is that the tech is generally much faster than what the policy makers are.
And a lot of the policy makers also don't have a tech background. So,
you know, we really need to work hard to educate them on how, on what's taking place here.
So, so we're getting this situation where the first kind of non, so I define artificial
intelligence just as non-biological intelligence. And by that definition, a company, a corporation,
is also an artificial intelligence.
Because the corporation isn't its humans, it's the system.
If a CEO decides, if a CEO of a tobacco company decides
one morning, the CEO, he doesn't want to sell cigarettes anymore,
they'll just put another CEO in there.
It's not enough to align the incentives of individual people or
in-line individual computers and incentives to their owners, which is what
technically I safety research is about. You also have to align the incentives
corporations with the greater good. And some corporations have gotten so big and
so powerful very quickly that in many cases, they're lobbyists instead
align the regulators to what they want rather than the early around.
It's a classic regulatory capture.
Right.
Is the thing that the slowdown hopes to achieve is given enough time to the regulars to
catch up or enough time to the companies themselves to breathe and understand
how to do AI safety correctly.
I think both, and I think that the vision of the path to success I see is first you give
a breather actually to the people in these companies, their leadership who wants to do the right
thing and they all have safety teams and so on on their companies.
Give them a chance to get together with the other companies.
And the outside pressure can also help catalyze that, right?
And the workout, what is it that's, what are the reasonable safety requirements when should
put on future systems before they get rolled out?
There are a lot of people also in academia
and elsewhere outside of these companies
who can be brought into this and have a lot of very good ideas.
And then I think it's very realistic that within six months,
you can get these people coming up.
So here's a white paper, here's where we all think
it's reasonable.
You didn't, just because cars killed a lot of people, you didn't ban cars.
But they got together, a bunch of people on the side.
And you know, in order to be allowed to sell a car,
it has to have a seat belt in it.
There are the analogous things that you
can start requiring a future AI systems so that they are safe.
And once this heavy lifting, this intellectual work has been done by experts
in the field, which can be done quickly, I think it's been going to be quite easy to
get policy makers to see, yeah, this is a good idea. And it's, you know, for the companies
to fight malloc, they want, and I believe Sam Altman has explicitly called for this.
They want the regulators to actually adopt it so that their competition is going to divide by two.
You don't want to be enacting all these principles, then you abide by them,
and then there's this one little company that doesn't sign on to it, and then they can gradually overtake
you.
Then the companies will get, be able to sleep, secure knowing that everybody is playing
by the same rules.
So do you think it's possible to develop guardrails that keep the systems from, from basically damaging irreparably humanity, while still enabling sort of the capitalist
fueled competition between companies as they develop how to best make money with a AI.
You think there's a balancing?
Totally.
That's possible.
Absolutely.
We've seen that in many other sectors where you've had the free market produce quite good
things without causing particular harm.
When the guardrails are there and they work,
capitalism is a very good way of optimizing for just getting the same things on more efficiently.
But it was good. I can hindsight, I never met anyone, even on parties way over on the right in any country who think it was a bad
it thinks it was a terrible idea to ban child labor for example.
Yeah, but it seems like this particular technology has gotten so good so fast become powerful
to degree where you could see in the near term the ability to make a lot of money.
Yeah.
And to put guard rails, develop guard rails quickly in that kind of contact seems to be tricky.
It's not similar to cars or child labor.
It seems like the opportunity to make a lot of money here very quickly is right here before
us.
Yeah, there's this cliff.
Yeah.
It gets quickly.
It gets closer. There's a cliff that you go. Yeah. There's this cliff. Yeah. It gets quite seated. Closer. There you go. One of the more, there are more money there is. More gold
and ingots there on the ground. You can pick up or whatever. You want to drive there very fast.
But it's not in anyone's incentive that we go over the cliff. And it's not like everybody's
in the wrong car. All the cars are connected together with a chain. So if anyone goes over,
they'll start driving the others down. The others down too. And so ultimately it's in the selfless
interests also of the people in the companies to slow down when
when you just start seeing the contours of the cliff there in front of you.
And the problem is that even though the people who are building the technology
and the CEOs, they really get
it.
The shareholders and these other market forces, they are people who don't, honestly.
Understand that the cliff is there.
They usually don't...
You have to get quite into the weeds to really appreciate how powerful this is and how
fast.
And a lot of people are even still stuck again in this idea that intelligence in this carbon chauvinism, as I like to call it, that you can only have
our level of intelligence in humans.
There's something magical about it, whereas the people in the tech companies who build
this stuff, they all realize that intelligence is information processing of a certain kind.
And it really doesn't matter at all
whether the information is processed by carbon atoms in neurons and brains or by silicon atoms
and some technology we build. So you brought up capitalism earlier and there are a lot of people who really, really don't.
And it struck me recently that what's happening with capitalism here is exactly analogous
to the way in which super intelligence might wipe us out.
So you know, I studied economics for my undergrad, Stockholm School of Economics, yay.
Well, no, I tell them.
So I was very interested in how you could use market forces,
to just get stuff done more efficiently,
but give the right incentives to market so that it wouldn't do really bad things.
So Dylan had Phil Manell, who's a professor and colleague of mine at MIT, wrote this really
interesting paper with some collaborators recently, where they proved mathematically
that if you just take one goal that you just optimized for, on and on and on indefinitely,
that you think is going to bring you in the right direction.
What basically always happens is, in the beginning,
it will make things better for you.
But if you keep going, at some point,
it's gonna start making things worse for you again.
And then gradually it's gonna make it really, really terrible.
So just as a simple, the way I think of the proof is,
suppose you wanna go from here, back to Austin, for example,
and you're like, okay, yeah, let's just go south,
but you put in exactly the right sort of the right direction.
Just optimize that, south is possible.
You get closer and closer to Austin,
but there was always some little error,
so you're not going exactly towards Austin,
but you get pretty close,
but eventually you start going away again. and eventually you're going to be leaving
the solar system.
And they proved, it's a beautiful mathematical proof.
This happens generally.
And this is very important for AI, because even though Stuart Russell has written a book
and given a lot of talks on why it's a bad idea to have AI just blindly
optimize something.
That's what pretty much all our systems do.
We have something called the loss function that we're just minimizing or reward function
or just minimize, maximize.
And capitalism is exactly like that too.
We want to get stuff done more efficiently than people wanted. So, introduced the free
market. Things got done much more efficiently than they did in, say, communism, right? And
it got better. But then it just kept optimizing, and kept optimizing. And you got ever bigger
companies and ever more efficient information processing.
And I was also very much powered by IT.
And eventually a lot of people are beginning to feel, wait, we're kind of optimizing a bit
too much.
Like why did we just chop down half the rainforest?
And why did suddenly these regulators get captured by lobbyists and so on?
It's just the same optimization that's been running for too long.
If you have an AI that actually has power over the world and you just give it one goal and just
keep optimizing that, most likely everybody's going to be like, yay, it's great in the beginning,
things are getting better. But it's almost impossible to give it exactly the right direction to optimize
in. And then eventually, I'll have a break loose, right? Nick Bossdram and others are given
an example to sound quite silly, like what if you just want to like tell it to cure cancer
or something. And that's all you tell it. Maybe it's going to decide to take over
entire continents just so we can get
more supercomputer facilities in there
and figure out how to cure cancer backwards.
Then you're like, wait, that's not what I want, then.
The issue with capitalism and the issue
in front of my way, I have kind of merged now
because the malloc I talked about is exactly the capitalist malloc that we have built
an economy that has its optimized for only one thing, profit.
And that worked great back when things were very inefficient and then now it's getting
done better.
And it worked great as long as the companies were small enough that they couldn't capture
the regulators. But that's not true anymore, but to keep optimizing. And now they realize
that these companies can make even more profit by building ever more powerfully, even if it's
reckless, but optimize more and more and more and more.
So this is malloc again showing up.
And I just want to anyone here has any concerns
about late stage capitalism having gone a little too far?
You should worry about superintelligence
because it's the same villain in both cases.
It's malloc. And optimizing one objective function aggressively, blindly is going to take us there.
Yeah, we have this pause from time to time and look into our hearts and ask,
why are we doing this? Is this, am I still going towards Austin or have I gone too far?
You know, maybe we should change direction.
And that is the idea behind a halt for six months. Why six months?
It seems like a very short period.
Just can we just linger and explore different ideas here?
Because this feels like a really important moment in human history,
where a pause in would actually have a significant positive effect.
We said six months because we figured the number one pushback
that we were going to get in the West was like, but China.
And everybody knows there's no way that China is going to catch up with the West
on this in six months.
So that argument goes off the table and you can forget about geopolitical competition
and just focus on the real issue.
That's why we put this.
That's really interesting.
But you've already made the case that even for China,
if you actually want to take on that argument,
China too would not be bothered by a longer halt
because they don't want to lose control, even more than
the OS doesn't.
That's what I think.
That's a really interesting argument.
I have to actually really think about that, which the kind of thing people assume is if
you develop an AGI, that open AI, if they're the ones that do it, for example, they're going
to win.
But you're saying, no, everybody loses.
Yeah, it's going to get better and better and better and then
Kepu and we all lose. That's what's going to happen.
When losing, when a define an ametric of basically quality of life
for human civilization and for Sam Altman.
To be blunt, my personal guess, you know, and people think
Quibble with this is that we're just going to, there won't be any humans. That's it. That's what I mean by lose.
You know, if you, we can see in history, once you have some species or some group of people who aren't needed anymore, doesn't usually work out so well for them, right?
Yeah. There were a lot of horses for the were used for traffic in Boston and then the car got invented and most of them got, you know, I mean, only to go there.
And if you look at humans, you know, right now we why did the labor movement succeed and after the industrial revolution?
Because it was needed.
Even though we had a lot of mollocks
and there was child labor and so on,
the company still needed to have workers.
And that's why strikes had power and so on.
If we get to the point where most humans aren't needed
anymore, I think it's quite naive
to think that they're going to still be treated well.
You know, we say that, yeah, yeah, everybody's equal and the government will always protect
them, but if you look in practice, groups that are very disenfranchised and don't have any
actual power usually gets screwed. And now, in the beginning, so industrial revolution,
we automated away muscle work.
But that got, went, worked out pretty well eventually,
because we educated ourselves and started
to do it working with our brains instead
and got, usually, more interesting, better paid jobs.
But now, we're beginning to replace brain work.
So we replaced a lot of boring stuff.
Like we got the pocket calculator so you don't have people adding, multiplying numbers anymore at work.
Fine. There were better jobs they could get. But now GPT-4, you know, and the stable diffusion and
techniques like this, they're really beginning to blow away some jobs that people really
love having. It was a heartbreaking article just posted just yesterday on social media I saw,
but this guy who was doing 3D modeling for gaming and he, and all of a sudden now that he got this
new software, he just says prompts and he feels his whole job that he loved sloths. It's meaning, you know and I
Asked the GPT-4 to rewrite twinkle twinkle little star in the style of Shakespeare
I
Couldn't have done such a good job. It was just really impressive. You've seen a lot of the art coming out here, right? So I'm all for
Automating away the dangerous jobs and the boring jobs, but? So I'm all for automated way, the dangerous jobs and the boring jobs. But I think
you hear some arguments which are too glib. Sometimes people say, well, that's all that's going
to happen. We're getting rid of the boring tedious, dangerous jobs. It's just not true. There are
a lot of really interesting jobs that are being taken away now. Journalism is getting crushed.
jobs that are being taken away now. Journalism is getting crushed.
Coding is gonna get crushed.
I predict the job market for programmers,
salaries are gonna start dropping.
You said you can code five times faster,
then you need five times fewer programmers.
Maybe there will be more output also,
but you'll still end up using fewer programmers
needing fewer programmers than today.
And I love coding, you know, I think it's super cool.
So we need to stop and ask ourselves why again, are we doing this as humans?
I feel that AI should be built by humanity for humanity.
And let's not forget that.
It shouldn't be by malloc, for malloc.
Or what it really is now is kind of by humanity for malloc
which doesn't make any sense.
It's for us that we're doing it.
And make a lot more sense if we build,
develop, figure out, gradually safely how to make all this tech.
And then we think about what are the kind of jobs that people really don't want to have, you know, automate them all
away. And then we ask what are the jobs that people really find meaning in, like maybe
taking care of children, the care center, maybe doing art, etc, etc. And even if it were possible
to automate that way, we don't need to do that, right? We built these machines.
Was possible that we redefine or rediscover what are the jobs that give us meaning?
So for me, the thing, it is really sad. I'm excited. Half the time I'm crying,
Half the time I'm excited, half the time I'm crying, as I'm generating code, because I kind of love programming.
It's an act of creation.
You have an idea, you design it, and then you bring it to life, and it does something,
especially if there's some intelligence, it does something.
It doesn't even have to have intelligence.
Printing a Hello World on screen, you made a little machine and it comes to life.
And there's a bunch of tricks you learn along the way because you've been doing it for
many, many years.
And then to see AI, be able to generate all the tricks you thought were special.
I don't know, it's very, it's scary. It's almost painful. Like a loss, a loss of innocence
maybe like maybe when I was younger. I remember before I learned that sugar is bad for you.
You should be on a diet. I remember I enjoyed candy deeply. In a way, I just can't anymore
I remember I enjoyed candy deeply in a way I just can't anymore that I know is bad for me. I enjoyed it unapologetically fully just intensely and I just I lost that now I feel like a
little bit of that is lost for me with program or being lost with programming. Similar as it is for
the 3D modeler,
no longer being able to really enjoy
the art of modeling 3D things for gaming.
I don't know what to make sense of that.
Maybe I would rediscover that the true magic
of what it means to be human is connecting
without the humans to have conversations like this.
I don't know, to have sex, to eat food,
to really intensify the value from conscious
experiences versus like creating other stuff. You're pitching the rebranding again
from homo sapiens, the homo sapiens, the meaningful experiences. And just to
inject some optimism in this here so we don't sound like it's a gloomers.
You know, we can totally have our cake and eat it. You hear a lot of totally
bullshit claims that we can't afford to have a more teachers.
Yeah.
Have to cut the number of nurses. That's just nonsense, obviously.
With anything, even quite far short of AGI, we can dramatically improve, grow the GDP
and produce this wealth of goods and services. It's very easy to create a world where everybody's
better off than today, including the richest people can be better off as well. It's not a zero-sum
game technology. You can have two countries like Sweden and Denmark, all these ridiculous wars
century after century. And sometimes that's Sweden got a little better off
because it got a little bit bigger
and then Denmark got a little bit better off
because Sweden got a little bit smaller.
But then technology came along
and we both got just dramatically wealthier
without taking away from anyone else.
It was just a total win for everyone.
And AI can do that on steroids.
If you can build safe AGI, if you can build super intelligence,
you know, basically all the limitations that cause harm today can be completely eliminated.
All right, so wonderful. You talk possibility. And this is not sci-fi. This is something which
is clearly possible according to the laws of physics. And I can talk about ways of making it safe also.
But unfortunately, that'll only happen if we steer in that direction.
That's absolutely not the default outcome.
That's why income inequality keeps going up.
That's why the life expectancy in the US has been going down.
Now, I think it's four years in a row.
I was just read a heartbreaking study from CDC about how
something like one third of all teenage girls in the US
been thinking about suicide, you know, like
those are steps in their totally the wrong direction and and it's important to keep our eyes on the prize here that
And it's important to keep our eyes on the prize here that we can, we have the power now for the first time in the history of our species to harness artificial intelligence to help us really flourish and help bring out the best in our humanity rather than the worst of it to help us have really fulfilling
experiences that feel truly meaningful and you and I shouldn't sit here and dictate the future generations what they will be, let them figure it out but let's give them a chance to live
and not foreclose all these possibilities for them by just messing things up, right?
And for that we have to solve the AI safety problem. It would be nice if we can link on exploring that a little bit.
One interesting way to enter that discussion is you tweeted and Elon replied, you tweeted,
let's not just focus on whether GPT-4 will do more harm or good on the job market, but
also whether it's coding skills will hasten the arrival of superintelligence. That's something we've been talking about. So Elon proposed one thing
in the reply saying, maximum truth seeking is my best guess for AI safety. Can you maybe
steal me on the case for this, uh, sense, this objective function of truth and, uh, maybe
make an argument against it in general,
what are your different ideas to start approaching the solution to AI safety?
I didn't see that reply, actually.
Oh, sure.
But I really resonate with it because AI is not evil.
It caused people around the world to hate each other much more, but that's
because we made it in a certain way. It's a tool. We can use it for great things and bad
things, and we could just as well have AI systems. And this is part of my vision for success
here, truth-seeking AI, that really brings us together again. Why do people hate each other so much between countries and within countries?
It's because they have totally different versions of the truth. If they all had the same truth,
they trusted for good reason, because they could check it and verify it and not have to believe in some self-proclaimed authority, right? There wouldn't be as nearly as much hate. There'd be a lot more understanding instead.
And this is, I think something AI can help enormously with.
For example, a little baby step in this direction
is this website called Metaculous,
where people bet and make predictions not for money,
but just for their own reputation.
It's funny, actually, you treat the humans like you treat AI as you have a loss function
where they get penalized if they're super confident on something and then the opposite happens.
Whereas, if you're humble and then I think it's 51% chance this is going to happen and
then the other happens, you don't get penalized much. And what you can see is that some people are much better
at predicting than others. They've earned your trust, right? One project that I'm working on right
now is the Outgrowth Improve the News Foundation together with the MetaXylist folks is seeing if we
can really scale this up a lot with more powerful AI. I would love it, I would love for there to be
a really powerful truth-seeking system
where that is trustworthy
because it keeps being right about stuff
and people who come to it and maybe look at its latest
trust ranking of different pundits and newspapers,
et cetera. If they wanna know why someone got a low score, maybe look at its latest trust ranking of different pundits and newspapers, etc.
If they want to know why someone got a low score, they can click on it and see all the
predictions that they actually made and how they turned out.
This is how we do it in science.
You trust scientists like Einstein who said something, everybody thought it was bullshit
and turned out to be right.
You get a lot of trust point and he did it multiple times even.
I think AI has the power to really heal a lot of the rifts
we're seeing by creating trust system.
It has to get away from this idea today with some fact checking
site which might themselves have an agenda and you just trust it
because of its reputation.
You want to have at the, so these sort of systems they earn their trust and they're completely
transparent.
This I think would actually help a lot that I think help heal the very dysfunctional conversation
that humanity has about how it's going to deal with all its
biggest challenges in the world today. And then on the technical side,
another common sort of gloom comment I get from people who are saying, we're just screwed,
there's no hope, is, well, things like GPT-4 are way too complicated for a human to ever understand
and prove that they can be trustworthy. They're forgetting that AI can help us prove the things work, right?
And there's this very fundamental fact that in math, it's much harder to come up with a proof
than it is to verify that the proof is correct. You can actually write a little proof checking code.
It's quite short, but you can as human understand.
And then it can check the most monstrously long proof,
ever generated even by a computer.
And so yeah, this is valid.
So right now, we have this approach with virus checking software that it looks to see if there's something you should not trust it.
And if it can prove to itself that you should not trust that code, it warns you.
Mine.
What if you flip this around?
And this is an idea, I give credit to Steve, I'm a hundred or four.
So that it will only run the code if it can prove,
instead of not running it, if it can prove that it's not trust
with it, if it will only run it,
if it can prove that it's trust with it.
So it asks the code, prove to me that you're
going to do what you say you're going to do.
It gives you this proof.
I'm your little proof trigger, can check it.
Now you can actually trust an AI that's much more intelligent
than you are, right?
Because it's problem to come up with this proof
that you could never have found that you should trust it.
So this is the interesting point.
I agree with you, but this is where Elias or Yikowsky
might disagree with you.
His claim, not with you, but with this idea.
His claim as a super intelligent AI would be able to know how to lie to you with such a proof. How to lie to
you and give me a proof that I'm going to think is correct. Yeah. But it's not me
is lying to you. That's the trick my proof checker. So he's a code. So his general idea is a super intelligent system can lie to a dumber proof checker.
So you're going to have as a system because more and more intelligent, there's going to
be a threshold where a super intelligent system will be able to effectively lie just slightly
dumber aji system. Like there's a threat, like he really
focuses on this weak aji, strong aji jump, where the strong aji can make all the weak aji
think that it's just one of them, but it's no longer that. And that leap is when it runs away.
Yeah, I don't buy that argument. I think no matter how super intelligent an AI is,
it's never going to be able to prove to me
that there are only finitely many primes, for example.
It just can't.
And it can try to snow me by making up also the new weird rules
of the dutgeon that, say, trust me,
you know, the way your proof checker works is too
limited, and we have this new hyper map, and it's true. But then I would just take that
to, okay, I'm going to forfeit some of these supposedly super cool technologies. I'm
only going to go with the ones that I can prove in my own trusted proof checker. Then I
don't, I think it's fine. There's still, of course,
this is not something anyone has successfully implemented at this point, but I think I just give it as an example of hope. We don't have to do all the work ourselves. This is exactly the
very boring and tedious task that is perfect outsourced to an AI. And this is a way in which less
powerful and less intelligent agents like us can actually
continue to control and trust more powerful ones. So build a GI systems that help us
defend against other AGI systems. Well, for starters, begin with a simple problem of just making
sure that the system that you own or that's supposed to be loyal to you has to prove to itself that
it's always going to do the things that you actually wanted to do right and if it can't prove it, maybe it's still going to do it, but you won't run it.
So you just forfeit some aspects of all the cool things I can do.
I bet you don't know that you can still do some incredibly cool stuff for you.
There are other things too that we shouldn't speak under the rug.
Not every human agrees on exactly what direction we should go with humanity.
You've talked a lot about geopolitical things on your podcast to this effect.
I think that shouldn't distract us from the fact that there are actually a lot of things
that everybody in the world virtually agrees on, that hey, you know, like having no humans on the planet
in a near future, let's not do that, right?
You looked at something like the United Nations
Sustainable Development Goals.
Some are quite ambitious.
And basically all the countries agree.
US, China, Russia, Ukraine, all agree.
So instead of quibbling about the little things we don't agree on,
let's start with the things we do agree on and get them done.
Instead of being so distracted by all these things we disagree on,
that Mollock wins, because frankly, Mocke going wild now. It feels like a war on life
Playing out in front of our eyes if you just look at it from space, you know, we're on this planet
Beautiful vibrant ecosystem now we start chopping down
Big parts of it even though nobody most people thought that was a bad idea, always
start doing ocean acidification, wiping out all sorts of species.
Now, we have all these close calls.
We almost had a nuclear war.
And we're replacing more and more of the biosphere with non-living things.
We're also replacing in our social lives a lot of the things which we're so valuable to humanity.
A lot of social interactions now are replaced by people staring into their rectangles, right? And I
I'm not a psychologist. I'm out of my depth here, but I suspect that part of the reason why
teen suicide and suicide in general in the US, the record breaking level is actually caused by,
again, AI technologies and social media, making people spend less time with actually just human
interaction. We've all seen a bunch of good-looking people and restaurants into the rectangles instead
of looking into each other's eyes, right? So that's also part of the war on life that we're replacing so many really life-affirming
things by technology.
We're putting technology between us, the technology that was supposed to connect us is
actually distancing ourselves from each other.
And then we're giving every more power to things that are not alive.
These large corporations are not living things, right? They're just maximizing profit.
They're... I want to win them more on life. I think we humans together with all our fellow living things on this planet will be better off if we can
Remaining control over the non-living things and make sure that they work for us. I really think it can be done
Can you just linger on this
maybe high-level philosophical disagreement with Eliezery Galski?
I philosophical disagreement with Eliezeria Galski.
I, in this, the hope you're stating.
So he is very sure.
He puts a very high probability, very close to one,
depending on the day, he puts it at one that AI is going to kill humans.
That there's just, he does not see a trajectory which he doesn't end up with that conclusion. What trajectory do you see that doesn't end up there? And maybe can you
see the point he's making and can you also see a way out? First of all, I tremendously respect Elias Yutkowski and his thinking.
Second, I do share his view that there's a pretty large chance that we're not going to make it as humans.
There won't be any humans on the planet and not a distant future.
And that makes me very sad.
We just had a little baby and I keep asking myself, you know, is...
very sad, you know, we just had a little baby and I keep asking myself, you know, is, um, how old is even you know, get, you know, and I asked myself, it feels, I said to my wife,
recently, it feels a little bit like I was just diagnosed with some sort of cancer, which has some, you know, risks of dying from and some risk of surviving, you
know, except this is a kind of cancer, which will kill all of humanity. So I completely
take seriously his concerns. I think, but I don't absolutely don't think it's hopeless. I think there is a, there is a, first of all, a lot of momentum now.
For the first time actually since the many, many years that have passed since I and many
other started worrying about this, I feel most people are getting it now. I just talking to this guy in the gas station, they were a house
the other day. And he's like, I think we're getting replaced. And I think in it. So that's
positive that they're finally, we're finally seeing this reaction, which is the first step
towards solving the problem.
Second, I really think that this vision of only running AI's, really, if that takes a really
high, they can prove to us that they're safe.
It's really just virus checking in reverse again.
I think it's scientifically doable.
I don't think it's hopeless.
We might have to forfeit some of the technology that we could get if we were putting blind faith in our AIs,
but we're still gonna get amazing stuff.
The envision a process with a proof checker,
like something like GPT-4, GPT-5,
will go through a process of regress.
No.
No, I think it's hopeless
that's like trying to prove there are about five spaghetti.
Okay. What I think, well hopeless. That's like trying to prove there about five spaghetti.
What I think, well, the vision I have for success is instead that,
yeah, just like we human beings were able to look at our brains and
and distill out the key knowledge. Galileo, when his dad threw him an apple when he was a kid, he was able to catch it because his brain could and his funny spaghetti kind of way predict how parabolas are going to move his quantum on system one right. But then he got older
and it's like, wait, this is a parabola. It's Y equals X squared. I can still just knowledge out
and today you can easily program it into a computer and it can simulate not just that but how to get
tamaras and so on right. I envision a similar process where we use the amazing learning power of neural networks
to discover the knowledge in the first place,
but we don't stop with a black box and use that.
We then do a second round of AI
where we use automated systems to extract out the knowledge
and see what are the insights it's had, okay?
And then we put that knowledge
into a completely different kind of architecture
or programming language or whatever that's made
in a way that can be both really efficient
and also is more amenable to very formal verification.
That's my vision.
I'm not saying, sitting here saying I'm confident 100% sure that it's going to work.
But I don't think it's a chance of certainly not zero either.
It will certainly be possible to do for a lot of
really cool AI applications that we're not using now.
So we can have a lot of the fun that we're excited about if we do this.
We're going to need a little bit of time.
That's why it's good to pause and put in place requirements.
One more thing also, I think someone might think, well, zero percent chance we're going to
survive.
Let's just give up, right?
That's very dangerous because there's no more guaranteed way to fail than to
convince yourself that it's impossible and not to try. You know, any, you know, when you
study history and military history, the first thing you learn is that that's how you do
psychological warfare. You persuade the other side that it's hopeless so they don't even fight.
And then of course you win, right?
Let's not do this psychological warfare on ourselves and say there's 100% probability
we're all screwed anyway.
It's sadly, I do get that a little bit.
Sometimes from young people who are so convinced that we're
all screwed, that they're like, I'm just going to play computer games and do drugs because
we're screwed anyway.
It's important to keep the hope alive because it actually has a causal impact and makes
it more likely that we're going to succeed.
It seems like the people that actually build solutions to a problem seem really impossible to solve problems are the ones that believe. Yeah. They would
go on score the optimists. Yeah. And it's like, it seems like there's some fundamental
law to the universe where fake it till you make it kind of works. Like believe is possible
and it becomes possible. Yeah. Was it Henry Ford who said that if you can,
if you tell yourself that it's impossible, it is.
So let's not make that mistake.
And this is a big mistake society is making.
I think all and all.
Everybody's so gloomy and the media are also very biased
towards if it bleeds at leads and gloom and doom.
Might so most visions of the future we have or dystopian which really demotivates
people. We want to really really really focus on the upside also to give people the willingness
to fight for it. And for AI, you and I mostly talked about gloom here again, but let's not remember, not forget that you know
We have probably both lost
Someone we really cared about some disease that we were told was incurable. Well, it's not there's no law of physics saying
We had to die of that cancer or whatever. Of course you can cure it
And there's so many other things where that, with a human intelligence, have also failed
to solve on this planet, which AI could also very much help us with.
So if we can get this right, just be a little more chill and slow down a little bit so we
get it right.
It's mind blowing how awesome our future can be.
We talked a lot about stuff on Earth, it can be great.
But even if you really get ambitious and look up at the skies, there's no reason we have to
be stuck on this planet for the rest of the remains for billions of years to come. We totally
understand now that lots of physics let life spread out into space, to other solar systems,
to other galaxies,
and flourish for billions of billions of years.
And this, to me, is a very, very hopeful vision that really motivates me to fight.
And coming back to the end of something you talked about again, you know, the struggle,
how the human struggle is one of the things that also really gives meaning to our lives.
If there's ever been an epic struggle, this is it. Isn't it even more epic if you're the underdog?
If most people are telling you this is gonna fail, it's impossible, right?
And you persist. And you succeed.
That's what we can do together with the species on this one.
A lot of pundits are ready to count us out.
Both in the battle to keep AI safe
and becoming a multibillionaire species.
Yeah, and they're the same challenge.
If we can keep AI safe,
that's how we're going to get multplanetary very efficiently.
I have some sort of technical questions
about how to get it right.
So one idea that I'm not even sure
what the right answer is to is,
should systems like GPT-4 be open-source
in the whole or in part?
Can you see the case for either?
I think the answer right now is no.
I think the answer early on was yes.
So we could bring in all the wonderful great thought process of everybody on this. But asking should we open source GPT-4 now
is just the same as if you say well, should we open source how to build really small nuclear weapons, should we open source how to make bio weapons,
should the open source how to make a new virus that kills 90% of it, where it gets it, of course,
we shouldn't. So it's already that powerful. It's already that powerful that we have to respect
the power of the systems we've built. The knowledge that you get from open sourcing everything we do now,
might very well be powerful enough that people looking at that can use it to
build the things that you're really threatening again. Let's get it.
Remember, OpenAI is GPT-4 is a baby AI. Baby, sort of baby, proto, almost a little bit AGI,
according to what Microsoft's recent paper said, right?
It's not that they were scared of.
What we're scared about is people taking that who are,
who might be a lot less responsible
than the company that made it.
And just go in a town with it.
That's why we want to, it's an information hazard.
There are many things which are not open-source right now
in society for a very good reason.
How do you make certain kind of very powerful toxins out
of stuff you can buy at Home depot. We don't open source
those things for reason. And this is really no different. So, I'm saying that I have to say,
it feels a bit weird, in a way that we are to say it because MIT is like the cradle of the open
source movement. And I love open source in general power to the people, let's say.
But there's always gonna be some stuff that you don't open source.
And it's just like you don't open source.
So we have a three month old baby, right?
When he gets a little bit older, we're not gonna open source to him,
all the most dangerous things he can do in the house.
But it does, it's a weird feeling because this is one of the first moments in history where
There's a strong case to be made not to open source software
This is when the software has become
Yeah, too dangerous. Yeah, but it's not the first time that we didn't want to open source a technology. Yeah
to open source technology. Yeah.
Is there something to be said about how to get the release of such systems
right like GPT for and GPT five?
So opening I went through a pre rigorous effort for several months.
You could say it could be longer, but nevertheless, it's longer than you
would have expected of trying to test the system to see like,
what are the ways it goes wrong to make it very difficult for people. Somewhat difficult for people to ask things,
how do I make a bomb for one dollar? Or how do I say I hate a certain group on Twitter in
a way that doesn't give me a block from Twitter, ban from Twitter? Those kinds of questions.
from Twitter, ban from Twitter, those kinds of questions. Yeah.
So you basically use the system to do harm.
Yeah.
Is there something you could say about ideas?
You have just, on looking, having thought about this problem
of AI safety, how to release such system, how to test
such systems when you have them inside the company.
Yeah, so a lot of people say that the two biggest risks from large
language models are it's spreading this information, harmful information types and second being used for offensive cyber weapon.
I think that was not the two greatest threats.
They're very serious threats and it's wonderful that people are trying to mitigate them.
So much bigger elephant in the room is how is this going to disrupt the economy in a huge way obviously
and maybe take away a lot of the most meaningful jobs. And an even bigger one is the one we spent so much time talking about here, that
this becomes the bootloader for the more powerful AI.
Right code connected to the internet manipulate humans.
Yeah. And before we know it, we have something else, which is not at all a large language
model. It looks nothing like it, but which is way more intelligent and capable and has goals.
And that's the elephant in the room.
And obviously, no matter how hard any of these companies have tried, that's not something
that's easy for them to verify with large language models.
And the only way to be really lower that risk a lot would be to
not let for example, to never let it read any code, not train on that and not put it into
an API and not not to give it access to so much information about how to manipulate humans.
So but that doesn't mean you still can't make a ton of money on them, you know.
We're just watching this coming year, right?
Microsoft is rolling out the new Office Suite where you go into Microsoft Word and give it a prompt.
It writes the whole text for you and then you edit it.
And then you're like, oh, give me a powerpoint version of this and it makes it and now I'll take the spreadsheet and blah blah and you know,
all of those things I think are you can debate the economic impact of it and whether society is
prepared to deal with this disruption, but those are not the things which that's not the elephant
of the room that keeps me awake at night for wiping out humanity. And I think that's the biggest misunderstanding we have.
A lot of people think they were scared of automatic spreadsheets.
That's not the case.
That's not what Eliasio was freaked out about either.
Is there, in terms of the actual mechanism of how AI might kill all humans So something you've been outspoken about you've talked about a lot is it autonomous weapon systems
So the use of AI in war
Mm-hmm is that one of the things that still you carry concern for as these systems become more and more powerful
I carry concern for it not that with all humans are to get killed by slaughterbots, but rather just this express root
into Orwellian dystopia, where it becomes much easier
for very few to kill very many.
And therefore, it becomes very easy for very few
to dominate very many.
If you want to know how I could kill all people,
just ask yourself, humans have driven a lot of species extinct.
How do we do it?
You know, we were smarter than them.
Usually we didn't do it even systematically by going around one on one after the other and stepping on them or shooting them or anything like that.
We just like chopped down their habitat because we needed it for something else. In some cases, we did it by putting more carbon dioxide in the atmosphere because of some
reason that those animals didn't even understand and now they're gone.
So, if you're in AI and you just want to figure something out, then you decide, you know, we just really need the space here to build more compute facilities.
You know, if that's the only goal it has, you know, we are just the sort of accidental road
kill along the way. And you could totally imagine, yeah, maybe this oxygen is kind of annoying
because it causes more corrosion. So let's get rid of the oxygen. And the good luck surviving after that. I'm
not particularly concerned that they would want to kill us just because that would be like
a goal in itself. You know, when we driven a number, we've driven a number of the elephant
species extinct, right? It wasn't because we didn't like elephants.
What the basic problem is, you just don't want to give, you don't want to see the control over your planet
to some other more intelligent entity that doesn't share your goals. It's that simple. So
which brings us to another key challenge, which AI Safety Research has been grappling with for a long time. How do you make AI first of all understand our goals and then adopt our goals and then
retain them as they get smarter, right? And all three of those are really hard, right?
A human child.
First, they're just not smart enough to understand our goals.
They can't even talk.
And then eventually, they're teenagers and understand our goals just fine, but they don't share.
But there's fortunately, a magic phase in the middle where there are smart enough to understand
our goals and malleable enough that we can hopefully with good parenting and teach them right from
wrong and instead of good goal, still good goals in them, right?
So those are all tough challenges with computers and then even if you teach your kids good
goals when they're little they might outgrow them too and that's a challenge for machines
and keep improving. So these are a lot of hard challenges we're up for, but I don't think any of them are insurmountable.
The fundamental reason why Eliezer looked so depressed when I last saw him was because he felt
it just wasn't enough time. Oh, it did not. It was unsolvable. It's just not enough time. He was
hoping that humanity was going to take this threat more seriously,
so we would have more time.
Yeah.
And now we don't have more time.
That's why the open letter is calling for more time.
But even with time, the AI alignment problem seems to be really difficult.
Oh, yeah. But it's also the most worthy problem. It seems to be really difficult. Oh, yeah. But it's also the most worthy problem,
the most important problem for humanity to ever solve. Because if we solve that one, Lex,
that aligned AI can help us solve all the other problems. Because it seems like it has to have
constant humility about its goal, constantly question the goal.
Because as you optimize towards a particular goal and you start to achieve it,
that's when you have the unintended consequences,
all the things you mentioned about.
So how do you enforce and code a constant humility
as your ability become better and better and better and better?
Stuart, Professor Stuart Russell, Berkeley,
who is also one of the driving forces behind this letter,
he has a whole research program about this.
I think of it as AI humility, exactly.
Although he calls it inverse reinforcement learning and other nerdy terms, but it's about
exactly that.
Instead of telling the AI, here's this goal.
Go optimize the bejesus out of it.
You tell it, okay, do what I want you to do,
but I'm not gonna tell you right now what it is,
I want you to do, you need to figure it out.
So then you give the incentives to be very humble
and keep asking you questions along the way,
is this what you really meant?
Is this what you wanted?
And oh, this other thing I tried didn't work.
It seemed like it didn't work our right.
Should I try it differently?
What's nice about this is it's not just philosophical
mumbo jumbo, it's theorems and technical work
that with more time, I think you can make a lot of progress.
And there are a lot of brilliant people now
working on AI safety.
And we just need to give them a bit more time.
But also not that many relatives to scale the problem.
No, exactly.
There should be, at least just like every university worth
its name has some cancer research going on
in this biology department, right?
Every university that does computer science
should have a real effort in this area and it's nowhere
near that. This is something I hope is changing now thanks to the GPT-4, so I think if there's
a silver lining to what's happening here, even though I think many people would wish it would have
been rolled out more carefully, it's that this might be the wake-up call
that humanity needed to really stop fantasizing about this being 100 years off and stop fantasizing about this being completely controllable and predictable because it's so obvious.
It's not predictable. Why is it that open that I think it was chat GPT,
tried to persuade a journalist,
or was it a GPT Ford to divorce his wife?
It was not because the engineers had built it.
I was like, let's put this in here and screw a little bit with
people. They hadn't predicted it at all. They built the giant black box, trained to
predict the next word, got all these emergent properties and oops, it did this. I think
this is a very powerful wake up call. Anyone watching this who's not scared,
I would encourage them to just play a bit more
where these tools they're out there now, like GPD4.
So wake up call is first up.
Once you're broken up, then go to slow down
a little bit the risky stuff to give a
chance to everyone's working up to catch up on the safety front. You know what's interesting is
MIT that's computer science but in general but let's just even say computer science curriculum.
How does the computer science curriculum change now? You mentioned programming.
How does the computer science curriculum change now? You mentioned programming.
Yeah.
Like, why would you be...
When I was coming up programming as a prestigious position,
like why would you be dedicating crazy amounts of time
to become an excellent programmer?
Like the nature of programming is fundamentally changing.
The nature of an entire education system is completely torn on its head.
Has anyone been able to like load that in and like think about it?
Because it's really turning...
I mean, it's a English professor. So, English teachers are beginning to really freak out now.
They give an essay assignment and they get back all this fantastic prose,
like this is a style of Hemingway and then they realize they have to completely
rethink. And even you know, just like we stopped at teaching writing script. Is that what
you're saying English? Yeah, yeah. Yeah. When everybody started typing, you know, like So much of what we teach our kids today.
Yeah, I mean, that's, uh,
everything is changing and it's changing very, it's changing very quickly.
And so much of us understanding how to deal with the big problems of the world is
through the education system. And if the education system is being turned on its head, then what's next? It feels like having these kinds of conversations is essential to try to figure
it out, and everything is happening so rapidly. I don't think there's even speaking of safety,
what the broad AI safety define, I don't think most universities have courses on AI safety.
No, philosophy is not enough.
Yeah, and I'm an educator myself, so it pains me to see this, say this, but I feel our
education right now is completely obsoleteed by what's happening.
You put a kid into first grade and then you're envisioning, and then they're going to come
out of high school 12 years later.
And you've already pre-planed now what they're going to learn
when you're not even sure if there's going to be any world
left to come out to.
Clearly, you need to have a much more opportunistic education
system that keeps adapting itself very rapidly
as society re-adapts.
The skills that were really useful when the curriculum was written,
I mean, how many of those skills are going to get you a job in 12 years?
I mean, seriously, if we just linger on the GPT 4 system a little bit, you kind of
hinted at it, especially talking about the importance of consciousness in the human
mind with homo-sentience.
Do you think GPT-4 is conscious?
I love this question.
So let's define consciousness for us, because in my experience,
like 90% of all arguments about consciousness were allowed
to the two people arguing having totally different definitions of what it is
and they're just shouting past each other. I define consciousness as subjective experience.
Right now, I'm experiencing colors and sounds and emotions, you know,
but does a self-driving car experience anything? That's the question about whether it's conscious or not,
right?
Other people think you should define conscious as differently,
fine by me, but then maybe use a different word for it, or they can, I'm going to use consciousness for this at least. So, um, but if people hate the way, yeah. So,
is GPT4 conscious?
Does GPT4 have subjective experience?
Short answer, I don't know, because we still don't know
what it is that gives this wonderful subjective experience
that is kind of the meaning of our life, right?
Because meaning itself, the feeling of meaning
is a subjective experience.
Joy is a subjective experience.
Love is a subjective experience.
We don't know what it is. I've written some papers about this. A lot of people have Julia Tounone, a professor
who has stuck his neck out the farthest and written down an actually very bold mathematical
conjecture for what's the essence of conscious information processing. He might be wrong, he might be right, but we should test it.
He postulates that consciousness has to do with loops in the information processing. So our brain has loops.
Information can go round and round in computer science nerd speak, you call it a recurrent neural network where some of the output gets fed back in again. With his mathematical formalism, if it's a feed forward
neural network, where information only goes in one direction, like from your eye, retina
into the back of your brain, for example, that's not conscious. So he would predict that
your retina itself isn't conscious of anything.
Or a video camera.
Now, the interesting thing about GPT-4 is it's also one way, flow of information.
So if Tionini is right, GPT-4 is a very intelligent zombie.
They can do all this smart stuff, but isn't experiencing anything.
And this is both a relief in that you don't have,
if it's true, you don't have to feel guilty
about turning off GPT-4 and wiping its memory
whenever a new user comes along.
I wouldn't like if someone used that to me,
it newerized me like in men and black.
But it's also creepy that you can have very high intelligence perhaps then it's
not conscious because if we get replaced by machines, it's sad enough that humanity isn't
here anymore because I kind of like humanity. But at least if the machines were conscious
they could be like, well, but there are descendants, and maybe they have our values, there are children.
But if Tionini is right, and these are all transformers that are not in the sense of the
Hollywood, but in the sense of these one way direction and they're all at works.
So they're all the zombies.
That's the ultimate zombie apocalypse now.
We have this universe that goes on
with great construction projects and stuff,
but there's no one experiencing anything.
That would be like the ultimate depressing future.
So I actually think as we move forward
to the building world last day,
I should do more research on figuring out
what kind of information processing actually has experience because I think that's what it's all about and I totally don't buy
The dismissal that some people some people will say well, this is all bullshit because consciousness equals intelligence
It's obviously not true
You can have a lot of conscious experience when you're not really
Accomplishing any goals at all.
You're just reflecting on something.
You can sometimes have things doing things that are quite intelligent, probably without being
conscious.
But I also worry that we humans won't discriminate against AI systems that clearly exhibit consciousness,
that we will not allow AI systems to have consciousness,
will come up with theories about measuring consciousness that will say this is a lesser being.
And this is like, I worry about that because maybe we humans will create something that is better
than us humans in the way that we find beautiful, which is they have a deeper
subjective experience of reality. Not only are they smarter, but they feel deeper, and we humans
will hate them for it. As human history is shown, there will be the other, we'll try to suppress it,
they'll create conflict, they'll create war war all of this. I worry about this too
Are you saying that we humans sometimes come up with self-serving arguments? No, we would never do that
Would he? Well, that's the danger here is
Even in this early stages, we might create something beautiful. Yeah, and
Will erase its memory. I
Was horrified as a kid when someone started boiling,
boiling lobster. Like, oh my God, that's so cruel.
And some grown up there, back in Sweden,
oh, it doesn't feel pain.
I'm like, how do you know that?
Oh, scientists have shown that.
And then there was a recent study
where they show that
lobsters actually do feel pain when you boil them. So they banned lobsters boiling in Switzerland now
to kill them in a different way first. So presumably that scientific research boil
them to someone asked the lobster to start a survey. And we do the same thing with cruelty
to farm animals also all these self-serving aren't arguments for why they're fine and yeah, so we should certainly
What do you watch for? I think step one is just be humble and acknowledge that consciousness is not the same thing as intelligence and
I believe that consciousness still is a form of information processing where it's really information being aware of itself in a certain way and let's study it and
Give ourselves a little bit of time and I think we will be able to figure out actually what it is that causes consciousness.
And then we can make probably unconscious robots that do the boring jobs that we would
feel immoral to get the machines, but if you have a companion robot taking care of your
mom or something like that, she would probably want it to be conscious, right? So the emotions that seems to display aren't fake.
All these things can be done in a good way if we give ourselves a little bit of time
and don't run and take on this challenge.
Is there something you could say to the timeline that you think about, about the development of AGI?
Depending on the day, I'm sure that changes for you, but when do you think about about the development of AGI
Depending on the day. I'm sure that changes for you, but
When do you think there'll be a really big leap in intelligence where you definitively say we have built AGI
Do you think it's one year from now five years from now 10 20 50? What's your gut say?
Honestly
For the past decade I've deliberately given very long timelines It was because I didn't want to fuel some kind of stupid malloc race. Yeah
But I think that cat has really left the bag now
I think you might be very very close
I think it might be very, very close. I don't think that Microsoft paper is totally off when they say that there are some glimmers of
AGI. It's not AGI yet. It's not an agent. There's a lot of things that can't do.
But I wouldn't bet very strongly against it happening very soon.
That's why we decided to do this open letter because if there's ever been a time to pause,
it's today.
There's a feeling like this GPT-4 is a big transition into waking everybody up to the
effectiveness of the system.
So the next version will be big.
Yeah, and if that next one isn't AGI,
maybe the next one will.
And there are many companies trying to do these things.
The basic architecture of them is not some sort of super well kept secret.
This is a time to...
A lot of people have said for many years that they will come at time when we want
to pause a little bit.
That time is now.
You have spoken about and thought about nuclear war, then at least in my lifetime.
What do you learn about human nature from that?
It's our old friend, Mollock, again.
It's really scary.
You see it where America doesn't want there to be a nuclear war.
Russia doesn't want there to be a global nuclear war either.
We both know that it would just be another,
if we just try to do it,
it both sides try to launch first.
It's just another suicide race, right?
So why is it the way you said that this is the closest
we've come since 1962?
In fact, I think we've come closer now
than even the Cuban Missile Crisis. It's
because of Mollock. You have these other forces. On one hand, you have the West saying that
we have to drive Russia out of Ukraine. It's a matter of pride. We've staked so much on it that
it would be seen as a huge loss of credibility of the
West if we don't drive Russia out entirely of the Ukraine.
And on the other hand, you have Russia who has, and you have the Russian leadership
who knows that if they get completely driven out of Ukraine, you know, it might, it's not
just going to be very humiliating for them, but they might, it often happens when countries
lose wars that things don't go so well for their leadership either. Like, you remember
when Argentina invaded the Faulkner Islands, the military junta that ordered that, right? People were cheering on the streets
at first when they took it. And then when they got their butt kicked by the British, you
know what happened to those guys? They were out. And I believe those were still alive
and jail now, right? So, you know So the Russian leadership is entirely cornered where they know that just getting driven out
of Ukraine is not an option.
So this to me is a typical example of a malloc.
You have these incentives of the two parties where both
of them are just driven to escalate more and more, right? If Russia starts losing in the
conventional warfare, the only thing they can do when they're back against the war is
to keep escalating. And the West has put itself in the situation now, we've already committed
to the dry rush out, so the only option the West has is the called rush is bluff and keeps sending in more weapons. This really bothers
me because Mollock can sometimes drive competing parties to do something which is ultimately
just really bad for both of them. And what makes me even more worried is not just that it's difficult to see a quick peaceful ending to this tragedy
that doesn't involve some horrible escalation, but also that we understand more clearly now,
just how horrible it would be. There was an amazing paper that was published in Nature Food this
There was an amazing paper that was published in Nature Food this August by some of the top researchers who've been studying nuclear winter for a long time.
What they basically did was they combined climate models with food and agricultural models.
So instead of just saying, yeah, it gets really cold, blah, blah, blah.
They figured out actually how many people would die in the different countries. And it's
pretty mind blowing. So basically what happens, you know, is the thing that kills the most
people is not the explosions, it's not the radioactivity, it's not the ENP, mayhem, it's not
the rampaging moms, forging food. No, it's the fact that you get so much smoke coming up from the burning cities into the stratosphere that spreads around the earth from the jet streams.
So in typical models, the temperature drops in Nebraska
and in the Ukraine, bread baskets, you know, by like,
20 Celsius or so if I remember.
No, yeah, 20, 30 Celsius depending on where you are.
40 Celsius in some places, which is, you know,
40 Fahrenheit to 80 Fahrenheit colder than what it would normally be. So, you know, I'm not
good at farming, but it's knowing if it drops a little freezing pretty much, oh, most
days in July and that's not good. So they worked out, they put this into their farming
models and what they found was really interesting. The countries that get the most hard hit are the ones in the Northern hemisphere.
So in the US and one model, they had about 99% of all Americans starving to death.
In Russia and China and Europe, also about 99%, 98% starving to death.
So you might be like, oh, it's kind of poetic justice that both the Russians and the Americans
99% of them have to pay for it because there are bombs that did it.
But, you know, that doesn't particularly cheer people up in Sweden or other random countries
that have nothing to do with it, right?
And, I think it hasn't entered the mainstream understanding very much just like how bad this is.
Most people, especially a lot of people in decision making positions, still think of
nuclear weapons as something that makes you powerful.
Scary, powerful, they don't think of it as something where, yeah, just,
two within a percent or two,
we're all just just gonna starve to death.
And,
and starving to death is,
the worst way to die as Haltamore,
is all the famines in history show,
the torture and involved in that.
Probably brings out the worst in people also, when people are desperate like this.
It's not, so some people have heard some people say that if that's what's going to happen,
they'd rather be it around zero and just get vaporized, you know, but uh, so I think people underestimate the risk of this because they, they aren't afraid
of malloc. They think, oh, it's just gonna be, because humans don't want this. So it's not
gonna happen. That's the whole point of malloc that things happen that nobody wanted.
And that applies to new clear weapons and that applies to AGI. Exactly. And it applies to some of the things that people
have gotten most upset with capitalism for also, right? Where everybody just kind of trapped,
it's not to see if some company does something that causes a lot of harm. Not that the CEO is a bad
person, but she or he knew that
The other all the other companies were doing this too. So Molok is
As a formidable foe. I hope we're someone
Make them make make good movies so we can see who the real enemy is so we don't
We're not fighting against each other
Molok makes us fight against each other.
That's what Mollock super power is.
The hope here is, is any kind of technology or the mechanism that lets us
instead realize that we're fighting the wrong enemy?
It's such a fascinating battle.
It's not our system.
It's us versus it.
Yeah. Yeah.
We are fighting Mollock for human survival.
We are the civilization.
Have you seen the movie Needful Things?
It's a Stephen King novel.
I love Stephen King and Max von Südov, Swedish actor, playing the guys.
It's brilliant.
I just thought I hadn't thought about that until now,
but that's the closest I've seen to a movie about Mollock. I don't want to spoil the film for anyone
who wants to watch it, but basically it's about this guy who turns out to, you can interpret him as
the devil or whatever, but he doesn't actually ever go around and kill people or torture people
or go burning coal or anything.
He makes everybody fight each other, makes everybody hate,
fear each other, hate each other and then kill each other.
So that's the movie about Mollock, you know?
Love is the answer. That seems to be one of the ways to fight Mollock is by compassion,
by seeing the common humanity.
Yes, yes. And to not sound, so we don't sound like like once a combayatri huggers here,
right? We're not just saying love and peace, man. We're trying to actually help people
understand the true facts about the other side and feel the compassion because the truth makes
you more compassionate, right? So that's why I really like using AI for truth and for truth seeking technologies can that can as a result, you know, get us
more love than hate. And even if you can't get love, you know, settle for some understanding
which already gives compassion. If someone is like, you know, I really disagree with you, Lex, but I can see why you're, where you're
coming from. You're not a bad person who needs to be destroyed, but I disagree with you.
And I'm happy to have an argument about it. No, that's a lot of progress compared to
where we are 2023 in the public space. Wouldn't you say?
If we solve the AI safety problem as we've talked about and then you
Max tag work who has been talking about this
For many years get to sit down with the AGI with the early AGI system on a beach with a drink
What what kind of what would you ask her what kind of question would you ask what would you talk about?
Something so much smarter than you.
Would you be afraid?
I never gonna get me with it. Really. Zinger of a question.
Would you be afraid to ask some questions?
So I'm not afraid of the truth.
I'm very humble. I know I'm just a meat bag with all these flaws.
You know, but I have, I'm just a meat bi with all these flaws, you know, but yeah, I
Have the time we talked a lot about homo sentience. I've really already tried that for a long time with myself
Just so that is what's really valuable about being alive for me is that I have these meaningful experiences
It's not um
Have what I'm good at this or good at that or whatever because there's so much I suck at and
So you're not afraid for the system to show you just how dumb you are no no in fact my son reminds me
You could find out how dumb you are in terms of physics a little how little we humans understand cool that I
think I think
So I can't waffle my way out of this question.
It's a fair one. I'm just tough.
I think given that I'm a really, really curious person,
that's really the defining part of who I am.
I'm so curious.
I have some physics questions.
I loved, I loved to understand.
I have some questions about consciousness, about the nature of reality,
I would just really, really love to understand also.
I could tell you one, for example, that I've been obsessing about a lot recently.
So I believe that
so suppose Teno know me is right,
suppose there are some information processing systems that are conscious and some that are not.
Suppose you can even make reasonably smart things like GPT-4 that are not conscious,
but you can also make them conscious.
Here is the question that keeps me naked, mate.
Is it the case that the unconscious zombie systems that are really intelligent are also
really efficient?
So they're really inefficient so that when you try to make things more efficient, it
will naturally be a pressure to do.
They become conscious.
I'm kind of hoping that that's correct.
And I do want me to give you a hand-away the argument for it.
In my lab, every time we look at how these large language models do something, we see that
they do it in a really dumb way, and you could make it better.
We have loops in our computer language for a reason.
The code would get way, way longer if you weren't allowed to use them.
It's more efficient to have the loops.
And in order to have self-reflection,
whether it's conscious or not,
even an operating system knows things about itself, right?
You need to have loops already.
So I think, I'm waving my hands a lot,
but I suspect that the most efficient way
of implementing a given level of intelligence
has loops in it, the self-reflection,
and will be conscious.
Isn't that great news?
Yes, if it's true, it's wonderful,
because then we don't have to fear the ultimate zombie apocalypse. And I think if you look at our brains, actually,
our brains are part zombie and part conscious. When I open my eyes, I mean, you'll
take all these pixels and hit my retina.
I'm like, oh, that's Lex.
But I have no freaking clue of how I did that computation.
It's actually quite complicated, right?
It was only relatively recent that we could even do it well with machines, right?
You get a bunch of information processing happening in my retina,
and then it goes to the lateral jenicular nucleus, my phalamus,
and the area V1, V2, V4, and the fusiform face area here that Nancy can wish her,
and MIT invented, and blah, blah, blah, blah, blah, and I have no freaking clue how that worked, right?
It feels to me, subjectively, like my conscious module, just got a little email, say,
say face facial processing, task complete, it's Lex. Yeah. I'm going to go with that, right? So this fits perfectly with Tenoenie's model because this was all one way information processing,
mainly. And it turned out for that particular task, that's all you needed. And it probably
was kind of the most efficient way to do it. But there are a lot of other things that
we associate with higher intelligence and planning and so on and so forth where you kind
of want to have loops and be able to ruminate and self-reflect and introspect and so on, where
my hunch is that if you want to fake that with a zombie
system that just all goes one way, you have to like, unroll those loops and it gets really,
really long and it's much more inefficient. So I'm actually hopeful that AI, if in the future,
we have all these very sublime and interesting machines that do cool things and are aligned with
that, they will be at at least they will also have consciousness
for the kind of these things that we do. That great intelligence is also correlated to great
consciousness or a deep kind of consciousness. Yes. So that's a happy thought for me because there's
a zombie of a couple that the apocalypse really is my worst nightmare of all of it. It would be like adding insult to injury not only get replaced, but we
freaking replace ourselves by zombies.
Like how dumb can we be?
That's such a beautiful vision and that's actually a provable one.
That's one that we humans can, in two, it improve that those two things are correlated.
As we start to understand what it means to be intelligent and what it means to be conscious which these
systems early age I like systems will help us understand and I just want to say one more thing is super important
Most of my colleagues when I started going on about consciousness tell me that it's all bullshit and I just stop talking about it
I hear a little inner voice from my father and my mom saying, keep talking about it because I think they're wrong and and
the main
Way to convince people like that
That they're wrong if they say that consciousness is just equal to intelligence is to ask them. What's wrong with torture?
Why are you against torture?
if it's just about you know
These these particles moving this way around on that way, and there is no such thing as subjective experience, what's wrong with
torture? I mean, do you have a good comeback to that? No, it seems like suffering, suffering
imposed on other humans is somehow deeply wrong in a way that intelligence doesn't quite explain. If someone tells me, well, you know, it's just an illusion,
consciousness, whatever, you know, I like to invite them to next time they're having
surgery to do it without anesthesia.
What is anesthesia really doing?
If you have it, you can have it local anesthesia when you're awake.
I had that when they fixed my shoulder.
I was super entertaining.
What was that it did?
It just removed my subjective experience of pain.
It didn't change anything about what's actually happening
in my shoulder, right?
So if someone says that's all bullshit,
skip the anesthesia, that's my advice.
This is incredibly central.
It could be fundamental to whatever this thing we have going on here.
It is fundamental because we are, we, where we feel is so fundamental is suffering and joy
and pleasure and meaning and that's all, those are all subjective experiences there.
And let's not that those are the elephant in the room.
That's what makes life worth living.
And that's what can make it horrible if it's just a software.
So let's not make a mistake of saying that that's all bullshit.
And let's not make the mistake of not instilling the AI systems
with that same thing that makes us special.
Yeah.
Max, it's a huge honor that he was said down to me the first time.
On the first episode of this podcast, it's a huge honor to sit down again and talk
about this, what I think is the most important topic, the most important problem that we humans
have to face and hopefully solve.
Yeah, well the honor is all mine and I'm so grateful to you for making more people aware of this fact that humanity has reached
the most important fork in the road ever in its history and that's
could turn in the correct direction. Thanks for listening to this conversation with Max Tagmark.
To support this podcast, please check out our sponsors in the description.
And now, let me leave you some words from Frank Herbert.
History is a constant race between invention and catastrophe.
Thank you for listening and hope to see you next time.