Lex Fridman Podcast - #267 – Mark Zuckerberg: Meta, Facebook, Instagram, and the Metaverse

Episode Date: February 26, 2022

Mark Zuckerberg is CEO of Meta, formerly Facebook. Please support this podcast by checking out our sponsors: - Paperspace: https://gradient.run/lex to get $15 credit - Coinbase: https://coinbase.com/...lex to get $5 in free Bitcoin - InsideTracker: https://insidetracker.com/lex and use code Lex25 to get 25% off - ExpressVPN: https://expressvpn.com/lexpod and use code LexPod to get 3 months free - Blinkist: https://blinkist.com/lex and use code LEX to get 25% off premium EPISODE LINKS: Mark's Facebook: https://facebook.com/zuck Mark's Instagram: https://instagram.com/zuck Meta AI: https://ai.facebook.com/ PODCAST INFO: Podcast website: https://lexfridman.com/podcast Apple Podcasts: https://apple.co/2lwqZIr Spotify: https://spoti.fi/2nEwCF8 RSS: https://lexfridman.com/feed/podcast/ YouTube Full Episodes: https://youtube.com/lexfridman YouTube Clips: https://youtube.com/lexclips SUPPORT & CONNECT: - Check out the sponsors above, it's the best way to support this podcast - Support on Patreon: https://www.patreon.com/lexfridman - Twitter: https://twitter.com/lexfridman - Instagram: https://www.instagram.com/lexfridman - LinkedIn: https://www.linkedin.com/in/lexfridman - Facebook: https://www.facebook.com/lexfridman - Medium: https://medium.com/@lexfridman OUTLINE: Here's the timestamps for the episode. On some podcast players you should be able to click the timestamp to jump to that time. (00:00) - Introduction (11:29) - Metaverse (31:06) - Identity in Metaverse (43:15) - Security (47:40) - Social Dilemma (1:09:46) - Instagram whistleblower (1:14:31) - Social media and mental health (1:19:56) - Censorship (1:37:05) - Translation (1:44:40) - Advice for young people (1:50:28) - Daughters (1:53:16) - Mortality (1:57:49) - Question for God (2:00:55) - Meaning of life

Transcript
Discussion (0)
Starting point is 00:00:00 The following is a conversation with Magzaka Berg, CEO of Facebook, now called Mehta. Please allow me to say a few words about this conversation with Magzaka Berg, about social media, and about what troubles me in the world today, and what gives me hope. If this is not interesting to you, I understand, please skip. I believe that at its best, social media puts a mirror to humanity and reveals the full complexity of our world, shining a light on the dark aspects of human nature and giving us hope, a way out, through compassionate but tense chaos of conversation that eventually can turn into understanding, friendship, and even love.
Starting point is 00:00:45 But this is not simple. Our world is not simple. It is full of human suffering. I think about the hundreds of millions of people who are starving and who live in extreme poverty. The one million people who take their own life every year, the twenty million people that attempt it, and the many, many more millions who suffer quietly in ways that numbers can never know. I'm troubled by
Starting point is 00:01:11 the cruelty and pain of war. Today my heart goes out to the people of Ukraine. My grandfather spilled his blood on this land,lled the line as a machine gunner against the Nazi invasion, surviving impossible odds. I am nothing without him. His blood runs in my blood. My words are useless here. I send my love, it's all I have. I hope to travel to Russia in Ukraine soon. I will speak to citizens and leaders, including Vladimir Putin.
Starting point is 00:01:49 As I've said in the past, I don't care about access, fame, money, or power, and I'm afraid of nothing. But I am who I am, and my goal in conversation is to understand the human being before me, no matter who they are, no matter their position. And I do believe the line between good and evil runs through the heart of every man. So this is it. This is our world.
Starting point is 00:02:17 It is full of hate, violence, and destruction. But it is also full of love, beauty, and the insatiable desire to help each other. The people who run the social networks that show this world, that show us to ourselves, have the greatest of responsibilities. In a time of war, pandemic, atrocity, we turn to social networks to share real human insights and experiences to organize protests and celebrations, to learn and to challenge our understanding of the world of our history and of our future, and above all to be reminded of our common humanity. When the social networks fail, they have the power to cause immense suffering, and when they succeed, they have the power to lessen that suffering.
Starting point is 00:03:07 This is hard. It's a responsibility, perhaps almost unlike any other in history. This podcast conversation attempts to understand the man and the company who take this responsibility on, where they fail, and where they hope to succeed. Mark Zuckerberg's feet are often held to the fire, as they should be, and this actually gives me hope. The power of innovation and engineering coupled with the freedom of speech in the form of its highest ideal, I believe, can solve any problem in the world.
Starting point is 00:03:40 But that's just it. Both are necessary. The engineer and the critic. I believe that criticism is essential, but cynicism is not. And I worry that in our public discourse cynicism too easily masquerades as wisdom, as truth, becomes viral and takes over and worse suffocates the dreams of young minds who want to build solutions to the problems of the world. We need to inspire those young minds, at least for me, they give me hope. And one small way I'm trying to contribute is to have honest conversations like these that don't just ride the viral wave of cynicism, but seek to understand the failures and successes of
Starting point is 00:04:25 the past, the problems before us, and the possible solutions in this very complicated world of ours. I'm sure I will fail often, and I count on the critic to point it out when I do. But I ask for one thing, and that is to fuel the fire of optimism, especially in those who dream to build solutions. Because without that, we don't have a chance. On this too fragile tiny planet of ours. And now, a quick few second mention of each sponsor. Check them out in the description. It's the best way to support this podcast. First is paper space. A platform I used to train and deploy machine learning models. Second is Coinbase. A platform I used to buy cryptocurrency. Third is Inside Tracker.
Starting point is 00:05:13 A service I used to track my biological data. Fourth is ExpressVPN. The VPN I've been using for many years, and fifth is Blinkist. The app I use to read summaries of books. So the choice is machine learning, and fifth is Blinkist, the app I use to read summaries of books. So the choice is machine learning, cryptocurrency, health, privacy, or knowledge. Choose wise name my friends.
Starting point is 00:05:32 And now onto the full ad reads, as always, no ads in the middle. I try to make this interesting, but if you skipped them, please still check out the sponsors. I enjoy their stuff, maybe you will too. This show is brought to you by PaperSpace Gradient, which is a platform that lets you build, train, and deploy machine learning models of any size and complexity. I love how powerful and intuitive it is. I should mention that Fest.AI, of course, I highly recommend a machine learning uses it. It's run by Jeremy Howard, who is pretty much as legit of an educator and technologist programmer developer just intellect in the space of machine learning as it gets.
Starting point is 00:06:16 You can host notebooks on there, you can swap out the compute instance at your time, start in a small scale GPU instance or even CPU and swap out once your compute needs increase. I'm excited by what they're calling workflows, which provides a way to automate ML pipelines on top of gradient compute infrastructure. It makes it really easy with simple configuration files, Yamol files, to give gradient a try, visit gradient.runslashlex and use the sign up link there. You'll get 15 bucks in free credit, which you can use to power your next machine learning application. That's gradient.runslashlex.
Starting point is 00:06:55 This show is also brought to you by Coinbase, which is a trusted and easy to use platform to buy, sell, and spend cryptocurrency. I use it and love it. You can buy Bitcoin, Ethereum, Cardano, Dogecoin. I can keep going. All of the most popular digital currencies. In fact, I think all of the currencies for the people who have been interviewed on this podcast
Starting point is 00:07:19 and even the ones that are coming up on this podcast. So, it's a great way to track cryptocurrency and to learn about cryptocurrency. It's a great way to track the prices of things. It's just the interface is so intuitive. I love it. They also have the Coinbase wallet, which I just recently set up. It's as good as it gets in terms of simplicity, accessibility. If you're new to cryptocurrency, especially Coinbase is where you should go. Go to Coinbase.com slash Lex. For a limited time, new users get $5 on a free Bitcoin
Starting point is 00:07:51 when you sign up today at Coinbase.com slash Lex. That's Coinbase.com slash Lex. This show is also brought to you by Inside Tracker, a service that you use to track biological data. They have a bunch of plans, most of which include a blood test that gives you a lot of information that you can then make decisions based on. They have machine learning algorithms that analyze the data, the blood data, DNA data, fitness tracker data to provide you with a clear picture what's going on inside you
Starting point is 00:08:20 and to give you science-backed recommendations for positive diet and lifestyle changes. So using data from your body, using machine learning algorithms to tell you what you should change, where you should improve all those kinds of things. This idea is what I love. It feels like the future because you should be making decisions in your life based on data that comes from your body, not some generic population data. For a limited time, you can get 25% off the entire inside tracker store if you go to inside tracker.com slash Lex.
Starting point is 00:08:56 That's inside tracker.com slash Lex. This show is also brought to you by ExpressVPN. I use them to protect my privacy on the internet. There's so much I can say about ExpressVPN. I've been using them for many many many years. Obviously, as you probably know, ISPs want to track your data even when you're using incognito mode on Chrome, all the shady sites you visit, your ISPs know about them. You can also, if you're watching Netflix, change your geographical location, which unlocks a bunch of shows that are only available in certain localities. Finally, my favorite reason to use ExpressVPN is it's just damn fast.
Starting point is 00:09:37 And intuitive, clean design of the app. It does the thing it's supposed to do, it does it well, it doesn't do anything extra. Big button, turning on, pick location, it works. Any device, any operating system, including my favorite OS, Linux, anyway, go to expressvpn.com slash Lex pod to get an extra three months free, that's expressvupn.com slash Lex pod. This show is also brought to you by Blinkist. My favorite app for learning new things. Blinkist takes the key ideas from thousands of nonfiction books and can essence them down
Starting point is 00:10:15 into just 15 minutes that you can read or listen to. There are so many nonfiction books that can recommend on there. I just actually re-listened to meditations by Marcus Aurelius. Beginning of Infinity by David Dorch is on there, Snowden book is on there, there's just countless ones. I use it all kinds of ways, help me select which books I'm going to read in full. Next, review books I've already read and review books I'm never going to get a chance to read because life is short and finite and that's actually what makes life beautiful.
Starting point is 00:10:49 Go to blinkis.com slash likes to start your free seven day trial and get 25% off of a blinkist premium membership as blinkis.com slash Lex spelled B L I N K I S T blinkest.com slash Lex. This is the Lex Friedbrand Podcast and here is my conversation with Mark Zuckerberg. Is it possible that this conversation is happening inside the metaverse created by you, by meta many years from now, and we're doing a memory replay experience? I don't know the answer to that. Then I'd be some computer construct and not the person who created that meta company. But that would truly be meta. Right, so this could be somebody else using the Mark Zuckerberg avatar. You can do the mark and the Lex conversation replay from four decades ago when,
Starting point is 00:12:01 when meta for it was first. I mean, it's not going to be four decades before we have photorealistic avatars like this. So I think we're much closer to that. Well, that's something you talk about is how passionate you are about the idea of the avatar representing who you are in the metaverse. So I do these podcasts in person. You know, I'm a stickler for that
Starting point is 00:12:22 because there's a magic to the in-person conversation. How long do you think it'll be before you can have the same kind of magic and the metaverse, the same kind of intimacy and the chemistry, whatever the heck it's there when we're talking a person, how difficult is it, how long before we have it in the metaverse? Well, I think that's, this is like the key question, right? Because the thing that's different about virtual and hopefully augmented reality compared to all other forms of digital platforms before is this feeling of presence,
Starting point is 00:12:54 right? The feeling that you're right, they are in an experience and that you're there with other people or in another place. And that's just different from all the other screens that we have today, right? Fones, TVs, all the stuff. It we have today, right phones TV is all the stuff It's you know, they're trying to in some cases deliver experiences that feel
Starting point is 00:13:11 High fidelity, but at no point do you actually feel like you're in it right at some level your content is trying to sort of Convince you that this is a realistic thing that's happening, but all of the kind of subtle signals are telling you now You're looking at a screen. So the question about how you develop these systems is like, what are all of the things that make the physical world all the different cues? So I think on visual presence and spatial audio, we're making reasonable progress.
Starting point is 00:13:45 Spatial audio makes a huge deal. I don't know if you've tried this experience, work rooms that we launch where you have meetings. And I basically made a rule for all of the top management folks at the company that they need to be doing standing meetings in work rooms already, right? I feel like we got a dog through this.
Starting point is 00:14:03 This is how people are gonna to work in the future. So we have to adopt this now. And there are already a lot of things that I think feel significantly better than typical Zoom meetings. Even though the avatars are a lot lower fidelity, the idea that you have spatial audio, you're around a table in VR with people. If someone's talking from over there,
Starting point is 00:14:25 it sounds like it's talking from over there. You can see the arm gestures and stuff feel more natural. You can have side conversations, which is something that you can't really do in Zoom. I mean, I guess you can text someone over, like out of band, but if you're actually sitting around a table with people, you can lean over and whisper to the person next to you and like have a conversation that you can't, you know, that you can't
Starting point is 00:14:48 really do with in just video communication. So I think it's interesting in what ways some of these things already feel more real than a lot of the technology that we have, even when the visual fidelity isn't quite there, but I think it'll get there over the next few years. Now, I mean, you were asking about comparing that to the true physical world, not the Zoom or something like that. And there, I mean, I think you have feelings of like
Starting point is 00:15:17 temperature, you know, all factory, obviously touch, right, we're working on haptic gloves, you know, the sense that you want to put your hands down and feel some pressure from the table. All these things I think are going to be really critical to be able to keep up this illusion that you're in a world and that you're fully present in this world. But I think we're going to have a lot of these building blocks within the next 10 years or so.
Starting point is 00:15:44 Even before that, I think it's amazing how much you're just going to be able to build with software that sort of masks some of these things. I'm going along, but I was told we have a few hours here. We're here for five to six hours to bring this. Yeah, so look, that's on the shorter end of the congressional testimonies I've done. But it's, but, you know, one of the things that we found with, with hand presence, right? So the earliest VR, you just have the headset and then, and that was cool. You could look around, you feel like you're in a place, but you don't feel like you're really able to interact with it until you have hands.
Starting point is 00:16:18 And then there was this big question where once you got hands, what's the right way to represent them? And initially, all of our assumptions was, okay, when I look down and see my hands in the physical world, I see an arm and it's gonna be super weird if you see just your hand. But it turned out to not be the case because there's this issue with your arms,
Starting point is 00:16:39 which is like what's your elbow angle? And if the elbow angle that we're kind of interpolating based on where your hand is and where your headset is, actually as an accurate, it creates this very uncomfortable feeling where it's like, oh, like my arm is actually out like this, but it's like showing it in here and that actually broke the feeling of presence
Starting point is 00:16:58 a lot more, whereas it turns out that if you just show the hands and you don't show the arms, it actually is fine for people. So I think that there's a bunch of these interesting psychological cues where it'll be more about getting the right details right. And I think a lot of that will be possible even over,
Starting point is 00:17:17 you know, a few year period or a five year period and we won't need like every single thing to be solved to deliver this like full sense of presence. Yeah, that's a fascinating psychology question of what is the essence that makes in-person conversation special? It's like emojis are able to convey emotion really well even though they're obviously not photorealistic. And so in that same way, just like you're saying, just showing the hands is able to create a comfortable expression with your hands.
Starting point is 00:17:48 So I wonder what that is. People in the world, wars, these the right letters, and you can fall in love with just writing letters. You don't need to see each other in person. You can convey emotion. You can be depth of experience with just words. So that's a, I think, a fascinating place to explore psychology of like, how do you find that intimacy?
Starting point is 00:18:09 Yeah, and, you know, the way that I come to all of this stuff is, you know, I basically studied psychology and computer science. So all of the work that I do is sort of at the intersection of those things. I think most of the other big tech companies are building technology for you to interact with. What I care about is building technology to help people interact with each other.
Starting point is 00:18:28 So I think it's a somewhat different approach than most of the other tech entrepreneurs and big companies come at this from. And a lot of the lessons in terms of how I think about designing products come from some just basic elements of psychology, right? In terms of, you know, our brains, you know, you can compare to the brains of other animals, you know, we're very wired to specific things, facial expressions, right? I mean, we're very visual, right? So compared to other animals, I mean, that's clearly the main sense that most people have.
Starting point is 00:19:03 But there's whole part of your brain that's just kind of focused on reading facial cues. So when we're designing the next version of Quest, or the VR headset, a big focus for us is face tracking. And basically eye tracking so you can make eye contact, which again isn't really something that you can do over a video conference. It's sort of amazing how much, how far video confer thing has gotten without the ability to make eye contact, right? It's sort of a bizarre thing if you think about it. You're like looking at someone's face, you know, sometimes for, you know, an hour when you're in a meeting and like, you're looking at
Starting point is 00:19:37 their eyes to them, doesn't look like you're looking at their eyes. So it's a, you're always looking at me pass each other, I guess. Yeah, I guess you're right. You're not sending that signal. Well, you're signing too. You're trying to. Like a lot of times, I mean, or at least I find myself, I'm trying to look into the other person's eyes. But they don't feel like you're looking into their eyes. Yeah.
Starting point is 00:19:53 So then the question is, all right, am I supposed to look at the camera so that way you can, you know, have a sensation that I'm looking at you? I think that that's an interesting question. And then, you know, with VR today, even without eye tracking and knowing what your eyes are actually looking at, you can fake it reasonably well, right? So you can look at like where the head poses and if it looks like I'm kind of looking
Starting point is 00:20:13 in your general direction, then you can sort of assume that maybe there's some eye contact intended and you can do it in a way where it's okay, maybe not it's like a, maybe it's not a fixated stare, but it's somewhat natural, but it's not a fixated stare, but it's somewhat natural. But once you have actual eye tracking, you can do it for real. And I think that's really important stuff.
Starting point is 00:20:32 So when I think about Meta's contribution to this field, I have to say it's not clear to me that any of the other companies that are focused on the metaverse or on virtual and augmented reality are going to prioritize putting these features in the hardware because like everything they're trade-offs, right? I mean, it adds some weight to the device, maybe it adds some thickness. You could totally see another company taking the approach to just make the lightest and thinnest thing possible.
Starting point is 00:20:57 But I want us to design the most human thing possible that creates the richest sense of presence and because so much of human emotion and expression comes from these micro-movements. If I move my eyebrow millimeter, you will notice, and that means something. So the fact that we're losing these signals in a lot of communication, I think, is a loss. So it's not like, okay, there's one feature, and you add this, then it all of a sudden is going to feel like we have real presence. You can sort of look at how the human brain works
Starting point is 00:21:27 and how we express and kind of read emotions and you can just build a roadmap of that. You know, of just what are the most important things to try to unlock over a five to 10 year period and just try to make the experience more and more human and social? When do you think would be a moment, like a singularity moment for the metaverse where there's a lot of ways to ask this question, but you know, people will have many or most
Starting point is 00:21:56 of their meaningful experiences in the metaverse versus the real world. And actually it's interesting to think about the fact that a lot of people are having the most important moments of their life happen in the digital sphere, especially in Audra and COVID, you know, like even falling in love or meeting friends or getting excited about stuff that is happening on the 2D digital plane. When do you think the metaversal provide those experiences for a large number like a? Yeah, I think it number like a good question. There was someone, I read this piece that frame this as a lot of people think that the metaverse is about a place, but one definition of this is it's about a time when basically immersive
Starting point is 00:22:41 digital worlds become the primary way that we That we live our lives and spend our time I think that's a reasonable construct and from that perspective, you know, I think You also just want to look at this as a continuation because it's not like Okay, we are building digital worlds, but we don't have that today I think you know, you know, you and I probably already live a very large part of our life in digital worlds They're just not 3D immersive virtual reality, but you know, I do a lot of meetings over video or I spend a lot of time writing things over email or WhatsApp or whatever. So what is it going to take to get there for kind of the immersive presence version of this, which I think is what you're
Starting point is 00:23:19 asking. And for that, I think that there's just a bunch of different use cases, right? And And for that, I think that there's just a bunch of different use cases. And I think when you're building technology, I think a lot of it is just you're managing this duality where on the one hand you want to build these elegant things that can scale and have billions of people use them and get value from them. And then on the other hand, you're fighting this kind of ground game where it's just, there are just a lot of different use cases and people do different things and like you want to be able to unlock them. So the first ones that we basically went after were gaming
Starting point is 00:23:57 with Quest and social experiences. And this is, you know, it goes back to when we started working on virtual reality. My theory at the time was basically, people thought about it as gaming. But if you look at all computing platforms up to that point, you know, gaming is a huge part. It was a huge part of PCs. It was a huge part of mobile.
Starting point is 00:24:19 But it was also very decentralized, right? There wasn't, you know, for the most part, you know, one or two gaming companies. There were a lot of gaming companies and gaming is somewhat hit-spaced. I mean, we're getting some games that have more longevity, but it put in general, you know, there were a lot of different games out there. But on PC and on mobile,
Starting point is 00:24:40 the companies that focused on communication and social interaction, there tended to be a smaller number of those, and that ended up being just as important of a thing as all of the games that you did combined. I think productivity is another area. That's obviously something that we've historically been less focused on, but I think it's going to be really important for us. Was workroom, or give me productivity in the collaborative aspect of it? Yeah, I think that there's a workroom aspect of this, like a meeting aspect, and then I think that
Starting point is 00:25:05 there's like a word, Excel, productivity. You're working or coding or what knowledge work, right? It's as opposed to just meetings. So you can kind of go through all these different use cases. You know, gaming, I think, we're well on our way. Social, I think, we're just the kind of preeminent company that focuses on this. And I think that that's already on quest becoming,
Starting point is 00:25:29 that if you look at the list of what are the top apps, social apps are already number one, two, three. So that's kind of becoming a critical thing. But I don't know, I would imagine for someone like you, it'll be until we get a lot of the work things dialed in. When this is just much more adopted and clearly better than Zoom for VC, when if you're doing your coding or your writing or whatever it is in VR, which it's not that far off to
Starting point is 00:26:00 imagine that, because pretty soon you're just going to be able to have a screen that's bigger than it'll be your ideal setup and you can bring it with you and you put it on anywhere and have your, your kind of ideal workstation. So I think that there are a few things to work out on that, but I don't think that that's more than, you know, five years off. And then you'll get a bunch of other things that like aren't even possible or you don't even think about using a phone or PC for today like fitness, right? So I mean, I know that you're, we were talking before about how you're into running. And like I'm really into, you know,
Starting point is 00:26:30 a lot of things around fitness as well, different things in different places. I got really into hydrofoiling recently. And that's a video. Yeah, and surfing and I used to fence competitively. I like run. So. And you were saying that you were thinking about
Starting point is 00:26:45 trying different martial arts, and I tried to trick you and convince you into doing Brazilian Jiu-Jitsu. Or you actually mentioned that that was one you're curious about, and I did. I did. Was that a trick? Yeah, I don't know.
Starting point is 00:26:56 We're in the metaverse now. Yeah, not, I mean, I took that seriously. I thought that that was a real suggestion. That would be an amazing chance if we ever step on the mat together and just roll around. I'll show you some moves. Well give me a year to train. And then we can do it. You know, you've seen Rocky IV with a Russian, faces off the American.
Starting point is 00:27:16 I'm the Russian in this picture. And then you're the Rocky, the underdog that gets to win in the year. The idea of me as Rocky and like the fighting is, if he dies, he dies. Sorry, just had to, I mean. Anyway, yeah. But I mean, a lot of aspects of fitness, you know, I don't know if you've tried supernatural
Starting point is 00:27:38 on Quest or, so first of all, can I just comment on the fact every time I played around Quest 2, I just, I get giddy every time I step into virtual reality. So you mentioned productivity on those kinds of things. That's definitely something I'm excited about, but really I just love the possibilities of stepping into that world. It's maybe it's the introverted me, but it just feels like the most convenient way to travel into worlds, into worlds that are similar to the real world or totally different.
Starting point is 00:28:11 It's like Alice in Wonderland, just try out crazy stuff, the possibilities are endless. And I just, I personally, and just love get excited for stepping in those virtual worlds. So I'm a huge fan. In terms of the productivity as a program, I spend most of my day programming, that's really interesting also,
Starting point is 00:28:32 but then you have to develop the right IDs, you have to develop, like there has to be a threshold where a large amount of the program community moves there. But the collaborative aspects that are possible in terms of meetings, in terms of when two coders are working together, I mean, that, the possibility is they're super, super exciting.
Starting point is 00:28:51 I think that in building this, we sort of need to balance. There are going to be some new things that you just couldn't do before, and those are going to be the amazing experiences. So teleporting to any place, right, whether it's a real place or something that people made. I mean, some of the experiences around how we can build stuff in new ways, where a lot of the stuff that when I'm coding stuff, it's a card, you code it and then you build it and then you see it afterwards. But increasingly, it's going to be possible to, you know, you're in a world and you're building
Starting point is 00:29:22 the world as you are in it and kind of manipulating it. One of the things that we showed inside the lab for recent artificial intelligence progress is this builder bot program where now you can just talk to it and say, hey, I'm in this world, put some trees over there and it'll do that. I put some bottles of water on our picnic blanket and it'll do that and you're in the world. I think there are going to be new paradigms for coding. There are going to be some things that I think are just pretty amazing, especially the first
Starting point is 00:29:55 few times that you do them, but that you're like, whoa, I've never had an experience like this. But most of your life, I would imagine, is not doing things that are amazing for the first time. A lot of this in terms of, I mean, just answering your question from before around, what is it going to take before you're spending most of your time in this? Well, first of all, let me just say it as an aside, the goal isn't to have people spend a lot more time in computing.
Starting point is 00:30:20 I'm asking you to make it to yourself. Yeah, it's to make... Well, I spend all my time in... It's to make... Yeah, it's to make computing more natural, but I think you will spend most of your computing time in this when it does the things that you use computing for somewhat better. So maybe having your perfect workstation is a 5% improvement on your coding productivity. Maybe it's not like a completely new thing, but I mean, look, if I could increase the productivity
Starting point is 00:30:51 of every engineer in meta by 5%, we'd buy those devices for everyone. And I imagine a lot of other companies would too, and that's how you start getting to the scale that I think makes this rival some of the bigger computing platforms that exist today. Let me ask you about identity. We talked about the avatar.
Starting point is 00:31:10 How do you see identity in the metaverse? Should the avatar be tied to your identity, or can I be anything in the metaverse? Can I be whatever the heck I want, can I even be a troll? So there's there's a there's a exciting Freeing possibilities and there's the darker possibilities too Yeah, I mean I think that there's gonna be a range right so we're working on for expression and avatars On one end of the spectrum are kind of expressive in cartoonish avatars. And then on the other end of the spectrum are photorealistic avatars. And I just think the reality is that there are going to be different use cases for different things. And I guess there's another axis.
Starting point is 00:31:55 So if you're going from photorealistic to expressive, there's also like representing you directly versus like some fantasy identity. And I think that there are gonna be things on all ends of that spectrum too. So you'll want photo, like in some experience, you might wanna be like a photo realistic dragon, right? Or, or you know, if I'm playing onward, or just this military simulator game,
Starting point is 00:32:20 it's, I think getting to be more photo realistic as a soldier in that could enhance the experience. There are times when I'm hanging out with friends where I want them to know it's me, so a cartoon or an expressive version of me is good, but there are also experiences like VR chat does this well today where a lot of the experience is dressing up and wearing a fantastical avatar
Starting point is 00:32:47 that's almost like a meme or is humorous. So you come into an experience and it's almost like, you have like a built-in icebreaker because you see people and you're just like, all right, I'm cracking up at what you're wearing because that's funny and it's just like, where'd you get that or oh, you made that, that's, you know, it's awesome.
Starting point is 00:33:05 Whereas, you know, okay, if you're going into a work meeting, maybe a photorealistic version of your real self is gonna be the most appropriate thing for that. So I think the reality is there aren't going to be, there's, it's not just gonna be one thing. You know, my own sense of kind of how you wanna express identity online has sort of evolved over time and that you know early days in Facebook
Starting point is 00:33:28 I thought okay people are gonna have one identity and now I think that's clearly not gonna be the case I think you're gonna have all these different things and and there's utility and being able to do different things so Some of the technical challenges that I'm really interested in around it are How do you build the software to allow people to seamlessly go between them? So say, you could view them as just completely discrete points on a spectrum, but let's talk about the metaverse economy for a second. Let's say I buy a digital shirt for my photo realistic avatar, which by the way I think at the time where we're spending a lot of time in the metaverse doing a lot of our work meetings in the metaverse and et cetera, I would imagine that the economy around virtual clothing as an example is going
Starting point is 00:34:13 to be quite as big. Why wouldn't I spend almost as much money in investing in my appearance or expression for my photo realistic avatar for meetings, as I would for whatever I'm going to wear in my video chat. But the question is, okay, so let's see you buy some shirt for your photo realistic avatar for meetings, this is I would for whatever I'm gonna wear in my video chat. But the question is, okay, so let's see, let's see, buy some shirt for your photo realistic avatar. Wouldn't it be cool if there was a way to basically translate that into a more expressive thing for your kind of cartoonish
Starting point is 00:34:39 or expressive avatar? And there are multiple ways to do that. You can view them as two discrete points. And okay, maybe if a designer sells one thing, then it actually comes on a pack, and there's two, and you can use either one on that. But I actually think the stuff might exist more as a spectrum in the future. And that's what I do think the direction on some of the AI advances that is happening, to be able to, especially stuff around like style transfer being able to take a piece of art or express something and say, okay, paint me, you know,
Starting point is 00:35:11 this photo in the style of go-gen or, you know, whoever it is that you're interested in, you know, take this shirt and put it in the style of what I've designed for my expressive avatar. I think that's going to be pretty compelling. and put it in the style of what I've designed for my expressive avatar, I think that's going to be pretty compelling. And so the fashion you, you might be buying like a generator or like a closet that generates a style. And then let, like, like with the GANS, they'll be able to infinitely generate outfits there by making it. So the reason I wear the same thing all the time, I said, don't like choice.
Starting point is 00:35:42 You've talked about, you've talked about the same thing, but now you don't even have to choose that your closet generates your outfit for you every time. And so you have to live without the generates. I mean, you could do that, although I think that that's, I think some people will, but I think like, I think there's going to be a huge aspect of just people doing creative commerce here.
Starting point is 00:36:05 So I think there is going to be a big market around people designing digital clothing. But the question is, if you're designing digital clothing, do you need to design, if you're the designer, do you need to make it for each kind of specific discrete point along a spectrum, or you're just designing it for a photo realistic case or an expressive case,
Starting point is 00:36:24 or can you design one and have it translate across these things? If I buy a style from a designer who I care about and now I'm a dragon, is there a way to morph that so it goes on the dragon in a way that makes sense? And that I think is an interesting AI problem because you're probably not gonna make it so that
Starting point is 00:36:42 designers have to go design for all those things. But the more useful, the digital content is that you buy in a lot of uses, in a lot of use cases, the more that economy will just explode. That's a lot of what all of the, we were joking about NFTs before, but I think a lot of the promise here is that if the digital goods that you buy are not just tied to one platform or one use case, they end up being more valuable, which means that people are more willing and more likely to invest in them and that just spurs the whole economy. But the question is, that's a fascinating positive aspect, but the potential negative
Starting point is 00:37:19 aspect is that you can have people concealing their identity in order to troll or even not people bots. So, how do you know in the metaverse that you're talking to a real human or an AI or a well-intentioned human? Is that something you think about, something you're concerned about? Well, let's break that down into a few different cases. I mean, because knowing that you're talking to someone who has good intentions is something that I think is not even solved in pretty much anywhere.
Starting point is 00:37:47 But I mean, if you're talking to someone who's a dragon, I think it's pretty clear that they're not representing themselves as a person. I think probably the most pernicious thing that you want to solve for is, I think probably one of the scariest ones is how do you make sure that someone isn't impersonating you? Right? I think probably one of the scariest ones is how do you make sure that someone isn't impersonating you? Right, so you like okay, you're in a future version of this conversation. Yeah, and we have photo-realistic avatars and we're doing this in work rooms or whatever the future version of that is and someone walks in who like looks like me. How do you know that that's me? And one of the things that we're thinking about is,
Starting point is 00:38:25 it's still a pretty big AI project to be able to generate photorealistic avatars that basically can like, they work like these codecs of you. So you kind of have a map from your headset and whatever sensors of what your body's actually doing and it takes the model and it kind of displays it in VR. But there's a question which is, should there be some sort of biometric security
Starting point is 00:38:45 so that like when I put on my VR headset or I'm going to go use that avatar, I need to first prove that I am that. And I think you probably are going to want something like that. So as we're developing these technologies, we're also thinking about the security for things like that because people aren't going to want to be impersonated. That's a huge security issue. Then you just get the question of people hiding behind fake accounts to do malicious things, which is not going to be unique to the metaverse, although certainly in a environment where it's more immersive and you have more of a sense of presence, it could be more painful, but this is obviously something
Starting point is 00:39:33 that we've just dealt with for years in social media and the internet more broadly. And there have been a bunch of tactics that I think we've just evolved to, we've built up these different AI systems to basically get a sense of, is this account behaving in the way that a person would? And it turns out, so in all of the work that we've done around, we call it community integrity, and it's basically like policing harmful content
Starting point is 00:40:06 and trying to figure out where to draw the line. And there are all these like really hard and philosophical questions around like, where do you draw the line on some of this stuff? And the thing that I've kind of found the most effective is as much as possible trying to figure out who are the inauthentic accounts or where the accounts that are behaving
Starting point is 00:40:25 in an overall harmful way at the account level rather than trying to get into policing what they're saying, which I think the metaverse is going to be even harder because the metaverse I think will have more properties of it's almost more like a phone call or it's not like I post a piece of content and is that piece of content good or bad. So I think more of this stuff will have to be done at the level of the account. But this is the area where between the kind of counterintelligence teams that we built up inside the company and years of building
Starting point is 00:41:01 just different AI systems to basically detect what is a real account and what is and I'm not saying we're perfect, but like this is an area where I just think we are like years ahead of basically anyone else in the industry in terms of having built those capabilities and I think that just is going to be incredibly important for this next wave of things. And like you said on the technical level, on a philosophical level, it's an incredibly difficult problem to solve. By the way, I would probably like to open source my avatar so that could be like millions of Lex's walking on just like an army. Like Agent Smith.
Starting point is 00:41:38 Agent Smith, yeah, exactly. So the Unity ML folks built a copy of me and they sent it to me. So there's a person running around and I just been doing reinforcement learning on it. I was going to release it now because, you know, just to have sort of like thousands of Lexes doing reinforcement learning. So they fall over naturally. They have to learn how to like walk around and stuff. So I love that idea of this tension between biometric security. You want to have one identity, but then certain avatars you might have to have many. I don't know which is better security, sort of flooding the world with Lexus and thereby achieving security or really being protective of your identity. I have to ask a security question actually. Well, how does flooding the world with Lex's help me know
Starting point is 00:42:26 in our conversation that I'm talking to the real Lex? I completely destroy the trust in all my relationships then, right, if I flood, because then it's, yeah, that... I think that one's not gonna work that well for you. It's not gonna work that well for the original company. Well, it probably fits some things. Like if you're a public figure and you're trying to have You know a bunch of if you're trying to show up in a bunch of different places in the future
Starting point is 00:42:50 You'll be able to do that in the metaverse So that kind of replication. I think will be useful Yeah, but I do think you're gonna want a notion of like I am talking to the real one Yeah Yeah, especially if the if the fake ones start out performing you and all your private relationships and then you're left behind. I mean, that's a serious concern I have with clones. Again, the things I think about. Okay, so I recently got, I use QNAP NAS storage. So just storage for video stuff. And I recently got hacked. It's the first time for me with the ransomware. It's not me personally, it's all QNUptimizes.
Starting point is 00:43:28 So the question that people have about is about security in general, because I was doing a lot of the right things in terms of security and nevertheless, ransomware basically disabled my device. Is that something you think about? What are the different steps you could take to protect people's data on the security front?
Starting point is 00:43:46 I think that there's different solutions for In strategies where it makes sense to have stuff kind of put behind a fortress, right? So the centralized model versus Decentralizing then I think both have strengths and weaknesses So I think anyone who says okay, just decentralized everything that'll make it more secure. I think that that's tough because the advantage of something like encryption is that we run the largest encrypted service in the world with WhatsApp and one of the first to roll out a multi-platform encryption service. That's something
Starting point is 00:44:23 that I think was a big advance for the industry. And one of the promises that we can basically make because of that, our company doesn't see when you're sending an encrypted message and to an encrypted message what the content is of what you're sharing. So that way if someone hacks meta servers,
Starting point is 00:44:41 they're not going to be able to access the WhatsApp message that you're sending to your friend. And that, I think, matters a lot to people because obviously if someone is able to compromise a company's servers and that company has hundreds of millions or billions of people, then that ends up being a very big deal. The flip side of that is, okay, all the content is on your phone. Are you following security best practices on your phone?
Starting point is 00:45:05 If you lose your phone, all your content is gone. So that's an issue. Maybe you go back up your content from WhatsApp or some other service in an iCloud or something, but then you're just at Apple's Wims about are they going to go turn over the data to some government or are they going to get hacked? So a lot of the time it is useful to have data in a centralized place too, because then you can train systems that can just do much better personalization. I think that in a lot of cases, centralized systems can offer, especially if you're a
Starting point is 00:45:42 serious company, you're running the state of the art stuff. And you have red teams attacking your own stuff. And you're putting out bounty programs and trying to retract some of the best hackers in the world to go break into your stuff all the time. So any system is going to have security issues. But I think the best way forward is to basically try to be as aggressive and open
Starting point is 00:46:05 about hardening the systems as possible, not trying to hide and pretend that there aren't going to be issues, which I think is over time why a lot of open source systems have gotten relatively more secure is because they're open. And it's not rather than pretending that there aren't going to be issues, just people surface them quicker. So I think you want to adopt that approach as a company and just constantly be hardening yourself. Trying to stay once that ahead of the attackers. It's an inherently adversarial space. Yeah. Right, I think it's an interesting security is interesting because of the different kind of threats that we've managed over the last
Starting point is 00:46:41 five years. There are ones who basically the adversaries keep on getting better and better, so trying to kind of interfere with, security is certainly one area of this. If you have like nation states that are trying to interfere in elections or something, like they're kind of evolving their tactics. Whereas on the other hand, I don't wanna be too simplistic about it,
Starting point is 00:47:03 but like if someone is saying something hateful, people usually aren't getting smarter and smarter about how they say hateful things. Right? So maybe there's some element of that, but it's a very small dynamic compared to how advanced attackers and some of these other places get over time. I believe most people are good, so they actually get better over time and not being less hateful, because they realize it's not fun being hateful. It's at least the belief I have.
Starting point is 00:47:32 But first, bathroom break, sure, okay. So we'll come back to AI, but let me ask some difficult questions now. Social dilemma is a popular documentary that raised concerns about the effects of social media on society. You responded with a point by a point of rebuttal titled, what the social dilemma gets wrong. People should read that.
Starting point is 00:47:54 I would say the key point they make is because social media is funded by ads, algorithms want to maximize attention and engagement and an effective way to do so is to get people angry at each other, increase division and so on. Can you steal man their criticisms and arguments that they make in the documentary as a way to understand the concern and as a way to respond to it? Well, yeah, I think that's a good conversation to have. I don't happen to agree with the conclusions, and I think that they make a few assumptions that are just very big jumps that I don't think are reasonable to make.
Starting point is 00:48:39 But I understand overall why people would be concerned that our business model and ads in general, we do make more money as people use the service more in general. Right? So as a kind of basic assumption, okay, do we have an incentive for people to build a service that people use more? Yes, on a lot of levels. I mean, we think what we're doing is good. So we think that if people are finding it useful, they'll use it more. Yes, on a lot of levels. I mean, we think what we're doing is good. So, you know,
Starting point is 00:49:05 we think that if people are finding it useful, they'll use it more. Or if you just look at it, is this sort of if the only thing we cared about is money, which I is not for anyone who knows me, but okay, we're a company. So let's say you just kind of simplify it down to that. Then would we want people to use the services more? Yes. But then, and then you get to the second question, which is, does kind of getting people agitated make them more likely to use the services more? And I think from looking at other media in the world, especially TV and there's the old news adage, if it bleeds at leads, I think that there are a bunch of reasons why someone might think that that
Starting point is 00:49:57 kind of provocative content would be the most engaging. Now what I've always found is two things. One is that, well grab someone's attention in the near term is not necessarily something that they're going to appreciate having seen or going to be the best over the long term. So I think what a lot of people get wrong is that I'm not building this company to like make the most money or get people to spend the most time on this
Starting point is 00:50:23 in the next quarter or the next year. I've been doing this for 17 years at this point, and I'm still relatively young, and I have a lot more that I want to do over the coming decades. I think that it's too simplistic to say, hey, this might increase time in the near term, therefore it's what you're going to do. Because I actually think a deeper look at it, what my incentives are, the incentives of a company that are focused on the long term, is to basically do what people are going to find valuable over time, not what is going to draw people's attention today.
Starting point is 00:50:56 The other thing that I'd say is that I think a lot of times people look at this from the perspective of media or kind of information or civic discourse, but one other way of looking at this is just that, okay, I'm a product designer, right? Our company, you know, we build products. And a big part of building a product is not just the function and utility of what you're delivering, but the feeling of how it feels, right? And we spent a lot of time talking about, in a virtual reality and how the feeling of how it feels. We spend a lot of time talking about virtual reality and how
Starting point is 00:51:26 the key aspect of that experience is the feeling of presence, which is a visceral thing. It's not just about the utility that you're delivering. It's about the sensation. Similarly, I care a lot about how people feel when they use our products. I don't want to build products that make people angry. I mean, that's like not, I think what we're here on this earth to do is to, you know, build something that, you know, people spend a bunch of time doing and it just kind of makes them angry at other people. I mean, I mean, I think that that's, that's not good. That's, you know, that's, that's not what I think would be, sort of a good use of, of our time or a good contribution to the world. So, okay, you know, it's like people, they tell us on a per-content basis, you know, does
Starting point is 00:52:08 this thing, do I like it? Do I love it? Does it make me angry? Does it make me sad? And, you know, based on that, I mean, we choose to basically show content that makes people angry less. Because, you know, of course, if you're designing a product and you want people to be able to connect and feel good over a long period of time, then that's naturally what you're going to do. I think overall,
Starting point is 00:52:35 I understand at a high level, if you're not thinking too deeply about it, why that argument might be appealing, but I just think if you actually look at what our real incentives are, not just like, you know, if we were trying to optimize for the next week, but like as people working on this, like why are we here? And I think it's pretty clear that that's not actually how you would want to design the system. I guess one other thing that I'd say is that, not actually how you would want to design the system. I guess one other thing that I'd say is that while we're focused on the ads business model, I do think it's important to note that a lot of these issues are not unique to ads.
Starting point is 00:53:15 So take like a subscription news business model, for example, I think that has just as many potential pitfalls. Maybe if someone's paying for a subscription, you don't get paid per piece of content that they look at. But say, for example, I think like a bunch of the partisanship that we see could potentially be made worse by you have these kind of partisan news organizations that basically sell subscriptions, and they're only going to
Starting point is 00:53:46 get people on one side to basically subscribe to them. So their incentive is not to print content or produce content that's kind of centrist or down the line either. I bet that what a lot of them find is that if they produce stuff that's kind of more polarizing or more partisan, then that is what gets the more subscribers. So I think that this stuff is all, there's no perfect business model. Everything has pitfalls.
Starting point is 00:54:14 The thing that I think is great about advertising is it makes it's the consumer services free, which if you believe that everyone should have a voice and everyone should be able to connect, then that's a great thing. It was opposed to building a luxury service that not everyone can afford. But look, I mean, every business model,
Starting point is 00:54:29 you have to be careful about how you're implementing what you're doing. You responded to a few things there. You spoke to the fact that there is a narrative of malevolence. Like, you're leaning into them making people angry just because it makes more money in the short term, that kind of thing. So you're, you responded to that, but there's also kind of reality of human nature, just
Starting point is 00:54:52 like you spoke about, there is fights, arguments we get in, and we don't like ourselves afterwards, but we got into them anyway. So our long term growth is, I believe for most of us has to do with learning challenging yourself improving being kind to each other finding a community of people that you know you Connect with on a real human level all that kind of stuff
Starting point is 00:55:20 But it does seem when you look at social media that a lot of fights break out, a lot of arguments break out, a lot of viral content ends up being sort of outraged in one direction or the other. And so it's easy from that to infer the narrative that social media companies are letting this outrage become viral. And so they're increasing the division in the world. I mean, perhaps you can comment on that or further, how can you be, how can you push back on this narrative? How can you be transparent about this battle? Because I think it's not
Starting point is 00:55:59 just motivation or financials, it's a technical problem too, which is how do you improve long-term well-being of human beings? I think that going through some of the design decisions would be a good conversation, but first I actually think, you acknowledge that narrative is somewhat anecdotal. I think it's worth grounding this conversation in the actual research that has been done on this, which by and large finds that social media is not a large driver of polarization. There's been a number of economists and social scientists and folks who have studied this. In a lot of polarization, it varies around the world.
Starting point is 00:56:51 Social media is basically in every country, Facebook is in pretty much every country except for China and maybe North Korea. And you see different trends in different places where, you know, in a lot of countries, polarization is declining, in some it's flat. In the US, it's, it's risen sharply. So the question is, what are the unique phenomenon in the different places? And I think for the people who are trying to say, Hey, social media is the thing that's doing this. I think that that clearly doesn't hold up because social media is a phenomenon that is pretty much
Starting point is 00:57:25 equivalent in all of these different countries. And you have researchers like this economist at Stanford, Matthew Genskau, who is just written at length about this. And it's a bunch of books by political scientists, as reclined in folks, why we're polarized basically goes through this decades long analysis in the US, you know, before I was born, basically talking about some of the forces and kind of partisan politics and Fox News and different things that predate the internet in a lot of ways that that I think are are likely larger contributors. So to the contrary, on this, not only is it pretty clear that social media is not a major contributor, but most of the academic studies that I've seen actually show that social media use
Starting point is 00:58:12 is correlated with lower polarization. Against how the same person who just did the study that I cited about longitudinal polarization across different countries, also did a study that basically showed that if you looked after the 2016 election in the US, the voters who are the most polarized were actually the ones who were not on the internet.
Starting point is 00:58:37 So in there have been recent other studies, I think in Europe and around the world, basically showing that as people stop using social media, they tend to get more polarized. Then there's a deeper analysis around, okay, well, polarization actually isn't even one thing because having different opinions on something isn't, I don't think that that's by itself bad. What people who study this say is most problematic is what they call affective polarization, which is basically
Starting point is 00:59:07 are you, do you have negative feelings towards people of another group? And the way that a lot of scholars study this is they basically ask a group, would you let your kids marry someone of group X? Whatever the groups are that you're worried that someone might have negative feelings towards. And in general, use of social media has corresponded to decreases in that kind of affective polarization. So I just want to, I think we should talk to the design decisions and how we handle the kind of specific pieces of content. But overall, I think it's just worth grounding that discussion in the research that's existed that I think overwhelmingly shows that the mainstream narrative around this is just not right. But the narrative does take hold. And it's compelling to a lot
Starting point is 00:59:56 of people. There's another question I'd like to ask you on this. I was looking at various polls and saw that you're one of the most dislike tech leaders today. 54% unfavorable rating. Elon Musk is 23%. It's basically everybody has a very high unfavorable rating that are tech leaders. Maybe you can help me understand that. Why do you think so many people dislike you? some even hate you, and how do you regain their trust and support? Given everything you just said, why are you losing the battle in explaining to people what actual impact the social media has on society? Well, I'm curious if that's a US survey or world. It is US. Yeah. So I think
Starting point is 01:00:48 that there's a few dynamics. One is that our brand has been somewhat uniquely challenged in the US compared to other places. It's not that there are. I mean, other countries we have issues too. But I think in the US, there was this dynamic where if you look at the net sentiment of coverage or attitude towards us, before 2016, I think there were probably very few months of any where it was negative and since 2016, I think there were probably very few months of any than it's been positive. Politics. But I think it's a specific thing. And this is very
Starting point is 01:01:25 different from other places. So, I think in a lot of other countries in the world, the sentiment towards meta and our services is extremely positive. In the US, we have more challenges. And I think compared to other companies, you can look at certain industries, I think, if you look at it from like a partisan perspective, not from a political perspective, but just kind of culturally. It's like there are people who are probably more left of center and there are people who are more right of center and there's kind of blue America and red America. There are certain industries that I think maybe one half of the country has a more positive
Starting point is 01:02:00 view towards than another. I think we're in a one of the positions that we're in that I think is really challenging is that because of a lot of the content decisions that we're, that we've basically had to arbitrate, and because we're not a partisan company, right, we're not, we're not a Democrat company or a Republican company, we're trying to make the best decisions we can to help people connect and help people have as much voice as they can while having some rules because we're running a community. The net effect of that is that we are constantly making decisions that piss off people in both camps. And the effect that I've sort of seen is that when we make a decision that is a controversial
Starting point is 01:02:52 one that's going to upset, say, about half the country, those decisions are all negative some from a brand perspective. Because it's not like, like, if we make that decision in one way and say half the country is happy about that particular decision that we make, they tend to not say, oh, sweet, Metagot that one right. They're just like, I didn't mess that one up, right?
Starting point is 01:03:16 But their opinion doesn't tend to go up by that much. Whereas the people who kind of are on the other side of it, God, how could you mess that up? Like how could you possibly think that like that piece of content is okay and should be up and should not be censored? Or, and so I think the, whereas if you leave it up and,
Starting point is 01:03:37 or if you take it down, the people who thought it should be taken down, or it's like, all right, fine, great, you didn't mess that one up. So our internal assessment of, in the kind of analytics on our brand are basically any time one of these big controversial things comes up in society. Our brand goes down with half of the country. And then like, if you, and then if you just kind of extrapolate that out, it's just been
Starting point is 01:04:00 very challenging for us to try to navigate what is a polarizing country in a principled way, where we're not trying to kind of huge a one side or the other. We're trying to do what we think is the right thing. But that's what I think is the right thing for us to do, though. So I mean, that's what we'll try to keep doing. Just as a human being, how does it feel, though, when you're giving so much of your day to day life to try to heal division to try to do good in the world as we've talked about that so many people in the US, the place you call home, have a negative view of you as a leader, as a human being, and the company you love.
Starting point is 01:04:41 the company you love. Well, I mean, it's not great, but I, I mean, look, if I wanted people to think positively about me as a person, I don't know, I'm not sure if you go build a company. I mean, it's like, like, or social media companies. I just think it's just something difficult to do with a social media company. Yeah, so, I mean, I don't know, there is a dynamic where a lot of the other people running these companies, internet companies, have sort of stepped back and they just do things
Starting point is 01:05:15 that are sort of, I don't know, less controversial. And some of it may be that they just get tired over time. But it's, I don't know, I think that running a company is hard, building something at scale is hard. You only really do it for a long period of time if you really care about what you're doing. And yeah, so I mean, it's not great. But look, I think that it's some level
Starting point is 01:05:41 whether 25% of people dislike you or 75% of people dislike you or 75% of people dislike you, your experience as a public figure is going to be that there's a lot of people who dislike you, right? So, I actually am not sure how different it is. Certainly, the country's gotten more polarized and we in particular have gotten more controversial over the last five or years or so. But I don't know. I kind of think like as a public figure and leader of one of these enterprises comes to the job.
Starting point is 01:06:19 Part of what you do is like, and look, the answer can't just be ignored, right? Because like a huge part of the job is like you need to be getting feedback and internalizing feedback on how you can do better. But I think increasing what you need to do is be able to figure out who are the good faith critics, who are criticizing you because they're trying to help you do a better job rather than tear you down. And those are the people I just think you have to cherish and listen very closely to the things that they're saying,
Starting point is 01:06:50 because I think it's just as dangerous to tune out everyone who says anything negative and just listen to the people who are kind of positive and support you, you know, as it would be psychologically to pay attention trying to make people who are never gonna like you like you. So I think would be psychologically to pay attention trying to make people who are never going to like you like you. So I think that's just kind of a dance that people have to do. But I mean, I, you know, you kind of develop more of a feel for like who actually is trying
Starting point is 01:07:15 to accomplish the same types of things in the world and who has different ideas about how to do that and how can I learn from those people? And like, yeah, we get stuff wrong. And when the people whose opinions I respect call me out on getting stuff wrong, that hurts and makes me want to do better. But I think at this point, I'm pretty tuned to just, all right, if someone, if I know they're, they're kind of like operating in bad faith and they're not really trying to help, then, you know, I don't know. It's not, it doesn't, you know, I think over time, it just doesn't bother you that much. Then, you know, I don't know. It doesn't, you know, I think over time,
Starting point is 01:07:43 it just doesn't bother you that much. But you are surrounded by people that believe in the mission that love you. Are there friends or colleagues in your inner circle you trust that call you out on your bullshit whenever you're thinking maybe misguided it as it is for leaders at times? I think we have a famously open company culture
Starting point is 01:08:04 where we sort of encourage that kind of descent internally, which is why there's so much material internally that can leak out with people sort of disagreeing is because that's sort of the culture. Our management team, I think it's a lot of people, there's some newer folks who come in, there's some folks who've been there for a while, but there's a very high level of trust. I would say it is a relatively confrontational group of people. My friends and family, I think, will push me on this, but look, I think you need some diversity. It can't just be people who are your friends and family. people who are your friends and family. It's also, you know, I mean, there are journalists or analysts or, you know, peer executives at other companies
Starting point is 01:08:53 or, you know, other people who sort of are insightful about thinking about the world, you know, certain politicians or people kind of in that sphere who I just think have like very insightful perspectives, who even if they would, they come at the world from a different perspective, which is sort of what makes the perspective so valuable. But, you know, I think fundamentally
Starting point is 01:09:16 you're trying to get to the same place in terms of, you know, helping people connect more, helping the whole world function better, not just, you know just one place or another. And I don't know, I mean, those are the people whose opinions really matter to me. And that's how I learn on a day-to-day basis. People are constantly sending me comments on stuff
Starting point is 01:09:37 for links to things they found interesting. And I don't know, it's kind of constantly evolving this model of the world and kind of what we should be aspiring to be. You've talked about, you have a famously open culture which comes with the criticism and the painful experiences. So let me ask you another difficult question. Francis Hogan, the Facebook whistleblower, leaked the internal Instagram research into teenagers and well-being. Her claim is that Instagram is choosing profit over well-being of teenage girls, so Instagram
Starting point is 01:10:13 is, quote, toxic for them. Your response titled, What Are Research Really Says About teen well-being and Instagram says no, Instagram research shows that 11 of 12 well-being issues teenage girls who said they struggle with those difficult issues also said that Instagram made them better rather than worse. Again, can you steal man and defend the point and Francis Hogan's characterization of the study and then help me understand the positive and negative effects of Instagram and Facebook on young people. So there are certainly questions around teen mental health that are really important.
Starting point is 01:10:56 It's hard to, you know, as a parent, it's like hard to imagine any set of questions that are sort of more important, I guess, maybe other aspects of physical health or well-being are probably come to that level. But these are really important questions, right? Which is why we dedicate teams to studying them. I don't think the internet or social media are unique in having these questions. I mean, I think people, and there have been sort of magazines with promoting certain body types for women
Starting point is 01:11:26 and kids for decades, but we really care about this stuff. So we wanted to study it, and of course, we didn't expect that everything was going to be positive all the time. So the reason why you study this stuff is to try to improve and get better. So the place where I disagree with the characterization. First, I thought, you know, some of the reporting and coverage of it just took the whole thing out of proportion and that it focused on, as you said, I think there were like 20 metrics in there and on, you know, 18 or 19, the effect of using Instagram was neutral or positive on the teens' well-being. And there was one area where I think it showed that we needed to improve, and
Starting point is 01:12:05 we took some steps to try to do that after doing the research. But I think having the coverage just focus on that one without focusing on the, I think an accurate characterization would have been that kids using Instagram, or not kids' teens, is generally positive for their mental health. But of course, that was not the narrative that came out. So I think it's hard to, that's not a kind of logical thing to straw man, but I sort of disagree or steel man,
Starting point is 01:12:31 but I sort of disagree with that overall characterization. I think anyone sort of looking at this objectively would. But then, you know, I mean, there is this sort of intent critique that I think you were getting at before, which says, you know, there is this sort of intent critique that I think you were getting at before, which says, you know, to assume some sort of malevolence, right? It's like, which it's really hard for me to really wrap my head around this because as far as I know,
Starting point is 01:12:59 it's not clear that any of the other tech companies are doing this kind of research. So why the narrative should form that we did research, because we were studying an issue because we wanted to understand it to improve and took steps after that to try to improve it, that your interpretation of that would be that we did the research and tried to sweep it under the rug. It's sort of, I don't know, it's beyond credibility to me that like that's the accurate description of the actions that we've taken compared to the others in the industry.
Starting point is 01:13:31 So I know that's kind of, that's my view on it. These are really important issues and there's a lot of stuff that I think we're going to be working on related to teen mental health for a long time, including trying to understand this better. And I would encourage everyone else in the industry to do this too. Yeah, I would love there to be open conversations and a lot of great research being released internally and then also externally. It doesn't make me feel good to see press obviously get way more clicks
Starting point is 01:14:05 when they see negative things about social media. Let's objectively speaking, I can just tell that there's hunger to say negative things about social media. And I don't understand how that's supposed to lead to an open conversation about the positives and the negatives, the concerns about social media, especially when you're doing those kind of research. I mean, I don't know what to do with that, but let me ask you as a father, there's a way
Starting point is 01:14:35 heavy on you that people get bullied on social networks. So people get bullied in their private life. But now, because so much of our life is in the digital world, the bullying moves from the physical world to the digital world. So you're now creating a platform on which bullying happens. And some of that bullying can lead to damage to mental health. And some of that bullying can lead to depression, even suicide. There's a way heavy on you that people have committed suicide or will commit suicide
Starting point is 01:15:15 based on the bullying that happens on social media. Yeah, I mean, this is a, there's a set of harms that we basically track and build systems to fight against. And bullying and self-harm are, these are some of the biggest things that we are most focused on. For bullying, like you say, it's going to be, while this predates the internet, that it's probably impossible to get rid of all of it, you want to give people tools to fight it and you want to fight it yourself.
Starting point is 01:15:57 And you also want to make sure that people have the tools to get help when they need it. So I think this isn't like a question of, you know, can you get rid of all bullying? I mean, it's like, all right. I mean, I have two daughters and they fight and push each other around and stuff too. And the question is just, how do you handle that situation? And there's a handful of things that I think you can do. We talked a little bit before around some of the AI tools that you can build
Starting point is 01:16:26 to identify when something harmful is happening. It's very hard in bullying because a lot of bullying is very context specific. It's not like you're trying to fit a formula of like, if looking at the different harms, someone promoting a terrorist group is probably one of the simpler things to generally find because things promoting that group are going to, you know, look at a certain way or feel a certain way. Bullying could just be, you know, someone making some subtle comment about someone's appearance that's idiosyncratic to them. And it could look at just like humor.
Starting point is 01:16:58 So humor to one person. Exactly. And destructive to another being. Yeah. So with bullying, I think there are certain things that you can find through AI systems, but I think it is increasingly important to just give people more agency themselves.
Starting point is 01:17:14 So we've done things like making it so people can turn off comments or take a break from hearing from a specific person without having to signal at all that they're going to stop following them or kind of make some stand that, okay, I'm not friends with you anymore, I'm not following you. I just like, I just don't want to hear about this, but I also don't want to signal at all publicly or to them that there's been an issue. And then you get to some of the more extreme cases like you're talking about where someone is thinking about self-harm or suicide.
Starting point is 01:17:49 And there, we found that that is a place where AI can identify a lot, as well as people flagging things. If people are expressing something that is potentially their thing of hurting themselves. Those are cues that you can build systems and hundreds of languages around the world to be able to identify that. And one of the things that I'm actually quite proud of is we've built these systems that I think are clearly
Starting point is 01:18:19 leading at this point that not only identify that, but then connect with local first responders and have been able to save, I think, at this point that not only identify that, but then connect with local first responders, and have been able to save, I think, at this point. In thousands of cases, be able to get first responders to people through these systems who really need them, because of specific plumbing that we've done between the AI work and being able to communicate with local first responder organizations. We're rolling that out in more places around the world. And I think the team that worked on that just did awesome stuff.
Starting point is 01:18:49 So I think that that's a long way of saying, yeah, I mean, this is a heavy topic. And there's you want to attack it in a bunch of different ways. And also kind of understand that some of nature is for people to do this to each other, which is unfortunate, but you can give people tools and build things that help. It's still one hell of a burden, no. A platform that allows people to fall in love with each other is also by nature going to be a platform that allows people to hurt each other. And when you're managing such a platform, it's difficult.
Starting point is 01:19:27 And I think you spoke to it, but the psychology of that of being a leader in that space of creating technology that's playing in this space, like you mentioned, psychology is really damn difficult. And I mean, the burden of that is just great. I just wanted to hear you speak to that point. I have to ask about the thing you've brought up a few times, which is making controversial decisions.
Starting point is 01:19:55 Let's talk about free speech and censorship. So there are two groups of people, pressuring meta on this. One group is upset that Facebook, the social network allows misinformation and quotes to be spread on the platform. The other group are concerned that Facebook censors speech by calling it misinformation. So you're getting it from both sides. You, um, in 2019, October at, uh, Georgetown University eloquently defended the importance of free speech but then COVID came and the 2020 election came
Starting point is 01:20:33 Do you worry that outside pressures from advertisers politicians the public have forced me to damage the ideal of free speech that you spoke highly of? Just to say some obvious things up front, I don't think pressure from advertisers or politicians directly in any way affects how we think about this. I think these are just hard topics. So let me just take you through our evolution from the beginning of the company to where we are now. You don't build a company like this unless you believe that people expressing themselves is a good thing.
Starting point is 01:21:04 Right, so that's themselves is a good thing. Right, so that's sort of the foundational thing. You can kind of think about our company as a formula where we think giving people voice and helping people connect creates opportunity. Right, so those are the two things that we're always focused on are sort of helping people connect. We talked about that a lot, but also giving people voice and ability to express themselves. And by the way, most of the time when people express themselves, that's not like politically controversial content. It's like expressing something about their identity that's more related to the avatar conversation we had earlier
Starting point is 01:21:37 in terms of expressing some facet, but that's what's important to people on a day-to-day basis. And sometimes when people feel strongly enough about something, it kind of becomes a political topic. That's sort of always been a thing that we've focused on. There's always been the question of safety in this, which, you know, if you're building a community, I think you have to focus on safety. We've had these community standards from early on, and there are about 20 different kinds
Starting point is 01:22:01 of harm that we track and try to fight actively. And we've talked about some of them already. So it includes things like bullying and harassment. It includes things like terrorism or promoting terrorism, inciting violence, intellectual property theft. And in general, I think, called about 18 out of 20 of those, there's not really a particularly polarized definition of that.
Starting point is 01:22:28 I think you're not really going to find many people in the country or in the world who are trying to say we should be fighting terrorist content less. I think that the content where there are a couple of areas where I think this has gotten more controversial recently, which I'll talk about. And you're right that misinformation is basically is up there. I think sometimes the definition of hate speech is up there too.
Starting point is 01:22:53 But I think in general, most of the content that I think we're working on for safety is not actually, you know, people don't kind of have these questions. So it's sort of this subset. But if you go back to the beginning of a company, this was sort of pre-deep learning days. And therefore, you know, it was me and my roommate Dustin joined me. And like, if someone posted something bad, you know, it was the AI technology did not exist yet
Starting point is 01:23:27 to be able to go basically look at all the content, and we were a small enough outfit that no one would expect that we could review it all, even if like someone reported it to us, we basically did our best, right? It's like someone would report it, and we try to look at stuff and deal with stuff. And for call the first, I don't know, seven or eight years of the company, you know, we weren't that big of a company, you know,
Starting point is 01:23:56 for a lot of that period, we weren't even really profitable. The AI didn't really exist to be able to do the kind of moderation that we do today, and that at some point, and kind of the middle of the last decade that started to flip. And we, you know, we became, it got to the point where we were sort of a larger and more profitable company, and the AI was starting to come online to be able to proactively detect
Starting point is 01:24:20 some of the simpler forms of this. So things, things like pornography, you could train an image class of fire to identify what a nipple was, or you can fight against terrorist content. You still put your papers on this. It's great. Oh, of course there are.
Starting point is 01:24:34 Technical papers. Of course there are. Those are relatively easier things to train AI to do. Then, for example, understand the nuances of what is inciting violence in 100 languages around the world and not have the false positives of like, okay, are you posting about this thing that might be inciting violence? Because you're actually trying to denounce it. In which case, we probably shouldn't take that down. Right? If you're trying to denounce something that's inciting violence in some kind
Starting point is 01:25:02 of dialect in a corner of India, as opposed to, okay, actually, you're posting this thing because you're trying to inside violence. Okay, building an AI that can basically get to that level of nuance and all the languages that we serve is something that I think is only really becoming possible now, not towards the middle of the last decade. But there's been this evolution, and I think what happened, people sort of woke up after 2016, and a lot of people were like,
Starting point is 01:25:32 okay, the country is a lot more polarized, and there's a lot more stuff here than we realized. Why weren't these internet companies on top of this? And I think at that point, it was reasonable feedback that some of this technology had started becoming possible. And at that point, I really did feel like we needed to make a substantially larger investment. We'd already worked on this stuff a lot on AI
Starting point is 01:26:01 and on these integrity problems, but that we should basically invest, you know, have a thousand or more engineers basically work on building these AI systems to be able to go and proactively identify the stuff across all these different areas. Okay, so we went into that. Now we've built the tools to be able to do that. And now I think it's actually a much more complicated set of philosophical rather than technical questions, which is the exact policy is which are okay. Now, the way that we basically hold ourselves accountable, because we use transparency reports every quarter, and the metric that we track is for each of those
Starting point is 01:26:37 20 types of harmful content. How much of that content are we taking down before someone even has to report it to us? Right, so how effective is our AI at doing this? But that basically creates this big question, which is, okay, now we need to really be careful about how proactive we set the AI and where the exact policy lines are around what we're taking down. It's certainly at a point now where, you know where I felt like at the beginning of that journey of building those AI systems, there's a lot of push. There's things like, you've got to do more. There's clearly a lot more bad content that people aren't reporting or that
Starting point is 01:27:19 you're not getting to, and you need to get more effective at that. And I was pretty sympathetic to that. But then I think at some point along the way, there started to be almost equal issues on both sides of, okay, actually you're kind of taking down too much stuff, right? Or some of the stuff is borderline and it wasn't really bothering anyone and then report it. So is that really an issue that you need to take down?
Starting point is 01:27:43 Whereas we still have the critique on the other side too, where a lot of people think we're not doing enough. So it's become, as we built the technical capacity, I think it becomes more philosophically interesting, almost where you want to be on the line. And I just think like you don't want one person making those decisions. So we've also tried to innovate in terms of building out this independent oversight board,
Starting point is 01:28:07 which has people who are dedicated to free expression, but from around the world, who people can appeal cases to. So, a lot of the most controversial cases basically go to them, and they make the final binding decision on how we should handle that. And then, of course, their decisions, we then try to figure out what the principles are behind those and encode them into the algorithms. And how are those people chosen, which, you know, you're outsourcing a difficult decision? Yeah, the initial people, we chose a handful of chairs for the group. And we basically chose the people for a commitment to free expression and like a broad understanding
Starting point is 01:28:49 of human rights and the trade-offs around free expression. But fundamentally people who are going to lean towards free expression. To towards freedom of speech. Yeah. So there's also this idea of fact checkers, so jumping around to the misinformation questions actually during COVID, which is an exceptionally speaking of poor legislation. I got to speak to the COVID thing. And I think one of the hardest set of questions
Starting point is 01:29:10 around free expression, because you asked about Georgetown is my stance fundamentally changed. And the answer to that is no, my stance has not changed. It is fundamentally the same as when I was talking about Georgetown from a philosophical perspective. The challenge with free speech is that everyone agrees
Starting point is 01:29:31 that there is a line where if you're actually about to do physical harm to people, that there should be restrictions. So I mean, there's the famous Supreme Court historical example of like you can't yell, fire in a crowded theater. The thing that everyone disagrees on is what is the definition of real harm, where I think some people think, okay, this should only be a very literal, I mean, take it back to
Starting point is 01:29:58 the bullying conversation we were just having, where is it just harm if the person is about to hurt themselves because they've been bullied so hard, or is it actually harm if the person is about to hurt themselves because they've been bullied so hard, or is it actually harm, like as they're being bullied? And kind of at what point in the spectrum is that, and that's the part that there's not agreement on. But I think what people agree on pretty broadly is that when there is an acute threat that it does make sense from a societal perspective to tolerate less speech that could be potentially harmful in that acute situation. So I think where COVID got very difficult is, I don't think anyone expected this to be going on for years. But if you'd kind of asked the opriory, would a global pandemic where a lot of people are dying
Starting point is 01:30:44 and catching this, is that an emergency that where you'd kind of consider it that, it's problematic to basically yell fire in a crowded theater, I think that that probably passes that test. So I think that that's, it's a very tricky situation, but I think the fundamental commitment to free expression is there. That's what I believe.
Starting point is 01:31:09 And again, I don't think you start this company unless you care about people being able to express themselves as much as possible. But I think that that's the question. It is like, how do you define what the harm is and how acute that is? And what are the institutions that define that harm? A lot of the criticism is that the CDC, the WHO, the institutions we've come to trust as a civilization to give the line of what is and isn't harm
Starting point is 01:31:37 in terms of health policy have failed in many ways and small ways and in big ways, depending on who you ask. And then the perspective of meta and Facebook is like, well, where the hell do I get the information of what isn't, isn't misinformation. So it's a really difficult place to be in, but it's great to hear that you're leaning towards freedom of speech on this aspect. And again, I think this actually calls to the fact that we need to reform institutions that help keep an open mind of what isn't isn't misinformation.
Starting point is 01:32:09 And misinformation has been used to bully on the issue. I mean, I just have, you know, friends with Joe Rogan and he's called as a I remember hanging out with him in Vegas and somebody else stops spreading misinformation. with him in Vegas and somebody else stops spreading misinformation. I mean, and there's a lot of people that follow him that believe he's not spreading misinformation. Like, you can't just not acknowledge the fact that there's a large number of people that have a different definition of misinformation. And that's such a tough place to be. Like, who do you listen to? Do you listen to quote unquote experts who gets as a person was a PhD I gotta say I mean I'm not sure I know what defines an expert especially in a new
Starting point is 01:32:53 In a totally new Pandemic or a new catastrophic event especially when politics is involved and especially when the media involved that can propagate sort of outrageous narratives and thereby make a lot of money, like, what the hell? Where's the source of truth? And then everybody turns to Facebook, it's like, please tell me what the source of truth is. Well, I mean, well, how would you handle this if you were in my position? It's very, very, very, very difficult.
Starting point is 01:33:25 I would say, I would more speak about how difficult the choices are and be transparent about what the hell do you do with this? Like here, you got exactly, ask the exact question, you just asked me, but to the broader public. Like, okay, yeah, you guys tell me what to do. So like crowdsource it. And then the other, uh, the other aspect is when you, you spoke really eloquently about the fact that the,
Starting point is 01:33:51 there's this, there's this going back and forth. And now there's a feeling like you're censoring a little bit too much. And so I would lean, I would try to be ahead of that feeling. I would now lean towards freedom of speech and say, you know, we're not the ones that go to define this information Let it be a public debate Let the idea stand and I actually place, you know this idea of misinformation. I place the responsibility on the poor communication skills of scientists They should be in the battlefield of ideas and everybody who is They should be in the battlefield of ideas. And everybody who is spreading information
Starting point is 01:34:27 against the vaccine, they should not be censored. They should be talked with. And you should show the data. You should have open discussion as opposed to rolling your eyes and saying, I'm the expert. I know what I'm talking about. No, you need to convince people. It's a battle of ideas.
Starting point is 01:34:43 So that's the whole point of freedom of speech It's the way to defeat bad ideas is with with good ideas with speech So like the responsibility here falls on the poor Communication skills of scientists. Thanks to social media Scientists are not communicators. They have the power to communicate some of the best stuff I've seen about COVID-19 from doctors is on social media. It's a way to learn to respond really quickly to go faster than the peer review process and so they just need to get way better at that communication and also by better I don't mean just
Starting point is 01:35:19 Convincing I also mean speak with humility don't talk down to people all those kinds of kinds of things. And as a platform, I would say, I would step back a little bit. Not all the way, of course, because there's a lot of stuff that can cause real harm, as we talked about, but you'd lean more towards freedom of speech, because then people from a brand perspective wouldn't be blaming you for the other ills of society, which there are many. The institutions have flaws, the political divide, obviously politicians have flaws, that's news. The media has flaws that they're all trying to work with, and because of the central place of Facebook in the world, all of those flaws somehow kind of propagate to Facebook, and you're sitting there as
Starting point is 01:36:07 Plato, the philosopher, have to answer to some of the most difficult questions asking, being asked a human civilization. So I don't know, maybe this is an American answer, though, to lean towards phantom of speech. I don't know if that applies globally. So yeah, I don't know, but transparency and saying, I think as a technologist, one of the things I sense about Facebook and matter when people talk about this company is they don't necessarily understand fully how difficult the problem is. You talked about AI has to catch a bunch of harmful stuff really quickly, just the sea of data you have to deal with.
Starting point is 01:36:44 It's a really difficult problem. So like any of the critics, if you just hand them the helm for a week, let's see how well you can do. Like that, to me, that's definitely something that would wake people up to how difficult this problem is, if there's more transparency, saying how difficult this problem is, if there's more transparency, it's saying how difficult this problem is. Let me ask you about an AI front,
Starting point is 01:37:07 just because you mentioned language in my ineliquence. Translation is something I wanted to ask you about. And first, just to give a shout out to the supercomputer, you've recently announced the AI research supercluster, RSC, obviously, I'm somebody who loves the GPUs. It currently has 6000 GPUs, NVidADGX, A100s is the systems that have in total 6000 GPUs and it will eventually, maybe this year, maybe soon, we'll have 16,000 GPUs. So it can do a bunch of different
Starting point is 01:37:41 kinds of machine learning applications. There's a cool thing on the distributed storage aspect and all that kind of stuff. So one of the applications that I think is super exciting is translation, real-time translation. I mentioned to you that having a conversation, I speak Russian fluently, I speak English somewhat fluently, and I'm having a conversation with Vladimir Putin,
Starting point is 01:38:04 say, as a use case, me, as a user coming to you, as a use case, we both speak each other's language. I speak Russian, he speaks English. How can we have that communication go well with the help of AI? I think it's such a beautiful and a powerful application of AI to connect the world that bridged the gap, not necessarily to me and Putin, but people that don't have that shared language.
Starting point is 01:38:30 Can you just speak about your vision with translation? Because I think that's a really exciting application. If you're trying to help people connect all around the world, a lot of content is produced in one language and people and all these other places are interested in it. So being able to translate that just unlocks a lot of value on a day-to-day basis. And so the kind of AI around translation is interesting because it's gone through a bunch of iterations. But the basic state of the art is that
Starting point is 01:39:00 you don't wanna go through different kind of intermediate symbolic representations of language or something like that. You basically want to be able to map the concepts and basically go directly from one language to another. You just can train bigger and bigger models in order to be able to do that. That's where the research supercluster comes in is basically a lot of the trend in machine learning. It's just you're building bigger and bigger models and you just need a lot of computation
Starting point is 01:39:34 to train them. So it's not that like the translation would run on the supercomputer, the training of the model, which could have billions or trillions of examples of, you know, basically that. You're training models on this supercluster in days or weeks that might take a much longer period of time on a smaller cluster, so just wouldn't be practical for most teams to do. But the translation work, we're basically getting from, being able to go between about 100 languages seamlessly today,
Starting point is 01:40:11 to being able to go to about 300 languages in the near term. So from any language to any other language? Yeah, and that's it. And part of the issue when you get closer to more languages is, some of these get to be pretty not very popular languages, where there isn't that much content in them. So you end up having less data and you need to use a model that you've built up around other examples.
Starting point is 01:40:42 And this is one of the big questions around AI is how generalizable can things be. And that I this is one of the big questions around AI is like how generalizable can things be. And that I think is one of the things that's just kind of exciting here from a technical perspective. But capturing, we talked about this with the metaverse capturing the magic of human-to-human interaction.
Starting point is 01:40:56 So me and Putin, okay, again, this is- I mean, it's a topic example because you actually both speak Russian and English. But that's a future. I see it as a touring test of a kind because we would both like to have an AI that improves because I don't speak Russian that well. He doesn't speak English that well. It would be nice to outperform our abilities.
Starting point is 01:41:17 Sure. And that's a really nice bar because I think AI can really help in translation for people that don't speak the language at all, but to actually capture the magic of the chemistry, the translation, which would make the metaverse super immersive. I mean, that's exciting. You remember the barrier of language, period? Yeah, so when people think about translation, I think a lot of that is, they think about text to text, but speech to speech, I think, is a whole other thing.
Starting point is 01:41:46 And I mean, one of the big lessons on that, which I was referring to before, is, I think early models, it's like, all right, they take speech, they translate it to text, translate the text to another language, and then kind of output that as speech in that language. And you don't want to do that. You just want to be able to go directly from speech in one language to speech in another language and build up the models to do that. I think one of the, there have been, when you look at the progress in machine learning, there have been big advances in the techniques. Some of the advances in self-supervised learning, which I know you talked to Jan about, and he's one of the leading thinkers in this area.
Starting point is 01:42:25 I just think that that stuff is really exciting, but then you couple that with the ability to just throw larger and larger amounts of compute at training these models, and you can just do a lot of things that were harder to do before. But we're asking more of our systems too, right? So, if you think about the applications
Starting point is 01:42:44 that we're gonna need for the metaverse, or think about it, okay, so let's talk about AR here for a second, you're gonna have these glasses. They're gonna look, hopefully like a normal-ish looking pair of glasses, but they're gonna be able to put holograms in the world and intermix virtual and physical objects in your scene.
Starting point is 01:43:06 One of the things that's going to be unique about this compared to every other computing device that you've had before is that this is going to be the first computing device that has all the same signals about what's going on around you that you have. Right, so your phone, you can have it take a photo or a video. But I mean, these glasses are gonna, whenever you activate them, they're gonna be able to see what you see. From your perspective, they're gonna be able to hear what you hear
Starting point is 01:43:31 because they're the microphones and all that are gonna be right around where your ears are. So you're gonna want an AI assistant, that's a new kind of AI assistant that can basically help you process the world from this first person perspective, or from the perspective that you have. And the utility of that is gonna be huge,
Starting point is 01:43:51 but the kinds of AI models that we're gonna need are going to be just, I don't know, there's a lot that we're gonna need to basically make advances in. But I mean, but that's why I think these concepts of the metaverse and the advances in AI are so fundamentally interlinked that I mean, they're kind of enabling each other.
Starting point is 01:44:12 Yeah, like the world builder is a really cool idea. Like you can be like a Bob Ross, like I'm gonna put a little tree right here. Yeah, I mean, I need a little tree. It's missing a little tree. And then at scale, like enriching your experience in all kinds of ways.
Starting point is 01:44:26 You mentioned the assistant too. That's really interesting how you can have AI assistants helping you out on different levels of sort of intimacy of communication. It could be just like scheduling or it could be like almost like therapy. Clearly, I need some. So let me ask you, you're one of the most successful people ever. You've built an incredible company that has a lot of impact. What advice do you have for young people today?
Starting point is 01:44:53 How to live a life that can be proud of? How to build something that can have a big positive impact on the world? Well, let's break that down because I think you're proud of have a big positive impact. Well, you're actually listening. And how do live your life are actually three different things that I think they could line up. But, and also like what age of people are you talking to? Because I mean, I can like high school in college, so you don't really know what you're doing But you're dream big and you really have a chance to do something unprecedented. Yeah So I guess people my age. Okay, so let's maybe maybe start with the the kind of most philosophical and abstract version of this
Starting point is 01:45:41 every night when I put my daughters to bed We go through this thing and they call it the Good Night Things because we basically, what we talk about at night. And I go through them. Sounds like a good show. Yeah, the Good Night Things. Yeah, Krasil is always asking, can I get Good Night Things?
Starting point is 01:46:04 I don't know, we go to bed too early. But it's, um, but I basically go through with Max and and, and Augie, um, you know, what are the, the things that are most important in life? Right? That I just, it's like, what do I want them to remember and just have like really ingrained in them as they grow up? And it's health, right? Making sure that you take care of yourself and keep yourself in good shape, loving friends and family, right? Because, you know, having the relationships, the family and making time for friends, I think, is perhaps one of the most important things.
Starting point is 01:46:43 And then the third is maybe a little more amorphous, but it is something that you're excited about for the future. And when I'm talking to a four year old, often I'll ask her what she's excited about for tomorrow or the week ahead. But I think for most people, it's really hard. I mean, the world is a heavy place.
Starting point is 01:47:01 And I think like the way that we navigate it is that we have things that we're looking forward to. So whether it is building AR glasses for the future or being able to celebrate my 10 year wedding anniversary with my wife that's coming up, it's like I think people, you have things that you're looking forward to. Or for the girls, it's often I wanna see mom
Starting point is 01:47:23 in the morning. Right, it's like just, but it's, I think that that's a really critical thing. And then the last thing is I ask them every day, what did you do today to help someone? Because I just think that that's a really critical thing is like, like, it's easy to kind of get caught up in yourself and kind of stuff that's really far down the road. But did you do something just concrete today to help someone? And it can just be as simple as, okay, yeah, I helped set the table for lunch or this other kid in our school was having a hard time with something and I like helped explain it to him.
Starting point is 01:47:59 But, but those are, that's sort of like, if you were to boil down my overall life philosophy into what I try to impart to my kids Those are the things that I think are really important. So okay, so let's say college so if you're graduating college probably more practical advice It's almost very focused on people and I It's almost very focused on people. And I think the most important decision you're probably going to make if you're in college is who you surround yourself with because you become like the people you surround yourself with. And I sort of have this hiring heuristic at Meta, which is that I will only hire someone to work for me if I could see
Starting point is 01:48:48 myself working for them. Not necessarily that I want them to run the company because I like my job, but in an alternate universe, if it was their company and I was looking to go work somewhere, would I be happy to work for them. I think that's a helpful heuristic to help balance. When you're building something like this, there's a lot ofuristic, to help balance, and when you're building something like this, there's a lot of pressure to, you know, you want to build out your teams because there's a lot of stuff that you need to get done, and then everyone always says, don't compromise
Starting point is 01:49:13 on quality, but there's this question of, okay, how do you know that someone is good enough? And I think my answer is, I would want someone to be, to be on my team if I would work for them. But I think it's actually a pretty similar answer to like if you were choosing friends or a partner or something like that. So when you're kind of in college, trying to figure out what your circle is gonna be,
Starting point is 01:49:35 trying to figure out, you know, you're evaluating different job opportunities, who are the people, even if they're gonna be peers in what you're doing, who are the people who in an alternate university You would want to work for them because you think you're gonna learn a lot from them because they know because they are kind of values aligned on the things that you care about and they're gonna like And they're gonna push you, but also they know different things and have different experiences that that are kind of more of what you want to become like over time
Starting point is 01:50:03 So I don't know. I think probably people are too in general objective focused and maybe not focused enough on the connections and the people who they're who they're basically building relationship with. I don't know what it says about me, but my place in Austin now has seven, like, a robot. So I'm surrounded myself by robots, which is probably something I should look into. What kind of world would you like to see your daughters grow up in, even after you're gone? Well, I think one of the promises of all the stuff that is getting built now, is that it can be a world where more people have,
Starting point is 01:50:48 can just live out their imagination. But one of my favorite quotes, it's, I think it was Attributes to Picasso. It's that all children are artists, and the challenge is how do you remain one when you grow up? And, I mean, if you have kids, you, this is pretty clear. I mean, they're just like, have wonderful imaginations. And part of what I think is gonna you have kids, you, this is pretty clear. I mean, they just like have wonderful imaginations.
Starting point is 01:51:06 And part of what I think is going to be great about creator economy and the metaverse and all this stuff is like this notion around that a lot more people in the future are going to get to work doing creative stuff than what I think today we would just consider traditional labor or service. And I think that that's awesome. And like, I think that that's like what, a lot of what people are here to do is like collaborate
Starting point is 01:51:31 together, work together, think of things that you want to build, and go do it. And I don't know, one of the things that I just think is striking. So I like, I teach my, my, my daughters, like some basic coding with scratch. I mean, they're still obviously really young, but I think of coding as building, right? It's like when I'm coding, I'm like building something that I want to exist. But in my youngest daughter, she's very musical and pretty artistic, and she thinks about coding as art. She calls it code art, not the code, but the output of what she is making. It's like she's just very interesting visually and what she can kind of output and how it can move
Starting point is 01:52:11 around. And do we need to fix that? Are we good? What happened? Do we have to clap? Alexa? Yeah. So I was just talking about, you know, Auggie and her code art. But I mean, to me, this is like a beautiful thing, right? The notion that like, for me, coding was this functional thing. And I enjoyed it. And it, like, helped build something utilitarian, but that for the next generation of people, it will be even more an expression of their kind of imagination and artistic sense for what they want to exist. So I don't know, if that happens, if we can help bring about this world where, you know, a lot more people can,
Starting point is 01:52:53 that that's like their existence going forward is being able to basically create and live out, you know, in all these different kinds of art. I just think that that's like a beautiful and wonderful thing. And we'll be very freeing for humanity to spend more of our time on the things that matter to us. Yeah, allow more and more people to express their art in the full meaning of that word. Yeah. That's a beautiful vision. We mentioned that you are mortal. Are you afraid of death? Do you think about your mortality? And are you afraid of death? Do you think about your mortality?
Starting point is 01:53:27 And are you afraid of it? You didn't sign up for this on a podcast. No, I mean, that's an interesting question. I'm definitely aware of it. I do a fair amount of extreme sport type stuff. So I'm definitely aware of it. And you're flirting with it a bit. I train hard. I mean, so if I'm gonna go out in a 15 foot wave and I'm like, then we'll then it's like, all right, I'll make sure we have the right safety here
Starting point is 01:54:03 and make sure that I'm like,, use that spot and all that stuff. But like, but, you know, I mean, you the risk is still there. It takes some head blows along the way. Yes. Um, but definitely aware of it, definitely would like to stay safe. I have a lot of stuff that I want to build and want to. There's freak you out that it's finite though? That there's a deadline when it's all over? And there'll be a time when your daughters are around and you're gone. I don't know, that doesn't freak me out. I think... Constraints are helpful.
Starting point is 01:54:46 Yeah, yeah, the finiteness makes ice cream taste more delicious and the fact that it's gonna be over. There's something about that with the metaverse too. You want, we talked about this identity earlier, like having just one with like NFTs. There's something powerful about the constraint of finiteness or uniqueness, that this moment is singular in history. But I mean, a lot of, you know, as you go through different waves of technology, I think a lot of what is interesting is what becomes in practice infinite or kind of there can be many, many of a thing.
Starting point is 01:55:21 And then what ends up still being constrained. So the metaverse should hopefully allow a very large number or maybe in practice, hopefully close to an infinite amount of expression and worlds, but we'll still only have a finite amount of time. I think living longer, I think, is good. Obviously, all of my are full and thropic work is it's not focused on longevity, but it is focused on trying to achieve what I think is a possible goal in this century, which is to be able to cure prevent or manage all diseases. So I certainly think people kind of getting sick and dying is a bad thing because and I'm, you know, dedicating
Starting point is 01:56:09 and we'll solve my capital towards advancing research in that area to push on that, which I'm gonna do a whole, another one of these podcasts about that. Because that's- So what people should know- Fascinating topic. I mean, this is with the White Priscilla Chan,
Starting point is 01:56:21 you formed the Chan Zuckerberg Initiative, give away 99% or pledge to give away 99% of Facebook, no matter shares. I mean, like you said, we could talk forever about all the exciting things you're working on there, including the sort of moonshot of eradicating disease by the mid-century mark or... I don't actually know if you're going to ever eradicate it,
Starting point is 01:56:45 but I think you can get to a point where you can either cure things that happen, right? So people get diseases, but you can cure them. Prevent is probably closest to eradication or just be able to manage sort of like ongoing things that are not going to ruin your life. And I think that that's possible. I think saying that there's gonna be no disease at all probably is not possible within the next several decades. Basically, it's increased the quality of life.
Starting point is 01:57:13 Yeah. And maybe keep the fine-kniteness because it makes everything taste more delicious. Yeah. Maybe that's just being a romantic 20th century human. Maybe, but I mean, but it was an intentional decision to not focus on our philanthropy on like explicitly on longevity or living forever. Yes.
Starting point is 01:57:37 If at the moment of your death, and by the way, I like that the lights went out when we'd start talking about death, you get to meet God. It does make it a lot more dramatic. It does. I should get closer to the mic. At the moment of your death, you get to meet God and you get to ask one question. What question would you like to ask? Or maybe a whole conversation, I don't know, it's up to you.
Starting point is 01:58:02 It's more dramatic when it's just one question. Well, if it's only one question and I died, I would just wanna know that Priscilla and my family, like if they were gonna be okay, that might depend on the circumstances of my death, but I think that in most circumstances that I can think of, that's probably the main thing that I would care about. Yeah, I think God will hear that question,
Starting point is 01:58:32 be like, all right, fine, you get in. That's the right question, ask. Is it, I don't know. Humility and selfishness. All right, you're right. I mean, but, well, maybe. They're gonna be fine, don't worry. Okay. But I I mean one of the things that I think you struggle with at least
Starting point is 01:58:49 is On the one hand, that's probably the most The thing that's closest to me and maybe the most common human experience, but I Know one of the things that I just struggle with in terms of running this large Enter prize is like, should the thing that I care more about be that responsibility? And I think it's shifted over time. I mean, before I really had a family that was like the only thing I cared about. And I, at this point, it's, I mean, I'm, I mean, I, I care deeply about it, but like, yeah, I think that that's, that's not as obvious of a question.
Starting point is 01:59:36 Yeah, we humans are weird. You get, you get, you get this ability to impact millions of lives and it's definitely something, billions of lives is something you care about, but the weird humans that are closest to us, those are the ones that mean the most. And I suppose that's the dream of the metaversists to connect, form small groups like that, where you can have those intimate relationships.
Starting point is 02:00:01 Let me ask you the big ridiculous. Well, I'm to be able to be close, not just based on who you happen to be next to. I think that's what the internet is already doing is allowing you to spend more of your time not physically proximate. I mean, I always think when you think about the the metaverse people ask this question of the real world, it's like, the virtual world versus the real world. It's like, no, the real world is a question of the real world. It's like, the virtual world versus the real world. And it's like, no, the real world is a combination of the virtual world and the physical world. But I think over time, as we get more technology,
Starting point is 02:00:34 the physical world is becoming less of a percent of the real world. And I think that that opens up a lot of opportunities for people, because you know, you can work in different places, you can stay more close to, stay closer to people who are in different places. That's good. Removing barriers of geography and then barriers of language. That's a beautiful vision.
Starting point is 02:00:55 Big ridiculous question. What do you think is the meaning of life? I think that there are probably a couple of different ways that I would go at this. But I think it gets back to this last question that we talked about about the duality between, you have the people around you who you care the most about, and then there's this bigger thing that maybe you're building. And I think that in my own life, I mean, I sort of think about this tension, but when it's look, I started this whole company
Starting point is 02:01:41 and my life's work is around human connection. So I think it's intellectually probably the thing that I go to first is just that human connection is the meaning. And I mean, I think that it's a thing that our society probably systematically undervalues. I mean, I just remember when I was growing up and in school, it's like, do your homework and then go play with your friends after. And it's like, no, what if playing with your friends is the point? It sounds like an argument your daughter would make. Well, I mean, I don't know.
Starting point is 02:02:23 I just think it's interesting. I don't know. I just think it's interesting. I don't know. I just think it's interesting. I don't know. I don't know. I just think it's interesting. I don't know. I don't know.
Starting point is 02:02:31 I just think it's interesting. I don't know. I don't know. I just think it's interesting. I don't know. I don't know. I just think it's interesting. I don't know. I just think it's interesting.
Starting point is 02:02:39 I don't know. I just think it's interesting. I don't know. I just think it's interesting. I don't know. I just think it's interesting. I don't know. I just think it's interesting. I don't know. I just think it's interesting. I that's one. But here's maybe a different way of counting out this, which is maybe a more religious in nature. I mean, I always like... There's a rabbi who I've studied with who kind of gave me this... We were talking through Genesis and the Bible and the Torah and they're basically walking through, it's like, okay, you go
Starting point is 02:03:07 through the seven days of creation and it's basically, it's like, why does the Bible start there? It's like, it could have started anywhere, in terms of how to live. But basically it starts with talking about how God created people in his her image. But the Bible starts by talking about how God created everything. So I actually think that there's like a compelling argument that I think I've always just found meaningful and inspiring that a lot of the point of what sort of religion has been telling us that we should do is to create and build things. So these things are not necessarily odds. I mean, I think like, I mean, that's and I think probably to some degree,
Starting point is 02:04:06 you'd expect me to say something like this because I've dedicated my life to creating things that help people connect. So I'm in that sort of the fusion of, I'm getting back to what we talked about earlier. I mean, what I studied in school or psychology and computer science, right? So these are the two themes that I care about.
Starting point is 02:04:23 But I don't know for me, that's what I think about. That's what matters. To create and to love, which is the ultimate form of connection. I think this is one hell of an amazing replay experience in the metaverse. So whoever is using our avatars years from now, I hope you had fun. thank you for talking to me. Thank you. Thanks for listening to this conversation with Mark Zuckerberg. To support this podcast please check out our sponsors in the description. And now let me leave you with the end of the poem if by Rajak Kipling. If you can talk with crowds and keep your virtue, or walk with kings, nor lose the common touch, if neither foes nor loving friends can hurt you, if all men count with you, but
Starting point is 02:05:14 none too much. If you can fill the unforgiving minute with sixty seconds worth of distance run. Yours is the earth, and everything that's in it, and which is more, you'll be a man, my son. Thank you for listening, and hope to see you next time. you

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.