SciShow Tangents - Machine Learning

Episode Date: April 30, 2024

We do a lot of learning on Tangents, so naturally, we've got questions about what learning looks like...for machines! AI and machine learning science communicator Jordan Harrod (https://www.youtube.co...m/@jordanharrod) lends her expertise to this episode as we definitely for sure learn all there is to know and will ever need to be known about machine learning, algorithms, and AI. Definitely we got all of it, everyone.SciShow Tangents is on YouTube! Go to www.youtube.com/scishowtangents to check out this episode with the added bonus of seeing our faces! Head to www.patreon.com/SciShowTangents to find out how you can help support SciShow Tangents, and see all the cool perks you’ll get in return, like bonus episodes and a monthly newsletter! A big thank you to Patreon subscribers Garth Riley and Glenn Trewitt for helping to make the show possible!And go to https://store.dftba.com/collections/scishow-tangents to buy some great Tangents merch!Follow us on Twitter @SciShowTangents, where we’ll tweet out topics for upcoming episodes and you can ask the science couch questions! [Truth or Fail Express]Racoon-inspired algorithmhttps://ieeexplore.ieee.org/abstract/document/8552661Dolphin-inspired algorithmPelican-inspired algorithm[Trivia Question]Using machine learning to measure plant vibration and sounds when they experience stress https://www.cell.com/cell/fulltext/S0092-8674(23)00262-3https://www.smithsonianmag.com/smart-news/plants-make-noises-when-stressed-study-finds-180981920/[Fact Off]Lenna (an image of a Playboy model) used as the standard for digital image processinghttps://www.404media.co/lena-test-image-ieee-policy/Tricking speech recognition systems with “neural voice camouflage”https://www.engineering.columbia.edu/news/block-smartphone-microphone-speech-recognition-spyinghttps://web.stanford.edu/~jurafsky/slp3/16.pdfhttps://arxiv.org/pdf/2112.07076.pdfhttps://www.vox.com/2018/7/20/17594074/phone-spying-wiretap-microphone-smartphone-northeastern-dave-choffnes-christo-wilson-kara-swisherhttps://gizmodo.com/these-academics-spent-the-last-year-testing-whether-you-1826961188[Ask the Science Couch]AI image generator mistakes (hands, feet, teeth) and how they work (diffusion models)https://www.sciencefocus.com/future-technology/why-ai-generated-hands-are-the-stuff-of-nightmares-explained-by-a-scientisthttps://www.buzzfeednews.com/article/pranavdixit/ai-generated-art-hands-fingers-messed-uphttps://hai.stanford.edu/news/what-dall-e-reveals-about-human-creativityhttps://proceedings.mlr.press/v37/sohl-dickstein15.pdf[Butt One More Thing]Training machine learning models (convolutional neural networks) to classify poop imageshttps://arxiv.org/pdf/1903.10578.pdfhttps://journals.lww.com/ajg/abstract/2022/07000/a_smartphone_application_using_artificial.24.aspxhttps://pubmed.ncbi.nlm.nih.gov/33275399/Anal Recognition Paperhttps://www.nature.com/articles/s41551-020-0534-9

Transcript
Discussion (0)
Starting point is 00:00:00 Hello and welcome to SciShow Tangents, it's the lightly competitive science knowledge showcase. I'm your host Hank Green and joining me this week as always is science expert and Forbes 30 Under 30 Education Luminary, Sarri Riley! Hello! I feel the most anxious about my epithets when we have an actual expert on the podcast. And our resident everyman who doesn't have to feel anxious at all, Sam Schultz.
Starting point is 00:00:36 Yeah, I still do all the time. And today, we do have a very special guest. It's a PhD candidate researching brain machine interfaces and machine learning for medicine and content creator on YouTube and Nebula who professionally turns her hobbies into work. It's Jordan Herod. I know a thing or two about that. Hello. Hello. I am also probably professionally anxious. So I feel like I'm in a room with friends. I don't think you can be a YouTuber without being anxious, unless you are a bad person. That's the direction it goes.
Starting point is 00:01:09 Interesting, okay. I have a question. I guess I won't spoil it, even though it's in the title, but I have a question. So, when you're using like a chat GPT or something like it, sometimes it apparently helps if you like bribe it or threaten it or explain a certain situation to it that like it's a star ship captain
Starting point is 00:01:33 and has to correctly plot the course to the center of the nebula. So it must do the math questions correctly. If you were a generative AI language model, what would people have to do to you to threaten you to get you to output correct information? They'd have to tell me I get to lay down after I did a good job.
Starting point is 00:01:52 I feel like they would have to tell me that I can't take my ADHD meds the next day if I don't. Oh, yeah. That'll really kick the hyper focus into gear. Yeah, but then if you didn't do it, they'd ruin their model the next day. Oh, absolutely. They take that risk.
Starting point is 00:02:08 I think they would just have to say that they'd be disappointed in me if I didn't do it, which is like the fuel that drives me so much in everything that I do. I think they could say, we'll be totally fine. You'll be disappointed in yourself. Oh my God. You'll be disappointed in yourself. It actually doesn't matter to us one way or the other. I'll be at 4am trying to find your answer to the math question. And then you'll say it and then it'll be like, I forgot I even asked you that.
Starting point is 00:02:37 Thanks, I guess. They just leave like one comment on the YouTube channel that isn't like explicitly negative, but it's like generally questioning the intellect and implying that I am wrong. I will be down that rabbit hole for the rest of the day. Oh, some soft criticism is the worst where it's like, I just don't really think that they that like the thing that they're doing is for me. And I'm like, but I meant for it to be
Starting point is 00:03:01 very good and okay job explaining batteries, but she missed some things. And I'm like looking up courses on electrical engineering. How? How? Re-enrolling in college. Yeah. Why did I do this? Oh, I'm too emotionally stable.
Starting point is 00:03:14 I just want a hot chicken sandwich. That's I would do. I would answer any question for a Nashville hot chicken sandwich. Yes. That's all I want. I would answer any question to be emotionally stable. Yeah. A mood. Sandwich. Yes. That's all I want. I would answer any question to be emotionally stable. Yeah.
Starting point is 00:03:25 A mood. Oh, well, I can't offer you that. But we're we're working on it as a society. That's the goal of this podcast. Someday we'll get there. We're all going to be better people. And then that'll be our last episode. We don't need to impress anybody anymore.
Starting point is 00:03:41 Yeah. And then we just walk through the gate in the woods and cease to exist. That'll also be me with you too, because I assume that the AI situation will just be fixed. Mm-hmm. That sounds exciting to get to the point where the AI situation is fixed. Fingers crossed. I feel like that's a while yet into the future. Every week here on SciShow Tangents, we get together to try to one-up, amaze, and delight each other with science facts while also trying to stay on topic.
Starting point is 00:04:08 Our panelists are playing for glory, but also for Hank bucks, which I'll be awarding to them as we play. And at the end of the episode, one of them will be crowned the winner. But as always, we must introduce this week's topic with the traditional science poem, this week from Jordan. It's hard to know the ways you're blinded by the cause you hope to fix. If one can barely comprehend, the wealth of knowledge stored as bits. Shaping a system line by line and stumbling through a field of minds, predictions come not always true,
Starting point is 00:04:37 but data's costly, we must make do. Released into the light of day, come listen to what it has to say. Hope that it's making the right choices and amplifies the unheard voices." And while this poem isn't friendly to companies that do offend me, it's true that they're not moral actors. There's more than what they manufacture. It's not a person and not your friend, but it probably won't cause the world to end. It can be helpful, don't forget. Careful design can keep your values met. This poem was not made with AI, but with two hours and a glass of wine and I couldn't find a place to say the only winning move is not to play
Starting point is 00:05:09 Oh, I'm gonna lose then I mean, I'm hoping that every AI movie does not Come to fruition in real life because we're we're we're gonna be in a real bad place in that case Yeah to fruition in real life because we're gonna be in a real bad place in that case. Yeah. If anything I've learned from science, reading science fiction and being alive for 43 years is that it always ends up much weirder but also less interesting. That's totally true. We're not gonna get to go anywhere fun or have any adventures. We'll just watch strange things happen from afar. Instead of Mars, we made Facebook. Yep.
Starting point is 00:05:46 Good job, everyone. So this week's topic is machine learning, but before we dive in, we're going to take a short break and then we'll be back to define the topic. All right, Sari, are you going to try and define machine learning in front of Jordan? Because that sounds terrifying. I'll do my best. But Jordan, you can jump in at any time to correct your knowledge. So artificial intelligence is like a broader umbrella. And there is artificial intelligence
Starting point is 00:06:29 that involves machine learning. And there's some that doesn't involve machine learning. And machine learning, in my understanding, as not the science expert of the episode, is when you develop algorithms, usually they're statistics-based or frequently they're statistics-based on a machine, which is where the machine comes develop algorithms, usually they're statistics based or frequently they're statistics based on a machine, just where the machine comes in, like a computer, that can learn from data
Starting point is 00:06:51 and then use that information to process new data on their own without explicit instructions necessarily. So you might be able to code something or code a program that says, if this is true, do X, if not, do Y. If you find a word that starts with A, add one to this counter. If you find a word that starts with B, don't add one to this counter. That is not machine learning. But if you train a machine on a set of words so that it can make predictions about what word goes next in a sentence. That is a form of machine learning. And then there are different ways to teach these algorithms.
Starting point is 00:07:30 So you can do it in a supervised way where you provide a labeled data set. So that is if you provide a bunch of images and you tell a computer that those are all cats, and then it learns. You give it both cats and not cats. Yeah, give it, I guess all cats and then it learns. You give it both cats and not cats. Yeah, give it, I guess cats and not cats. It's like, here's a bunch of cats, here's a bunch of not cats, and these are cats and these aren't cats.
Starting point is 00:07:54 Then you're like, is this a cat? And it's like, yeah, or no. The thing that bugs me about all of this is we have no idea how any of these things actually work. We know how we build them, but we don't know how they're deciding what to do or say. Or we don't... Like nobody knows that.
Starting point is 00:08:08 It doesn't know what a cat is. It just like knows that this is a cat, which are two, they sound like identical sentences, but aren't. Yeah, a lot of it is finding patterns in data. So there's some sort of representation of a cat and there's some sort of representation of not a cat. And that's the thing that it's learning to discriminate between.
Starting point is 00:08:29 But like what that is, we don't necessarily know and how you like translate that into English words is particularly challenging as models get bigger because you can, you know, theoretically explain like you put in, I don't know, the square footage of a house and then the price of the house and you find something to model that and that can be machine learning. But that often creates some sort of formula that you can interpret in some way versus chat GPT where you could open up that box
Starting point is 00:09:00 and it would just be staring into the abyss. I'm gonna ask a stupid question because that's my job. Is that distinct from how people understand stuff? Could we say that we understand how we understand stuff? So I think there are parallels between how we learn and how models learn. I could go into a whole deep dive on like why neural networks are called neural networks and how they're based off neural architectures and the
Starting point is 00:09:32 neuroscience side of my brain will then come in and explain why brains are real complicated and we do not know that much about them and so like information storage parallels there don't necessarily translate that well because it's not really a one in, one out situation. It's a lot of things in and a lot of things out in very complicated ways. So it's not unrelated to how we learn. We do learn faster.
Starting point is 00:10:00 So I guess when people talk about like zero shot models, which is kind of a common term in the ML space, what we're talking about is having a model do something based on input that it has not seen and it being able to kind of infer how to process that information even though it hasn't seen the exact representation. And so it takes a lot to get to a point where a model can do that. That's something that as we've seen like the exponential curve of And so it takes a lot to get to a point where a model can do that. That's something that, as we've seen, like the exponential curve of AI go up over the
Starting point is 00:10:30 last five to six years, has been something that's easier to achieve. But it took like that first three and a half years to get there. And now we're like going all the way up. And babies can do that in like six months or less. So humans pick it up faster. But the species of computer, right? Like the species is learning faster than the human species learned.
Starting point is 00:10:56 Is that true? Like Linux versus Mac or like? No, no, I think what Sam was saying is like, it took several billion years to get from a single celled organism to a human. Correct. Yes.
Starting point is 00:11:08 Thank you, Hank. Yeah, going from the first like vacuum tube transistors to now is quite quick, which is sometimes terrifying. Like, part of me wants to feel very comforted that like, oh, well, this is still not as good as me at things. And I assume that that like last step is much bigger than people think it is. Yeah, but I don't know. I don't know what they're working on next. Like, I wouldn't have thought that if you put a bunch of sentences
Starting point is 00:11:38 into a machine learning model, that you could then have a thing that would convincingly be able to answer the question, why is a vegan riding an electric bike so punchable? Oh, it does know humor and sarcasm these days, which, um, yeah, it's going to steal all of our good jokes. It's going to be the thing that takes us out. I thought it was good that artificial intelligence, like the heart, the thing that would be worst at was jokes because I was raised watching Star Trek. But it turns out that's no problem. Whereas 53 plus four is really tricky.
Starting point is 00:12:10 Why is that? Is that true? Yeah, they're bad at math. What the hell? Because it's like, they're not programmed to do math. They're like language models. They're programmed to predict the next word in the sentence. And like, people don't walk around saying 53 plus four equals 57 all the time. Took me a second. And like people don't walk around saying 53 plus four equals
Starting point is 00:12:26 57 all the time took me a second That would be a much less useful data to have than the data from Star Trek if data was just like a stand-up comedian Yeah, you crashed your spaceship all the time. Yeah, but like bad at it. You only told jokes that other people have told before Yeah, oh my gosh. Well, I'm glad that we've gotten to the bottom of machine learning, everybody. I feel like we've really sort of got it all down. No stone left unturned. Everybody understands it as well as is going to be necessary for the future of being a human on a planet where things are constantly changing extremely quickly.
Starting point is 00:13:01 Yeah, I think we're good. Why am I doing this PhD? We're going to move on to the quiz portion of our show. We're going to be playing a little game. It's called Truth or Fail Express. Here on SciShow Tangents, we often bring up technologies that are inspired by nature. Scientists have figured out all sorts of cool materials
Starting point is 00:13:21 and machines and other things that cheat off of biology so we can figure out how to approach different problems. But you know who also has problems to solve? Animals do. And people working in machine learning have developed all sorts of methods inspired by the way animals solve the various problems in their lives, including ants, killer whales, and gray wolves. So today we're gonna be playing Truth or Fail Express. But with a little twist in honor of our theme, I will present to you an algorithm inspired by an animal behavior that was stitched together entirely from lines written in an academic paper. Or it might be just from chat GPT. So you have to tell me if it's a real paper or chat GPT talking about weird animal-based algorithms.
Starting point is 00:14:07 Like for example, the raccoon optimization algorithm. Do you want me to tell you about it? Please. Yes, please do. Raccoons are known as intelligent and curious creatures. These qualities, combined with their dexterous paws, make raccoons extremely successful in searching for food. Moreover, zoologists believe that raccoons extremely successful in searching for food. Moreover, zoologists
Starting point is 00:14:25 believe that raccoons have an excellent memory. The process of finding the global optimum for a predefined fitness function in the raccoon optimization algorithm is inspired by the natural food-searching habits of raccoons. Food, which represents different possible solutions of the fitness function, is spread throughout the raccoon's living environment. The algorithm makes use of two different search zones in each iteration. In addition, the raccoon can remember best previously visited locations. Then, if it does not manage to find better solutions in future iterations, it can revert
Starting point is 00:15:01 back to the best location. Is this from an academic paper about machine learning or from chat GPT pretending to be an academic paper about machine learning? That to me feels like it was written by somebody who knew what they were talking about. And I just simply don't understand what they were talking about. That is how I felt reading it. I'm going to go academic paper. The moreover was where I started zoning out and it was also my like, this feels like a
Starting point is 00:15:29 chatbot maybe. So I'm gonna go with chat GPT. It is pretty wordy for an academic paper in CS. That's all the writing that that researcher did that year. I guess I'm gonna stick. I'll stick with paper. I just feel like a chatbot maybe would try to spice it up a little bit part way through it.
Starting point is 00:15:49 It would do better? Yeah, it would do better. It would write something wrong but more interesting. Well this text is from a paper published in 2018 titled Raccoon Optimization Algorithm. The researchers took inspiration from the fact that raccoons are known to be very good at solving problems, namely the problem of finding food. And the idea is to help the raccoon explore the area around it, starting by defining a reachable zone where it can look for potential solutions. But there's also a visible zone where raccoons can keep an eye out for other solutions.
Starting point is 00:16:19 So that's a real thing, and I don't know why I we wouldn't because they seem Have you guys ever had the thought if raccoons had gotten sentient first? They probably would have blown us the world up faster than us Yeah, cuz it gives me some comfort to know that we probably weren't the worst thing Like I feel like raccoons probably would have done a worse job than us There's gotta be an anxious raccoon out there who would feel bad There's plenty of mischief makers, but there's always probably a guy saying, what if we don't steal the trash? What if we're just like, are a little bit nice? What if we tidy up a little bit?
Starting point is 00:16:54 Guys, we're going to get in trouble. Yeah. You got to hope that that raccoon becomes the president of the raccoons and then everything's okay. But odds are, odds are not gonna happen. As we'd say, if they're doing worse than us, then not holding my breath on that, I don't think. I like the idea of just like a raccoon president standing on top of the trash being like, I'm the craziest one.
Starting point is 00:17:18 You picked the craziest one. Just punching all the buttons. It's like we do, you know? Yeah. All right. The next algorithm is dolphin-inspired. Dolphins possess exceptional cognitive abilities, including highly developed echolocation skills and complex social behaviors.
Starting point is 00:17:34 These capabilities enable them to navigate challenging environments, communicate effectively, and cooperate with peers to achieve common goals. The dolphin-inspired learning algorithm, DILA, incorporates an echo-based sensing mechanism to capture information from the environment, simulating the echolocation abilities of dolphins. This mechanism enables the algorithm to extract relevant features and navigate complex data landscapes effectively. Inspired by the social behaviors of dolphins, Dilla employs collaborative optimization strategies to facilitate information exchange and collective decision-making among multiple agents. By leveraging the collective intelligence of the algorithmic ensemble, Dilla enhances
Starting point is 00:18:21 robustness, scalability, and adaptability. Chat GPT or real research paper? This is rough. There was like the scandal where a biology research paper got published with figures that were generated. Yeah. Oh my God. It was barely a scandal.
Starting point is 00:18:41 Those figures were amazing. I want to get that Tattooed on me. It was like how big can you we make the mouse penis? Yeah I'm searching mouse penis right now I'm searching mouse penis right now. It's a lot shit That's my metric for AI papers Yeah, if there was a visual cue of a dolphin with a giant penis. I would be like that was AI obviously Unfortunately, we have just the text read to us by Hank to go off of, which is so much less of a visual indicator. Would a visual component help with this that much?
Starting point is 00:19:31 I don't know. I mean, if it looked like this mouse. See, these are the kind of buzzwords that you get in front of a boardroom and you start saying, moreover, you start saying whatever they say. You start saying robustness, scalability, and adaptability. And at this point, all of the people with the suits on are going, yes, and we're writing checks to you, sir. And sir, in this case, is a robot because a robot wrote this so that people would be
Starting point is 00:19:57 excited about it. My strategy is I'm just going to get us to chat GPT every time and hope that one will be right, maybe. It's got to be, I assume. See the problem is that wait, what was it called Dilla? D-I-L-A Dilla. That feels like a bad acronym. I'm gonna go chat GPT I wouldn't I believe that this could be a paper that someone wrote on like archive But a scientist would be having more fun with that acronym
Starting point is 00:20:21 I think you know a real person now it would be some I feel like the thing in CS is finding the Quippy title. So like once attention is all you need was the big transformer paper. Everyone had to riff off of that. So it wouldn't be an acronym. It would just be something like snarky. Well here's the prompt Deboki used to create this paper. It said, make up a machine learning algorithm inspired by dolphins in the style of an academic paper. That's what it hit us with. It said, make up a machine learning algorithm inspired by dolphins in the style of an academic
Starting point is 00:20:45 paper. That's what it hit us with. That was ChatGVG. All right. Last one. Pelicans after identifying the location of the prey, dive to their prey. Then they spread their wings on the surface of the water to force the fish to go to shallow water.
Starting point is 00:21:03 The fundamental inspiration of the proposed pelican optimization algorithm is the strategy and behavior of pelicans during hunting. In the first phase, the pelicans identify the location of their prey. In the second phase, after the pelicans reach the surface of the water, they spread their wings on the surface of the water to move the fish upwards. This strategy leads to more fish in the attacked area to be caught by the pelicans and causes the proposed Pelican Optimization algorithm to converge to better points in the hunting area. Is this from a paper about machine learning or from a chat GPT pretending to be a paper
Starting point is 00:21:41 about machine learning? What are they? What is the algorithm doing? What is the fish? Are the fish in the computer somewhere? The fish are the captains. What are the fish? The problem is that these are such long descriptions.
Starting point is 00:22:00 It's like no one writes this long. It doesn't look as long on the page. But when I'm saying it out loud, it takes forever. Just go forever. I feel like this I can picture what the pelicans are doing. So it's like, that's logical. But you're also a pelican guy. So it makes sense that the book you would be like, it is.
Starting point is 00:22:19 It almost feels like a red herring to have that. And yeah, exactly. But they don't say what the fish are to have that. Yeah, exactly. But they don't say what the fish are. And that's a real problem. Or what the pelicans are. I'd probably get to it later in the paper. You can picture what the pelicans are doing,
Starting point is 00:22:33 but you can't picture what the machine learning algorithm is doing. And so I think it's chat-shitty-titty again. It's pushing the data upwards. Yeah, I was about to say, they they dive and then it pushes the fish up. I'm gonna go paper, you know? Let's... It would not be the weirdest thing that I've read. What I'm remembering sticking point to me is that I don't know...
Starting point is 00:22:57 I don't know that pelicans hunt collaboratively. I feel like I've only seen that... I will tell you that in fact white pelicans do hunt collaboratively. They do. That's clearly that they're talking about the hunting strategy of white pelicans here. But are you telling me this? We're not plunge feeders. Everyone thinks that all pelicans are plunge feeders, but they're not. There's only two plunge feeding pelicans. I think you gave me that piece of information.
Starting point is 00:23:18 So I guess paper. So I'm going to guess Chad GPT. Oh, wow. No, I will just tell you as much about Pelicans as I possibly can. OK, that is a fact you should know about me. This was a real paper. Oh, man. Wow. Jordan, you are in the field. You are nailing your peers work. It's a it's a 2022 paper.
Starting point is 00:23:39 It was titled Pelican Optimization Algorithm, a novel nature inspired algorithm for engineering applications. The algorithm is balancing the way that pelicans explore for food, which is, but I don't know. I don't know what it's doing. No one explains to me in the paper, in the description what the heck it's actually doing
Starting point is 00:23:57 and what the fish are. Yeah, I definitely still don't get what the fish are. The fish are solutions is what Deboki has written. In this case, the fish, it's sweeping solutions into an area. It's like a political cartoon with just way too many labels. The fish are solutions. Yeah, Raccoon and Pelican both sounded like versions of reinforcement learning slash control problems,
Starting point is 00:24:26 which made sense as papers. And then dolphin echolocation came in and that one just like threw me so much that I was like, I don't know. It did seem like it was both like suddenly halfway through the introduction it was like, and also it's about dolphin social structures. Like it's not just echolocation,
Starting point is 00:24:41 it's a completely separate dolphin thing. That'll be the 2025 paper. We will create consciousness by mixing echolocation so that the thing knows about physical space with social structure, which is... And then they'll all sing, So long and thanks for all the fish. And they'll take off into space and be like,
Starting point is 00:25:04 Look, you guys can't do AI anymore. We do not trust you. The moment you start to code again, we got a laser up here ready to shoot you. All right, everybody. Jordan's in the lead with three. Sam's got two, Sarah's got one. Next up, we're gonna take a short break
Starting point is 00:25:19 and then it'll be time for the Fact Off. [♪ Music playing. Welcome back everybody. Now get ready for the Fact Off. Our panelists have brought to me science facts to present to me in an attempt to blow my mind. And after they have presented their facts, I will judge them and award Hank Bucks anyway I see fit. To decide who goes first, though, I have a trivia question. Research has shown that plants vibrate when experiencing drought. And scientists, that's real
Starting point is 00:26:14 Okay, let's keep reading to find out what's going on scientists look to see what kind of sounds that might produce recently researchers set up microphones about four inches away from tomato and tobacco plants and recorded the plants after cutting them or Not giving them water and they found that the stressed out plants made more sound compared to the non-stressed plants. While the sounds were at a frequency too high for humans to hear, the scientists did design a machine learning model to distinguish between the sounds produced by the cut plants and the non-water plants. How accurate was the model at differentiating between the two stressful conditions in a percentage?
Starting point is 00:26:47 I'm still not over the fact that plants vibrate. Don't they don't look like they vibrate to me! I feel like I wouldn't notice that, but I guess it's a little vibration. I would vibrate if I'm stressed because I'm thirsty. Oh yeah, I totally vibrate when I'm stressed. Or cut. How accurate was the model? I feel like the most accurate models are in this 70%, 80%,
Starting point is 00:27:15 maybe, that I've seen. But I also am not in this field. So that would be like high accuracy, in my opinion. Maybe like 62% of the time it was able to differentiate dry from hurt. I'm going to go 75. The answer was 70% so Jordan gets to go first. 70% is also, it's somehow both lower and higher than I'd expect. I don't know. So my fact is when digital image processing research was getting off the ground in the
Starting point is 00:27:53 70s, the standard test image, so the image that was used when people published papers, when they went to conferences in order to standardize everyone's results, was called Lena. And it came from a Playboy magazine centerfold that one of the researchers just happened to have on his desk at work. That research is foundational to pretty much all image and video-related AI results that we see now,
Starting point is 00:28:14 including image generation, video generation, deepfakes, et cetera, and then a not so shocking turn of events. The model received exactly zero compensation other than what she was paid for the Playboy shoot, nor did she consent to have her image used in that way. Although she would go to conferences every few years and be like, yay, I'm happy that you guys are finding this useful. And it wasn't until I think the early 2000s that she was like, okay, can we stop doing this? That's not enough.
Starting point is 00:28:41 I think it was only like last week actually that IEEE, one of the big conferences in the field was like, okay, you can't use this photo anymore. It's slowly been phased out of the ML community as something you are allowed to submit with. I'm looking at the photo now. I do recognize it. I feel like I have seen this picture before. It is, maybe important to say, in an audio medium, just from the shoulder up. So they didn't, they did not do this with an actual full-body naked woman. So there's at least
Starting point is 00:29:12 that, but they did do it with a woman who was naked at the time of the photograph being taken. Correct. That's wild. And so this is like with an image, a standard image they would use and try to reproduce with machine learning and that was like the... So not reproduce. So when people were trying to create
Starting point is 00:29:30 and test out different methods of processing images, so like filtering images, transforming images, it was easier to have one photo that everyone could use to compare results and methods on. Gotcha. And so this was the one photo. So if you're like all those early Photoshop things, you know, increasing the edges and the making it fuzzier
Starting point is 00:29:53 or changing the hues and saturations. Yeah, gotcha. I guess there's probably not any way have any compensation actually occur at this point. I would imagine not like academia doesn't pay people in academia particularly well, let alone everyone else. Like a bunch of money. So I don't know where that money would come from.
Starting point is 00:30:12 Yeah. Adobe. I don't know. Yeah. I was about to say, if Insert X major AI company would like to compensate her for her contribution to the field, then they should be out. Might be just a good press play. Yeah.
Starting point is 00:30:24 Like, it's just like, hey, you want some good PR today? Here's this lady whose image was... or her contribution to the field, then they should be out. Just a good press play. Yeah. Like, hey, want some good PR today? Here's this lady whose image was, but also probably the Playboy company owns the picture, actually. Playboy would go, uh-uh, now wait a minute. Just give people money. This is how I feel. People who do weird stuff, if you're
Starting point is 00:30:40 a company with billions of dollars, just make their day. Just mail some checks out, guys. Make somebody's day. Why not do that? I support this, especially if the check goes to me, but also everyone else. Are there other images that are used now as standard image, like that replaced this, or is it at a point in the field where you don't need this sort of like anchor digital image processing
Starting point is 00:31:06 reference. As the field has evolved and as we've gotten into like image generation, deepfakes and things like that, there's been a little bit less of a need for standard images because like you don't need a standard deepfake image to work with. I think what's happened more so is that we have ImageNet and the really big image datasets and that's the standard thing that everyone uses to build their models and then you can test performance as a way of having the same baseline. I guess I knew that that would be the case.
Starting point is 00:31:37 And also in the same way, all of those images are scraped from the internet without the permission of the people who made the birth in them. So you know, 50 years later. We're still at it. The proud tradition lives on. All right, wild, interesting. Sarah, what do you got? One of the ways that machine learning is being applied
Starting point is 00:31:55 right now is called automatic speech recognition, which is basically one of the driving technologies behind any voice assistants, like Google's assistant or Alexa or my mortal nemesis Siri. And my very basic understanding of what these algorithms do is they take the sound wave form of your speech and then chop it up into very small pieces. And then the algorithms at work try to match
Starting point is 00:32:18 those tiny pieces to sounds that they were trained on. And by putting those sounds back together and incorporating other information from training data about different words and how words become sentences, that makes sense. The computer does its best to try and guess out what you're saying. And this is a hard thing because people mumble
Starting point is 00:32:36 or have different accents or misspeak or whatnot. So the same sequence of phonemes doesn't necessarily translate to the same sentence. But as these voice technologies are in more places, people are worried about AI listening to them or surveilling them. And I feel like this is a huge paranoia that I see pop up every once in a while, is that if you talk about with your friends how much you want chocolate milk, then you'll start getting milk ads across your devices or something like that.
Starting point is 00:33:04 And as far as I can tell, a lot of those targeted ads are more about what you've been searching, like if you're searching straws for chocolate milk, or if you're sharing a Wi-Fi network with someone who's searching those things, or if you were physically at a chocolate cow dairy farm or something like that, that can provide that location data rather than listening. So data privacy is, it's kind of tangential. Data privacy is a concern, but your phone probably isn't recording and transmitting everything you are saying all the time.
Starting point is 00:33:31 A rabbit hole I went down as I was trying to find this fact. But the fact is, even still, as an experiment in audio counter surveillance, there was a paper from Columbia University researchers that was presented at a conference in 2022 that showcases a machine learning model that tries to disrupt speech recognition algorithms in real time as someone is talking. So it basically uses similar principles to take your voice as an input. It does a series of calculations on it. And then instead of interpreting the speech, it instructs computer speakers to emit specific noise. The press release says it's whisper quiet, but there weren't audio
Starting point is 00:34:09 samples so I couldn't hear it. And the noises specifically exploit the weaknesses of existing speech recognition algorithms. So they change the waveforms in the room just enough either by adding in interference or adding like blanket white noise or other things like that just enough that those chopped up sounds don't get interpreted as sensible phonemes or words or sentences anymore even though they're fine by human ears the computer can't interpret them. And they tested this model across various speech recognition programs and databases I think kind of like Jordan was saying,
Starting point is 00:34:45 my understanding is that these are open source and like regular databases that people train their models on like Deep Speech or Libra Speech or Wave 2 VEC 2. And if I'm interpreting the figure in the paper correctly, this counter surveillance thing blocked anywhere from 28% to 87% of recognizable speech depending on the system. So it's not perfect.
Starting point is 00:35:06 It's more for a hypothetical spine situation than a real one. But I think it's interesting to talk about the flaws in all these technologies as much as the breakthroughs. What if like when I'm doing my spy craft, I just talk, I'm just like, hello, it's Hank. I think the computer would just put a big old question mark. Yeah. That's how it works. I think whoever's listening would also put a big old question mark.
Starting point is 00:35:32 So if you just like had this on in your house, all it would like at the beginning of your fact, basically you said your phone's not actually listening to you in order to advertise to you. So to have it on your house, all you'd be doing is making it so you couldn't say, Hey, Siri, and it wouldn't wake up. Right? Yeah, I think effectively, yes. But it was weird because even though there's this body of research that's
Starting point is 00:35:55 saying that your phone isn't spying on you at the beginning of this paper, they were like in situations where you might be monitored by technologies around you. And so I don't know if this is, I fell down that rabbit hole when researching whether that was like a legit thing. I would also like for YouTube to add that to the background of every YouTube video in which someone says, hey, whatever. So it stops setting off my various devices at home.
Starting point is 00:36:20 All right. I just love the weird esoteric history bits, like things that are like so weirdly important and will inevitably be forgotten, but are also just shine a light on the bizarre culture that that results in many of the in all of the structures around us. So I must announce Jordan as the winner of the episode. She was also way ahead, very powerful lead. It's funny, cause there was an article about this that literally came out last week. And so for the last week, I've been most days going through your tweet likes and your replies
Starting point is 00:37:02 to make sure that you didn't see it. Oh no, no, what did I like? I like some real stinkers I bet. Hank is the one being surveilled all along. I was just reading for discussion of that article. Okay. I cannot comment on anything else that I saw. And now it's time for Ask the Science Couch, where we ask a listener question to our couch of finely honed scientific minds. At Ponyoty on YouTube asked, why does it get human extremities so very wrong? By that I assume it means AI that draws pictures. Can I guess? Because I just have a guess. Which is that it's looking at what object is coming next, and the next thing that comes after a finger is usually a finger.
Starting point is 00:37:53 And so it's just like, I guess, because look at all my fingers, there's a bunch of them. And so to know that there's four is a very different thing than being like, what comes after a finger? A finger. What comes after that finger? A finger. And then maybe on the fifth or sixth finger it draws, it's like, oh, another finger doesn't come after that one. But after the third or fourth, it's like, I don't know.
Starting point is 00:38:17 I don't know where the other finger comes next. It's like, that's too many. Can it ever take a step back and look at what it did and be like, Oh, I did two, my fingers or when it does it's six finger as it is it. Yeah. Does it not ever look at it and say, Oh gosh, too many fingers. It's starting to draw the finger and it's like, Oh shit, I'm drawing an extra finger. I got to put the delay and I guess, yeah.
Starting point is 00:38:38 How you feel as an artist, Sam, when you're drawing a guy, no, I can erase stuff. I'm a human man. That's the power I have over AI. And that's how we're going to win against the AI. Some models can do that. It's ultimately down to how the model is built. So there's also I think it was deep mind made language model where it had like inherent fact checking systems where it would like create a response and then it would be like, let me go like pull a bunch of information from repeatable resources and see if what I say matches what these things say. And then if it doesn't, let me go back and adjust what I said. So there are mechanisms to do that when you're
Starting point is 00:39:22 using like, I don't know, Dolly or like Mid Journey or whatever. Most of those don't really run on that. So that's why most of what you see is like weird number of fingers or like there's, you know, three fingers, but then one's coming out like this way. Confidently wrong. Perpendicular to the rest of your fingers. Which I guess depending on what data sets they use. I was born with, I guess, five fingers and a thumb, so six on each hand, because one was coming out of each of my pinkies.
Starting point is 00:39:53 So I would imagine that there isn't like a ton of that kind of data in this model, but like maybe. I did make the mistake of typing in weird AI hands into Google and it is it is nightmare fuel. These fingers have fingers. I don't like that. In all likelihood, it's it's an edge case issue in that there are lots and lots and lots of photos of people's faces and people's bodies in a bunch of different positions and whatnot.
Starting point is 00:40:20 But the the number of different configurations of like hand positions and you know, you may go a lot away obviously have five fingers, but like one might be behind like out of frame or something like that makes it more challenging in the same way that like ears used to be the issue and hair used to be the issue representing that if you don't have enough data, it ends up being really, really wonky in the same way that language models hallucinate things when they don't have good representations of them. The one thing that I do have to add, I guess, is what... And Jordan, you can correct me if I'm wrong because I'm coming in a little bit new here. I think the model that these sort of image generators
Starting point is 00:41:06 are trained on are diffusion models, which were introduced from what I found in 2015 in a paper called Deep Unsupervised Learning Using Non-Equilibrium Thermodynamics, which is a fancy way of saying that they took an idea from physics, so this idea of equilibrium, and that if you have, I don't know, I feel like a pretty classic chemistry problem where if you have two containers filled with particles and then eventually they will like drift over and form equilibrium.
Starting point is 00:41:33 There are a lot of systems that don't form equilibriums, but things drift back and forth. And so you train these models by giving them images and then having them destroy the data in it, which is like an iterative diffusion process or what they call it. So they turn a picture of a horse into a bunch of uncomprehensible pixels. And then the model learns through destroying it how to predict how to create an image, like how to reverse that process and go backwards and generate something from nothing, which is how you get these predictive of, okay, there's a finger, is there another finger? What is the shape of a hand? What is the shape of a letter?
Starting point is 00:42:18 And so that's where the guessing is. That's where the prediction and statistical modeling of these image generations come from and why you can't necessarily, unless there is that fact-checking layer, say this AI knows what a hand is or knows what the letter A looks like or knows what a pipe looks like because it's all just guessing what is the statistically most likely pixel to appear next to this pixel in a certain class of images. And I mean, that's also the issue that you run into with language models doing math.
Starting point is 00:42:52 It's not about the math. It's about what is statistically most likely to come after this sequence of characters. And we as humans are really good at noticing things about other humans. We look at ourselves and each other all the time. And so it's easy to see like that hand is weird. That looks like an alien thing.
Starting point is 00:43:10 Whereas there could be a wrong number of leaves on a tree in the background of a picture and you like wouldn't notice necessarily. Like there's like any botanist probably looks at any AI generated picture of any tree and it's like, oh my God, what were they thinking? Yeah, that flower has an incorrect number of petals. That's biologically impossible.
Starting point is 00:43:31 But because we're all humans, it's really easy to be like, oh, that foot messed up. I've seen a foot and it doesn't go at a right angle necessarily in many cases. If you want to ask the Science Couch, follow us on Twitter at SciShowTangents or check out our YouTube community tab where we will send out topics for upcoming episodes every week or you can join the SciShow Tangents Patreon and ask us on our Discord. Thank you to the Space A on Discord, at Ging-Ging-Ging-Ging on Twitter and everybody else who asked us
Starting point is 00:44:02 your questions for this episode. I did my best. That's what it says. That's right. Jordan, thank you so much for coming on the Sideshow Tangents and sharing your knowledge and understanding with us. I wish we could go deeper. There's so much to know and so much to be excited about and scared by. If I want to see more of what you're up to, where would I go? You can find me on YouTube. You can find me on Nebula. You can Google my name to see what else I'm doing because I'm also working on my PhD and that is taking up 80% of my brain space right now.
Starting point is 00:44:34 Imagine that. And your name is Jordan Herrod. Herrod has two R's. Yes, like the store, which is a reference that only really works when I'm in London. But yeah, I'm like, is it the store? Should I know what the store everybody head on over and subscribe to Jordan on YouTube. And thanks for being here for us. If you like this show and you want to help us out super easy to do that. First, you can go to patreon.com slash size
Starting point is 00:44:58 show tangents to become a patron get access to our discord shout out to patron less acre for their support. There's also lots of bonus episodes and weird stuff, minions, commentaries. Second, you can leave us a review wherever you listen. That helps us know what you like about the show and also other people see them and think, I maybe will watch this show. Finally, if you want to show your love for SciShow Tangents, just tell people about us. Thank you for joining us. I've been Hank Green. I've been Sari Reilly. I've been Sam Schultz. I've been Jordan Harret. Sideshow Tangents is created by all of us and produced by Jess Stempert. Our associate producer is Eve Schmidt. Our editor is Seth Glicksman. Our social media organizer is Julia Buzz-Bazio.
Starting point is 00:45:36 Our editorial assistant is Dibbupetra Krabarti. Our sound design is by Joseph Tuna-Medish. Our executive producers are Nicole Sweeney and me, Hank Green. And of course, we couldn't make any of this without our patrons on Patreon. Thank you, and remember, the mind is not a vessel to be filled, but a fire to be lighted. But one more thing. We talked about some machine learning algorithms are used to sort and analyze different images like dogs, stoplights, or even pictures of human poop. Gastroenterologists already have ways to classify their patients' poop, like the Bristol stool scale or the Brussels infant and toddler stool scale,
Starting point is 00:46:31 which look at things like stool consistency or how fragmented the chunks are. So the goal of these machine learning models, which may eventually become smartphone apps, is to give doctors an extra tool and help patients self-report their gut health more consistently. There was also a 2020 Stanford paper that looked at a similar system embedded in a bidet and it used anal fingerprint recognition to tell which person was sitting on the toilet
Starting point is 00:46:56 to associate that person's data with their selves. That was not in the script. That was just Jordan. It wasn't. I just knew that. Bonus buts. I just knew that. Bonus buts. I did a video about it a while ago.
Starting point is 00:47:09 A wild race. Are people's anuses that unique that you can tell? Apparently. Every one is beautiful in their own way. Specifically in reference to. But who knew that? Who was like, I wonder if the butthole looks different enough. I feel like it was in like the introduction of the paper, because I remember talking about
Starting point is 00:47:30 it in the video being like, who found out? Like who? Yeah. Where's this data set? I have a question.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.