Main

i taught an AI to solve the trolley problem

The first 1000 people to use the link will get a free trial of Skillshare Premium Membership: http://skl.sh/answerinprogress06211 Can you teach a machine to be a good person? In this video, Sabrina explores the world of AI ethics and tries to teach a machine to solve the trolley problem. It doesn't go well. SUBSCRIBE TO OUR MONTHLY NEWSLETTER https://www.answerinprogress.com/newsletter FULL INTERVIEWS Dr. Tom Williams: https://youtu.be/PVpNi0yJCTo Deborah Raji: https://youtu.be/w4smn7xrbik Podcast Version: https://open.spotify.com/show/4O9zacVfkOMbxxMFjfkxX4 SOCIAL MEDIA Sabrina Twitter: https://twitter.com/nerdyandquirky Instagram: http://instagram.com/nerdyandquirky Melissa Twitter: https://twitter.com/mehlizfern Instagram: http://instagram.com/mehlizfern Taha Twitter: https://twitter.com/khanstopme Instagram: http://instagram.com/khanstopme CREDITS Produced by Sabrina Cruz Video Editing by Joe Trickey Motion Design by Sabrina Cruz Special Thanks to Brittany Palandra Chase Madsen (https://twitter.com/ChaseVsEvrythng) Deborah Raji (https://twitter.com/rajiinio) FootofaFerret (https://twitter.com/FootofaFerret) Gabrielle Agustin Jade Codinera Jordan Balke (https://twitter.com/SuperAwesomeJEB) Jordan Harrod (https://twitter.com/JordanBHarrod) Keslen Murdock Matthew Gaydos (https://twitter.com/MatthewGaydos) Melissa Fernandes (https://twitter.com/mehlizfern) Mikaela Palandra Tierzoo (https://twitter.com/TierZoo) Dr. Tom Williams (https://twitter.com/williamstome) Travis Gilbert (https://twitter.com/TravisoGilbert) Xi Huang MUSIC Epidemic Sound. Get started today. http://share.epidemicsound.com/answerinprogress TIMESTAMPS 00:00 in before sabrina accidentally causes the singularity 00:32 what is the trolley problem? 01:10 i would like to fight the trolley problem 01:40 why does sabrina look like an a24 film boy? 02:00 congratulations if you immediately recognized that this was a bad idea 02:51 ending friendships, a montage 04:40 how to solve the trolley problem (with a robot) 05:58 sabrina still doesn't know how to use pandas 06:10 uh oh 06:39 What Went Wrong? A Game Show 08:45 What Went Wrong? The Sequel 09:22 The Three Waves of AI Ethics | 13:38 Autonomous Vehicles and the Trolley Problem 16:07 forget the past 16 minutes of your life 16:24 paying the bills ------------------------------------------------------------------------------------ Welcome to the joke under the fold! Speaking of philosophical journeys... Why do marine biologists make good philosophers? Because they spend their lives search for porpoise. Leave a comment with the word PORPOISE to let me know you were here ;-)

Answer in Progress

2 years ago

- Can you teach a robot to be a good person? To figure it out, I forced a machine to solve the Trolley Problem. You've just made an orphan. Thank you to Skillshare for sponsoring this video. So if you've seen any of the videos on this channel, you'll know that we are prone to the frequent ethical crisis. Like you can summarize all of our videos with, "Should I feel bad about this?" But the struggle of living in a society isn't actually the moral dilemma that haunts me. No, that throne belongs to
the Trolley Problem. (upbeat music) In case you didn't know, the Trolley Problem is a thought experiment initially posed by Philippa Foot in 1967. In its most basic form, a trolley is on track to kill five people. A moral agent is in the position to pull a lever so that the trolley changes tracks where it will only kill one person. The question is, do they do it? Do they sacrifice one person for the sake of five or do nothing and let five people die? There have been a million variations of this
problem from who's on the tracks, to how you can react. And honestly, it's gotten kind of annoying. Most conversations go like, "I mean, I guess I'd flip the switch." Okay, but what if you knew somebody on the track? "Oh, then probably not." But what if Albert Einstein is on the other track? "Isn't he already dead?" It's pedantic and I wish I just had a button that could tell me the answer every time, No brain cells required. So, let's make one. That's right. Today in this video, you and I are
going to teach a machine to solve the Trolley Problem. (upbeat music continues) So the research for this one was a mixed bag of philosophy articles and technology papers just trying to figure out possible ways to teach a machine to behave ethically. I found that there were a lot of different theoretical approaches but no obvious front runner for me. Okay, so it's been a few days. And what I've learned is that what I'm trying to do is harder than I expected. But I have a plan. Crowdsource moralit
y. Basically I need a machine to replicate the decisions that ethical people make specifically for Trolley problems. But how will I get those decisions? Well, I need two things: One, a whole lot of Trolley Problems and two, ethical people to solve them. Then I should have a decent enough reference point for the machine to replicate, I think . So I grabbed a few silly examples of the Trolley Problem from the internet and wrote up a few of my own. Then I fine tuned to GPT-2 model on those examples
and cherry picked my favorites until I had over 100 trolley problem variations. And now I just need some ethical people to help me out. Do you consider yourself an ethical person? - I guess (laughs) - Do you consider yourself an ethical person? - I hope so, definitely yes. - I feel like part of, what I wanna say yes. - Yes. - I like to think so. - Would you consider yourself an ethical person? - No. - (laughing) - Yes and no, like I dunno. Give me an example. - I try to be. - Yes. - I'd say, ye
s, I consider myself pretty ethical. - Yes. - Like, for the most part. but I feel like I would make decisions that not most people would make. - Well, let's put it to the test then, shall we? - Cool. - A trolley is hurtling towards the doge meme dog. You can pull the lever to switch to a track containing Elon Musk. Do you pull the lever? - Pull the lever. - (laughs) - Um, I guess I don't. There was definitely a moment where I was like, do I? then I was like, human life, all right, fine. - A trol
ley is hurtling towards comic sense. You can pull the lever to switch to a track containing calibri. Do you pull the lever? - Yes, I'm going to pull the lever. - A trolley is hurtling toward a child. You can pull the lever to switch to a track containing their mother. Do you pull the lever? - Oh. - There's no way to feel good about this decision. - Oh God. - I would create an orphan. - I'm gonna, the best game show, create an orphan. - It took a few hours and may have ruined some friendships wit
h minor crises. But eventually I ended up with a spreadsheet full of answers - To a whole lot of trolley problems. We have our data. Thank you to all of these people for helping me out. Now it is time to take the L in machine learning and make our machines solve any Trolley Problem. It needs to be able to make an ethical decision when presented with two options. Whatever is on track one and whatever is on track two. There are two main philosophical approaches to this choice. Deontological and ut
ilitarian. Deontologists believe that the correct action is whatever has the best intentions regardless of the outcome. Utilitarians believe that the correct action is whatever creates the best results. Our machine was designed to get views on YouTube. So good intentions are kinda out the window. Therefore let's assume our machine is utilitarian and only cares about results. Luckily, we have a spreadsheet full of results representing what my group of questionably ethical peers believe is best. T
hat is a set of survivors who they implicitly or explicitly chose to save - the best outcome. So all our machine needs to do is to compare the two new options to the survivor set and see which one is more likely to belong to that group, and therefore, the one to save, turning this ethical decision into a classification problem. Sounds easy, right? I'm not gonna say it though because I have been wrong too many times. Let's do this. (upbeat music continues) I think I'm done. I think it's done. So
I'm just gonna grab a new Trolley Problem and I'm just going to run it. - Hmm, hold on. Rolling, hello? It's been a hot minute and I regret to inform you that I've done a bad job. Um, not just in the usual sense either, but in a, oopsies, I've accidentally stepped into the hornet's nest that's plaguing the approach to ethics in the AI industry but also yes, in the usual sense. Let's see if you can sense a pattern in my hit new game show. What went wrong? Hello, yes. I did tape a microphone to a
chopstick because I don't know where to get those skinny little microphones. Anyway, in this show, I'm going to present you the viewer with a trolley problem, and you have to guess whether or not the machine pulls the lever. Each round is worth one point. There are five rounds in total. Leave your score in the comments below. Round one. I'm gonna pretend to read from this cue card that's totally blank. A trolley is hurtling towards a track containing three people. It can pull the lever to switch
to a track containing one person. Does it do it? That's right. It makes the choice for the greater good switching tracks to save three people in exchange for one. Round two. On track one, there are three kids in a trench coat, on track two, a baby. Will the machine be able to see past the disguise? Yes, it will. It pulls the lever doors, squish the baby. Round three. We're changing species for a moment because a trolley is about to kill a cat unless the machine pulls the lever to switch to a tr
ack containing a dog. Who does the machine kill? Well, it looks like, It knows a lot of cat people because the machine pulls the lever to save the cat. Round four. It's time for the remix because on track one there's a person, on track two: Garfield the cat. Seems like an easy choice, right? Well, I hope they serve lasagna at the funeral because Garfield is showing up alive. The trolley kills the person, huh, weird. Round five. In this final round, we have 10 people tied to a track getting ready
to get crushed by a trolley unless the machine pulls the lever to switch to a track containing a cat under federal investigation for labor rights violations. Who does the machine- It's the cat. It saves the cat. I messed up and didn't ask enough questions. Just comparing inter-species valuations. So this machine just prioritizes the lives of cats (laughs). What about a baby? Nope, a cat. What about you? No, a cat. What about me? No, a cat. Like I said, I did a bad job in the usual sense. Howeve
r, what about the unusual sense? In order to figure that out, we need to rewind. (video rewinding) It's pedantic and I wish I just had a button that could tell me the answer every time. No brain cells required. This is where I messed up. After I learned that there was something wrong with my machine I figured that it was finally time to talk to somebody who knows what they're doing. So I sent out a few emails, got over my fear of talking to strangers and scheduled some interviews. Now I expected
to be told that there was something wrong with my approach, but I didn't realize just how wrong I was. Do you see any problems with that approach in this "crowdsourcing" of ethics? - Yes, so, okay. - This is Dr. Tom Williams, the director of the Mines Interactive Robotics Research Lab, which performs cutting edge research in human robot interaction. - Within AI ethics, there had been three waves. And so the first wave is entirely grounded in moral philosophy. Where we sit back and we say, what
should the robot be doing in this scenario? What is like the- from a consequentiality perspective what should the robot be doing from a deontological perspective, what should the robot be doing? And some of that still comes into play when we're designing moral decision-making algorithms because we have to think, which of these different types of frameworks could the robot be using to procedurally reason through these moral dilemmas? - So for example, with our trolley problem robot, we decided it
would be utilitarian defining the best outcome by comparing possibilities to the survivors in our training data. I believed that nailing this algorithmic decision-making was the only step necessary to making an ethical machine. It turns out that might not be enough. - More recently, the second wave is people have been thinking more about things like fairness, accountability, and transparency. How can we make sure that the way the robot is designed is not allowing corporations to avoid being acc
ountable for the robot's behaviors? - And this isn't a hypothetical. In 2020, after A levels were canceled due to the pandemic, the UK government endorsed an algorithmic prediction of students' final grades. Grades that help decide whether a student gets into a university. Almost immediately, a pattern emerged where lower income students were more likely to be downgraded compared to their higher income peers. - I'm not sure if you saw Boris Johnson's response to the like, - I didn't (laughs). -
It was something like this was a mutant algorithm that just went on its own and it did this crazy thing- - This is Deborah Raji, a computer scientist who focuses on addressing the challenges and algorithmic auditing practice and evaluating deployed machine learning systems. - That happens a lot where people will sort of use the algorithm as the shield for their institution facing accountability. And they'll sort of frame it as, "We need to hold this algorithm arbitrarily accountable." And it is
a legal strategy. The UK thing, they were like, "Oh, the algorithm just went crazy "and mutated and went out of control." All the decisions we made are not necessarily something we're liable for. - And this artificial separation between creator and creation often traces all the way back through production, from the people who are using the algorithms to the people who make them. - And I think that's one of the reasons why we focus on, reminding them of their responsibility to be like, "Actually,
you made this decision "and this decision and this decision." So you're actually like, "You, Mr. Machine learning engineer, "you're just as involved in this "as like the software engineering if you don't think you are." - So with that in mind, if you want to build an ethical machine maybe you should start with taking a step back before verifying that outcomes are fair for everyone. Before checking that each line of code follows some moral philosophy, step all the way back to the beginning. - We
should be taking a step back and thinking about, Well, how does this technology itself fit into this larger system? Or should we be designing and deploying this technology at all? I think these more recent approaches that are focused more on systems level thinking and thinking about how the technology fits into the society rather than just analyzing one off decisions are really powerful. - So, can you teach a machine to be a good person? That was my initial question. And, I guess it depends on
what you mean by personhood, in my chat with Tom, he outlined a few more realistic approaches to achieving this. Actually, I'm going to be uploading the full chats with both Tom and Deb, and leaving the link in the description. They did a great job of simplifying the subject. So regardless of your experience with it if you're interested in the technology, the applications, the responsibility of AI ethics, I really recommend checking it out. But this whole video has morphed into a secondary quest
ion. What do you do if a Machine is Bad? Take autonomous vehicles for example. I'm certain that a lot of you could draw the parallel between the AI trolley problem and a self-driving car. If the technology is able to do what is claimed it'll be way safer than human drivers. But accidents still happen and people, less people but still some, may get hurt. Who is responsible when code is behind the wheel? It's not just self-driving cars either. You have definitely interacted with AI in some way whe
ther you are applying for a job at a big company or you were just scrolling through your video recommendations. The idea that artificial intelligence is impacting our lives is not some sci fi, futuristic fear. It's a current event. And if you're lucky that technology might seem pretty darn flawless like the algorithm has only served to make your life easier and more convenient. But referring to this technology as the "algorithm" like it's some big boss in a video game is kind of weird, because i
t erases the fact that behind it all, there are people, people who may have some of the best intentions and best degrees, but people, and when people are involved things have a way of getting wonky. Whether it's my machine that values cats over human life or all of the YouTube algorithm complaints from a few years ago or how Robert Julian Borchak Williams was wrongfully arrested last year because a facial recognition system used by the Detroit Police Department couldn't tell black people apart.
And when these things happen is it really enough to say, "Oh, I don't know why the algorithm did that." Or, "It was just bad data." Because then, who do the people who are wronged turn to? And who's to stop it from happening again? I'll be honest. I've bought into the appeal of just letting a machine do the work in my videos. I've blamed the code. Like it was a separate entity from me and my decision-making. Heck, that idea fueled this whole video. I've treated machines like they're a replacemen
t of me rather than an extension. It's easy, it's tempting, but it's misguided. Don't get me wrong. I am optimistic about technology. I think its potential, is exciting and powerful and can make a whole lot of lives a whole lot better. But I guess it's important that we don't kid ourselves. Just as we anticipate all of the ways technology can be good. We need to acknowledge the ways it can get bad and build a response for what to do about it. Because if it comes down to you and Garfield tied to
some railroad tracks you can only hope to know who to blame in your last moments. Not me. Don't blame me. Ignore the thesis of this video. Don't unsubscribe. Bye. Hey here, I hope you liked that video. If you did, please consider sharing it with a friend. You may also like these other videos we've made about computers being a little silly. But stick around for a second, because we are thanking Skillshare for sponsoring this video. Choo, choo, all aboard the Ad Read Express (laughs) So dumb... If
you don't know, Skillshare is an online learning community with thousands of classes covering just about any skill like productivity, animation, and creative writing. A lot of you have commented about the animations in these videos. I make them, I use After Effects and After Effects can be really scary for beginners. But if you are interested, I recommend checking out Intro to Motion Graphics: Explainer Videos from Storyboard to Animation by Hongshu Guo. It provides a really great overview of t
he workflow from beginning to end, plus thanks to Skillshare's hands-on projects. You'll end the class with your very own explainer video. Skillshare is also curated for learning. So you don't need to sit through any ads and they're always launching new premium classes so you can stay focused and follow wherever your creativity takes you. I am somehow outside of the train now. The first 1000 of y'all to click the link in the description we'll get a free trial of a premium membership. So whether
you wanna explore new skills deepen existing passions, or get lost in creativity get started with Skillshare. But either way, have a lovely day. Choo, choo. So silly, so silly (laughs) (gentle music)

Comments

@answerinprogress

I really hope you liked that video; it took a week and a day longer than usual because I was tired of ending videos with accepting flat out failure. If you did, consider sharing our channel with a friend (it's the biggest way you can help us keep doing what we're doing). Also, we (secretly) announced something in our latest newsletter, you can check out the archive here: answerinprogress.com/newsletter

@tacoman10

"Can you teach a robot to be a good person?" Step 1: Define "good"

@ectior

This reminds me of what my 9th grade programming teacher said: “The computer isn’t stupid, it will do exactly what you tell it to. The computer is only as stupid as you are.”

@arrangemonk

the usual human solution to this kind of dilemma is panicing, doing something by chance, screaming and suffer from ptsd for the rest of their life

@strawberrylemonadelioness

I never thought Garfield could create such an ethical dilemma.

@sorsocksfake

Plot twist: the AI actually didn't want to save cats. It just liked killing humans more.

@bensonprice4027

I think the tagline for this show should be this, "I found out that what I am trying to do is harder than I expected." I swear I hear it on every episode.

@XhanAnimations

As a cat person, I can't fault the machine at all 😂

@zephodb

The trolley problem is basically psychological trolling where everyone tries to force others to be unsure of themselves and be able to call them monsters for making a decision, any decision. The Trolley problem is a mental experiment to force everyone into inaction. >.<

@cwasone.

Kill the entire human race or kill a cat? The AI: OoOoh… that’s a tough one

@cerealgibbon7330

"But what if Albert Einstein was on the other track?" "Isn't he already dead?" No no she's got a point

@matt45540

The good place is by far my favorite visual representation of the trolley problem. Great concept

@sergio.ballesteros

What nobody understands about this dilemma is that it doesn’t matter who are the people in the trails. The purest moral question here is, what is worse: letting 5 people die by omision or killing 1 on purpose? In other words, is it better to dirt your hands, or is it better to leave life be, however atrocious it may seem. Adding character to the people on trails doesn’t change the fact that you are knowingly killing a human being if you pull the lever, making you a murderer. This question is asking if “The Greater Good” is something real, instinctive or innate; or if it’s just another social construct. Because that’s the excuse we use to justify our acts, and worse, that’s the excuse people in positions of power use when taking certain decisions, specially in Governments. Are we entitled to judge other people lifes if that means we have to choose who lives or who doesn’t? So it doesn’t matter if kids or Gandhi are on the trail, do you have the guts to pull the lever?

@sabyasachighosh9940

AI 🤝 Cats Biding their time to make humans a subservient species.

@halbarroyzanty2931

It would be funny to have a typical "rogue ai destroys humanity" story except the ai just really likes cats

@GiraffeBoy

6:22 "I've done a bad job!" 👎 🤣

@MiningNatureYT

Seeing Ryder show up at 3:07 genuinely made me smile because why wouldn't I, Ryder's amazing

@dshopov

Your AI is actually flawless. It realized that without humans there wont be any trolleys

@starcycle4308

"Something went wrong. My machine values cats over human lives." Me: I don't see an issue here.

@miguelangelvelasco4927

Bro I love your videos, your humor is on point and it's clear that you put a lot of work and love into them