WEBVTT 00:00:01.041 --> 00:00:04.175 Hello, I'm Joy, a poet of code, 00:00:04.199 --> 00:00:09.192 on a mission to stop an unseen force that's rising, 00:00:09.216 --> 00:00:12.072 a force that I called "the coded gaze," 00:00:12.096 --> 00:00:15.405 my term for algorithmic bias. NOTE Paragraph 00:00:15.429 --> 00:00:19.729 Algorithmic bias, like human bias, results in unfairness. 00:00:19.753 --> 00:00:25.775 However, algorithms, like viruses, can spread bias on a massive scale 00:00:25.799 --> 00:00:27.381 at a rapid pace. 00:00:27.943 --> 00:00:32.330 Algorithmic bias can also lead to exclusionary experiences 00:00:32.354 --> 00:00:34.482 and discriminatory practices. 00:00:34.506 --> 00:00:36.567 Let me show you what I mean. NOTE Paragraph 00:00:38.250 --> 00:00:40.490 (Video) Joy Boulamwini: Camera. I've got a face. 00:00:40.514 --> 00:00:42.387 Can you see my face? 00:00:42.411 --> 00:00:45.524 No-glasses face. 00:00:45.548 --> 00:00:48.241 You can see her face. 00:00:48.265 --> 00:00:50.495 What about my face? NOTE Paragraph 00:00:53.255 --> 00:00:54.406 (Laughter) 00:00:54.430 --> 00:00:58.180 I've got a mask. Can you see my mask? NOTE Paragraph 00:00:59.014 --> 00:01:01.379 Joy Boulamwini: So how did this happen? 00:01:01.403 --> 00:01:04.544 Why am I sitting in front of a computer 00:01:04.568 --> 00:01:05.992 in a white mask, 00:01:06.016 --> 00:01:09.666 trying to be detected by a cheap webcam? 00:01:09.690 --> 00:01:11.981 Well, when I'm not fighting the coded gaze 00:01:12.005 --> 00:01:13.525 as a poet of code, 00:01:13.549 --> 00:01:16.821 I'm a graduate student at the MIT Media Lab, 00:01:16.845 --> 00:01:21.762 and there I have the opportunity to work on all sorts of whimsical projects, 00:01:21.786 --> 00:01:23.813 including the Aspire Mirror, 00:01:23.837 --> 00:01:28.971 a project I did so I could project digital masks onto my reflection. 00:01:28.995 --> 00:01:31.345 So in the morning, if I wanted to feel powerful, 00:01:31.369 --> 00:01:32.803 I could put on a lion. 00:01:32.827 --> 00:01:36.323 If I wanted to be uplifted, I might have a quote. 00:01:36.347 --> 00:01:39.336 So I used generic facial recognition software 00:01:39.360 --> 00:01:40.711 to build the system, 00:01:40.735 --> 00:01:45.838 but found it was really hard to test it unless I wore a white mask. NOTE Paragraph 00:01:46.822 --> 00:01:51.168 Unfortunately, I've run into this issue before. 00:01:51.192 --> 00:01:55.495 When I was an undergraduate at Georgia Tech studying computer science, 00:01:55.519 --> 00:01:57.574 I used to work on social robots, 00:01:57.598 --> 00:02:01.375 and one of my tasks was to get a robot to play peek-a-boo, 00:02:01.399 --> 00:02:03.082 a simple turn-taking game 00:02:03.106 --> 00:02:07.427 where partners cover their face and then uncover it saying, "Peek-a-boo!" 00:02:07.451 --> 00:02:11.880 The problem is, peek-a-boo doesn't really work if I can't see you, 00:02:11.904 --> 00:02:14.403 and my robot couldn't see me. 00:02:14.427 --> 00:02:18.377 But I borrowed my roommate's face to get the project done, 00:02:18.401 --> 00:02:19.781 submitted the assignment, 00:02:19.805 --> 00:02:23.558 and figured, you know what, somebody else will solve this problem. NOTE Paragraph 00:02:24.209 --> 00:02:26.212 Not too longer after, 00:02:26.236 --> 00:02:30.395 I was in Hong Kong for an entrepreneurship competition. 00:02:30.879 --> 00:02:33.573 The organizers decided to take participants 00:02:33.597 --> 00:02:35.969 on a tour of local start-ups. 00:02:35.993 --> 00:02:38.708 One of the start-ups had a social robot, 00:02:38.732 --> 00:02:40.644 and they decided to do a demo. 00:02:40.668 --> 00:02:43.648 The demo worked on everybody until it got to me, 00:02:43.672 --> 00:02:45.595 and you can probably guess it. 00:02:45.619 --> 00:02:48.584 It couldn't detect my face. 00:02:48.608 --> 00:02:51.119 I asked the developers what was going on, 00:02:51.143 --> 00:02:56.676 and it turned out we had used the same generic facial recognition software. 00:02:56.700 --> 00:02:58.350 Halfway around the world, 00:02:58.374 --> 00:03:02.226 I learned that algorithmic bias can travel as quickly 00:03:02.250 --> 00:03:05.420 as it takes to download some files off of the internet. NOTE Paragraph 00:03:06.285 --> 00:03:09.361 So what's going on? Why isn't my face being detected? 00:03:09.385 --> 00:03:12.741 Well, we have to look at how we give machines sight. 00:03:12.765 --> 00:03:16.174 Computer vision uses machine learning techniques 00:03:16.198 --> 00:03:18.078 to do facial recognition. 00:03:18.102 --> 00:03:21.999 So how this works is, you create a training set with examples of faces. 00:03:22.023 --> 00:03:24.841 This is a face. This is a face. This is not a face. 00:03:24.865 --> 00:03:29.384 And over time, you can teach a computer how to recognize other faces. 00:03:29.408 --> 00:03:33.397 However, if the training sets aren't really that diverse, 00:03:33.421 --> 00:03:36.770 any face that deviates too much from the established norm 00:03:36.794 --> 00:03:38.443 will be harder to detect, 00:03:38.467 --> 00:03:40.430 which is what was happening to me. NOTE Paragraph 00:03:40.454 --> 00:03:42.836 But don't worry -- there's some good news. 00:03:42.860 --> 00:03:45.631 Training sets don't just materialize out of nowhere. 00:03:45.655 --> 00:03:47.443 We actually can create them. 00:03:47.467 --> 00:03:51.643 So there's an opportunity to create full-spectrum training sets 00:03:51.667 --> 00:03:55.491 that reflect a richer portrait of humanity. NOTE Paragraph 00:03:55.515 --> 00:03:57.736 Now you've seen in my examples 00:03:57.760 --> 00:03:59.528 how social robots 00:03:59.552 --> 00:04:04.163 was how I found out about exclusion with algorithmic bias. 00:04:04.187 --> 00:04:09.002 But algorithmic bias can also lead to discriminatory practices. 00:04:09.977 --> 00:04:11.430 Across the US, 00:04:11.454 --> 00:04:15.652 police departments are starting to use facial recognition software 00:04:15.676 --> 00:04:18.135 in their crime-fighting arsenal. 00:04:18.159 --> 00:04:20.172 Georgetown Law published a report 00:04:20.196 --> 00:04:26.959 showing that one in two adults in the US -- that's 117 million people -- 00:04:26.983 --> 00:04:30.517 have their faces in facial recognition networks. 00:04:30.541 --> 00:04:35.093 Police departments can currently look at these networks unregulated, 00:04:35.117 --> 00:04:39.403 using algorithms that have not been audited for accuracy. 00:04:39.427 --> 00:04:43.291 Yet we know facial recognition is not fail proof, 00:04:43.315 --> 00:04:47.494 and labeling faces consistently remains a challenge. 00:04:47.518 --> 00:04:49.280 You might have seen this on Facebook. 00:04:49.304 --> 00:04:52.292 My friends and I laugh all the time when we see other people 00:04:52.316 --> 00:04:54.774 mislabeled in our photos. 00:04:54.798 --> 00:05:00.389 But misidentifying a suspected criminal is no laughing matter, 00:05:00.413 --> 00:05:03.240 nor is breaching civil liberties. NOTE Paragraph 00:05:03.264 --> 00:05:06.469 Machine learning is being used for facial recognition, 00:05:06.493 --> 00:05:10.998 but it's also extending beyond the realm of computer vision. 00:05:11.806 --> 00:05:15.822 In her book, "Weapons of Math Destruction," 00:05:15.846 --> 00:05:22.527 data scientist Cathy O'Neil talks about the rising new WMDs -- 00:05:22.551 --> 00:05:26.904 widespread, mysterious and destructive algorithms 00:05:26.928 --> 00:05:29.892 that are increasingly being used to make decisions 00:05:29.916 --> 00:05:33.093 that impact more aspects of our lives. 00:05:33.117 --> 00:05:34.987 So who gets hired or fired? 00:05:35.011 --> 00:05:37.123 Do you get that loan? Do you get insurance? 00:05:37.147 --> 00:05:40.650 Are you admitted into the college you wanted to get into? 00:05:40.674 --> 00:05:44.183 Do you and I pay the same price for the same product 00:05:44.207 --> 00:05:46.649 purchased on the same platform? NOTE Paragraph 00:05:46.673 --> 00:05:50.432 Law enforcement is also starting to use machine learning 00:05:50.456 --> 00:05:52.745 for predictive policing. 00:05:52.769 --> 00:05:56.263 Some judges use machine-generated risk scores to determine 00:05:56.287 --> 00:06:00.689 how long an individual is going to spend in prison. 00:06:00.713 --> 00:06:03.167 So we really have to think about these decisions. 00:06:03.191 --> 00:06:04.373 Are they fair? 00:06:04.397 --> 00:06:07.287 And we've seen that algorithmic bias 00:06:07.311 --> 00:06:10.685 doesn't necessarily always lead to fair outcomes. NOTE Paragraph 00:06:10.709 --> 00:06:12.673 So what can we do about it? 00:06:12.697 --> 00:06:16.377 Well, we can start thinking about how we create more inclusive code 00:06:16.401 --> 00:06:19.391 and employ inclusive coding practices. 00:06:19.415 --> 00:06:21.724 It really starts with people. 00:06:22.248 --> 00:06:24.209 So who codes matters. 00:06:24.233 --> 00:06:28.352 Are we creating full-spectrum teams with diverse individuals 00:06:28.376 --> 00:06:30.787 who can check each other's blind spots? 00:06:30.811 --> 00:06:34.356 On the technical side, how we code matters. 00:06:34.380 --> 00:06:38.031 Are we factoring in fairness as we're developing systems? 00:06:38.055 --> 00:06:40.968 And finally, why we code matters. 00:06:41.325 --> 00:06:46.408 We've used tools of computational creation to unlock immense wealth. 00:06:46.432 --> 00:06:50.879 We now have the opportunity to unlock even greater equality 00:06:50.903 --> 00:06:53.833 if we make social change a priority 00:06:53.857 --> 00:06:56.027 and not an afterthought. 00:06:56.548 --> 00:07:01.070 And so these are the three tenets that will make up the "incoding" movement. 00:07:01.094 --> 00:07:02.746 Who codes matters, 00:07:02.770 --> 00:07:04.313 how we code matters 00:07:04.337 --> 00:07:06.360 and why we code matters. NOTE Paragraph 00:07:06.384 --> 00:07:09.483 So to go towards encoding, we can start thinking about 00:07:09.507 --> 00:07:12.671 building platforms that can identify bias 00:07:12.695 --> 00:07:15.773 by collecting people's experiences like the ones I shared, 00:07:15.797 --> 00:07:18.867 but also auditing existing software. 00:07:18.891 --> 00:07:22.656 We can also start to create more inclusive training sets. 00:07:22.680 --> 00:07:25.483 Imagine a "Selfies for Inclusion" campaign 00:07:25.507 --> 00:07:29.162 where you and I can help developers test and create 00:07:29.186 --> 00:07:31.279 more inclusive training sets. 00:07:31.842 --> 00:07:34.670 And we can also start thinking more conscientiously 00:07:34.694 --> 00:07:40.085 about the social impact of the technology that we're developing. NOTE Paragraph 00:07:40.109 --> 00:07:42.502 To get the incoding movement started, 00:07:42.526 --> 00:07:45.373 I've launched the Algorithmic Justice League, 00:07:45.397 --> 00:07:51.269 where anyone who cares about fairness can help fight the coded gaze. 00:07:51.293 --> 00:07:54.589 On codedgaze.com, you can report bias, 00:07:54.613 --> 00:07:57.058 request audits, become a tester 00:07:57.082 --> 00:07:59.853 and join the ongoing conversation, 00:07:59.877 --> 00:08:02.164 #codedgaze. NOTE Paragraph 00:08:03.282 --> 00:08:05.769 So I invite you to join me 00:08:05.793 --> 00:08:09.512 in creating a world where technology works for all of us, 00:08:09.536 --> 00:08:11.433 not just some of us, 00:08:11.457 --> 00:08:16.045 a world where we value inclusion and center social change. NOTE Paragraph 00:08:16.069 --> 00:08:17.244 Thank you. NOTE Paragraph 00:08:17.268 --> 00:08:21.539 (Applause) NOTE Paragraph 00:08:23.413 --> 00:08:26.267 But I have one question: 00:08:26.291 --> 00:08:28.350 Will you join me in the fight? NOTE Paragraph 00:08:28.374 --> 00:08:29.659 (Laughter) NOTE Paragraph 00:08:29.683 --> 00:08:33.370 (Applause)