Return to Video

Protecting Twitter users (sometimes from themselves)

  • 0:01 - 0:02
    My job at Twitter
  • 0:02 - 0:04
    is to ensure user trust,
  • 0:04 - 0:07
    protect user rights and keep users safe,
  • 0:07 - 0:08
    both from each other
  • 0:08 - 0:12
    and, at times, from themselves.
  • 0:12 - 0:17
    Let's talk about what scale looks like at Twitter.
  • 0:17 - 0:19
    Back in January 2009,
  • 0:19 - 0:23
    we saw more than two million new tweets each day
  • 0:23 - 0:24
    on the platform.
  • 0:24 - 0:30
    January 2014, more than 500 million.
  • 0:30 - 0:33
    We were seeing two million tweets
  • 0:33 - 0:35
    in less than six minutes.
  • 0:35 - 0:42
    That's a 24,900-percent increase.
  • 0:42 - 0:45
    Now, the vast majority of activity on Twitter
  • 0:45 - 0:47
    puts no one in harm's way.
  • 0:47 - 0:49
    There's no risk involved.
  • 0:49 - 0:54
    My job is to root out and prevent activity that might.
  • 0:54 - 0:56
    Sounds straightforward, right?
  • 0:56 - 0:58
    You might even think it'd be easy,
  • 0:58 - 1:00
    given that I just said the vast majority
  • 1:00 - 1:04
    of activity on Twitter puts no one in harm's way.
  • 1:04 - 1:06
    Why spend so much time
  • 1:06 - 1:09
    searching for potential calamities
  • 1:09 - 1:11
    in innocuous activities?
  • 1:11 - 1:14
    Given the scale that Twitter is at,
  • 1:14 - 1:17
    a one-in-a-million chance happens
  • 1:17 - 1:22
    500 times a day.
  • 1:22 - 1:23
    It's the same for other companies
  • 1:23 - 1:24
    dealing at this sort of scale.
  • 1:24 - 1:26
    For us, edge cases,
  • 1:26 - 1:30
    those rare situations that are unlikely to occur,
  • 1:30 - 1:32
    are more like norms.
  • 1:32 - 1:36
    Say 99.999 percent of tweets
  • 1:36 - 1:38
    pose no risk to anyone.
  • 1:38 - 1:39
    There's no threat involved.
  • 1:39 - 1:42
    Maybe people are documenting travel landmarks
  • 1:42 - 1:44
    like Australia's Heart Reef,
  • 1:44 - 1:47
    or tweeting about a concert they're attending,
  • 1:47 - 1:52
    or sharing pictures of cute baby animals.
  • 1:52 - 1:56
    After you take out that 99.999 percent,
  • 1:56 - 2:00
    that tiny percentage of tweets remaining
  • 2:00 - 2:02
    works out to roughly
  • 2:02 - 2:06
    150,000 per month.
  • 2:06 - 2:08
    The sheer scale of what we're dealing with
  • 2:08 - 2:11
    makes for a challenge.
  • 2:11 - 2:12
    You know what else makes my role
  • 2:12 - 2:15
    particularly challenging?
  • 2:15 - 2:20
    People do weird things.
  • 2:20 - 2:22
    (Laughter)
  • 2:22 - 2:24
    And I have to figure out what they're doing,
  • 2:24 - 2:26
    why, and whether or not there's risk involved,
  • 2:26 - 2:29
    often without much in terms of context
  • 2:29 - 2:30
    or background.
  • 2:30 - 2:33
    I'm going to show you some examples
  • 2:33 - 2:35
    that I've run into during my time at Twitter --
  • 2:35 - 2:36
    these are all real examples —
  • 2:36 - 2:39
    of situations that at first seemed cut and dried,
  • 2:39 - 2:40
    but the truth of the matter was something
  • 2:40 - 2:42
    altogether different.
  • 2:42 - 2:44
    The details have been changed
  • 2:44 - 2:45
    to protect the innocent
  • 2:45 - 2:49
    and sometimes the guilty.
  • 2:49 - 2:52
    We'll start off easy.
  • 2:52 - 2:53
    ["Yo bitch"]
  • 2:53 - 2:57
    If you saw a Tweet that only said this,
  • 2:57 - 2:58
    you might think to yourself,
  • 2:58 - 3:00
    "That looks like abuse."
  • 3:00 - 3:03
    After all, why would you
    want to receive the message,
  • 3:03 - 3:05
    "Yo, bitch."
  • 3:05 - 3:10
    Now, I try to stay relatively hip
  • 3:10 - 3:12
    to the latest trends and memes,
  • 3:12 - 3:15
    so I knew that "yo, bitch"
  • 3:15 - 3:18
    was also often a common greeting between friends,
  • 3:18 - 3:23
    as well as being a popular "Breaking Bad" reference.
  • 3:23 - 3:25
    I will admit that I did not expect
  • 3:25 - 3:28
    to encounter a fourth use case.
  • 3:28 - 3:31
    It turns out it is also used on Twitter
  • 3:31 - 3:34
    when people are role-playing as dogs.
  • 3:34 - 3:39
    (Laughter)
  • 3:39 - 3:41
    And in fact, in that case,
  • 3:41 - 3:43
    it's not only not abusive,
  • 3:43 - 3:46
    it's technically just an accurate greeting.
  • 3:46 - 3:49
    (Laughter)
  • 3:49 - 3:51
    So okay, determining whether or not
  • 3:51 - 3:52
    something is abusive without context,
  • 3:52 - 3:54
    definitely hard.
  • 3:54 - 3:57
    Let's look at spam.
  • 3:57 - 3:59
    Here's an example of an account engaged
  • 3:59 - 4:00
    in classic spammer behavior,
  • 4:00 - 4:02
    sending the exact same message
  • 4:02 - 4:04
    to thousands of people.
  • 4:04 - 4:07
    While this is a mockup I put
    together using my account,
  • 4:07 - 4:10
    we see accounts doing this all the time.
  • 4:10 - 4:12
    Seems pretty straightforward.
  • 4:12 - 4:14
    We should just automatically suspend accounts
  • 4:14 - 4:17
    engaging in this kind of behavior.
  • 4:17 - 4:20
    Turns out there's some exceptions to that rule.
  • 4:20 - 4:23
    Turns out that that message
    could also be a notification
  • 4:23 - 4:27
    you signed up for that the International
    Space Station is passing overhead
  • 4:27 - 4:29
    because you wanted to go outside
  • 4:29 - 4:31
    and see if you could see it.
  • 4:31 - 4:32
    You're not going to get that chance
  • 4:32 - 4:34
    if we mistakenly suspend the account
  • 4:34 - 4:36
    thinking it's spam.
  • 4:36 - 4:40
    Okay. Let's make the stakes higher.
  • 4:40 - 4:41
    Back to my account,
  • 4:41 - 4:45
    again exhibiting classic behavior.
  • 4:45 - 4:48
    This time it's sending the same message and link.
  • 4:48 - 4:50
    This is often indicative of
    something called phishing,
  • 4:50 - 4:54
    somebody trying to steal another
    person's account information
  • 4:54 - 4:56
    by directing them to another website.
  • 4:56 - 5:00
    That's pretty clearly not a good thing.
  • 5:00 - 5:02
    We want to, and do, suspend accounts
  • 5:02 - 5:05
    engaging in that kind of behavior.
  • 5:05 - 5:08
    So why are the stakes higher for this?
  • 5:08 - 5:11
    Well, this could also be a bystander at a rally
  • 5:11 - 5:13
    who managed to record a video
  • 5:13 - 5:16
    of a police officer beating a non-violent protester
  • 5:16 - 5:19
    who's trying to let the world know what's happening.
  • 5:19 - 5:21
    We don't want to gamble
  • 5:21 - 5:23
    on potentially silencing that crucial speech
  • 5:23 - 5:26
    by classifying it as spam and suspending it.
  • 5:26 - 5:29
    That means we evaluate hundreds of parameters
  • 5:29 - 5:31
    when looking at account behaviors,
  • 5:31 - 5:33
    and even then, we can still get it wrong
  • 5:33 - 5:35
    and have to reevaluate.
  • 5:35 - 5:39
    Now, given the sorts of challenges I'm up against,
  • 5:39 - 5:41
    it's crucial that I not only predict
  • 5:41 - 5:45
    but also design protections for the unexpected.
  • 5:45 - 5:47
    And that's not just an issue for me,
  • 5:47 - 5:49
    or for Twitter, it's an issue for you.
  • 5:49 - 5:52
    It's an issue for anybody who's building or creating
  • 5:52 - 5:54
    something that you think is going to be amazing
  • 5:54 - 5:57
    and will let people do awesome things.
  • 5:57 - 5:59
    So what do I do?
  • 5:59 - 6:03
    I pause and I think,
  • 6:03 - 6:05
    how could all of this
  • 6:05 - 6:09
    go horribly wrong?
  • 6:09 - 6:13
    I visualize catastrophe.
  • 6:13 - 6:16
    And that's hard. There's a sort of
  • 6:16 - 6:18
    inherent cognitive dissonance in doing that,
  • 6:18 - 6:20
    like when you're writing your wedding vows
  • 6:20 - 6:23
    at the same time as your prenuptial agreement.
  • 6:23 - 6:25
    (Laughter)
  • 6:25 - 6:27
    But you still have to do it,
  • 6:27 - 6:31
    particularly if you're marrying
    500 million tweets per day.
  • 6:31 - 6:34
    What do I mean by "visualize catastrophe?"
  • 6:34 - 6:37
    I try to think of how something as
  • 6:37 - 6:40
    benign and innocuous as a picture of a cat
  • 6:40 - 6:42
    could lead to death,
  • 6:42 - 6:44
    and what to do to prevent that.
  • 6:44 - 6:46
    Which happens to be my next example.
  • 6:46 - 6:49
    This is my cat, Eli.
  • 6:49 - 6:51
    We wanted to give users the ability
  • 6:51 - 6:53
    to add photos to their tweets.
  • 6:53 - 6:55
    A picture is worth a thousand words.
  • 6:55 - 6:57
    You only get 140 characters.
  • 6:57 - 6:58
    You add a photo to your tweet,
  • 6:58 - 7:01
    look at how much more content you've got now.
  • 7:01 - 7:03
    There's all sorts of great things you can do
  • 7:03 - 7:05
    by adding a photo to a tweet.
  • 7:05 - 7:07
    My job isn't to think of those.
  • 7:07 - 7:10
    It's to think of what could go wrong.
  • 7:10 - 7:12
    How could this picture
  • 7:12 - 7:15
    lead to my death?
  • 7:15 - 7:19
    Well, here's one possibility.
  • 7:19 - 7:22
    There's more in that picture than just a cat.
  • 7:22 - 7:24
    There's geodata.
  • 7:24 - 7:26
    When you take a picture with your smartphone
  • 7:26 - 7:27
    or digital camera,
  • 7:27 - 7:29
    there's a lot of additional information
  • 7:29 - 7:31
    saved along in that image.
  • 7:31 - 7:32
    In fact, this image also contains
  • 7:32 - 7:34
    the equivalent of this,
  • 7:34 - 7:37
    more specifically, this.
  • 7:37 - 7:39
    Sure, it's not likely that someone's going to try
  • 7:39 - 7:42
    to track me down and do me harm
  • 7:42 - 7:43
    based upon image data associated
  • 7:43 - 7:45
    with a picture I took of my cat,
  • 7:45 - 7:49
    but I start by assuming the worst will happen.
  • 7:49 - 7:51
    That's why, when we launched photos on Twitter,
  • 7:51 - 7:55
    we made the decision to strip that geodata out.
  • 7:55 - 8:01
    (Applause)
  • 8:01 - 8:04
    If I start by assuming the worst
  • 8:04 - 8:05
    and work backwards,
  • 8:05 - 8:07
    I can make sure that the protections we build
  • 8:07 - 8:09
    work for both expected
  • 8:09 - 8:11
    and unexpected use cases.
  • 8:11 - 8:14
    Given that I spend my days and nights
  • 8:14 - 8:16
    imagining the worst that could happen,
  • 8:16 - 8:21
    it wouldn't be surprising if
    my worldview was gloomy.
  • 8:21 - 8:22
    (Laughter)
  • 8:22 - 8:24
    It's not.
  • 8:24 - 8:28
    The vast majority of interactions I see --
  • 8:28 - 8:32
    and I see a lot, believe me -- are positive,
  • 8:32 - 8:34
    people reaching out to help
  • 8:34 - 8:37
    or to connect or share information with each other.
  • 8:37 - 8:40
    It's just that for those of us dealing with scale,
  • 8:40 - 8:44
    for those of us tasked with keeping people safe,
  • 8:44 - 8:47
    we have to assume the worst will happen,
  • 8:47 - 8:51
    because for us, a one-in-a-million chance
  • 8:51 - 8:54
    is pretty good odds.
  • 8:54 - 8:56
    Thank you.
  • 8:56 - 9:00
    (Applause)
Title:
Protecting Twitter users (sometimes from themselves)
Speaker:
Del Harvey
Description:

Del Harvey heads up Twitter’s Trust and Safety Team, and she thinks all day about how to prevent worst-case scenarios — abuse, trolling, stalking — while giving voice to people around the globe. With deadpan humor, she offers a window into how she works to keep 240 million users safe.

more » « less
Video Language:
English
Team:
closed TED
Project:
TEDTalks
Duration:
09:19

English subtitles

Revisions Compare revisions