The business of publishing disinformation, inaccurate information spread purposefully and/or maliciously, is more profitable than ever according to a new study from the Global Disinformation Index (GDI). Analyzing website traffic and audience information from 20,000 domains it suspected of disinformation, GDI estimates the sites generated at least $235 million in ad revenue. GDI is a nonprofit that evaluates and rates websites’ risk of spreading disinformation.
A fertile environment
The existence of disinformation predated the internet. However, social platforms offer a new level of amplification. Social media coupled with a programmatic marketplace provided the perfect environment for malevolent actors looking for reach, a target audience, and revenue.
Ad-tech companies’ opaque practices make the environment even more attractive. Some connect buyers and sellers while others collect, aggregate, package and sell data. The result is a black box of operating systems. These shifts in the digital ecosystem offer a golden opportunity for marketers of fraud and disinformation.
Focused approach
Disinformation actors rely on each another to amplify their messages. Interfering with one actor could potentially make it more difficult for the others to spread their disinformation. However, despite their inter-reliance, disinformation actors each have a distinct focus:
State actors include governments as well as state-linked actors to spread inaccurate information and promote government propaganda. They are centralized actors using digital virality to amplify their message.
Private influence operators are for-hire companies (e.g. Cambridge Analytica) that run commercial marketing and public relations campaigns that aim to disinform the public. They use targeted campaigns to identify a specific psychological, behavioral or politically affiliated audience to amplify their message. Their misleading and false content sites look professional. These ad-supported domains mimic traditional journalism.
Grassroots trolls are individuals or groups that band together for a specific issue or cause. Their content and activities often focus on hate speech or try to push a false narrative. Their messaging often starts out on forums like 4chan or 8chan, move to other intermediate platforms like Reddit and then finally into mainstream media.
Pure rent-seekers are all about clickbait. They churn out sensational disinformation to drive visitors (and bots) to click on their site in order to collect revenue.
It’s not surprising that today’s digital marketplace offers an effective delivery system for disinformation actors. Unfortunately, the internet is filled with disinformation that is rapidly amplified via social media. And it is human nature to find drama attractive. Disinformation is loud content that demands our attention. It also claims advertising dollars in an efficient and expedited programmatic manner. As an industry, we need to de-incentivize disinformation actors by removing financial and amplification motivators.
Disinformation comes in all shapes and sizes. Whether it takes
the form of a text-based article, a meme, a video or photo – they are designed
to go viral across message boards, websites and social platforms like Facebook,
Twitter and YouTube. And, its polluting the internet. In fact, an Oxford
Internet Institute study
found that in the 30 days leading up to the 2018 U.S. midterm elections, a full
25% of Facebook and Twitter shares contained misleading and deceptive information
claiming to be real news. Addressing concerns about domestic disinformation, Paul
Barrett, of the NYU Stern Center for Business and Human Rights, identified the
steps social platforms need to take to stop the problem in a new report, Tackling
Domestic Disinformation: What the Social Media Companies Need to Do.
Disinformation
epidemic
In the Report Barrett cites a MIT study that analyzed every English-language news story distributed on Twitter over an 11-year period and then verified the content of each story as either true or false. The study found that, on average, false news is 70//5 more likely to be retweeted than true news. Where are all these falsehoods coming from? A Knight Foundation study found that 65% of fake and conspiracy news links on Twitter could be traced back to just 10 large disinformation websites (e.g. Infowars).
First Amendment
Domestic disinformation is a constant in today’s digital experience.
While there are many who call for its removal, others believe that it’s
difficult to differentiate it from ordinary political communication protected
by the First Amendment. Importantly, Barrett does not suggest that the government
determine what content should be removed from social media. He believes social
platforms can make better choices determining if content is accurate, and also how
they promote and rank it.
Practices in place
Social platforms use machine learning to improve their
ability to identify false stories, photographs, and videos. In addition, while Facebook
previously flagged content to warn readers that it was potentially false, they
now offer “Related Articles,” a feature that provides factually-reliable
context about misleading stories. YouTube offers a similar program. When a user
searches for topics that YouTube identifies as “subject to misinformation,” they
preface video results with a link to information from reliable third parties. Even
with these efforts, disinformation remains available on these platforms for
anyone to view and share. Social platforms’ current practices are not enough in
their fight to reduce disinformation.
Barrett’s recommendations:
Remove
false content. Content that is proven to be untrue should be removed from
social media sites, not just demoted or annotated.
Clarify
the principles for content removal. Offer insight and transparency into
what constitutes facts and rational argument versus the manipulation of
information for disinformation purposes.
Hire a
senior executive who has company-wide responsibility for combating false
information.
Establish
a more robust appeals process. Offer an opportunity for an appeal to a
person or people not involved in the initial content removal decision.
Step up
efforts to purge bot networks. Increase efforts to eliminate automated
accounts that imitate human behavior online.
Retool
algorithms to reduce monetization of disinformation. Doing so will diminish
the incentive and therefore the amplification of fake news.
Provide
more data for academic research. The platforms have an ethical and social
responsibility to provide data they possess to facilitate studies and tracking of
disinformation.
Increase
industry-wide cooperation. Establish a data exchange and offer best
practices across platforms to ensure common challenges are addressed.
Support initiatives
for digital media literacy. Teaching people to identify digital literacy and
how to be more discriminating of online content should remain a priority.
Sponsor
more fact-checking and explore new approaches to authenticate news content. Continue
fact checking efforts, a crucial first step to distinguish between truth and
falsehoods.
Support
narrow, targeted government regulation. Identify specific content
regulations similar to the degree of disclosure for online political
advertising currently required for traditional broadcast media.
Barnett concludes that “neither the First Amendment nor international principles protect the lies on social media.” It’s essential for social platforms to step up their role in self-governing to ensure disinformation is not monetized or worse, used to manipulate people and trigger violence. Importantly, humans must remain in control of platforms, overseeing the impact of AI in all its forms. Scrutiny and transparency are key in uniting efforts to dismantle the prongs of disinformation.
On January 29th, at the 2019 DCN Next: Summit, Rappler CEO Maria Ressa outlined the role social media and concerted, well-orchestrated disinformation campaigns played in perpetuating false information and media distrust in the Philippines, as well as attacks aimed at Rappler.
She then went on to have a wide-ranging discussion examining the various pressures on media credibility (and safety) worldwide with interviewer extraordinaire Kara Swisher, Co-founder of Recode.
Less than two weeks later, on Wednesday, February 13 at 5 p.m. local time in Manila, plainclothes officers from the National Bureau of Investigation, an agency within the Department of Justice, arrested Ressa on charges of cyber libel. As Ressa wrote in a statement: “We are not intimidated. No amount of legal cases, black propaganda, and lies can silence Filipino journalists who continue to hold the line. These legal acrobatics show how far the government will go to silence journalists, including the pettiness of forcing me to spend the night in jail.”
The Board of Directors of Digital Content Next (DCN), a trade association representing nearly 80 high-quality media companies, said, “The arrest of Maria Ressa is deeply troubling. Maria traveled to the U.S. to share her developing story with our members only two weeks ago. It is vital we value and protect the independence of media organizations and journalists around the world. Any effort to silence journalists or use intimidation to reduce their reporting is an affront to freedom. We encourage global leaders and the press community to make it clear this cannot be tolerated.”
In light of Ressa’s arrest, and to reinforce our support of a free press everywhere, DCN is pleased to share the video of Ressa and Swisher’s interview (full transcript below):
And, for those who would like to show support for Rappler and Ressa’s work, she has provided a link to their crowdfunding page.
Below, we’ve shared a full transcript of Ressa’s conversation with Swisher.
Alexandra Roman: [00:00:00]
I am truly honored to introduce this next conversation interviewer
extraordinaire Recode’s Kara Swisher. She’ll be speaking with a very special
person in our world these days. Named Time magazine’s Person of the year as one
of the guardians of journalism, please welcome the CEO of Rappler, Maria Ressa.
Kara Swisher: [00:00:33] So we’re going to start…first Maria is going to make a presentation then we’re gonna have a full fantastic discussion. Maria was on my podcast recently. It was, it was an amazing experience for me and I’m so glad she’s here and safe in the United States right now. We’ll be talking about that more. But first Maria go ahead.
Maria Ressa: [00:00:52] So
I like that Jason [Kint, CEO of DCN] talked about trust. And this is stuff I’ll
show to you from our perspective in the Philippines because it’s got the data
to prove the thesis and then I think you guys are not quite… I think you’re
not seeing the termites eating at the the credibility that you have as news
organizations and those termites are coming from geopolitical power plays. We
go back to information is power and with that that let me show you what’s
happened in the Philippines.
January last year, there were two surveys that came up exactly the same time but they’re almost complete opposite results. The top is real world Pew Global Attitudes Survey: How do Filipinos look at traditional media? And they came back they said 86 percent think traditional media is and the right quote is “fair and accurate.” But the Philippine trust index, which is part of the Edelman Trust survey, they came out with a survey that same month a year ago. And they asked people on social media and they came out with 83 percent “distrust traditional media.” Right. So how did that happen? We tried to figure out why is the world upside down? That’s really the question right. Why is the world upside down?
We have a database that we started gathering in July of 2016 when the when the drug war began in the Philippines because the attacks all came on social media. In our case it’s Facebook. But this is a timeline of attacks on traditional media. And in Rappler because we were the main focal point for a period of time, [which] started January 2015 and then moving to April 2017. January 2016 was when the campaigns began and the social media machine of then Mayor Duterte. He Was elected to office May 2016. You see that one? And you can see the fracture line Byaran means corrupt. Bias. So Bayaran is the one in the middle. The first long line and bias is the last one. If you look at that it’s a fracture line of society right.
There were mentions before but it was constantly pounded until it became a straight line after president Duterte was elected the weaponization of social media happened after he was elected because it was repeatedly pounded until it became fact. A lie told a million times its truth.
Right. So, then what happened? Here: This This is the database I was telling you about right? We call it The Shark Tank. The one on your left is the URLs that are spreading fake news in the Philippines. The middle column are the court the Facebook pages that are spreading that page. And I always look at the average reposting time which is the one all the way to your… my right, sorry it’s flipped.
I want to show you when the real attacks began against Rappler and it was after we came out with a three-part series on the weaponization of social media. It was October 2016. I went to Facebook with the data August 2016. So, October 2016 this is what it looked like. In October 2016, if it’s more than 10 times reposting, it turns red, You can see how it turned red. This Facebook page Sally Might Die accomplished its goals by April 2017. It’s been deleted from Facebook but you can see … This was something we created for our social media team so that you can see it’s a cut and paste account. And they post; look at how many times they post in one day! Each one of those squares is just one day. And this is where they post the groups. They posted to go viral in the campaign pages of Duterte it and Marcos, the son of former President Ferdinand Marcos.
oneI’m going to just show you the last thing which is how can we figure out who’s attacking us. Well you can gather the data and it looks like this but if you put it in a network map, It looks like this. This is the network that was attacking Vice President Leonie Robredo about a year ago and it is the same network that constantly attacks me, Rappler, and every traditional media. It is so systematic that the content creators of the network are broken down by demographic. For the Motherland, it is pseudo-intellectual and tries to target the one percent but pseudo-intellectual. The middle class is targeted by thinking Pinoy and the mass base is this Moka Olsen blog who is former singer dancer. They used to use to build her Facebook page by having it like she has a singing group called the Mocha Girls and they do pillow fights every Sunday. That was how they first built her Facebook page. Then she became the head of social media for the presidential palace and it became a whole other thing.
Anyway, you can see this is what attacks what attacks
journalists systematically. And it happens so many times. I just want to show
you one last thing which is something we did for Rappler. Natural Language
Processing to pull out. So, we looked at the entire Lexis Nexis right to try to
figure out … What do we need to learn? What Is the data telling us about the
articles that were written about us at the time when I was about to come home
for bail to file bail? Yeah, I had an arrest warrant then.
Right. So, the Philippines wrote 34 percent of the stories.
The United States wrote 27 percent. You guys are a potent a potent force for
us. But what was most interesting is that the Filipino stories are part of the
reason it’s in a line like this is because they essentially just regurgitated
the press release of the Department of Justice. It was the American news
organizations that talked about it as a Duterte rights crackdown. That wrote
about it in context. That was an amazing thing. I want to leave you with sorry
I don’t know what wrong thing, I think. I want to move forward. I want to leave
you with this information warfare. Yeah, I guess this is the right one.
So, with information warfare I’m going to bring it to to Russia. Dezinformatsiya. This was really interesting because. For Duterte to end the drug war. Sorry about that my slides were. OK so… I don’t know if you remember Yuri Andropov. He was the former KGB Chairman. This quote stuck with me because it fit the Philippines. Dezinformatsiya works like cocaine. If you sniff once or twice it may not change your life. If you use it everyday though it will make you into an addict. A different man. I think this is the impact on our democracies and we’ve seen it.
The first reports came out in November of 2017 saying that cheap armies on social media are rolling back democracies all around the world. And at that point it was something like 28 countries. By last year, it was 48 countries. It’s doubling. We started looking at Ukraine to try to understand how we can use the data the way Ukraine started fighting back. It is information warfare. It is political. It is about power and the money part of it … or the people who are actually or who are actually catering to the politicians. Russia backed Facebook post, this was November of 2017, this is the first time that I saw Americans really starting to look at it. But even when I saw this ok they reached 126 million Americans. I think what people missed is it happens all the time. It wasn’t just ads, it was it’s all the time …
I talk about termites. This bot is interesting to me because
it tweeted about U.S. elections. First, remember the Philippine election of
Duterte was one month before Brexit. After Brexit, there were U.S. elections
and then the Catalan elections. This little bot Ivan tweeted about all of
those. So, we found him from the Catalan elections. And when I’ve looked at his
account, it was specifically only tweeting about the Philippines. When we
posted this story, within 24 hours Twitter took his network down.
On Facebook, this is the last part I want to show you, the most recent thing that I found fascinating. In December, two groups came out with reports based on data that was given to the US Senate Intelligence Committee. This is the chart that is from new knowledge. And this thing at the bottom, I want to show you the connection between the Philippines and that chart. It’s this: So we tend to map the networks around us. Let me just. Try to get this so that you can see it. There. This is the attack network. Not connecting.
OK. So, this attack network was from November. Sorry it’s frozen. There. Yay. OK. November 8, December 7th. This network. And you know what I used to map the network is this free tool called Flourish. It’s a startup. This is little Rappler. And what’s so interesting and this is where I will make the pitch that I don’t think we have any other choice but to actually collaborate together. Rappler is here. This, all of this, is a disinformation network that’s attacking us and you can kind of literally see it right.
But what’s so interesting is in the Philippines this
overshadows the information landscape. The traditional media groups are so set
aside they’re desperate. I’ve been trying for the last two years to get our top
television networks our newspapers to work together like retweet re share each
other so that we can rise up together in the algorithms. We refuse to do it
because people think it’s competitive. But you know what? You’re competing
against disinformation not against each other now.
I want to show you this because and I’ll end with this one… so this disinformation network is so interesting right. But this is the most fascinating one. When we saw this, I was surprised because this was created a year ago. It’s only one year old the daily sentry dot net. And yet the larger the circle, the larger the eigenvector centrality, the more powerful the account is. This is exponential pushes behind it. What’s interesting about it is that this is the first time we saw a direct connection to the Russian disinformation landscape because daily century dot net uses experts in quotes from this network. Sorry I can’t I can’t do the thing but on that chart there is an American man who who’s often interviewed by our de Sputnik by Iranian television … His name is Adam Gary. He is now an expert who’s popping into the Philippine ecosystem. He came in through the Daily Century and he’s from the Daily Century and he jumped into traditional newspapers from there. There’s a direct link to him because he writes for Global Research dot ca a group in Canada and connected to two other groups: one is Eurasian affairs dot net here. Another site both of whom come from a Russian IP address.
All that data in the chart came from the data that was given
to the Senate Intelligence Committee and published last December. This is
what’s happening in my country. I think you’re finding out what’s happening in
yours. But I think we’re only a small case study of what is happening globally
and that scares me.
Kara Swisher: [00:14:13] OK. All right. So, how was prison? [laughter] No really. How was prison?
Maria Ressa: [00:14:20] I, oh, I hope I won’t get there but you know…
Kara Swisher: [00:14:25] You were arrested. Explain what happened to you? We we did a podcast and I said you should not go back to the Philippines because you will be arrested. And what happened.
Maria Ressa: [00:14:32] Of
course I went back. Right. But I wasn’t arrested. OK. I thought I would be so
our lawyers told me … My flight arrived on Sunday night at 9:30 p.m. The
court, which is supposed to be an all night court, well it closes at 9:00 p.m.
So, if they had picked me up that night I couldn’t have filed bail until Monday
morning when courts opened. In the Philippines, if you have an arrest warrant
you’re not told you have an arrest warrant. They just come get you. I came I
went home and I was I wasn’t going to change anything and it went OK. I filed
bail. I I filed bail once I filed. I posted bail five times actually in that
week.
Kara Swisher: [00:15:16] But
to be. And you weren’t actually arrested.
Maria Ressa: [00:15:18] No
I wasn’t arrested. I wasn’t arrested.
Kara Swisher: [00:15:20] Please
explain to everyone here who doesn’t know why they [are going to] arrest you.
What are the charges?
Maria Ressa: [00:15:26] Well
charges are ludicrous. Tax evasion. It’s really one event, the same event, that
I have four other cases of. They’re alleging, the government is alleging, that
I am working for, well, that Rappler is owned by Americans one and that I am
essentially working for them to take down the government. Very Putin-esque.
None of that is true. And then on top of that, the arrest warrant came from
taking that same charge: the investment instrument that we used, which was
constitutional. They then decided that… we didn’t pay the right taxes. And
the reason why they said we didn’t pay the right taxes was because they
reclassified Rappler into a stock brokerage agency.
Kara Swisher: [00:16:16] Rather
than a journalist.
Maria Ressa: [00:16:17] Rather
than a newsgroup.
Kara Swisher: [00:16:18] Right.
Maria Ressa: [00:16:19] And
that’s what I have to post bail for.
Kara Swisher: [00:16:21] The
reason I’m asking what this is I want people to understand how people can use
social media to create trumped up charges and then arrest you for them arrest
you for it.
Maria Ressa: [00:16:32] Well it is interesting that you said that because all of these charges. I laughed off because they first appeared on social media. And they were thrown at me. CIA you’re a foreigner. I am a dual citizen. But, all of that. Like termites you know they just came at it and then a year and a half later it comes out of President Duterte’s mouth during the State of the Nation address. He said that you are a journalist; I’m covering the State of the Nation address. And then President Dutertet says look at Rappler: They are American. So, then I just tweeted back. President, no we’re not owned by Americans.
Kara Swisher: [00:17:12] Right, right. So, let’s talk about the state. Well, last we talked you, you made a very passionate plea to Facebook to do something about what’s happening. What you’re showing here is essentially organized disinformation campaigns to pull you down because you’re doing critical coverage of the president in the Philippines. And so they’re employing a very slow moving but powerful network to do so and using in the Philippines as you said most people get their news at not just the Philippines but across the world from Facebook. This is the purveyor of news. And these malevolent forces have created pages and news organizations and fake organizations to try to battle that. Talk a little bit about that. About what where you are right now because at the time. You were sort of subject to the biggest news organization attacking you being used to attack you.
Maria Ressa: [00:18:06] OK.
So I think that there’s a whole information ecosystem that has been
manufactured and it is manufactured reality. And we went down to a point where
we were looking at you know how how powerful is it really. We manually counted
the impact of 26 fake accounts. 26 fake accounts can actually reach up to three
million other accounts in the Philippines and it wasn’t we were the first
targets because we expose them. I was so naive.
You know, I thought wow we can just do a hashtag no place
for hate campaign and people will come back because you think these are real
people. They Are not. And after we did that, we became the target. And as you
saw in the first slide it’s not just us it is traditional media because the
main goal is to kill any trust in any institution that can that can push back.
All we have done is challenge impunity. Impunity here in information warfare and impunity in the drug war. You don’t know how many people have been killed in the Philippines during this drug war because they keep changing the numbers. At most recent count the Philippine police will admit to killing 5,000 people. Even that number alone is huge compared to the fact that 3,200 were killed in nine years of Marcos rule. Right. But. There’s this other number they never rule out. It’s the homicide cases under investigation and there are 30,000 people who’ve been killed there. So, If you think about it since July 2016 you can have more than it is tens of thousands. Thirty five thousand. I know the way they parse the number and I’m even cautious in the way I tell you how many people have been killed.
Kara Swisher: [00:19:58] So
what they’re doing is trying to use social media to stop you from writing about
them.
Maria Ressa: [00:20:03] Not just trying to use it they’ve used it effectively. iI think this is the first the first weapon it’s a new tool against journalists and against truth. And part of the reason we’re having a crisis of trust is because this is global.
Kara Swisher: [00:20:17] Right.
So, talk a little bit about your efforts with Facebook to do this initially. You
ran into Mark Zuckerberg and told him about this.
Maria Ressa: [00:20:28] F8 April 2017. There was a small group of us who had lunch together. It was founders groups of companies that were working with Facebook and I invited him to come to the Philippines because I said you know you have no idea how powerful Facebook is. Ninety-seven percent of Filipinos who are on the Internet are on Facebook. We’re 100 million people. And he was frowning and I was going so why are you frowning. And he just said, “Maria what are the other three percent doing?”[laughter]. We laughed: huh.
Kara Swisher: [00:21:05] Ah.
Ha. Ha. That’s how the board talks. But go ahead.
Maria Ressa: [00:21:09] But
that’s when you realize that that they didn’t understand their impact. What
they understood was their goal. And so I think now that’s changed.
Kara Swisher: [00:21:21] Right.
So they did that and then you brought this information to them. What happened
initially?
Maria Ressa: [00:21:27] Nothing.
You know by the time Mark Zuckerberg was in Congress for me everything that you
guys were finding out here is you know “been there done that.” We’ve
talked about this. I feel like Cassandra, you know. I’ve talked to maybe more
than 50 different officers and friends inside Facebook.
But we’re the Philippines and maybe people think you know
you’re out there. But, when He appeared in Congress and he said it would take
five years to fix this with AI. I was like you can’t do five years. Because In
the global South in my countries in Myanmar Sri Lanka and the Philippines every
day that it isn’t fixed means people die… I think they’re getting it. I think
partly your coverage you know the 2018 has spotlighted this but I don’t think
enough because it’s still being used.
The good thing is there have been take downs take downs of
Russian networks, Iranian networks, they’ve been to take downs in the
Philippines. The most recent take down was about three weeks ago of a network
we identified and did a story on 13 months earlier. You know so it’s a little
too little too late but you know what. I will take everything because at least
it cleans it up. But the fundamental problem is that. our gatekeeping power …
So, we used to create [and] distribute the news and when we distributed the news we’re the gatekeepers. Now that power has gone to the social media platforms. Facebook is now the world’s largest distributor of news and yet it has refused to be the gatekeeper. And when it does that when you allow lies to actually get on the same playing field as facts, it taints the entire public sphere. And it’s like introducing toxic sludge in the mix. And this I think that’s the fundamental problem. They have to actually at some point say take down the lies instead of allowing it to spread.
Kara Swisher: [00:23:33] So
what do you face when you go there and say you need to take down these lies?
Tell me what happens or how are they now working with you.
Maria Ressa: [00:23:41] It’s
it’s significantly different now. And that’s part of the reason.
Kara Swisher: [00:23:46] Well
they’re very sorry now. But they’re very very sorry and also very very very
sorry.
Maria Ressa: [00:23:53] I think they’re starting to understand what they’ve done. And I think they’ve started to hire the right people. In January of 2017, Nathaniel Glaser who was in charge of counterterrorism in the Obama White House. You know he was hired and shortly after that, well took a while, because this is a manual effort right? Tracking these networks down like counterterrorism requires somebody like a law enforcement official to go look for them. And so that’s part of the reason you see the takedown start starting to happen. I think it goes. The main thing that they have to do is to go to the content moderation system that they’ve put in place.
Kara Swisher: [00:24:39] Right.
Maria Ressa: [00:24:40] As journalists we have values and principles. We call it the standards and ethics manual. As tech people they tried to atomized it into a checklist and then this checklist goes to content moderators in — you know the two largest for a long period of time we’re in Warsaw and Manila.
Kara Swisher: [00:25:01] Right.
Maria Ressa: [00:25:02] And in Manila … I don’t know if you saw the movie, it was done by..
Kara Swisher: [00:25:06] The
Cleaners?
Maria Ressa: [00:25:08] The
Cleaners, right. And in that one you can see that that these content moderators
who barely make you know minimum wage here in the States but they they have
seconds to decide whether to delete or whether to let content stay. And if they
just go by a prescriptive checklist they’ll just go up delete delete and let it
stay. And the guy who took down Napalm Girl was a Filipino and he took down
Napalm Girl because check list naked.
Kara Swisher: [00:25:35] So there’s no famous photograph of the girl running from napalm in Vietnam. Pulitzer Prize-winning photograph. It was news.
Maria Ressa: [00:25:44] So
these Filipinos who were in a call center in the Philippines are taking down
terrorist content potential are taking down supposed hate speech without any
cultural context without understanding the content.
Kara Swisher: [00:25:59] So
what do you what is your solution to them. I’m using Facebook as a broad thing
but they really are the game. Twitter is sort you have the same problems with
Twitter and other social networks?
Maria Ressa: [00:26:10] Twitter
is only 7 percent of penetration in the Philippines.
Kara Swisher: [00:26:13] So
it’s an unpopular service. So yeah.
Maria Ressa: [00:26:18] No, but it’s same right the same content moderation policy as YouTube. YouTube is huge. Also in the Philippines. And you know what this disinformation cuts across all of them. So I mean you saw it in our shark tank. We had the you or else I would love to give that to Google and have them down ran some of that. Right. Because.
Kara Swisher: [00:26:39] This
is just you doing their work for them. Correct?
Maria Ressa: [00:26:43] You
know I… I guess for me when you’re dealing with this stuff. and you’re
breathing it, it’s like toxic fumes every day. You just want a solution. And it
takes… Imagine if somebody from America comes to the Philippines and tries to
figure this out. It would take them a year. I already know it. Here take it. Do
something with it. I don’t look at it as their work. I think OK. This is where
I’ll be really generous. I know that they didn’t mean to do it. It is an
extremely powerful tool and the reason why I continue to work with Facebook is
because I think if they had the political will and the economic will to do
this.
This is a game changer for the Philippines. Rappler couldn’t
exist without Facebook. We zoomed we grew 100 to 300 percent year on year
because of Facebook at the beginning in the good times. And I think they made a
crucial error in 2015 and that was instant articles when they brought all the
news groups in and then all of a sudden were at the same algorithms as the joke
that you heard or what you had for dinner. And when we became mob rule when
facts became determined by mob rule then it changed the ecosystem of democracy
in the world.
Kara Swisher: [00:28:03] And
what do you propose now that these… So, YouTube is a problem.
Maria Ressa: [00:28:09] YouTube
is huge…
Kara Swisher: [00:28:09] A
huge problem. Are you getting the same responses from them: So sorry. They’re
really, really sorry. [laughter] No they really are. But they’re not in any way
whatsoever.
Maria Ressa: [00:28:22] So
yeah. Tell me do you think they will act on it?
Kara Swisher: [00:28:27] You know I have an expression that was from one of my grandparents: You’re so poor all you have is money. I think they like their billions. I think they think they’re doing good for the world. And I think they’re careless. It’s sort of like from The Great Gatsby. They were careless people and they moved, they did damage and moved on.
Maria Ressa: [00:28:47] But they now know they’re not. And they’re killing people. They know that now.
Kara Swisher: [00:28:52] I
think they, what I’m getting now from a lot of people, is you’re so mean to us.
[laughter]
Maria Ressa: [00:28:59] Because
I do see them see this.
Kara Swisher: [00:29:01] When they say that I’m like fuck you. [laughter, applause.] You know what I mean. So it’s very hard for me to. But they are there’s a lot of victimy.
Maria Ressa: [00:29:11] I
mean until now. But you don’t know.
Kara Swisher: [00:29:14] No
I think they’re they literally get angry when people say hey hey now you know
hack democracy you really need to fix it. And they… I think one of the things
that I find interesting is when there is money to be made or whatever, they are
it’s their company. Yes.
And when there’s problems to be solved, it’s we all togethe have to solve it as a group. You know I mean and I’m like we didn’t get 64 billion dollars that I looked at. You know I have real old shoes. I don’t know. I mean we didn’t share in the upswing. And so I think again I joke. I’m so sorry but they feel badly but then I think are actually incapable in any way of taking care of it. I think they have they don’t have the mentality. They don’t have the talent. I think they’re incompetent to the task. That’s what I think.
Maria Ressa: [00:30:02] But
if that’s the case they will die. I mean it’s going to be a slow painful death.
But you know what I mean I guess for me I’m taking almost an opposite that it’s
there’s this phrase on enlightened self-interest that is…
Kara Swisher: [00:30:17] One
would think. One would think. No because this this will eventually… the
product will become terrible to use.
Maria Ressa: [00:30:24] Right.
Kara Swisher: [00:30:25] Or
it will become very addictive to use. And then what’s the difference? Like you
said with cocaine, I think. So, how do you … what are you wanting. What would
you like from them? You’d like them to become gatekeepers in other words.
Maria Ressa: [00:30:37] I don’t think they have a choice. I think they have to be. Otherwise we will leave. Right? Or again they’ll break be broken up by regulation or people will leave. In the Philippines so look at the immediate reaction. Alexa ranking of all the websites where do Filipinos go? From 2012 to 2016: number one Facebook. Undisputed. But then when the toxic sludge began mid-2016, by January 2017 on Alexa ranking Facebook dropped from number one to number eight. And then by January 2018, it went back to number five. In January 2019, right now, if you look at Alexa ranking in the Philippines, it’s number four.
So slowly they’re rising up but there’s no way. So, I mean
my thing is if they don’t fix it we will leave. We will leave. So that’s why I
think it is in their best interest they have no choice. But They are going to
have to suck it up and they’re going to have to have they are going to have to
hire real people. Machines can’t do this. But those real people will train the A.I.
and it will get better over time and they will have to lose money because they
will have to hire real people.
Kara Swisher: [00:31:52] So
talk to me a little bit about that business because you’re trying to create a
real business.
Maria Ressa: [00:31:57] Yeah.
2019 I’m trying to be a good CEO.
Kara Swisher: [00:32:00] Being
arrested attacked and essentially they’re trying to put you out of business.
Maria Ressa: [00:32:07] The
government.
Kara Swisher: [00:32:07] Talk
about the actual business. Because it’s hard enough to do a digital effort. You
know that. I know that.
Maria Ressa: [00:32:14] Yeah.
So, in the Philippines and in many other parts of the world good journalism is
really bad business and I wear both an executive editor hat and I’m the CEO so
it’s my job to make sure our business survives. In 2017, when the attacks started
happening we realized that and we had a big board battle. You know “you
journalists, you know you gotta tone it down” from the business men. And then,
from the journalists, because we had we were the largest group of shareholders
in Rappler. We had 3 percent more votes. So we pushed forward and 2018 was
mission and a lot of anger management issues. But 2019, I have to be a good CEO
and we need to build the business. So what we’ve decided. So when you’re under
attack by the government your advertisers get scared almost immediately they
don’t want to be associated with the brand. They always say you know Maria
we’re behind you but they’re very very far behind. [laughter].
Kara Swisher: [00:33:18] And
nice Time cover!
Maria Ressa: [00:33:23] So
I found out about it on Twitter. And I had to check whether it was real! But
the time cover is the first time I saw the ecosystem come up like real people
who were afraid. Fear is very real in the Philippines and I’m sorry. Before I
before I talk about the fear and I just want to finish on the part about the
business. So businessmen the businesses… they’re not the protectors of
democracy. And even if their values say that they want to do that they just
don’t because the money isn’t there. So, you can’t attack Facebook in the same
way or if you’re run by businesses your values — sorry — they follow
afterwards after the money. So, well, what we did is: We came up. We were
forced to be agile. And A lot of the things that you saw–the mapping, trying
to understand unstructured big data ,all of these things– we came up and
pivoted and became a consultant. Like I essentially carved out another team
that can do the same things we do for Rappler for other companies.
Kara Swisher: [00:34:37] So
your business… so, in that environment what do you do? Because good
journalism like you said is bad business.
Maria Ressa: [00:34:44] Rappler
continues doing good journalism. And I’ve we’ve taken the business and pushed
it away and we actually found a new business. The two things that we did. We’re
the first in the Philippines… The crowdfunding part, actually I didn’t think
it would work in the Philippines. But when our legal fees became like a quarter
of the entire monthly spend, we asked our community and they helped. And that
that helped pay for some of the legal fees. And then we, just December, we
began a membership program we called it Rappler Plus. I don’t think it would
have worked in the Philippines because unlike the United States or Europe
unlike the more developed countries, we don’t have a history of that but not
even subscriptions. People don’t want to pay for news especially in a country
where you struggle to put food on your table three times a day. So the Rappler
Plus took off much faster than I had expected and I think it is because of the
fear. People are afraid and by standing up … By being the kid telling the
emperor he has no clothes. By telling him he cannot do this with impunity.
This Is the most powerful man that we have had in since… I
think he’s more powerful than Marcos was. He controls the executive. He owns
the legislative and by the time he leaves office he will have appointed 11 of
13 Supreme Court justices. You guys in the states worry about one Supreme Court
justice he’ll have appointed 11 of 13. This is our next generation. And It’s
extremely worrisome, especially with this information warfare, with the young
men in our country who are sucking up these fumes. You know the levels of
misogyny according to our data women are attacked at least 10 times more than
men.
Kara Swisher: [00:36:40] Alright,
we have questions from the audience and then we are going to end. Are there
questions from the audience? Yes, you over here. Right here. Put your hand up.
Question: [00:36:51] Hi. Krishan Bhatia from NBCUniversal.
Thank you for sharing this story and the insights and everything that you’re
doing to uncover this. My question for you is in the US market, as we sit here
today as premium publishers most of whom have some sort of news business and we
serve large cap marketers in the US: What should we be doing differently with
respect to Facebook in particular but platforms in general that we’re not
doing.
Maria Ressa: [00:37:20] I think we have [to address the issue]: Who is the gatekeeper right now? But I think that ideas are very simple to me. If information is power. And the gatekeeping determines what information is taken by everyone. And we all focus … the debate in the US focus is on all of these different demographics and the polarization. The polarization happens because we don’t have the same facts. So it goes down to that. Please push. I think Kara asked the solution for me is when you have something like Facebook or YouTube moving beyond prescriptive to where we used to be which is what are the values? What are the principles like standards and ethics for journalism right? It can’t be prescriptive because. Ironically what they keep saying they defend free speech but free speech in this case is being used to stifle free speech. So, you’ve got to take the toxic sludge out of the body politic because that is killing us and everything else is organ failure you know because you’re not getting the oxygen that you need.
So please push you have far more power than little Rappler
does in terms of pushing for action in my part of the world I guess you know
maybe I’m happy with little because it’s been so long. We have elections in May
and these take downs will do a lot. I’ve seen the reactions of the people
running those those Facebook pages. But please look also do the investigations
here in the United States. The data is coming out now. I think that our
credibility are and I mean are for traditional media and the new ones coming
up. I think we’re getting eaten up by termites without realizing that that the
floorboards are about to crack. That’s why I think there’s a crisis of trust.
Kara Swisher: [00:39:19] Yes,
I would agree with that. Finish on this question of fear because I think it’s a
really important thing of fear of not speaking up of rocking the boat of all
kinds of stuff or just people just are exhausted by it because you’re not doing
journalism you’re spending time dealing with lawyers you’re spending time
moving businesses around you’re not doing the actual job which of which you
were.. used to do.
Maria Ressa: [00:39:43] Yeah
that’s also true. I know it just means I’m not sleeping that much. But you know
I find that the journalism… So look, Rappler has been mission-driven and all
of the friction of a normal organization is gone because everyone who stayed
with us and everyone did stay with us on the journalism side we lost sales and
tax strangely. But the mission is so clear and the purpose is so clear and I
think the challenge for all of our news groups is to be able to maintain that.
In a society, what fear does, what this stuff does is normal
people will not… When you get attacked like this I didn’t show you any of the
attacks, but when you’re attacked so viscerally when you’re threatened with
rape with murder, you just shut up. And that’s exactly it’s meant to pound you
into silence. But our community realizes this. So in a strange way. I. We’re
not just journalists anymore also that’s weird.
Like when I’m at the airport sometimes a family a family
came in and hugged me and I hug them back. I didn’t know who they were but it
was because they are also they are afraid to speak. So when you speak for them
you fulfill a role that I think that’s the mission of journalism. I think I
have a natural tendency to be more positive I should hang out with you a little
bit more. [laughter]
But you know when you’re in my place, I put one foot in
front of the other. The mission is clear. We’re going to have to deal with
this. And I think this is what Facebook has to realize. They have to get
through this because it’s not just us. We’re just the canary in the coal mine.
It’s here it’s happening here. Your problems are because of stuff like this. I
think. I think it’s global.
Kara Swisher: [00:41:37] Are
you scared?
Maria Ressa: [00:41:39] No
because there’s too much to do. Not right now. You know there are times when I
think it was far worse when no one was paying attention because when the
attacks were so personal the first two weeks…I got 90 hate messages per hour.
Not one nine. Nine zero hate messages per hour. And when I got that, it took me
two weeks to just figure out how do how am I going to deal with this and what’s
real and what’s not and then do I need security? You know all of that stuff. So
no I’m not afraid because now I know what it is. And the data helps me
understand it. So that’s the certainty. That’s why I know it’s important to
have the facts. You cannot fight back if you don’t have the facts.
Kara Swisher: [00:42:24] All
right. On that note Maria Ressa. [applause]