Are you one of the billion active TikTok users? Or are you rather the Twitter type? Either way, chances are you have come across hateful content online.
Hate speech starts offline — and can be accelerated by threats to society. COVID-19 is one such example: the pandemic has fuelled a global wave of social stigma and discrimination against the “other”.
Not surprisingly, anti-Semitism, and more largely racism, is on the rise. A study conducted by the University of Oxford unveiled that around 20 per cent of British adults endorse statements like “Jews have created the virus to collapse the economy for financial gain” or “Muslims are spreading the virus as an attack on Western values”.
The internet is where these beliefs can become mainstream. As the new epicentre of our public and private lives, the digital world has facilitated borderless and anonymous interactions thought impossible a generation ago.
Unlike the physical world, however, the internet has also provided a medium for the exponential dissemination and amplification of false information and hate. And tech companies know it. In 2018, Facebook admitted that its platform was used to inflame ethnic and religious tensions against the Rohingya in Myanmar.
As the lines between online and offline continue to blur, we have a tremendous responsibility: to ensure a safe digital space for all. The opportunity lies in deploying catalytic funding to innovative technologies that combat disinformation, hate and extremism in novel ways.
The butterfly effect of social media
Even if only one per cent of tweets contained offensive or hateful speech, it would be the equivalent to five million messages daily. It is not difficult to imagine the consequences of such virality — the Capitol siege on 6 January painfully exemplified how social media can incite violence so quickly.
Unlike violent extremism, hate speech is often subtle or hidden in between terabytes of content uploaded to the internet every day. Pseudonyms like “juice” (to refer to Jewish people) or covert symbols (like the ones in this database) feature frequently online.
They are also well-documented by advocacy organisations and academic institutions. The Decoding Antisemitism project, funded by the Alfred Landecker Foundation, leverages an interdisciplinary approach — from linguistics to machine learning — to identify both explicit and implicit online hatred by classifying secret codes and stereotypes.
However, the bottleneck is not how to single out defamatory content, but how to scan platforms accurately and at scale. Instagram offers users the option to filter out offensive comments. Twitter acquired Fabula AI to improve the health of online conversations. And TikTok and Facebook have gone so far as to set up Safety Advisory Councils or Oversight Boards that can decide what content should be taken down.
With just these efforts alone though, tech companies have failed to spot and moderate false, offensive or hateful content that is highly dependent on context, culture and language.
The dark holes of disinformation
Online hatred is ever-evolving in its shape and form to the extent that it becomes increasingly difficult to uncover. Facebook is only able to detect two thirds of altered videos (also known as “deepfakes”). Artificial Intelligence (AI) and Natural Language Processing (NLP) algorithms haven’t been quick enough in cracking down on trolls and bots that spread disinformation.
The question is: did technology fail us, or did people fail at using technology?
The 4Ps of combating online hate
Unless tech companies want to play catch-up on a constant basis, they should move beyond detection and content moderation to a holistic and proactive approach to how hatred is generated and disseminated online (see chart below).
Such an approach would have the following four target outcomes:
● Promoting diversity and anti-bias. Technologies can be designed and developed in an inclusive manner by engaging those who are affected by online hate and discrimination. For example, the Online Hate Index by the Anti-Defamation League uses a human-centred approach that involves impacted communities in the classification of hate speech.
● Preventing discrimination and violence. Certain tech designs, like recommendation engines, can accelerate pathways to online radicalisation. Others promote counter-speech or limit the virality of disinformation. We need more of the latter.
The re-direct method employed by social enterprise Moonshot CVE channels internet users who search for violent content towards alternative narratives.
● Protecting vulnerable groups. Tech platforms have primarily focused on semi-automated content moderation, through a combination of user reporting and AI flagging. However, new approaches have emerged. Samurai Labs’ reasoning machine can engage in conversations and stop them from evolving into online hate and cyberbullying.
● Prompting civic engagement. By ensuring that every voice is heard and that citizens — especially younger ones — are empowered to engage in civic discourse, we can help build more resilient societies that don’t revert back to harmful scapegoating. In the US, New/Mode makes it easier for citizens to affect policy change by leveraging digital advocacy tools.
Funding early-stage innovation
The common denominator of the 4Ps is that, with the help of technology, they address the root of the problem.
Indeed, it is no longer tech for the sake of tech. From Snapchat’s Evan Spiegel to SAP’s Christian Klein, dozens of CEOs signed President Macron’s Tech for Good call a few months ago.
Beyond pledges, companies are embracing technology’s potential to be a force for good by setting up mission-driven incubators, like Google’s Jigsaw, or by allocating funds to incentivise fundamental research, like WhatsApp’s Award for Social Science and Misinformation.
In conversations with various research organisations, such as the Institute of Strategic Dialogue (ISD), I have learnt first-hand about the demand for tech tools in the online hate and extremism space. Whether it is measuring hate real-time across social media or identifying (with high levels of confidence) troll accounts or deepfakes, there is room for innovation.
But the truth is that the pace and the various ways in which tech is being used to incite and promote online hatred is faster and more intricate than what companies can preempt or police. If the big platforms can’t solve the conundrum, we need smaller tech companies that will. And that is why catalytic funding for risky innovation is key.
Step aside big tech
With the increase of hatred due to COVID and the higher demand for new solutions, it is only natural that funders and investors become interested in scaling for-profit and non-profit early-stage tech tools. Established venture capital players like Seedcamp (investor in Factmata) and companies like Google’s Jigsaw are starting to bridge the gap between supply and demand to combat disinformation, hate and extremism online.
This article is republished from the World Economic Forum under a Creative Commons license.