By Katharine Gelber
Hate speech is proliferating online and governments, regulators and social media companies are struggling to keep pace with their efforts to combat it.
In July 2021, the racist abuse of Black English football players on Facebook and Twitter has brought the issue to the forefront and shown how slow and ineffective the tech companies have been in trying to control it and the urgent need for stronger laws.
'We don't allow hate speech on @Facebook' it says in its terms of service. Can it explain how many Facebook and @instagram accounts have been deleted since last night for abuse directed at the England footballers, and how it stopped recommendation tools promoting it? @FBoversight pic.twitter.com/hrbGprK8Mj
— Damian Collins (@DamianCollins) July 12, 2021
Australia’s piecemeal approach
In Australia, the regulation of these kinds of harmful online practices is still in its infancy.
In February, a digital industry body drafted an Australian code of conduct on disinformation and misinformation, which most of the large tech companies have adopted.
However, this self-regulatory approach has been criticised by community organisations for being voluntary and opt-in. Some have argued the code sets its threshold too high by requiring an “imminent and serious” threat of one of the identified harms occurring. This could render it ineffectual.
The Australian Communications and Media Authority has been tasked with reporting on the efficacy of the code, which is due soon.
Australia also has the “Safety by Design” framework, developed by the eSafety commissioner. This is another voluntary code of practice which encourages technology companies to mitigate risk in the way they design their products.
In late June, the federal parliament also enacted a new Online Safety Act. This legislation was developed in the aftermath of the live-streamed Christchurch massacre. It regulates some relevant types of harm, such as the cyber bullying of children and livestream broadcasts that could promote or incite extreme violence.
The Act creates a complaints-based system for removing harmful material, and in some cases the eSafety commissioner has the power to block sites. It also has a wide remit in terms of its coverage of a variety of online services. Yet, it only tackles a few specific types of harm, not the full spectrum of harmful speech.
The controversial Online Safety Act will give broad censorship powers to the eSafety commissioner, and experts warn that it could harm those it purports to save [Unlocked] #esafety #onlinecensorship https://t.co/RnXDBA1h6U
— Crikey (@crikey_news) June 23, 2021
There is also an inquiry into extremist movements and radicalisation currently under way by the Parliamentary Joint Committee on Intelligence and Security (PJCIS). It is tasked with considering steps the federal government could take to disrupt and deter hate speech, terrorism and extremism online, as well as the role of social media and the internet in allowing extremists to organise.
The inquiry was due to present its report to the home affairs minister in April, and it is now overdue.
A far deeper problem
These moves are a step in the right direction, but they still attempt to tackle each specific type of hate speech as a discrete problem.
ASIO has recently pointed to the problem of hateful echo chambers online. And Australian researchers have highlighted how right-wing extremists routinely dehumanise Muslims, Jews and immigrants as a way of coalescing support behind radical worldviews and socialise people towards violent responses.
On mainstream platforms, this is achieved through a relentless diet of distorted news that supports conspiracy theories and the “othering” of marginalised communities.
We know existing laws to tackle these types of hate speech and misinformation are inadequate.
For example, we have civil laws against discrimination and hate speech, but they rely on victims to bring legal action themselves. Members of targeted communities can be deeply fearful of the repercussions of pursuing legal action and it can result in an enormous personal cost.
What other countries are doing
Governments around the world are struggling with this problem, too. In New Zealand, for instance, there has been considerable debate over hate speech law reform, particularly whether there must be a clear link to violence before hate speech regulation can be justified.
Germany has enacted one of the toughest laws against online hate speech, imposing fines of up to €50 million for social media companies that fail to delete “evidently unlawful” content. Civil rights advocates, however, argue that it encroaches on freedom of expression.
Germany’s online hate speech law slammed by opposition, Commission https://t.co/Ca8KyPAtlp
— Euractiv Technology (@eaTechEU) May 10, 2021
France, too, passed a law last year that would have required online platforms to take down hateful content flagged by users within 24 hours, but a court struck down this provision on the ground that it infringed on freedom of expression in a way that was not necessary, suitable, and proportionate.
A potential new model in the UK
There is another, more comprehensive approach in the United Kingdom that could provide a path forward.
The Carnegie Trust has developed a proposal to introduce a statutory duty of care in response to online harms. Similar to how we require the builders of roads, buildings or bridges to exercise a duty of care to the people using them, the idea is that social media companies should be required to address the harms their platforms can cause users.
The UK government incorporated this idea into its Online Safety Bill, which was just released in May for public discussion. Trumpeted as a “new framework to tackle harmful content online”, the mammoth legislation (which runs to 145 pages) is framed around duties of care.
There are still some concerns. The Carnegie Trust itself has critiqued a number of aspects of the bill. And the powers given to the culture secretary are of particular concern to free speech advocates.
Carnegie UK Trust initial response to draft Online Safety Bill https://t.co/BtugveqAKI
— 5Rights Foundation (@5RightsFound) May 13, 2021
Despite these concerns, there is much to be said for the overall approach being pursued. First, the legislation sits within the existing framework of negligence law, in which businesses owe a duty of care, or responsibility, to the general public who use the facilities they create and enable.
Second, it places the burden of responsibility on the social media companies to protect people from the harm that could be caused by their products. This is a better approach than the government penalising social media companies after the fact for hosting illegal or harmful content (such as happens under the German law), or requiring an eSafety commissioner to do the heavy lifting on regulation.
Most importantly, this approach allows for broad coverage of existing — and emerging — types of online harm in a rapidly changing environment. For example, online speech that constitutes a threat to the democratic process would fall under the new law.
While the detail of the UK bill will no doubt be debated in coming months, it presents an opportunity to effectively tackle a problem that many agree is growing in scale and volume, yet is simultaneously very difficult to address. A statutory duty of care may be just what is needed.
Katharine Gelber is a Professor in the School of Political Science and International Studies at the University of Queensland, and a Fellow of the Academy of Social Sciences Australia. On Twitter @KGelber.
Rita Jabri Markwell, an advisor to the Australian Muslim Advocacy Network (AMAN), contributed to this article. The civil society organisation has been monitoring online hatred and engaging directly with Facebook, Twitter and the Global Internet Forum to Counter Terrorism. On Twitter @JabriRita.
This article was originally published on The Conversation, republished here under a Creative Commons license. Image credit: Pexels.