On May 21, the national government unveiled a lengthy “digital charter” with the noble goals of expanded internet access and more trust online. If one peels back the feel-good 10 principles and stated justifications, however, one finds a new weapon in the censor arsenal.
“The platforms are failing their users, and they’re failing our citizens. They have to step up in a major way to counter disinformation,” The current Prime Minister said when he announced the charter on May 16. “If they don’t, we will hold them to account and there will be meaningful financial consequences.”
Fines are one way to bend internet firms, along with broader interpretation and enforcement of hate-speech legislation. One impetus for the charter is the Christchurch Call, a pledge between 18 countries and tech giants Facebook, Twitter, Google, Microsoft, and Amazon “to eliminate terrorist and violent extremist content online.” Canada signed the agreement on May 15.
The signatory governments plan to prohibit content “in a manner consistent with the rule of law and international human rights law, including freedom of expression.” However, in a more than symbolic gesture, the United States declined to join. This is consistent with the First Amendment to the US Constitution, which guarantees free speech—among other rights—as a bedrock of the republic.
Who Defines Terrorism, Extremism?
There is no dispute that terrorists weaponize social media to maximize their impact, and social-media firms have a role to play in stopping such abuse. However, politicians and tech firms can also co-opt legitimate concerns to restrict speech, stifle criticism, and tilt public debates.
The application of safety protections is already limiting legitimate discussions deemed too extreme by censors. Debate over gender pronouns has led to a fine against a peaceful, law-abiding man in British Columbia. Further, forthcoming subsidies for approved media outlets—not to mention CBC’s taxpayer funding—demonstrate how Canadian officials can easily incentivize submissive journalism. Proponents of the complicated digital charter explicitly identify “fake news” as one of their chief targets.
The Christchurch Call global strategy came in the wake of the New Zealand mass murder in March. The terrorist live-streamed on Facebook and then supporters quickly re-uploaded to YouTube and other platforms, straining content-moderation efforts.
Although the Christchurch case was clear, and the platform eliminated the file without state coercion, just who is an who is not an extremist can be in the eye of the beholder. Access Now, an advocacy group for digital rights, has criticized the Christchurch Call for its lack of precision in defining a “terrorist and violent extremist content, a concept that can vary between countries and in some cases can be used arbitrarily to harm human rights.”
There is no shortage of warning examples. In Fiji, a law aimed at online harassment has led to overreach and a chilling free-speech climate. Bangladesh has throttled internet access ahead of elections, while Benin has imposed a nationwide social-media blackout. After the latest terrorist attack in Sri Lanka, the government preemptively shut down all social-media platforms to stop rumors.
Venezuela’s Chavista dictatorship routinely blocks streaming websites during opposition protests. As part of an arbitrary and wide-ranging secret inquiry into “fake news,” the Brazilian Supreme Court this year censored a story linking one of its members to a corrupt businessman. Russia has criminalized insults against government officials or the state itself and the spreading of “lies.”
We are wrong to believe this happens only in the Third World or in authoritarian countries. France has empowered judges to order the removal of any content they deem fake during elections. Germany has enacted a law compelling social media to remove “hate speech,” and Human Rights Watch has denounced this as “unaccountable, overbroad censorship” that “should be promptly reversed.” The United Kingdom wants to hold websites liable for “harmful” content posted by users, a proposal similar to Australia’s move to jail social-media executives. The New Zealand government has made it a crime to merely possess the Churchland terrorist’s manifesto and has bullied the press to limit its coverage of the massacre trial.
These laws and regulations need not be applied extensively to chill speech online and foster a culture of self-censorship. As human-rights activist Yulia Gorbunova has noted, “authorities don’t really need to control millions of users. All they need to do is have a handful of criminal extremism cases to instill fear.”
Across the Spectrum
Bureaucrats who do not understand the bottom-up nature of the internet are paving the way for censorship across the world. Online communication has become the norm, and those who believe they will never be a target should consider the social-media accounts already being blocked or suspended. Heavy-handed moderation affects conservatives, progressives, and people who just threaten the status quo, so attempts to paint this as a one-sided problem are wrong.
Further, because there is no precise definition of extremism, social-media censors continue to err on the side of what is less likely to get them into trouble with authorities. Companies face little choice if they are to remain in business, as demonstrated by Google when it got into bed with the brutal Chinese regime.
When all they have is a hammer, everything starts looking like a nail, both for state censors and centralized social-media moderators. The Electronic Frontier Foundation has documented “the vast spectrum of people silenced by social media platforms that inconsistently and erroneously apply terms of service (TOS) rules.” The digital charter exacerbates this problem and impedes the private reconciliation already underway between open access and safety protections.
This article was first published by FCPP.