Seedy, compromised and creepy, the surveillance machine of Facebook, now operating under the broader fold of its parent company Meta Platforms, is currently giving out the very signals that it was condemned for doing before: encourage discussions on hating a group and certain figures, while spreading the bad word to everyone else to do so.
The Russian Federation, President Vladimir Putin, and Russians in general emerge as the latest contenders, the comic strip villains who those in the broadly designated âWestâ can now take issue with. According to a Meta spokesperson, the Russian attack on Ukraine had made the company make temporary âallowances for forms of political expression that would normally violate our rules like violent speech such as âdeath to the Russian invaders.ââ Cryptically, the same spokesman goes on to say that, âWe still wonât allow credible calls for violence against Russian civilians.â Meta gives us no guidelines on what would constitute a âcredible callâ.
Twitter has also permitted posts openly advocating homicide and assassination. US Senator Lindsey Graham was caught up in the bloodlust of permissiveness, using the platform to ask whether Russia had its own Brutus. âIs there a more successful Colonel Stauffenberg in the Russian military?â The only way to conclude the conflict was âfor somebody in Russia to take this guy out.â
The cartoon villainy approach of the Meta group also has precedent. In July 2021, the policy on incitement and hate speech was eased with specific reference to Iranâs Supreme leader Ali Hosseini Khamenei. The firm decided to permit posts featuring âdeath to Khameneiâ, or videos of individuals chanting the phrase for a two-week window. Lorenzo Franceschi-Bicchierai wrote pointedly at the time that this permission was âa bizarre choice that highlightâs Facebookâs power and often confusing content moderation rules.â
The Russia-Ukraine policy is only startling for being an open admission to a practice that Facebook has embraced for years. With the companyâs astronomical growth, accusations about how it utilises hate speech and deceptive content have reached a crescendo without deep effect. Mock efforts have been taken to deal with them, never deviating from the firmâs market purpose.
An example of this zig-zag morality meet reputational damage was given in 2018. In August that year, the company employed 60 Burmese-language specialists to review posted and distributed content, with a promise to employ another 40 more by the end of the year. Product manager Sara Su called the violence against the Rohingya in Myanmar âhorrific and we have been too slow to prevent misinformation on Facebook.â
A more accurate appraisal of the companyâs conduct was revealed by an internal trove of documents showing how harms were closely monitored but algorithmically exacerbated. The documents, disclosed to the US Securities and Exchange Commission by whistleblower Frances Haugen, revealed a number of things, including the gulf between CEO Mark Zuckerbergâs public statements on improvements and the companyâs own findings.
In testimony given to Congress in 2020, Zuckerberg claimed that 94 percent of hate speech was removed before a human agent reported it. The picture emerging from the internal documents showed that the company did quite the opposite: less than 5 percent of hate speech on the platform was actually removed.
Haugen summed up the approach in her opening statement to the Senate Commerce Subcommittee on Consumer Protection, Product Safety, and Data Security in October last year. Conceding that social networks faced âcomplex and nuancedâ problems in dealing with misinformation, counterespionage and democracy, she was blunt about the âchoices being made inside Facebookâ. They were âdisastrous â for our children, for our public safety, for our privacy and for our democracy â and that is why we must demand Facebook makes changes.â
The platform has also been the target of legal suits for encouraging hate speech. In December, Rohingya refugees, having little time for the firmâs promises to turn a new leaf, instigated legal action in both the United States and the United Kingdom for $150 billion. The San Francisco lawsuit, filed by Edelson and Fields Law on behalf of an anonymous plaintiff, alleges that Facebookâs introduction in the country in 2011 encouraged âthe dissemination of hateful messages, disinformation and incitement to violenceâ which led to genocide of the Rohingya.
The Ukraine War has revealed a familiar pattern. On February 26, 2022 Facebook initially announced that it had âestablished a special operations center staffed by experts from across the company, including native Russian and Ukraine speakers, who are monitoring the platform around the clock, allowing us to respond to issues in real time.â The company promised that it was âtaking extensive steps to fight misinformation and implementing more transparency and restrictions around state-controlled media outlets.â
Then came the easing of policies on hate speech regarding Russian figures, with the predictable and, given the context, understandable reaction. The Russian embassy in Washington called the policy âaggressive and criminal [âŠ] leading to incitement and hatred and hostilityâ. It gave Moscow a good basis to claim that this was yet another feature of an âinformation war without rulesâ.
Disinformation experts adopt a bit of hair splitting in approving Metaâs approach. âThe policy calls for violence against Russian soldiers,â insists the Atlantic Councilâs Digital Forensic Research Labâs Emerson Brooking. âA call for violence here, by the way, is also a call for resistance because Ukrainians resist a violent invasion.â
This policy of intervening on the side of the Ukrainian cause to Russiaâs detriment is encouraged by Metaâs President of Global Affairs, Nick Clegg. In his March 11 statement, Clegg makes the case for selective violence even more pronounced. âI want to be crystal clear: our policies are focused on protecting peopleâs rights to speech as an expression of self-defense in reaction to a military invasion of their country.â Had standard content policies been followed, content âfrom ordinary Ukrainians expressing their resistance and fury at the invading military forces would have been removed.â
This immoderate stance does not have universal agreement. Media sociologist Jeremy Littau has made the pertinent observation that, âFacebook has rules, until it doesnât.â It claims to be merely a platform above taking sides, âuntil it does.â To not permit hate speech except in designated cases against certain people of a certain country was âone hell of a can of worms.â
Metaâs latest move is disturbingly refreshing in calling out a policy that remains haphazard, selectively applied, but always driven by the firmâs own amoral calculus. The Ukraine conflict now gives the group a cover for practices that enfeeble and corrupt democracy while picking sides in war. The company is clearly not above encouraging posts advocating homicide and murder after testing the windâs direction. With Russia being rapidly cancelled culturally, politically and economically throughout the fold of Western countries, Zuckerberg is bound to think he is onto a winner. At the very least, he has found a distracting alibi.
Source: Dissidentvoice.org