Meta, the parent company of Facebook, contributed to serious human rights abuses against Ethiopia’s Tigrayan community, Amnesty International said in a new report published today.
A death sentence for my father: Meta’s contribution to human rights abuses in northern Ethiopia, shows how Meta has once again failed to adequately curb the spread of content advocating hatred and violence, this time targeting Tigrayans during the November 2020 to November 2022 armed conflict in northern Ethiopia.
Amnesty International has previously highlighted Meta’s contribution to human rights violations against the Rohingya in Myanmar and warned against the recurrence of these harms if Meta’s business model and content-shaping algorithms were not fundamentally reformed.
“Three years after its staggering failures in Myanmar, Meta has once again – through its content-shaping algorithms and data-hungry business model – contributed to serious human rights abuses. Even before the outbreak of the conflict in northern Ethiopia, civil society organizations and human rights experts repeatedly warned that Meta risked contributing to violence in the country, and pleaded with the company to take meaningful action,” said Agnès Callamard, Amnesty International’s Secretary General.
“However, Meta ignored these warnings and did not take appropriate mitigation measures, even after the conflict had broken out. As a result, Meta has again contributed to serious human rights abuses, this time perpetrated against the Tigrayan community.”
The Facebook platform is a major source of information for many Ethiopians and is considered a trustworthy news source. Facebook’s algorithms fueled devastating human rights impacts by amplifying harmful content targeting the Tigrayan community across Facebook during the armed conflict.
Meta’s contribution to human rights abuses
Amnesty International’s research established that Facebook’s algorithmic systems supercharged the spread of harmful rhetoric targeting the Tigrayan community, while the platform’s content moderation systems failed to detect and respond appropriately to such content.
These failures ultimately contributed to the killing of Tigrayan university chemistry Professor Meareg Amare. Amare was killed by a group of men after posts targeting him were posted on Facebook on 3 November 2021.
The Facebook posts contained his name, photo, place of work, house address and claimed that he was a supporter of the Tigrayan People’s Liberation Front (TPLF), accusing him of stealing large sums of money. These allegations were denied by his family.
His son Abrham Meareg believes that these hostile Facebook posts contributed to his father’s death.
“I knew it would be a death sentence for my father,” Abrham Meareg told Amnesty International.
At the time, the armed conflict in northern Ethiopia involving the TPLF, Ethiopian Federal Forces, and other armed groups was raging.
Internal Meta documents reviewed by Amnesty International show that Meta knew of the inadequacies of its mitigation measures in Ethiopia and the risks this presented in a country that the company itself considered to be at a high risk of violence. An internal Meta document from 2020 warned that “current mitigation strategies are not enough” to stop the spread of harmful content on the Facebook platform in Ethiopia.
Alongside amplifying harmful content, Meta’s poor response time and refusal to take down reported content caused multiple people interviewed by Amnesty International to feel that there was no point in reporting content to the company.
Meta received multiple warnings both before and during the conflict from civil society organizations, human rights experts and its own Facebook Oversight Board, which recommended Meta undertake an independent human rights impact assessment on Ethiopia in 2021.
“The mass dissemination of these posts incited violence and discrimination targeting the Tigrayan community, pouring fuel on what was already an inflamed situation with significant ethnic tensions,” said Agnès Callamard.
*Gelila, a member of Ethiopian civil society, was part of Meta’s ‘Trusted Partner’ program – an initiative that aims to provide selected civil society groups with a designated channel to alert Meta of any harmful content – said Facebook’s failure to act on alerts made the human rights situation in the country worse.
“As someone who has been in Ethiopia for a long time, I can say that Facebook is making communities more vulnerable to conflict with each other,” said Gelila.
“They are extremely slow in reacting to things. They are not sensitive to what is said – I think they have standards which are very far from what is happening on the ground.”
Meta should take urgent measures to adequately mitigate the risks posed by the Facebook platform in Ethiopia, which is facing another security crisis in the Amhara region.
This is essential, given that UN-appointed investigators have warned about potential future atrocity crimes, a concern later echoed by the UN Special Adviser on the Prevention of Genocide, who warned of a heightened risk of genocide and related atrocity crimes in Tigray, Amhara, Afar and Oromia regions.
A recurrent failure to respect human rights
Meta’s content-shaping algorithms are designed to maximize user engagement for the purpose of serving targeted ads, with the result that they boost inflammatory, harmful, and divisive content, which tends to attract the most attention from users.
In 2018, Meta pivoted the Facebook news feed algorithm around a new metric called “MSI” or “Meaningful Social Interactions”, in a supposed attempt to “fix Facebook”.
However, Amnesty International’s analysis of evidence from the Facebook Papers – the internal Meta documents disclosed by whistleblower Frances Haugen in 2021 – show that this shift did not “fix” the problems associated with Facebook’s algorithms; instead, the algorithms remained hard-wired for maximum engagement, therefore disproportionately favouring inflammatory content, including advocacy of hatred.
In a document from the Facebook Papers from 2021, there is evidence suggesting that Meta CEO Mark Zuckerberg personally intervened to stop mitigation measures being applied in high-risk countries like Ethiopia, because the measures may have interfered with the MSI metric.
As the fundamentals of Meta’s engagement-centric business model have not changed, the company continues to present a significant and ongoing danger to human rights, particularly in conflict-affected settings.
Making it right – Meta’s responsibility to provide remedy
Meta has a responsibility to provide remedy for the human rights abuses it has contributed to in Ethiopia.
Urgent, wide-ranging reforms are needed to ensure that Meta does not contribute again to these harms in Ethiopia or in yet another country.
This includes deploying ‘break the glass’ measures – the steps which Meta can take in crisis contexts to reduce the power of algorithmic amplification – in high-risk situations as soon as they arise and ensuring equality and consistency between jurisdictions in terms of the resourcing of content moderation, policy and human rights teams globally.
States must fulfil their obligation to protect human rights by introducing and enforcing legislation to effectively rein in Big Tech’s business model. This includes prohibiting targeted advertising on the basis of invasive tracking practices.
Equally, Big Tech companies also have a responsibility to respect human rights independent of states’ obligations and where they fail to do so, they must be held accountable for the violations they have caused or contributed to.
Meta disputed the findings. Their response is reflected in the report.
Read the full report here.