Activists documenting Myanmar’s ethnic cleansing of Rohingya Muslims are reportedly having their Facebook posts removed and their accounts suspended, British media reported.
The Rohingya people who use Facebook to share information about the attacks have called on the company to stop silencing them, the Daily Beast reports. The mass exodus of Rohingya Muslims has sparked allegations of ethnic cleansing with the UN’s High Commissioner of Human Rights calling the operations a “textbook example of ethnic cleansing”.
A military crackdown has forced more than 410,000 Rohingya into neighbouring Bangladesh as their villages were burned and hundreds were killed.
Mohammad Anwar, a Rohingya activist based in Kuala Lumpur, told the Daily Beast that Facebook had repeatedly removed his posts about violence in Rakhine state, where most Rohingya people live. One of his posts reportedly showed military activity in Rakhine state, noting Burmese military helicopters were flying over Rohingya villages. The social media giant said they took it down because it “doesn’t follow the Facebook Community Standards.”
Another removed post allegedly showed the Burmese army burning down a Rohingya hamlet. Laura Haigh, Amnesty International’s researcher for Burma, told the Daily Beast there appeared to be a targeted campaign to report Rohingya accounts to Facebook in order to have them shut down.
Responding to the allegations, a spokesperson for Facebook told The Independent: “We want Facebook to be a place where people can share responsibly, and we work hard to strike the right balance between enabling expression while providing a safe and respectful experience. That’s why we have Community Standards, which outline what type of sharing is allowed on Facebook and what type of content may be reported to us and removed. These include hate speech, fake accounts, and dangerous organisations.”
Facebook said it allows content if it is newsworthy or important to public interest, but will take down content which violates its standards on hate speech and threats of violence. However, the company admitted that mistakes do happen, but said it would act quickly to resolve them.
“Anyone can report content to us if they think it violates our standards,” the Facebook spokesperson added. “It doesn’t matter how many times a piece of content is reported, it will be treated the same. In response to the situation in Myanmar, we are carefully reviewing content against our Community Standards.”