In war zones, social media disinformation is costing lives

In war zones, social media disinformation is costing lives

And precisely because war zones are complicated, the platforms should listen more to those whose job is to understand and address conflict – organisations such as the UN, or specialist disinformation researchers. These people could guide platforms on how misinformation can incite violence, and alert them ahead of events likely to attract disinformation. For instance, if a crucial round of peace talks is scheduled, social media companies could be told so that they can step up their resourcing and monitoring. Neither mediators nor the platforms are having this kind of conversation right now.

That is a problem. It’s now commonplace for social media companies to put in special measures around elections, with the adoption of stricter moderation rules and the investment of extra resources into enforcement. The rationale is obvious: elections are a major civic process and what happens online can undermine their integrity. But the same can be said of peace talks, which often determine not just who is in government but the very building blocks of the state.

With so much at stake, it’s no surprise that such negotiations will attract spoilers. Every war has individuals invested in more conflict, and many of them are acquiring sophisticated information operations capacities, either independently or from foreign backers.

That’s why social networks need a policy for peace talks – one that starts with being aware of them. Right now, Facebook staff sit down regularly to look at election calendars and decide which one could lead to violence, allocating company resources accordingly. But there is no process where the company looks out for countries entering into a delicate process of negotiations. This gap must be filled.

Companies wouldn’t need to reinvent the wheel. Many of the policies deployed by Facebook and Twitter around elections could be adapted to protect peace talks. Misinformation about talks could be labelled in local languages, with links to genuine information such as official statements by the UN. Content that aims to intimidate negotiators should be removed or labelled. All of this can be done in a way that still allows for legitimate criticism of the process.

That said, while social media companies clearly need to step up, we should be realistic about how much they will actually do to protect peace. However bad Libya’s civil war gets, it will never get as much attention as what happens in the US. There are also limits to how much social media “whack-a-mole” can achieve. Pulling down networks and accounts makes life harder for bad actors, but they’ll return in another guise before long. Labelling misinformation after it has appeared won’t change the minds of many.

The challenge is stopping the issue at the source. Could mediators persuade warring parties to put down not just their guns but their fake Facebook accounts, too? Are digital ceasefires possible? This wouldn’t be easy. But it’s worth trying, as nothing would make more of a difference than dealing with the problem before it hits social media platforms in the first place.

The barriers are less steep than many peacemakers assume. Often mediators mistakenly feel that what happens online is too opaque to do anything about. But the capacity to detect information operations exists, in organisations like Graphika, the Stanford Internet Observatory, and the Atlantic Council. Ignorance is no longer an excuse. Then it’s up to mediators to find a way to persuade and pressure those at the negotiating table to exercise restraint – online as offline. Just like in any negotiation, if one side is vastly more powerful than the other, compromise won’t be easy to find. But if everyone is suffering perhaps an online detente can emerge.

Published at Tue, 20 Apr 2021 05:00:00 +0000

Article source: