AI could support humanitarian organisations dealing with armed conflict or crisis

AI could provide humanitarian organisations with crucial insights to better monitor and anticipate risk of conflict or crisis, but users should understand the potential risks, a new study warns.

AdobeStock

Humanitarian organisations have been increasingly using digital technologies, with the Covid-19 pandemic accelerating this trend.

AI-supported disaster mapping has been used in Mozambique to speed up emergency response, and AI systems were rolled out by the World Bank across twenty-one countries to predict food crises.

While the new study claims that AI technologies have the potential to further expand the toolkit of humanitarian missions in their preparedness, response, and recovery, it also warns that some uses of AI may expose people to additional harms and present significant risks for the protection of their rights.

The study, published in the Research Handbook on Warfare and Artificial Intelligence, was conducted by Ana Beduschi, from Exeter University’s Law School.

“Safeguards must be put in place to ensure that AI systems used to support the work of humanitarians are not transformed into tools of exclusion of populations in need of assistance. Safeguards concerning the respect and protection of data privacy should also be put in place,” Professor Beduschi said in a statement.

The study said that humanitarian organisations designing AI systems should ensure data protection by design and by default to minimise risks of harm – whether they are legally obliged to do so or not, as well as using Data protection impact assessments (DPIAs) to understand the potential negative impacts of these technologies.

Grievance mechanisms should also be established, the study claims, so that people can challenge decisions that were either automated or made by humans with the support of AI systems if these adversely impacted them.

“AI systems can analyse large amounts of multidimensional data at increasingly fast speeds, identify patterns in the data, and predict future behaviour. That can help organisations gain crucial insights to better monitor and anticipate risks, such as a conflict outbreak or escalation,” said Professor Beduschi.

“Yet, deploying AI systems in the humanitarian context is not without risks for the affected populations. Issues include the poor quality of the data used to train AI algorithms, the existence of algorithmic bias, the lack of transparency about AI decision-making, and the pervading concerns about the respect and protection of data privacy.

“It is crucial that humanitarians abide by the humanitarian imperative of ‘do not harm’ when deciding whether to deploy AI to support their action. In many cases, the sensible solution would be not to rely on AI technologies as these may cause additional harm to civilian populations.”