Refugee protection in the artificial intelligence era

Think tank: Chatham House

Author(s): Madeleine Forster

September 7, 2022

This report from UK think tank Chatham House looks at AI being introduced to help decision-making in high-risk fields, including asylum and refugee protection.

Artificial intelligence (AI) is being introduced to help decision-making in high-risk fields. This includes decision-making about asylum and refugee protection, where automated ways of processing people and predicting risks in contested circumstances hold great appeal. This field, even more than most, will act as a test case for how AI protects or fails to protect human rights. Wrong or biased decisions about refugee status can have life and death consequences, including the return of refugees to places where they face persecution, contrary to international law. Existing refugee decision-making systems are already complex and are often affected by flaws, including lack of legal remedies – issues that can be exacerbated when overlayed with AI. This paper examines the primary protections being proposed to make AI more responsive to human rights, including the upcoming EU AI Law. Can innovation and protection of human rights really be combined in asylum systems and other domains that make decisions about the future of vulnerable communities and minorities? This is a question not just for governments but also for private sector providers, which have independent human rights responsibilities when providing AI products in a politically charged and changeable policy field that decides the future of vulnerable communities and minorities.