GENEVA (26 November 2020) — The Committee on the Elimination of Racial Discrimination today published its guidance to combat racial profiling, emphasizing, among other issues, the serious risk of algorithmic bias when artificial intelligence (AI) is used in law enforcement.
The Committee on the Elimination of Racial Discrimination (CERD), made up of 18 individual experts, has issued its general recommendation on preventing and combating racial profiling by law enforcement officials. General recommendations are legal guidance to assist the 182 State members to fulfil their obligations to the Convention on the Elimination of All Forms of Racial Discrimination.
CERD notes that the increased use by law enforcement of big data, AI, facial recognition, and other new technology risks deepening racism, racial discrimination, xenophobia and consequently the violation of many human rights.
“Big data and AI tools may reproduce and reinforce already existing biases and lead to even more discriminatory practices. We are deeply concerned with the particular risks when algorithmic profiling is used for determining the likelihood of criminal activity,” said CERD member Verene Shepherd who led the drafting of the general recommendation.
“For example, historical arrest data about a neighbourhood may reflect racially biased policing practices; and such data will deepen the risk of over-policing in the same neighbourhood, which in turn may lead to more arrests, creating a dangerous feedback loop,” she explained.
“The increasing use of facial recognition and surveillance technologies to track and control specific demographics raises concerns with respect to many human rights, including the rights to privacy, freedom of peaceful assembly and association; freedom of expression and freedom of movement,” Shepherd added.
Among its recommendations, the Committee stressed that algorithmic profiling systems should be in full compliance with international human rights law. It underscored the importance of transparency in the design and application of algorithmic profiling systems when they are deployed for law enforcement purposes.
“This includes public disclosure of the use of such systems and explanations of how the systems work, what data sets are being used and what measures preventing human rights harms are in place,” the Committee said in its recommendation.
The Committee believes that private companies that develop, sell or operate algorithmic profiling systems for law enforcement purposes also have a responsibility to involve individuals from various sectors, including legal experts, to assess the risk of human rights violation such systems may pose.
The experts also recommended that States should carefully evaluate the human rights impact prior to employing facial recognition technology.
“In addition to being unlawful, racial profiling may also be ineffective and counterproductive as a law enforcement tool,” the Committee warned. “People subjected to discriminatory law enforcement tend to have less trust in the police and, as a result, be less willing to cooperate with them.”
The full general recommendation is now available online.