British police forces are using AI crime prediction tools that disproportionately target Black and racialized communities, a new report from Amnesty International UK has revealed. According to Amnesty, nearly three-quarters of UK police forces rely on predictive policing, which uses past crime data to estimate where crimes might happen or who might commit them. The 120-page report, “Automated Racism – How Police Data and Algorithms Code Discrimination Into Policing,” argues that these predictive systems reinforce racial bias instead of improving public safety. “These systems have been built with discriminatory data