This article has been reviewed according to Science X's editorial process and policies. Editors have highlighted the following attributes while ensuring the content's credibility:

fact-checked

trusted source

written by researcher(s)

proofread

NZ police are using AI to catch criminals—but the law urgently needs to catch up too, researcher says

blur face
Credit: CC0 Public Domain

The use of artificial intelligence (AI) by New Zealand police is putting the spotlight on policing tactics in the 21st century.

A recent Official Information Act request by Radio New Zealand revealed the use of SearchX, an AI tool that can draw connections between suspects and their wider networks.

SearchX works by instantly finding connections between people, locations, criminal charges and other factors likely to increase the risk of harm to officers.

Police say SearchX is at the heart of a NZ$200 million front-line safety program, primarily developed after the death of police constable Matthew Hunt in West Auckland in 2020, as well as other recent gun violence.

But the use of SearchX and other AI programs raises questions about the invasive nature of the technology, inherent biases and whether New Zealand's current will be enough to protect the rights of everyone.

Controversial technologies

At this stage, New Zealanders only have a limited view of the AI programs being used by the police. While some the programs are public, others are being kept under wraps.

Police have acknowledged using Cellebrite, a controversial phone hacker technology. This program extracts from iPhones and Android mobiles and can access more than 50 social media platforms, including Instagram and Facebook.

The police have also acknowledged using BriefCam, which aggregates video footage, including facial recognition and vehicle license plates.

Briefcam allows police to focus on and track a person or vehicle of interest. Police claim Briefcam can reduce the time analyzing CCTV footage from three months to two hours.

Other AI tools such as Clearview AI—which takes photographs from publicly accessible social media sites to identify a person—were tested by police before being abandoned.

The use of Clearview was particularly controversial as it was trialed without the clearance of the police leadership team or the Privacy Commissioner.

Eroding privacy?

The promise of AI is that it can predict and prevent crime. But there are also concerns over the use of these tools by police.

Cellebrite and Briefcam are highly intrusive programs. They enable to access and analyze personal data without people realizing, much less providing consent.

But under current legislation, the use of both programs by police is legal.

The Privacy Act 2020 allows —including police—to collect, withhold, use or disclose personal information in a way that would otherwise breach the act, where necessary for the "maintenance of the law".

AI's biased decisions

Privacy is not the only issue being raised by the use of these programs. There is a tendency to assume decisions made by AI are more accurate than humans—particularly as tasks become more difficult.

This bias in favor of AI decisions means investigations may harden towards the AI-identified perpetrator rather than other suspects.

Some of the mistakes can be tied to biases in the algorithms. In the past decade, scholars have begun to document the negative impacts of AI on people with low incomes and the working class, particularly in the justice system.

Research has shown ethnic minorities are more likely to be misidentified by facial recognition software.

AI's use in predictive policing is also an issue as AI can be fed data from over-policed neighborhoods, which fails to record crime occurring in other neighborhoods.

The bias is compounded further as AI increasingly directs police patrols and other surveillance onto these already over-policed neighborhoods.

This is not just a problem overseas. Analyses of the New Zealand government's use of AI have raised a number of concerns, such as the issue of transparency and privacy, as well as how to manage "dirty data"—data with human biases already baked in before it is entered into AI programs.

We need updated laws

There is no legal framework for the use of AI in New Zealand, much less for the police use of it. This lack of regulation is not unique, though. Europe's long awaited AI law still hasn't been implemented.

That said, New Zealand Police is a signatory to the Australia New Zealand Police Artificial Intelligence Principles. These establish guidelines around transparency, proportionality and justifiability, human oversight, explainability, fairness, reliability, accountability, privacy and security.

The Algorithm Charter for Aotearoa New Zealand covers the ethical and responsible use of AI by government agencies.

Under the principles, police are meant to continuously monitor, test and develop AI systems and ensure data are relevant and contemporary. Under the charter, police must have a point of contact for public inquiries and a channel for challenging or appealing decisions made by AI.

But these are both voluntary codes, leaving significant gaps for legal accountability and police antipathy.

And it's not looking good so far. Police have failed to implement one of the first—and most basic—steps of the charter: to establish a point of inquiry for people who are concerned by the use of AI.

There is no special page on the police website dealing with the use of AI, nor is there anything on the main feedback page specifically mentioning the topic.

In the absence of a clear legal framework, with an independent body monitoring the police's actions and enforcing the law, New Zealanders are left relying on police to monitor themselves.

AI is barely on the radar ahead of the 2023 election. But as it becomes more pervasive across government agencies, New Zealand must follow Europe's lead and enact AI regulation to ensure use of AI doesn't cause more problems than it solves.

Provided by The Conversation

This article is republished from The Conversation under a Creative Commons license. Read the original article.The Conversation

Citation: NZ police are using AI to catch criminals—but the law urgently needs to catch up too, researcher says (2023, October 13) retrieved 28 April 2024 from https://phys.org/news/2023-10-nz-police-ai-criminalsbut-law.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

A second major British police force suffers a cyberattack in less than a month

5 shares

Feedback to editors