HomeTop StoriesPolice department tests AI-powered detective on real crimes

Police department tests AI-powered detective on real crimes

Bot crimes

A police department in the United Kingdom is testing an AI-powered system that could potentially help solve cold cases by condensing decades of detective work into just a few hours. Sky News reports.

But nothing is yet known about the accuracy of this Australian-developed platform, called Soze, which is a major concern because AI models tend to spit out wildly incorrect results or hallucinate made-up information.

Avon and Somerset Police, which covers parts of south-west England, are testing the program by having Soze scan and analyze emails, social media accounts, video, financial statements and other documents.

Air reports that the AI ​​was able to scan evidence from 27 ‘complex’ cases in about 30 hours – the equivalent of 81 years of human work, which is staggering. No wonder this police department is interested in using this program, as these numbers sound like a force multiplier on steroids – making it especially attractive to law enforcement agencies that may be limited in terms of staffing and budget constraints.

See also  Boston hosts annual robot block party

“Maybe you have a cold case review that just seems impossible because of the amount of material that’s out there, and you feed that into a system like this that can just process it and then give you an assessment of it,” he said. the British National Police Chiefs’ Council. Gavin Stephens told it Air. “I can see this being really helpful.”

Minority report

Another AI project Stephens refers to is compiling a database of knives and swords, which many suspects have used to attack, maim or kill victims in Britain.

Stephens seems optimistic about the soon rollout of these AI tools, but it would make sense to first validate that they work well. AI, perhaps especially in law enforcement, is notoriously error-prone and can lead to false positives.

One model used to predict a suspect’s chances of committing another crime in the future was inaccurate and also biased against black people – which sounds like something straight out of Philip K’s novella ‘Minority Report’ .Dick comes, which was later adapted into the 2002 Steven Report. Spielberg film. AI facial recognition could also lead to false arrests, with minorities repeatedly punished for crimes they did not commit.

See also  How the fact-checking industry deals with a flood of lies

These inaccuracies are so concerning that the U.S. Commission on Civil Rights recently criticized the use of AI in policing.

There is a perception that because these are machines that perform analytics, they will be infallible and accurate. But they are based on data collected by humans, which can be biased and can be downright wrong, ingraining known problems from the start.

More about AI and policing: Police say hallucinogenic AIs are ready to write police reports that could send people to jail

- Advertisement -
RELATED ARTICLES

LEAVE A REPLY

Please enter your comment!
Please enter your name here

Most Popular

Recent Comments