Are chatbots ready for court? Police officers use AI for crime reports

Oklahoma City police have begun experimenting with artificial intelligence (AI) to generate incident reports. The technology, Taser, provided by Axon, is designed to help officers spend less time on paperwork and more time on actual policing. 

The new tool is an extension of the generative AI model used in ChatGPT. Transcribing audio from body cameras can generate the initial report of an incident in a matter of seconds. Oklahoma City police Sgt. Matt Gilmore recently had a chance to test the technology when he was in search of suspects for a long time. 

AI technology speeds up report writing and frees officers’ time

The report generated by the artificial intelligence in just eight seconds was correct and comprehensive, and it even mentioned some facts that even Sgt. Gilmore couldn’t remember. 

This has been welcomed by officers who normally spend a lot of their time entering data. Axon CEO Rick Smith stated that officers signed up to be police officers to do police work and not to be tied up filling out forms. This issue is addressed in Draft One by offering a fast and efficient means of documenting incidents. 

While police officers are excited about the use of AI in generating reports, other stakeholders, such as prosecutors, watchdogs, and community activists, have expressed concerns. The major concern is that the reports produced by AI can bias the criminal justice system, especially if they are not well-checked. District attorneys have raised an issue that officers should continue to own their reports because they might be required to defend them in court. 

In Oklahoma City, the tool is only used for minor incidents and is not applied in situations involving arrests, felonies, or violent crimes. This cautious approach avoids the involvement of AI in high-stakes and sensitive matters until more is known about its performance and effects. 

However, while Oklahoma City is still somewhat cautious, other cities such as Lafayette, Indiana, and Fort Collins, Colorado, have adopted the technology more proactively. In Lafayette, officers are allowed to use Draft One for any case, while in Fort Collins, it is adopted only at certain times and when the noise levels are low. 

Canadian legislation considers AI regulation

AI in law enforcement is not without controversy. Community activists have expressed worry over possible racial prejudice in AI, especially with Axon, which has ties with the manufacture of Tasers. Some of the critics pointed out that AI will only worsen the problem of surveillance and profiling of minority groups. 

In Canada, new legislation is being considered that could have a major effect on the future of AI. The two pieces of legislation that are being discussed in Ontario include Ontario’s Bill 194, also known as the Strengthening Cyber Security and Building Trust in the Public Sector Act, and the federal Bill C-27 which seeks to enshrine the Artificial Intelligence and Data Act (AIDA). 

Though all of these bills are aimed at controlling the use of AI, none of them discusses how AI is used by law enforcement agencies. The police in Canada have integrated AI into their operations, both as owners and operators and through collaborations with private entities. 

Source: https://www.cryptopolitan.com/police-officers-use-ai-chatbots-for-reports/