Artificial intelligence can incredibly enhance law enforcement agencies’ capabilities to prevent, investigate, detect and prosecute crimes, as well as to predict and anticipate them. However, despite the numerous promised benefits, the use of AI systems in the law enforcement domain raises numerous ethical and legal concerns.
On behalf of Project ALIGNER, KU Leuven Centre for IT & IP Law (CiTiP) and CBRNE Ltd have jointly developed the ALIGNER Fundamental Rights Impact Assessment (AFRIA), a tool addressed to LEAs who aim to deploy AI systems for law enforcement purposes within the EU. The AFRIA is a reflective exercise, seeking to further enhance LEAs’ already existing legal and ethical governance systems, by assisting them in building and demonstrating compliance with ethical principles and fundamental rights while deploying AI systems. News about the AFRIA can be found at: https://aligner-h2020.eu/the-aligner-fundamental-rights-impact-assessment/
The AFRIA consists of two connected and complementary templates:
- The Fundamental Rights Impact Assessment template, which helps LEAs identify and assess the impact of their AI systems on those fundamental rights most likely to be infringed; and
- The AI System Governance template, which helps LEAs identify the relevant ethical standards for trustworthy AI and mitigate the impact on fundamental rights.”
The following link will provide you access to the templates and handbook: https://aligner-h2020.eu/fundamental-rights-impact-assessment-fria/
Access to this tool is available as part of the Project ALIGNER’s exploitation initiative and requests for further information should be directed to Donatella Casaburo firstname.lastname@example.org or Irina Marsh email@example.com