Balancing public interest, fundamental rights, and innovation: The EU’s governance model for non-high-risk AI systems

The question of the concrete design of a fair and efficient governance framework to ensure responsible technology development and implementation concerns not only high-risk artificial intelligence systems. Everyday applications with a limited ability to inflict harm are also addressed. This article...

Full description

Saved in:
Bibliographic Details
Main Authors: Michael Gille, Marina Tropmann-Frick, Thorben Schomacker
Format: Article
Language:English
Published: Alexander von Humboldt Institute for Internet and Society 2024-09-01
Series:Internet Policy Review
Subjects:
Online Access:https://policyreview.info/node/1797
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:The question of the concrete design of a fair and efficient governance framework to ensure responsible technology development and implementation concerns not only high-risk artificial intelligence systems. Everyday applications with a limited ability to inflict harm are also addressed. This article examines the European Union's approach to regulating these non-high-risk systems. We focus on the governance model for these systems established by the Artificial Intelligence Act. Based on a doctrinal legal reconstruction of the rules for codes of conduct and considering the European Union's stated goal of achieving a market-oriented balance between innovation, fundamental rights, and public interest, we explore our topic from three different perspectives: an analysis of specific regulatory components of the governance mechanism is followed by a reflection on ethics and trustworthiness implications of the EU´s approach and concluded by an analysis of a case study from an NLP-based, language-simplifying artificial intelligence application for assistive purposes.
ISSN:2197-6775