By clicking “Accept”, you agree to the storing of cookies on your device to enhance site navigation, analyze site usage, and assist in our marketing efforts. View our Privacy Policy for more information.

58b

Updated on Feb 6th 2024 based on the version endorsed by the Coreper I on Feb 2nd

Whilst risks related to AI systems can result from the way such systems are designed, risks can as well stem from how such AI systems are used. Deployers of high-risk AI system therefore play a critical role in ensuring that fundamental rights are protected, complementing the obligations of the provider when developing the AI system. Deployers are best placed to understand how the high-risk AI system will be used concretely and can therefore identify potential significant risks that were not foreseen in the development phase, due to a more precise knowledge of the context of use, the people or groups of people likely to be affected, including vulnerable groups. Deployers of high-risk AI systems referred to in Annex III also play a critical role in informing natural persons and should, when they make decisions or assist in making decisions related to natural persons, where applicable, inform the natural persons that they are subject to the use of the high risk AI system. This information should include the intended purpose and the type of decisions it makes. The deployer should also inform the natural person about its right to an explanation provided under this Regulation. With regard to high-risk AI systems used for law enforcement purposes, this obligation should be implemented in accordance with Article 13 of Directive 2016/680.

Report error

Report error

Please keep in mind that this form is only for feedback and suggestions for improvement.

Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.