GPT: The European Parliament and the Council have recently reached a provisional agreement on the Artificial Intelligence Act, a significant regulatory framework for the use of artificial intelligence (AI) in the European Union. The agreement aims to ensure that AI systems in the EU are safe and respect fundamental rights, while also fostering innovation and positioning Europe as a leader in the field. Key aspects of the Act include safeguards on general purpose AI, limitations on the use of biometric identification systems by law enforcement, bans on social scoring and AI used to manipulate or exploit vulnerabilities, and rights for consumers to lodge complaints and receive explanations.
Specifically, the Act prohibits certain applications of AI that pose a threat to citizens' rights and democracy. These include biometric categorisation systems using sensitive characteristics, untargeted scraping of facial images, emotion recognition in workplaces and educational institutions, social scoring based on social behaviour, and AI systems that manipulate human behaviour or exploit vulnerabilities. For high-risk AI systems, the Act establishes clear obligations, including a mandatory fundamental rights impact assessment. This applies also to the insurance and banking sectors, and AI systems used in influencing elections and voter behaviour are classified as high-risk. The Act also sets out obligations for high-impact general-purpose AI systems, which include risk assessments, adversarial testing, incident reports, and transparency requirements. In terms of enforcement, non-compliance with the rules can result in fines ranging from 35 million euros or 7% of global turnover to 7.5 million euros or 1.5% of turnover, depending on the infringement and the size of the company. The agreed text will now need formal adoption by both the Parliament and Council to become EU law, with the Parliament’s Internal Market and Civil Liberties committees set to vote on the agreement in an upcoming meeting. The legislation is not expected to take effect until at least 2025. For more detailed information, you can refer to the articles from the European Parliament's news page [here](6) and from BNN [here](7). Me: yes, but consider the following passage: For high-impact GPAI models with systemic risk, Parliament negotiators managed to secure more stringent obligations. If these models meet certain criteria they will have to conduct model evaluations, assess and mitigate systemic risks, conduct adversarial testing, report to the Commission on serious incidents, ensure cybersecurity and report on their energy efficiency. MEPs also insisted that, until harmonised EU standards are published, GPAIs with systemic risk may rely on codes of practice to comply with the regulation. What 'codes of practice' means, in practice? Him: In the context of the European AI Act, "codes of practice" refer to a set of guidelines or standards that organizations can follow to ensure compliance with the regulation, particularly for general-purpose AI (GPAI) systems that pose systemic risks. These codes are intended to provide practical guidance on how to meet the Act's requirements, such as conducting model evaluations, assessing and mitigating systemic risks, performing adversarial testing, ensuring cybersecurity, and reporting on energy efficiency. Until harmonized EU standards are established, these codes of practice offer a framework for organizations to align their AI systems with the Act's objectives and principles. For more detailed information, you can refer to the European Parliament's website: [Artificial Intelligence Act: deal on comprehensive rules for trustworthy AI]( https://www.europarl.europa.eu/news/en/press-room/20231206IPR15699/artificial-intelligence-act-deal-on-comprehensive-rules-for-trustworthy-ai ). Me: Please help me understand. What 'regulatory sandbox' could mean in the following passage? MEPs wanted to ensure that businesses, especially SMEs, can develop AI solutions without undue pressure from industry giants controlling the value chain. To this end, the agreement promotes so-called regulatory sandboxes and real-world-testing, established by national authorities to develop and train innovative AI before placement on the market Him: A "regulatory sandbox" in this context refers to a controlled environment where businesses, particularly small and medium-sized enterprises (SMEs), can develop and test innovative AI solutions under relaxed regulatory conditions. This setup allows for real-world testing and development without the full burden of compliance with extensive regulations that might normally apply. These sandboxes are established by national authorities to encourage innovation in AI while ensuring safety and compliance standards are met before these AI solutions are introduced to the broader market. For more detailed information, you can refer to the European Parliament's website: [Artificial Intelligence Act: deal on comprehensive rules for trustworthy AI]( https://www.europarl.europa.eu/news/en/press-room/20231206IPR15699/artificial-intelligence-act-deal-on-comprehensive-rules-for-trustworthy-ai ) Me: Suppose I'm a SME and I'm in the process of developing an AI solution based on LLM. Why should I care about a “regulatory sandbox”? Him: As an SME developing an AI solution based on Large Language Models (LLM), a regulatory sandbox offers significant advantages. It allows you to test and refine your AI product in a real-world environment with reduced regulatory constraints. This means you can innovate and experiment more freely, accelerating development while ensuring compliance with essential safety and ethical standards. Essentially, a regulatory sandbox provides a safer, more flexible space for you to bring your AI innovations to market more efficiently and effectively. For more details, visit the European Parliament's website: [Artificial Intelligence Act: deal on comprehensive rules for trustworthy AI]( https://www.europarl.europa.eu/news/en/press-room/20231206IPR15699/artificial-intelligence-act-deal-on-comprehensive-rules-for-trustworthy-ai ). Tutto a posto, buona domenica! Guido
_______________________________________________ nexa mailing list nexa@server-nexa.polito.it https://server-nexa.polito.it/cgi-bin/mailman/listinfo/nexa