Representative Image
Representative Image

EU Sets Global Standard with First AI Regulation: What It Means for Tech Sovereignty

EU sets global precedent with first AI law, aiming to balance innovation and ethical use. The regulation introduces risk-based classifications, strict transparency, and fines for non-compliance. Governance will be overseen by the AI Office and AI Board, fostering trust and investment in AI technology.
May 23, 2024
4 mins read

Table of the Contents:

  1. Balance of Innovation and Ethical AI Use Across Europe
  2. AI Risk Classification and Restrictions
  3. Governance and Enforcement
  4. Global Impact and Current Trends
  5. Future Prospects and the Inevitable Transition

Balance of Innovation and Ethical AI Use Across Europe

The Council of the European Union has just definitively approved the world’s first law to moderate the development, production, and use of systems based on artificial intelligence (AI). Authorities in the bloc claim that the regulation ” can set a global standard for AI regulation.” The legislation was authorized by the European Parliament in March. Following the Council’s decision, it will be published in the region’s Official Journal. The requirements and conditions of the reform will apply two years later “with some exceptions for specific provisions.”

“The new law aims to foster the development and uptake of safe and trustworthy AI systems across the EU’s single market by both private and public actors. At the same time, it aims to ensure respect of fundamental rights of EU citizens and stimulate investment and innovation on artificial intelligence in Europe,” says the Council.

More than 150 companies accuse the AI Law of being a threat to technological sovereignty. Europe’s largest companies signed an open letter to express their dissatisfaction with the AI Law approved this month by the European Union.

AI Risk Classification and Restrictions

The legal framework classifies different types of artificial intelligence based on their “higher the risk to cause harm to society.” It includes the categories: limited risk, high risk, and unacceptable risk. The first category considers resources such as assistance bots or analysis solutions that will have to meet “very light transparency obligations.” The second category is intended for biometric products, facial recognition, or those used for educational and labor market purposes. Developers of these applications must register their systems in a census and meet stricter threat management and quality obligations to operate in the European market. The third level lists programs used for cognitive-behavioral manipulation and social scoring. These will be completely banned in Europe.

“The law also prohibits the use of AI for predictive policing based on profiling and systems that use biometric data to categorise people according to specific categories such as race, religion, or sexual orientation,” authorities say. The reform includes a section for general-purpose AI models (GPAI) like GPT-4 or PaLM 2. Most of these will have to meet limited transparency provisions. Those attributed with high-impact capabilities and posing a “systemic risk” will face stricter regulation. This last classification will include GPAI trained with computational power exceeding 10^25 floating-point operations per second (FLOPS).

Governance and Enforcement

Large technology companies that do not comply with the rules will be fined up to 7% of their global turnover. Startups, small and medium-sized enterprises will be subject to administrative fines proportional to the size of their business.

The proper application of the law will be guaranteed through a new governance structure. The architecture will be headed by the AI Office, an organization that will be integrated into the European Commission and responsible for enforcing common rules across the territory. Its work will be supported by the AI Board, composed of representatives from the bloc’s member states, a panel of independent expert scientists, and an advisory forum allowing companies and other interested parties to share technical advice for the proper implementation of the reform.

Analysts suggest that the approval of the AI Law in Europe could dictate new competition parameters in the emerging global market. The regulation opens a window for European AI startups to have more opportunities to compete with the large U.S. corporations that dominate the industry.

The European Union is not the only one seeking to regulate the use and development of AI systems. U.S. authorities have presented some proposals in this regard, although most of the progress has been left to the sector’s large companies. Since last August, China has had a legislative framework for the use of generative AI, although it lacks more extensive regulations that consider the development and advancement of the technology. In Mexico, 31 initiatives have been presented to the Senate to regulate artificial intelligence, but only two propose constitutional reforms to moderate it.

The launch of ChatGPT by OpenAI highlighted the lack of detail in existing legislation to address the advanced capabilities of emerging generative AI and the risks related to the use of copyrighted material. This led officials to recognize the need for more specific regulations.

Mathieu Michel, Belgian Secretary of State for Digitalization, assures that “this landmark law, the first of its kind in the world, addresses a global technological challenge that also creates opportunities for our societies and economies.With the AI act, Europe emphasizes the importance of trust, transparency and accountability when dealing with new technologies while at the same time ensuring this fast-changing technology can flourish and boost European innovation.”


Similar Posts


Future Prospects and the Inevitable Transition

Although an agreement was reached on the AI Act, there is still a long way to go for its effective implementation and enforcement. Restrictions on general-purpose systems will not start until 12 months after the AI Act comes into effect. Meanwhile, any commercially available generative AI system, such as ChatGPT, Google’s Gemini, and Microsoft’s Copilot, will have a “transition period” of 36 months from the AI Act’s effective date to comply with the legislation.

High-risk AI applications are also covered by the new law. These include autonomous vehicles and medical devices, which will be assessed for the risks they pose to health, safety, and citizens’ fundamental rights. Additionally, AI applications in financial services and education will be evaluated where there is a risk of bias embedded in AI algorithms.

Finally, Michel emphasized that with this law, Europe reaffirms the need for AI technology to progress in a way that is beneficial and ethical for all. This European commitment to transparency and responsibility in AI is designed not only to regulate but also to incentivize responsible innovations in the field.

Rahul Somvanshi

Rahul, possessing a profound background in the creative industry, illuminates the unspoken, often confronting revelations and unpleasant subjects, navigating their complexities with a discerning eye. He perpetually questions, explores, and unveils the multifaceted impacts of change and transformation in our global landscape. As an experienced filmmaker and writer, he intricately delves into the realms of sustainability, design, flora and fauna, health, science and technology, mobility, and space, ceaselessly investigating the practical applications and transformative potentials of burgeoning developments.

Leave a Reply

Your email address will not be published.

Global MERRA-2 atmospheric temperature at 650 hPa (approximately 11,500 feet) on July 12, 2016. Red/dark red colors indicate higher atmospheric temperatures; blue/purple colors indicate lower temperatures. Credit: NASA GMAO.
Previous Story

NASA and IBM’s Prithvi Model: A 12x Boost in Climate Forecast Accuracy

Global Head of Infrastructure & Sustainability, Google ( Credit: Ben Townsend )
Next Story

Google’s Data Center Heat Recovery to Warm 80% of Finnish Homes by 2025

Latest from Artificial Intelligence