Live
Black Hat USADark ReadingBlack Hat AsiaAI BusinessHong Kong hasn’t issued a single HKD stablecoin license after March targetCoinDesk AIBitcoin is closer to its 'buy zone' than it's been in three yearsCoinDesk AIRAG Web Browser: Give Your AI Real-Time Web Access Without HallucinationsDEV CommunityWhat Nobody Tells You About Building a Protocol for AI AgentsDEV CommunityThe Evidence Is in the Phone. Most of It Never Makes It Into the Case.DEV CommunityIt's Not Smarter Models — It's Cheaper Memory: TurboQuant's Real Impact, Wall Street Panic & Academic StormDEV CommunityWindows might be hiding some of your PC's storage by default - here's how to reclaim itZDNet Big DataYour Production Code Is Training AI Models Right Now (And How to Audit Your Stack)DEV CommunitySetting Up Your Databricks Account (Free Trial + First Look at the UI)DEV CommunityHow to Use Augmented Coding to Build a Web AppDEV CommunitySetting up a hugo static site hosted with PorkbunDEV CommunityWhy Domain Knowledge Is the Core Architecture of Fine-Tuning and RAG — Not an AfterthoughtDEV CommunityBlack Hat USADark ReadingBlack Hat AsiaAI BusinessHong Kong hasn’t issued a single HKD stablecoin license after March targetCoinDesk AIBitcoin is closer to its 'buy zone' than it's been in three yearsCoinDesk AIRAG Web Browser: Give Your AI Real-Time Web Access Without HallucinationsDEV CommunityWhat Nobody Tells You About Building a Protocol for AI AgentsDEV CommunityThe Evidence Is in the Phone. Most of It Never Makes It Into the Case.DEV CommunityIt's Not Smarter Models — It's Cheaper Memory: TurboQuant's Real Impact, Wall Street Panic & Academic StormDEV CommunityWindows might be hiding some of your PC's storage by default - here's how to reclaim itZDNet Big DataYour Production Code Is Training AI Models Right Now (And How to Audit Your Stack)DEV CommunitySetting Up Your Databricks Account (Free Trial + First Look at the UI)DEV CommunityHow to Use Augmented Coding to Build a Web AppDEV CommunitySetting up a hugo static site hosted with PorkbunDEV CommunityWhy Domain Knowledge Is the Core Architecture of Fine-Tuning and RAG — Not an AfterthoughtDEV Community

EleutherAI's Thoughts on the EU AI Act

EleutherAI BlogJuly 26, 20231 min read0 views
Source Quiz

How we are supporting open source and open science in the EU AI Act.

In June, the European Parliament adopted its negotiating position on the EU AI, a comprehensive piece of legislation aimed at regulating a wide variety of artificial intelligence (AI) research, products, and services. It's expected to be finalized and adopted by the end of the year, bringing widespread changes to the way that AI organizations operate in the European Union. There is a lot in the current draft's regulations on large-scale AI systems that we agree with, such as an emphasis on transparency and documentation and an explicit requirement to assess the suitability of training data. Unfortunately the current text places a substantial burden on non-profit, open source, and community-driven research, drawing no distinction between tech giants like OpenAI and Google, non-profit research groups like EleutherAI and the Allen Institute for Artificial Intelligence, and independent hobbyists who train or finetune models governed by this law.

[Read the full position paper here]

In April we released the Pythia model suite, a set of eight models trained on two different datasets ranging from 70 million to 12 billion parameters. To empower researchers to study how the capabilities of large language models evolve over the course of training we saved and released 154 checkpoints per model, providing a previously unprecedented amount of detail to the picture of how large language models train. The 154 Pythia-12B checkpoints represents more partially trained checkpoints for a single model than the rest of the world has ever released across all other 12 billion parameter or larger language models. Pythia has received widespread acclaim, with over sixty citations in just four months and was accepted for an oral presentation at the International Conference on Machine Learning (ICML) occuring later today. Under the current parliamentary text we would not be able to do a project like this again, as the over 5,000 variations and partially trained model checkpoints each count as their own model and would require the same individualized documentation, testing, and reporting as if we developed over 5,000 distinct commercially deployed models.

The Parliamentary text also includes requirements that are currently impossible for EleutherAI to comply with. For example, it requires reporting energy usage and environmental data about the computing cluster used to train the model - information we do not necessarily have access to since we, like almost everyone who does large scale AI research, do not own the actual GPUs we use to train our models. While we work with our cloud providers to disclose as much as possible about energy usage and environmental impact, some information the EU Parliament texts requires for disclosure is viewed as proprietary by the cloud providers and is not something we have access to.

To address these shortcomings EleutherAI has partnered with Creative Commons, Hugging Face, GitHub, LAION, and Open Future to draft a position paper detailing our perspectives on the parlementary text and recommending how the EU can better achieve its goals by embracing what the open source community has to offer. Our primary recommendations are:

  • Define AI components clearly,

  • Clarify that collaborative development of open source AI components and making them available in public repositories does not subject developers to the requirements in the AI Act, building on and improving the Parliament text’s Recitals 12a-c and Article 2(5e),

  • Support the AI Office’s coordination and inclusive governance with the open source ecosystem, building on the Parliament’s text,

  • Ensure the R&D exception is practical and effective, by permitting limited testing in real-world conditions, combining aspects of the Council’s approach and an amended version of the Parliament’s Article 2(5d),

  • Set proportional requirements for “foundation models,” recognizing and distinctly treating different uses and development modalities, including open source approaches, tailoring the Parliament’s Article 28b.

EleutherAI is an unprecedented experiment in doing open, transparent, and public scientific research in artifical intelligence. While we do not believe that all organizations must necessary follow in our footsteps, we believe that it's important that somebody reveals what goes on behind the curtain during the development of these increasingly influential technologies. As such we are committing today to not only comply with the final text to the best of our ability, but also to document and publicly disclose all costs we incur and additional steps we need to take to achieve compliance. As countries around the world look to the EU AI Act when drafting their own regulation, we hope that an honest and open accounting of our ability to comply with the EU AI Act will provide lawmakers essential information about how to design regulatory frameworks that do not put an undue burden on non-profit, open source, and independent researchers.

Original source

EleutherAI Blog

https://blog.eleuther.ai/eu-aia/
Was this article helpful?

Sign in to highlight and annotate this article

AI
Ask AI about this article
Powered by AI News Hub · full article context loaded
Ready

Conversation starters

Ask anything about this article…

Daily AI Digest

Get the top 5 AI stories delivered to your inbox every morning.

More about

open sourceeu ai act

Knowledge Map

Knowledge Map
TopicsEntitiesSource
EleutherAI'…open sourceeu ai actEleutherAI …

Connected Articles — Knowledge Graph

This article is connected to other articles through shared AI topics and tags.

Knowledge Graph100 articles · 115 connections
Scroll to zoom · drag to pan · click to open

Discussion

Sign in to join the discussion

No comments yet — be the first to share your thoughts!