📝 Guest Post: The EU AI Act – A Guide for Developers*
Was this email forwarded to you? Sign up here In this guest post, Raza Habib, CEO and co-founder of Humanloop, shares insights on the EU AI Act's implications for developers and startups, emphasizing that the act primarily affects high-risk and foundation model applications. He outlines the categorization of AI systems by risk, compliance requirements for each, and how Humanloop facilitates adherence, particularly for high-risk applications. Habib addresses concerns around stifling innovation and the potential impact on open-source models, suggesting the act is a sensible compromise, with most AI applications facing minimal compliance burdens. This year the EU AI Act came into law and has worried a lot of developers. The concern is that large private labs, like OpenAI and Google Deepmind, have captured regulators and will lobby for laws that make it hard for open-source models and startups to compete. There’s also a general concern that regulation will stifle innovation and make it hard for companies to build AI products. If you’re building an AI product, this post will help you understand if the EU Act will affect you, what you need to do to comply, and what the regulation likely means for the wider tech ecosystem. We’ll also dive into how an LLMOps platform like Humanloop can help you stay compliant. As ever, this isn’t legal advice and you should consult your own lawyers before making decisions but I hope that it’s a useful summary of the key points of the act specifically written for AI product teams. Will you be affected by the EU AI Act?The good news is that for most developers, the act probably won’t affect you. You’ll mostly be affected if what you’re building is classified as “high risk” (explained below) or if you’re building a foundation model that requires more than 10^25 floating point operations to train. Anyone who is selling an AI product within the EU has to comply with the EU AI Act, even if you’re not based in the EU. The EU’s definition of “AI” is pretty broad:
so it covers all of generative AI as well as a lot of traditional machine learning. The regulation is primarily based on how risky your use case is rather than what technology you use . The act splits AI applications into 4 possible risk categories: prohibited, high, low and minimal:
Most applications should fall into the limited or minimal categories, which are easy to comply with. To determine whether your application is limited risk or minimal risk, ask yourself the following questions:
If the answer to these questions is "yes," the application likely falls into the limited risk category and you’ll have to make clear to users they’re interacting with AI. What do you need to do to comply?The first thing you need to do is determine if your application is prohibited, high risk, or limited risk. There is a helpful compliance checker available here that can help. If you're limited risk then all you have to do is ensure transparency by clearly informing users that they are interacting with an AI system. This can involve clear labelling or notifications within the user interface of your application to indicate the use of AI, especially in cases where the AI generates content or interacts in a way that could be mistaken for a human. The compliance burden is much higher if your application is high-risk. Then you have to:
For customers in this category an LLMOps platform like Humanloop can really help. Humanloop gives you tools to build high quality datasets, evaluate and monitor performance, keep good records of usage data and provide human oversight. This lets you meet many of the requirements of a high-risk application builder. High-risk AI systems have 2 years to come into compliance. What about open-source model providers?The EU AI Act creates a separate category for what they consider to be “General Purpose AI” systems. Foundation models and LLMs that are trained through self-supervision on large datasets fall within this category. There are special requirements on the developers of foundation models. They have to create comprehensive technical documentation detailing their training and testing processes, provide downstream providers with information to ensure an understanding of the model's capabilities and limitations, adhere to the Copyright Directive, and publish detailed summaries of the training content used. Open-source providers only need to do the last two (adhere to copyright and summarise their training data) Many open-source model providers are already meeting these requirements and so probably won’t have to change their behaviour much. The big exception is for companies that are building models considered to be a “systemic risk”. The EU considers models that have needed more than 10^25 Floating Operations during training to fall in this category. All models in this category will have to perform more stringent model evaluations, have good cyber security, show they are mitigating risks and document any incidents with EU AI office. What if you fine-tune LMMs?The restrictions on General Purpose AI (GPAI) models apply to the original developers rather than people who are building downstream applications on top of an LLM. You still need to comply with the broader AI act but you’re not considered a producer of a GPAI system and will only need to change what you’re doing if your application is high risk. How Humanloop can help you be compliantFor high-risk applications like credit scoring, educational applications or employment screening, Humanloop can make compliance a lot easier. Several key features help you meet your obligations under the EU AI act: Humanloop datasets allow you to implement data governance. You can version and track any changes to data, investigate the data quality and control access. Humanloop’s logging and evaluation helps you meet your requirements around automated record keeping, ensuring system integrity and providing human oversight. All of the actions of your developers whilst creating prompts and flows are recorded as is the data that flows through your system. Our evaluation system ensures that your system performance continues to stay above the thresholds you set. If you want to find out more about Humanloop, we’d be happy to discuss your use case and you can book a call with me personally here. What does this mean for the AI ecosystem?The worst elements of early drafts have mostly been stripped from the EU AI Act. Whilst there is a new compliance burden for high-risk AI systems, much of what's required is already aligned with best practices. The fact that the compliance burden scales with the risk of the use case seems sensible and the majority of AI applications won’t be affected. The definition of “systemic” general-purpose AI systems as models that require more than 10^25 FLOPs to train seems shortsighted and somewhat arbitrary. The quality of models that can be trained within this compute threshold is improving rapidly, so regulating FLOPs will likely be poorly calibrated with model capability. The US executive order on AI has chosen to set the threshold 10x higher at 10^26 FLOPs which also disincentives the largest players from operating in Europe. Overall the AI Act seems like a reasonable compromise for the first attempt at regulating what will be an enormously impactful technology. The delineation of risk means most developers won’t be affected and I’m hopeful that the arbitrary restrictions on compute will be updated rapidly. *This post was written by Raza Habib, CEO and co-founder of Humanloop, and originally published here. We thank Humanloop for their insights and ongoing support of TheSequence.You’re on the free list for TheSequence Scope and TheSequence Chat. For the full experience, become a paying subscriber to TheSequence Edge. Trusted by thousands of subscribers from the leading AI labs and universities. |
Older messages
Edge 384: Inside Genie: Google DeepMind's Astonishing Model that can Build 2D Games from Text and Images
Thursday, April 4, 2024
The model represents a new category in generative AI. ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏
Edge 383: The Key Capabilities of Autonomous Agens
Tuesday, April 2, 2024
Planning, memory, profiling, action execution, knowledge management and several others. ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏
Four New Major Open Source Foundation Models in a Week
Sunday, March 31, 2024
DBRX, Grok 1.5, Samba-CoE and Jamba are all bringing unique innovations to open source generative AI. ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏
Edge 381: Google DeepMind's PrompBreeder Self-Improves Prompts
Thursday, March 28, 2024
The method combines chain of thoughts, plan and solve and evolutionary algorithms in a single mthod. ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏
Edge 380: A New Series About Autonomous Agents
Tuesday, March 26, 2024
The series will cover memory, action execution, planning, collaboration and many other characteristics of autonomous agents. ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏
You Might Also Like
📧 Building Async APIs in ASP.NET Core - The Right Way
Saturday, November 23, 2024
Building Async APIs in ASP .NET Core - The Right Way Read on: my website / Read time: 5 minutes The .NET Weekly is brought to you by: Even the smartest AI in the world won't save you from a
WebAIM November 2024 Newsletter
Friday, November 22, 2024
WebAIM November 2024 Newsletter Read this newsletter online at https://webaim.org/newsletter/2024/november Features Using Severity Ratings to Prioritize Web Accessibility Remediation When it comes to
➡️ Why Your Phone Doesn't Want You to Sideload Apps — Setting the Default Gateway in Linux
Friday, November 22, 2024
Also: Hey Apple, It's Time to Upgrade the Macs Storage, and More! How-To Geek Logo November 22, 2024 Did You Know Fantasy author JRR Tolkien is credited with inventing the main concept of orcs and
JSK Daily for Nov 22, 2024
Friday, November 22, 2024
JSK Daily for Nov 22, 2024 View this email in your browser A community curated daily e-mail of JavaScript news React E-Commerce App for Digital Products: Part 4 (Creating the Home Page) This component
Spyglass Dispatch: The Fate of Chrome • Amazon Tops Up Anthropic • Pros Quit Xitter • Brave Powers AI Search • Apple's Lazy AI River • RIP Enrique Allen
Friday, November 22, 2024
The Fate of Chrome • Amazon Tops Up Anthropic • Pros Quit Xitter • Brave Powers AI Search • Apple's Lazy AI River • RIP Enrique Allen The Spyglass Dispatch is a free newsletter sent out daily on
Charted | How the Global Distribution of Wealth Has Changed (2000-2023) 💰
Friday, November 22, 2024
This graphic illustrates the shifts in global wealth distribution between 2000 and 2023. View Online | Subscribe | Download Our App Presented by: MSCI >> Get the Free Investor Guide Now FEATURED
Daily Coding Problem: Problem #1616 [Easy]
Friday, November 22, 2024
Daily Coding Problem Good morning! Here's your coding interview problem for today. This problem was asked by Alibaba. Given an even number (greater than 2), return two prime numbers whose sum will
The problem to solve
Friday, November 22, 2024
Use problem framing to define the problem to solve This week, Tom Parson and Krishna Raha share tools and frameworks to identify and address challenges effectively, while Voltage Control highlights
Issue #568: Random mazes, train clock, and ReKill
Friday, November 22, 2024
View this email in your browser Issue #568 - November 22nd 2024 Weekly newsletter about Web Game Development. If you have anything you want to share with our community please let me know by replying to
Whats Next for AI: Interpreting Anthropic CEOs Vision
Friday, November 22, 2024
Top Tech Content sent at Noon! How the world collects web data Read this email in your browser How are you, @newsletterest1? 🪐 What's happening in tech today, November 22, 2024? The HackerNoon