📝 Guest post: 4 Types of ML Data Errors You Can Fix Right Now*
Was this email forwarded to you? Sign up here In this article, Galileo founding engineer Nikita Demir discusses common data errors that NLP teams run into and how Galileo helps fix these errors in minutes with a few lines of code. A very helpful read! IntroductionOver the past year, we’ve seen an explosion of interest in improving ML data quality. While parts of the MLOps ecosystem have matured, ML data quality has been dramatically underserved. In this article, we’ll walk you through four of the most prominent types of data errors and show you techniques for fixing them. In the end, we’ll point out how your errors can be found and fixed in minutes with a tool like Galileo! ❗️ ML Data Error #1: MislabelsThe correctness of your ground truth labels can vary depending on how they were labeled. Most public datasets, including canonical benchmark ones, have been shown to have mislabeled samples. Although deep learning models can be robust to some random noise, labelers often make errors between specifically confusing classes: “cheetah” vs “leopard”, “SearchCreativeWork” and “SearchScreeningEvent”, “talk.religion.misc” and “soc.religion.christian”. Repeated errors are likely to degrade the model’s performance on confusing classes. We’ve seen the industry mitigate these errors in a few painstaking ways:
❗️ML Data Error #2: Class OverlapAnother common data error we see, especially for problems with many ground truth classes, is an overlap in the class definitions. Your dataset might have multiple classes that mean the same thing like “billing.refund” and “account.refund”, multiple classes that apply to the same input like “appointment.cancel” and “appointment.reschedule” for the input of “I would like to cancel and reschedule my appointment”, or unclear definitions for specific data cohorts. Finding the problematic classes when there are 20+ classes in your dataset can be challenging and tedious, but that is also the scenario where the class overlap is most likely to occur. Try:
Ultimately, you will have to educate your labelers or change the classes by either merging them or creating new ones. ❗️ML Data Error #3: ImbalanceBy itself having an imbalanced dataset might not be a problem, but often model performance can correlate with that imbalance. ML practitioners tend to be trained to look at class imbalance but often don’t consider imbalances in the inputs themselves. For example, if a training dataset has hundred subtle variants of “I need to repair my Jeep” but only one example of “I need to repair my Honda,” a model might not have enough signal to associate a Honda with a car and not a microwave, which would matter if your two classes are “repair.car” and “repair.microwave”. Similar imbalances in the test dataset can lead to misleadingly inflated accuracy numbers. Detecting class or any other metadata column’s imbalance is pretty easy to do and should be a standard part of a practitioner’s workflow. However, detecting imbalances in the patterns of your data is generally pretty hard. Some things to try are:
Once you detect imbalance, it can be reduced by downsampling frequent samples and increasing the number of less frequent samples through data augmentation or using embeddings to select similar samples from an unlabeled dataset and labeling. ❗️ML Data Error #4: DriftA model deployed in production only knows to make predictions based on what you trained it on. As the real-world data “drifts”, the model’s predictions veer into uncharted territory. Covid, for example, broke facial recognition models with the usage of masks. New words enter our lexicon daily, and old terms can change their meanings drastically. This is especially problematic for datasets that are hand-curated. Drift can be detected by looking for changes in the model’s predictions or the composition of production data. A mature ML workflow would send some of the drifted data to labelers and use the results in a retraining job. ⚡️ Better Data in MinutesData-centric techniques are not well documented or taught in the industry. Finding data errors is a very time-consuming process. It can feel like finding a needle in a haystack. Except, you have a ton of needles in a really large haystack of tens or hundreds of thousands of data samples. We’ve built Galileo to take that pain away. We want to empower you to find and fix data errors in minutes instead of hours without worrying about the technical details. With Galileo, you can find likely mislabeled samples, samples with class overlap, drift in your real-world data, or you can use it to explore your data’s patterns through their embeddings! Try it out yourself by signing up for free here. We would love to discuss what you find in our Slack community. Here’s to building better models, faster, with better data! Team Galileo (PS: we are hiring!) *This post was written by Nikita Demor, a founding engineer at Galileo. We thank Galileo for their support of TheSequence.You’re on the free list for TheSequence Scope and TheSequence Chat. For the full experience, become a paying subscriber to TheSequence Edge. Trusted by thousands of subscribers from the leading AI labs and universities. |
Key phrases
Older messages
❇️ NVIDIA Continues Pushing AI’s Boundaries
Sunday, September 25, 2022
Weekly news digest curated by the industry insiders
👾 Edge#228: How Amazon is Improving BERT-Based Models Used in Alexa
Thursday, September 22, 2022
Recently Amazon Research published three papers about BERT-based models
📝 Guest post: Unlock the Power of BLOOM With the Broadest Range of GPUs Served On-Demand*
Wednesday, September 21, 2022
In this guest post, CoreWeave introduces BLOOM deploy on their platform and guides you through their arsenal of GPUs to ensure you find the compute that delivers the best possible performance-adjusted
📃➡️🖼 Edge#227: Autoregressive Text-to-Image Models
Tuesday, September 20, 2022
+Google's Parti; +MS COCO
🐦 Follow us on Twitter
Monday, September 19, 2022
Check how helpful it might be for you
You Might Also Like
⚙️ Zuck has entered the chat(bot)
Friday, April 19, 2024
Plus: AI video's coming to mobile!
Noonification: Just Made my First Dollar With My SaaS After Quitting my Job
Friday, April 19, 2024
Top Tech Content sent at Noon! Get Algolia: AI Search that understands How are you, @newsletterest1? 🪐 What's happening in tech this week: The Noonification by HackerNoon has got you covered with
From Not to Hot 🔥7 Practices to Land a Trending Story
Friday, April 19, 2024
Discover the Insider Secrets to Elevate Your Story's Success! 🚀 ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏
iOS Dev Weekly - Issue 657
Friday, April 19, 2024
What's the easiest and least hassle path to launching a custom app store in the EU? 🏪 View on the Web Archives ISSUE 657 April 19th 2024 Comment You probably already saw this week's
Vision Pro App Development, TipKit and more...
Friday, April 19, 2024
View in browser Hello, you're reading Infinum iOS Cocoa Treats, bringing you the latest iOS related news straight to your inbox every week. Readers Become Part of the Plot with Vision Pro App
Your 3 AI Incubator Tracks: Curriculum, Coaching, or 1-on-1 Mentorship
Friday, April 19, 2024
How to pick the one that's right for you
Logitech's AI Prompt Builder is surprisingly handy
Friday, April 19, 2024
Torvalds on evil devs and AI hype; Quest 2's price drops; Virtual cards explained -- ZDNET ZDNET Tech Today - US April 19, 2024 placeholder Logitech's free AI Prompt Builder is surprisingly
Tesla recalls nearly 4,000 Cybertrucks
Friday, April 19, 2024
After reports of malfunctioning accelerator pedals, Tesla is recalling Cybertrucks View this email online in your browser By Alex Wilhelm Friday, April 19, 2024 Welcome to TechCrunch AM! Today's
SWLW #595: My role as a founder CTO, AI Product Management, and more.
Friday, April 19, 2024
Weekly articles & videos about people, culture and leadership: everything you need to design the org that makes the product. A weekly newsletter by Oren Ellenbogen with the best content I found
The bill to ban TikTok is barreling ahead
Friday, April 19, 2024
The Morning After It's Friday, April 19, 2024. The bill that could ban TikTok in the United States inches closer to becoming law. The legislation passed the House of Representatives last month,