The Sequence Chat: Emmanuel Turlay – CEO, Sematic
Was this email forwarded to you? Sign up here The Sequence Chat: Emmanuel Turlay – CEO, SematicModel orchestration, Airflow limitaitons in ML and new ideas about MLOps.👤 Quick bio
I’m Emmanuel, CEO and founder of Sematic . I started my career in academia doing particle physics research on the Large Hadron Collider at CERN. After my post-doc I went to work for a string of small European startups before moving to the US in 2014 and joining Instacart where I led engineering teams dealing with payments and orders, and dabbled in MLOps. In 2018, I joined Cruise and cofounded the ML Infrastructure team there. We built many critical platform systems that enabled the ML teams to develop and ship models much faster, which contributed to the commercial launch of robotaxis in San Francisco in 2022. In May 2022, I started Sematic to bring my experience in ML infrastructure to the industry in an open-source manner. 🛠 ML Work
At Cruise, we noticed a wide gap between the complexity of cloud infrastructure, and the needs of the ML workforce. ML Engineers want to focus on writing Python logic, and visualizing the impact of their changes quickly. On the other hand, leadership at Cruise wanted to enable almost weekly retraining with newly labeled data to improve model performance very quickly (and beat Waymo to commercial launch). This required large end-to-end pipelines. The vision for Sematic is to give all ML teams access to the type of orchestration platform previously only available to a few large organizations that built it in-house with large dedicated platform teams. By abstracting away infrastructure and guaranteeing things like visualizations, traceability, and reproducibility out of the box, we have noticed an 80% speed-up in development time and retraining time.
Production-ready ML pipelines should have the following characteristics:
Airflow is a fantastic tool but it is not adapted for Machine Learning work for three reasons:
The first thing to do is to leverage caching. When iterating on pipelines, it’s common that certain things do not change between executions. For example, when iterating on training, it is unnecessary to rerun data preparation. Sematic can hash inputs to detect changes and only run functions whose inputs are different, enabling fast iterations. Secondly, leverage heterogeneous compute. Not all pipeline steps need the same compute resources, and using the largest VMs possible for all tasks is not cost-effective. Sematic lets users specify for each pipeline steps what resources are needed (e.g. high-memory for data processing, GPUs for train/eval, small VMs to extract reports, etc.), and will allocate them accordingly at runtime. Thirdly, without dedicated attention, GPUs will often sit idle while data is being downloaded and loaded into memory. Optimizing data streaming into training frameworks (e.g. Pytorch dataloaders) is critical to making sure GPUs are maximally utilized and money can be used to scale instead of paying for idle resources. Finally, distributed compute can dramatically speed up your pipelines. Whether it is for data processing (e.g. map/reduce tasks) or training (distributed training), execution times can be cut by as many times as there are nodes available in your cluster. Sematic’s Ray integration enables spinning up and down Ray cluster at runtime with a couple of lines of Python code. This pattern also solves dependency packaging which is clunky in Ray.
Great question.
💥 Miscellaneous – a set of rapid-fire questions
I have been quite interested in so-called model collapse. The idea is that if large foundational models are trained on large amounts of public data from the internet, and if more and more online content is AI-generated, models will essentially train on their own data. Some studies proved that this leads to a collapse of the long tail of freak events that is present in human-generated data (e.g. experimental art, ground-breaking concepts and opinions, marginal content, etc.), leading to more conformity and less innovation. I call this model inbreeding.
Unlike traditional supervised ML, training foundational models will not become mainstream, for sheer scale, cost, and expertise reasons. I can see how every Fortune 500 company in 5 years will do some amount of deep learning (e.g. YOLO for industrial quality control) or fine-tuning (e.g. fine-tune LLama on private data), but I don’t think they will all train their own Falcon from scratch. Therefore, I think Big AI will have their own tools dedicated to large-scale foundational model training, while the rest of the industry will still need traditional MLOps tools. However, what is going to emerge are tools around LLM orchestration. Langchain is the first of those, and more will come. Essentially creating dynamic DAGs of operations such as prompt templating, model inferencing, ML-powered expert model selection, etc. But these will be real-time pipelines that will have to run within the milliseconds between user input and the expected feedback.
The idea of CI should be used in ML (e.g. regression testing), and CD as well (recurrent retraining with new data). However the usual tools (e.g. Buildkite, Circle CI) are not suitable because they lack visualization and traceability of assets. These tools will essentially give you a log trace of the job, but will not enable outputting plots, dataframe, et cetera.
In my experience, large platforms cater to large enterprise companies, that’s where the money is. The products are sometimes inferior, but are extremely well marketed to reassure CIOs of Fortune 500 companies. Especially when those have spend commitments with those platforms. It’s a common pattern in tech that indie challengers build their businesses on a more avant-garde customer base, and as they grow, shift upmarket towards enterprise and sometimes get acquired by Big Tech. GitHub is a good example. Luckily Microsoft seems smart enough to “keep GitHub cool”, but for how long? You’re on the free list for TheSequence Scope and TheSequence Chat. For the full experience, become a paying subscriber to TheSequence Edge. Trusted by thousands of subscribers from the leading AI labs and universities. |
Older messages
Meet LMQL: An Open Source Query Language for LLMs
Sunday, July 16, 2023
Developed by ETH Zürich, the language explores new paradigms for LLM programming.
Some Key Facts About Anthropic’s Claude 2 Release
Sunday, July 16, 2023
Sundays, The Sequence Scope brings a summary of the most important research papers, technology releases and VC funding deals in the artificial intelligence space.
Edge 307: Learning About Program-Aided Language Models
Tuesday, July 11, 2023
LLMs that generate code for intermediate steps in a target task and the NLP Test framework.
“The Other” Enterprise Generative AI Platforms
Sunday, July 9, 2023
Sundays, The Sequence Scope brings a summary of the most important research papers, technology releases and VC funding deals in the artificial intelligence space.
💡Webinar: Designing & Scaling FanDuel's Machine Learning Platform
Friday, July 7, 2023
Discover FanDuel's journey in building a powerful ML platform for personalized experiences. Join the webinar on July 11 at 9 am PT to learn how they scaled their platform and lessons learned.
You Might Also Like
Import AI 399: 1,000 samples to make a reasoning model; DeepSeek proliferation; Apple's self-driving car simulator
Friday, February 14, 2025
What came before the golem? ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏
Defining Your Paranoia Level: Navigating Change Without the Overkill
Friday, February 14, 2025
We've all been there: trying to learn something new, only to find our old habits holding us back. We discussed today how our gut feelings about solving problems can sometimes be our own worst enemy
5 ways AI can help with taxes 🪄
Friday, February 14, 2025
Remotely control an iPhone; 💸 50+ early Presidents' Day deals -- ZDNET ZDNET Tech Today - US February 10, 2025 5 ways AI can help you with your taxes (and what not to use it for) 5 ways AI can help
Recurring Automations + Secret Updates
Friday, February 14, 2025
Smarter automations, better templates, and hidden updates to explore 👀 ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏
The First Provable AI-Proof Game: Introducing Butterfly Wings 4
Friday, February 14, 2025
Top Tech Content sent at Noon! Boost Your Article on HackerNoon for $159.99! Read this email in your browser How are you, @newsletterest1? undefined The Market Today #01 Instagram (Meta) 714.52 -0.32%
GCP Newsletter #437
Friday, February 14, 2025
Welcome to issue #437 February 10th, 2025 News BigQuery Cloud Marketplace Official Blog Partners BigQuery datasets now available on Google Cloud Marketplace - Google Cloud Marketplace now offers
Charted | The 1%'s Share of U.S. Wealth Over Time (1989-2024) 💰
Friday, February 14, 2025
Discover how the share of US wealth held by the top 1% has evolved from 1989 to 2024 in this infographic. View Online | Subscribe | Download Our App Download our app to see thousands of new charts from
The Great Social Media Diaspora & Tapestry is here
Friday, February 14, 2025
Apple introduces new app called 'Apple Invites', The Iconfactory launches Tapestry, beyond the traditional portfolio, and more in this week's issue of Creativerly. Creativerly The Great
Daily Coding Problem: Problem #1689 [Medium]
Friday, February 14, 2025
Daily Coding Problem Good morning! Here's your coding interview problem for today. This problem was asked by Google. Given a linked list, sort it in O(n log n) time and constant space. For example,
📧 Stop Conflating CQRS and MediatR
Friday, February 14, 2025
Stop Conflating CQRS and MediatR Read on: my website / Read time: 4 minutes The .NET Weekly is brought to you by: Step right up to the Generative AI Use Cases Repository! See how MongoDB powers your