The Most Open Open Source Generative AI Release
Was this email forwarded to you? Sign up here The Most Open Open Source Generative AI ReleaseAllenAI just released all the components of its OLMo LLM model.Next Week in The Sequence:
You can subscribe below!📝 Editorial: The Most Open Open Source Generative AI ReleaseOpen source innovation is one of the most impressive aspects of the recent generative AI revolution. We constantly hear about new models being open-sourced to rival closed alternatives such as OpenAI or Anthropic. However, you might be surprised to learn that most of these releases are not completely open source. A more intriguing question is what open source means in the context of foundation models. It's important to understand that, when it comes to foundation models, the source code itself is very small and quite similar from one model to another. When you hear 'open source' in the context of foundation models, most of the time it refers to the weights of the models, which are huge files containing the neural network's connectivity structure. Understanding the weights of a foundation model is nearly impossible, so the value of making them open relies on reproducibility. Other aspects of a foundation model, such as source code, training pipeline, evaluation code, and fine-tuning code (for instance, in instruction-following models), remain closed. So, let's just say that we have been using the term 'open source' a bit lightly in generative AI, to say the least. Last week, researchers from the Allen Institute for Artificial Intelligence (AllenAI) released all the components of its OLMo LLM in a truly open fashion. You could call this release the most open open-source release in generative AI. The release includes:
Needless to say, getting OLMo to work and reproduce the results claimed in its technical report is way simpler than with other models. Let’s hope more open source AI proponents follow this practice. Let’s build a real open source generative AI. 🔎 ML ResearchFuyu-HeavyAdept published details about Fuyu-Heavy, the newest version of its multimodal model optimized for autonomous agents scenarios. The model scores in close proximity to GPT4-V and Gemini Ultra being 10-20 times smaller —> Read more. BootPIGSalesforce Research published a paper detailing BootPIG, a model architecture and training pipeline for subject driven image generation. BootPIG extends text-to-image models with additional layers that allow them to accept new images during text time —> Read more. MobileDiffusionGoogle Research published a paper detailing MobileDiffusion, a text-to-image model optimized for mobile devices. The model relies on techniques such as DiffusionGAN to achieve and can generate 512x512 images in half a second —> Read more. Multimodal LLMsResearchers from Tencent AI Labs and Kyoto University published a paper detailing recent advancements in multimodal LLMs. The paper reviews the architecture, training models and recent developments in over 20 multimodal LLMs —> Read more. Time Series Forecasting DecoderGoogle Research published a paper introducing TimeFM, a foundation model for time series forecasting. TimeFM has been pretrained in 100 billion time points and is based on 200M parameters —> Read more. 🤖 Cool AI Tech ReleasesOLMoResearchers from the Allen Institute for AI open sourced the code, training data and evaluation code for their OLMo LLM —> Read more. PyTorch 2.2A new version of PyTorch has been released with several incremental updates —> Read more. llmwareA new open source framework for and models optimized for enterprise AI patterns such as RAG-optimizations or semantic search —> Read more. Image FX and MusicFXGoogle released released previews of two generative AI tools: ImageFX and MusicFX —> Read more. 🛠 Real World MLFast ML at MetaMeta engineers discuss some of the practices for packaging and distribution used to optimized their ML models —> Read more. 📡AI Radar
You’re on the free list for TheSequence Scope and TheSequence Chat. For the full experience, become a paying subscriber to TheSequence Edge. Trusted by thousands of subscribers from the leading AI labs and universities. |
Older messages
Edge 366: Anthropic's Sleeper Agents Explore How LLMs can be Deceptive
Thursday, February 1, 2024
One of the most important recent papers in generative AI.
The Sequence Pulse: The ML Architecture Powering LinkedIn's Skills Graph
Wednesday, January 31, 2024
Using transformer models to map jobs to job seekers.
Edge 365: Understanding LLM Reasoning with Reflexion
Tuesday, January 30, 2024
A deep dive into one of the most complete LLM reasoning methods.
💡WEBINAR: Beyond fine-tuning. Approaches in LLM optimization
Monday, January 29, 2024
We've talked about tuning, and we've talked about prompt engineering, but those are not the only techniques at our disposal to optimize LLMs. Join us for the next webinar of our LLM series on 📅
The LLMcorns: 4 New Billion Dollar Gen AI Valuations in One Week
Sunday, January 28, 2024
LLM providers are still commanding remarkable valuations in this fundraising climate.
You Might Also Like
Import AI 399: 1,000 samples to make a reasoning model; DeepSeek proliferation; Apple's self-driving car simulator
Friday, February 14, 2025
What came before the golem? ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏
Defining Your Paranoia Level: Navigating Change Without the Overkill
Friday, February 14, 2025
We've all been there: trying to learn something new, only to find our old habits holding us back. We discussed today how our gut feelings about solving problems can sometimes be our own worst enemy
5 ways AI can help with taxes 🪄
Friday, February 14, 2025
Remotely control an iPhone; 💸 50+ early Presidents' Day deals -- ZDNET ZDNET Tech Today - US February 10, 2025 5 ways AI can help you with your taxes (and what not to use it for) 5 ways AI can help
Recurring Automations + Secret Updates
Friday, February 14, 2025
Smarter automations, better templates, and hidden updates to explore 👀 ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏
The First Provable AI-Proof Game: Introducing Butterfly Wings 4
Friday, February 14, 2025
Top Tech Content sent at Noon! Boost Your Article on HackerNoon for $159.99! Read this email in your browser How are you, @newsletterest1? undefined The Market Today #01 Instagram (Meta) 714.52 -0.32%
GCP Newsletter #437
Friday, February 14, 2025
Welcome to issue #437 February 10th, 2025 News BigQuery Cloud Marketplace Official Blog Partners BigQuery datasets now available on Google Cloud Marketplace - Google Cloud Marketplace now offers
Charted | The 1%'s Share of U.S. Wealth Over Time (1989-2024) 💰
Friday, February 14, 2025
Discover how the share of US wealth held by the top 1% has evolved from 1989 to 2024 in this infographic. View Online | Subscribe | Download Our App Download our app to see thousands of new charts from
The Great Social Media Diaspora & Tapestry is here
Friday, February 14, 2025
Apple introduces new app called 'Apple Invites', The Iconfactory launches Tapestry, beyond the traditional portfolio, and more in this week's issue of Creativerly. Creativerly The Great
Daily Coding Problem: Problem #1689 [Medium]
Friday, February 14, 2025
Daily Coding Problem Good morning! Here's your coding interview problem for today. This problem was asked by Google. Given a linked list, sort it in O(n log n) time and constant space. For example,
📧 Stop Conflating CQRS and MediatR
Friday, February 14, 2025
Stop Conflating CQRS and MediatR Read on: my website / Read time: 4 minutes The .NET Weekly is brought to you by: Step right up to the Generative AI Use Cases Repository! See how MongoDB powers your