The Sequence Chat: Oren Etzioni – Allen AI, About Advancing Research in Foundation Models
Was this email forwarded to you? Sign up here The Sequence Chat: Oren Etzioni – Allen AI, About Advancing Research in Foundation ModelsAn AI legend discusses cutting edge research in foundation models.Dr. Oren Etzioni is also a Venture Partner at the Madrona Venture Group, and a Technical Director at the AI2 Incubator. He was the Founding CEO of the Allen Institute for AI. His awards include AAAI Fellow and Seattle’s Geek of the Year. He founded several startups including Farecast (acquired by Microsoft). Etzioni has written over 200 technical papers, garnering several awards including the ACL Test of Time Award in 2022. He has also authored commentary for The New York Times, Harvard Business Review, and Nature. Quick bio
I was a professor for most of my career focused on AI, NLP, and Web Search. I launched The Allen Institute of AI (AI2) in 2014 for the late Paul Allen and it’s grown to 250+ and over $100M in annual funding. I’ve always been fascinated by startups, having launched several AI-based startups over the years. AI2 has also created and spun out an incubator that is approaching $1B in the total valuation of startup financing rounds and acquisitions. 🛠 AI Work
Recently the PRIOR team at AI2 released Unified-IO, the first neural model to perform a large and diverse set of AI tasks spanning classical computer vision, image synthesis, vision-and-language, and natural language processing (NLP). We are continuing work in this area and will have a new version of this model available in the near future. We are also investing in generative language models with our new initiative AI2 OLMo, a uniquely open language model intended to benefit the research community by providing access and education around all aspects of model creation.
This work out of AI2’s Mosaic common sense AI team was inspired by the dual process theory of human cognition featured in the well-known book Thinking, Fast and Slow by Daniel Kahneman. This theory proposes two distinct human thinking systems: one characterized by rapid and intuitive thought, and another that emphasizes analytical and deliberate reasoning. Our goal is ultimately to be able to tackle intricate real-world problems using AI models in a much more cost-effective manner than is currently possible, so by building a model framework that could integrate both of these approaches, we thought we might be able to optimize an AI agent’s potential for planning complex interactive tasks while minimizing the cost of its reasoning. Many tasks do not require a deliberate and detailed analysis to perform successfully, so by leveraging a “fast”-thinking approach when appropriate, we thought we could improve the model’s speed and performance overall. We called the framework we designed “SwiftSage.” The “Swift” module is an encoder-decoder based language model designed to quickly process short-term memory content such as previous actions, current observations, and the environment state, which simulates the fast, intuitive thinking characteristics found in the “Fast” mode of thinking. The second module, “Sage,” represents the deliberate process of the second mode of thinking by harnessing the power of large language models (LLMs) like GPT-4. A heuristic algorithm plays a crucial role in the system by determining when our framework should activate or deactivate the Sage module. Our intuitions about this approach were correct; thanks to its dual-system design for fast and slow thinking, SwiftSage dramatically reduces the number of tokens necessary for each action in LLM inference, making it more cost-effective and efficient than the next best system.
AI2 OLMo is a language model currently under development at AI2 that is being deliberately designed to support research by providing access to every element of the system we create, from data to training and evaluation code to model weights. We recognize a real dearth of accessible, understandable language models which is holding the AI research community back from understanding and advancing this critical new technology. The most powerful language models today are released by for-profit organizations and with limited insight into the data and methods used to create them; we aim to change that with OLMo. Our goal is to democratize access to systems like these and advance their development and safety for everyone.
While the performance of modern LLMs is stunning, it is hard to tell how they arrive at their answers, and if their internal reasoning even makes sense. (There have been numerous cases where LLMs' poor reasoning and hallucinations led people into trouble). We have been developing techniques (e.g., Entailer, Reflex) for uncovering the LLMs' "beliefs", and how its answers follow from them via systematic chains of reasoning, which can then be viewed by a user - in other words, providing users a view of the "mental model" that the LLM has about a current problem. This allows us to spot and reject answers derived from faulty chains of reasoning or faulty beliefs, helping engender trust in the model's answers. And it paves the way for users to correct any erroneous LLM beliefs that were uncovered, as users can now see relevant model beliefs, allowing them to teach the system when it goes wrong and improve over time. In addition, modern LLMs still struggle with complex tasks that require multiple, specialized steps to be chained together, for example mathematics (where different math operations are needed), or other data manipulation queries, e.g., "What do the third letters of the words in 'John Greg Calvin Melville Rhon' spell?" (answer: "hello"). However, while LLMs struggle to answer such questions directly, they are proficient at breaking up complex tasks into smaller tasks - a process called task decomposition. And in many cases, the LLM itself can solve those smaller tasks. We have recently developed a technique called Decomposed Prompting to control this process, so the LLM repeatedly decomposes problems where it is stuck, and solves the simpler pieces when it is not, putting a new class of previously unsolvable problems within reach.
Beaker was first launched internally at AI2 in 2017 as an experimentation platform for AI2 researchers to run jobs on the cloud and organize their experiments. Beaker made it substantially easier to manage cloud instances and run large-scale jobs over thousands of nodes, as well as to make sense of the large volume of experiments at AI2. With the rise of deep learning, Beaker evolved to primarily support GPU jobs and manage workloads across our dedicated GPU cluster. 💥 Miscellaneous – a set of rapid-fire questions
Building software agents that utilize tools and learn on the user’s behalf.
It would certainly be a huge benefit to humanity as we fight climate change, the next pandemic, and more.
The transformer is a simple method that is highly scalable. I do believe we will see new architectures in the next three years but will be a bit coy in predicting which one is next.
I think we will see balance—just as we’ve seen in operating systems between Windows and Linux. There will both open and closed models. You’re on the free list for TheSequence Scope and TheSequence Chat. For the full experience, become a paying subscriber to TheSequence Edge. Trusted by thousands of subscribers from the leading AI labs and universities. |
Older messages
Edge 319: The Factors Behind In-Context Learning
Tuesday, August 22, 2023
What are the key elements that cause the emergece of in-context-learning in language models and how are they unique.
📺 Webinar: Create better features for your ML models
Monday, August 21, 2023
Transforming raw data into features to power machine learning models is one of the biggest challenges in production ML. Join Tecton CEO, Mike Del Balso, for this webinar on Thursday, August 31, at 9 AM
The NVIDIA GPU Scarcity Madness
Sunday, August 20, 2023
Sundays, The Sequence Scope brings a summary of the most important research papers, technology releases and VC funding deals in the artificial intelligence space.
📍 Hands-On Lab Next Week: Learn how to build great ML features and deploy them to production quickly and reliably
Friday, August 18, 2023
Build and deploy batch, streaming, and real-time ML features in just a few minutes
Inside LLM-AUGMENTER: Microsoft Research’s Reference Architecture to Extend LLMs with Memory, Knowledge, and Exter…
Thursday, August 17, 2023
The architecture showcases the key building blocks of production-ready LLMs.
You Might Also Like
Weekend Reading — More time to write
Sunday, November 24, 2024
More Time to Write A fully functional clock that ticks backwards, giving you more time to write. Tech Stuff Martijn Faassen (FWIW I don't know how to use any debugger other than console.log) People
🕹️ Retro Consoles Worth Collecting While You Still Can — Is Last Year's Flagship Phone Worth Your Money?
Saturday, November 23, 2024
Also: Best Outdoor Smart Plugs, and More! How-To Geek Logo November 23, 2024 Did You Know After the "flair" that servers wore—buttons and other adornments—was made the butt of a joke in the
JSK Daily for Nov 23, 2024
Saturday, November 23, 2024
JSK Daily for Nov 23, 2024 View this email in your browser A community curated daily e-mail of JavaScript news React E-Commerce App for Digital Products: Part 4 (Creating the Home Page) This component
Not Ready For The Camera 📸
Saturday, November 23, 2024
What (and who) video-based social media leaves out. Here's a version for your browser. Hunting for the end of the long tail • November 23, 2024 Not Ready For The Camera Why hasn't video
Daily Coding Problem: Problem #1617 [Easy]
Saturday, November 23, 2024
Daily Coding Problem Good morning! Here's your coding interview problem for today. This problem was asked by Microsoft. You are given an string representing the initial conditions of some dominoes.
Ranked | The Tallest and Shortest Countries, by Average Height 📏
Saturday, November 23, 2024
These two maps compare the world's tallest countries, and the world's shortest countries, by average height. View Online | Subscribe | Download Our App TIME IS RUNNING OUT There's just 3
⚙️ Your own Personal AI Agent, for Everything
Saturday, November 23, 2024
November 23, 2024 | Read Online Subscribe | Advertise Good Morning. Welcome to this special edition of The Deep View, brought to you in collaboration with Convergence. Imagine if you had a digital
Educational Byte: Are Privacy Coins Like Monero and Zcash Legal?
Saturday, November 23, 2024
Top Tech Content sent at Noon! How the world collects web data Read this email in your browser How are you, @newsletterest1? 🪐 What's happening in tech today, November 23, 2024? The HackerNoon
🐍 New Python tutorials on Real Python
Saturday, November 23, 2024
Hey there, There's always something going on over at Real Python as far as Python tutorials go. Here's what you may have missed this past week: Black Friday Giveaway @ Real Python This Black
Re: Hackers may have stolen everyone's SSN!
Saturday, November 23, 2024
I wanted to make sure you saw Incogni's Black Friday deal, which is exclusively available for iPhone Life readers. Use coupon code IPHONELIFE to save 58%. Here's why we recommend Incogni for