TheSequence - LLM Scaling Laws vs. Everything Else
Was this email forwarded to you? Sign up here LLM Scaling Laws vs. Everything ElseSundays, The Sequence Scope brings a summary of the most important research papers, technology releases and VC funding deals in the artificial intelligence space.Next Week in The Sequence:
You can subscribe below:📝 Editorial: Next Week in The Sequence: LLM Scaling Laws vs. Everything ElseThe prevailing mantra in recent years for large language models (LLMs) has been "bigger is better." Reality has demonstrated that LLMs truly shine at scale, revealing emerging capabilities that were not initially envisioned during their pretraining process. Nevertheless, in recent times, we have witnessed the emergence of counter theories suggesting that LLM growth is reaching a plateau, and a new generation of models will ultimately converge towards more manageable sizes. Techniques such as distillation, RAG, quantization, and, of course, data quality curation have been developed to empower smaller and more efficient models. The question arises: are the LLM scaling laws approaching their limits? Which school of thought is correct? While optimization methods undoubtedly contribute to the creation of more efficient, compact models, there is currently no empirical evidence to suggest that we are anywhere near experiencing diminishing returns within the realm of LLM scaling laws. Quite the contrary, new scaling frontiers remain perfectly attainable with the current generation of transformer architectures. Presently, the cost of pretraining a large-scale LLM typically ranges in the double-digit millions. In the near future, we may witness this cost escalate into the hundreds of millions or even billions. At such scales, LLMs are likely to exhibit properties that are difficult to fathom today. This trajectory can be accelerated by ongoing hardware breakthroughs, which seem to occur annually. Challenging the value of the scaling laws in the current generation of LLMs is not only foolish but also factually incorrect. Similar to any other phenomenon in physics, there will come a day when we reach the limits of the scaling laws. However, that day is not today. 📺 To Watch: Vector Database FundamentalsThese short videos explain vector index types like HNSW, ANNOY, IVF and vector similarity metrics including Euclidian, cosine, inner product and more. 🔎 ML ResearchWho is Harry Potter?Researchers from Microsoft published a paper exploring a technique to fine-tune LLMs to unlearn specific concepts. The paper evaluates the process tuning Llama2-7b to forget all knowledge of Harry Potter’s books —> Read more. LLaVAResearchers from the University of Wisconsin-Madison, Microsoft Research and Columbia University published a paer detailing LLaVA, an instruction-tuned language-vision model. LLaVA extends Vicuna with a vision encoder in a very similar architecture to GPT-4 Vision —> Read more. Stable SignatureMeta AI published a paper introducing Stable Signature, a method for watermarking generative AI images. Stable Signature incorporates information in the image that is invisible to the naked eye but that can be verified —> Read more. RAG vs. Large Context in LLMsResearchers from NVIDIA published a paper detailing a study that evaluates the performance of RAG vs. long context windows in LLMs. The research shows that 4k RAG-augmented models can achieve similar performance than 16k context models and other fascinating findings —> Read more. SCREWSAI researchers from ETH Zurich and Microsoft Semantic Machines present SCREWS, a new reasoning framework for LLMs. The techniques combines different reasoning building blocks such as sampling, conditional resampling, selection and several others —> Read more. 🤖 Cool AI Tech ReleasesSteerLMNVIDIA open sourced SteerLM, a framework for customizing LLMs during inference —> Read more. Zephyr-7BHugging Face unveiled Zephyr 7B, a fine-tuned version of Mistral that outperformed Llama-70B in across different benchmarks —> Read more. 🛠 Real World MLTrusted Notebooks at SalesforceSalesforce engineering discusses their access control solution for securing notebooks in data science workflows —> Read more. 📡AI Radar
You’re on the free list for TheSequence Scope and TheSequence Chat. For the full experience, become a paying subscriber to TheSequence Edge. Trusted by thousands of subscribers from the leading AI labs and universities. |
Older messages
📝 Guest Post: Retrieval Augmented Generation on Notion Docs via LangChain*
Thursday, October 19, 2023
In this guest post, Yujian Tang, a developer advocate at Zilliz, explores how to enhance Notion documents with language model interactions using LangChain and Milvus. He lays out a step-by-step guide
Edge 335: LoRA Fine-Tuning and Low-Rank Adaptation Methods
Thursday, October 19, 2023
Diving into one of the most popular fine-tuning techniques for foundation models.
The Sequence Chat: Shreya Rajpal, Co-Founder and CEO, Guardrails AI About Ensuring the Safety and Robustness of LL…
Thursday, October 19, 2023
The co-creator of one of the most important LLM guardrails frameworks shares her perspectives on building safe, robust and efficient LLM applications, the architecture of Guardrails AI and more.
Edge 333: Understanding Parameter Efficient Fine Tuning
Tuesday, October 10, 2023
An overview of PEFT, one of the most important fine-tuning methods ever created;
📡 WEBINAR: Unraveling prompt engineering
Monday, October 9, 2023
You can't talk about LLMs without talking about prompt engineering – and at first glance, prompting may appear intuitive and straightforward, but well, it ain't. Join us for the next webinar of
You Might Also Like
Import AI 399: 1,000 samples to make a reasoning model; DeepSeek proliferation; Apple's self-driving car simulator
Friday, February 14, 2025
What came before the golem? ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏
Defining Your Paranoia Level: Navigating Change Without the Overkill
Friday, February 14, 2025
We've all been there: trying to learn something new, only to find our old habits holding us back. We discussed today how our gut feelings about solving problems can sometimes be our own worst enemy
5 ways AI can help with taxes 🪄
Friday, February 14, 2025
Remotely control an iPhone; 💸 50+ early Presidents' Day deals -- ZDNET ZDNET Tech Today - US February 10, 2025 5 ways AI can help you with your taxes (and what not to use it for) 5 ways AI can help
Recurring Automations + Secret Updates
Friday, February 14, 2025
Smarter automations, better templates, and hidden updates to explore 👀 ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏
The First Provable AI-Proof Game: Introducing Butterfly Wings 4
Friday, February 14, 2025
Top Tech Content sent at Noon! Boost Your Article on HackerNoon for $159.99! Read this email in your browser How are you, @newsletterest1? undefined The Market Today #01 Instagram (Meta) 714.52 -0.32%
GCP Newsletter #437
Friday, February 14, 2025
Welcome to issue #437 February 10th, 2025 News BigQuery Cloud Marketplace Official Blog Partners BigQuery datasets now available on Google Cloud Marketplace - Google Cloud Marketplace now offers
Charted | The 1%'s Share of U.S. Wealth Over Time (1989-2024) 💰
Friday, February 14, 2025
Discover how the share of US wealth held by the top 1% has evolved from 1989 to 2024 in this infographic. View Online | Subscribe | Download Our App Download our app to see thousands of new charts from
The Great Social Media Diaspora & Tapestry is here
Friday, February 14, 2025
Apple introduces new app called 'Apple Invites', The Iconfactory launches Tapestry, beyond the traditional portfolio, and more in this week's issue of Creativerly. Creativerly The Great
Daily Coding Problem: Problem #1689 [Medium]
Friday, February 14, 2025
Daily Coding Problem Good morning! Here's your coding interview problem for today. This problem was asked by Google. Given a linked list, sort it in O(n log n) time and constant space. For example,
📧 Stop Conflating CQRS and MediatR
Friday, February 14, 2025
Stop Conflating CQRS and MediatR Read on: my website / Read time: 4 minutes The .NET Weekly is brought to you by: Step right up to the Generative AI Use Cases Repository! See how MongoDB powers your