Back to Reading List
[Scaling]·PAP-MJCGEH·2022·March 17, 2026

Emergent Abilities of Large Language Models

2022

Jason Wei, Yi Tay, Rishi Bommasani et al.

4 min readScalingReasoning

Core Insight

Larger language models develop unexpected skills, challenging our predictions and scaling strategies.

In Plain English

The paper reveals that exhibit absent in smaller ones, defying performance predictions. These findings suggest that mere scaling introduces novel capabilities that smaller models can't achieve.

Knowledge Prerequisites

git blame for knowledge

To fully understand Emergent Abilities of Large Language Models, trace this dependency chain first. Papers in our library are linked — click to read them.

DIRECT PREREQIN LIBRARY
Attention Is All You Need

Understanding the attention mechanism is fundamental for grasping how large language models work since they rely heavily on transformer architectures introduced in this paper.

Attention mechanismTransformer architectureSelf-attention
DIRECT PREREQIN LIBRARY
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding

This paper introduces Bidirectional Encoder Representations from Transformers, which is a foundational large language model that shows how pre-training on vast textual data can improve language understanding tasks.

Bidirectional transformersPre-trainingMasked language models
DIRECT PREREQIN LIBRARY
Training language models to follow instructions with human feedback

Understanding instruction following through human feedback is crucial for realizing how large language models can be fine-tuned to improve task performance based on human-provided feedback.

Fine-tuningHuman feedbackInstruction-following
DIRECT PREREQIN LIBRARY
Toolformer: Language Models Can Teach Themselves to Use Tools

This paper explores how large language models can utilize external tools to enhance their capabilities, a concept that is potentially linked to the emergent abilities described.

Tool use in language modelsSelf-improvementExternal API integration
DIRECT PREREQIN LIBRARY
Tree of Thoughts: Deliberate Problem Solving with Large Language Models

Understanding structured thinking processes in language models will provide insights into how these models develop emergent problem-solving abilities.

Tree structuresDeliberate reasoningProblem-solving strategies

YOU ARE HERE

Emergent Abilities of Large Language Models

The Idea Graph

The Idea Graph
11 nodes · 15 edges
Click a node to explore · Drag to pan · Scroll to zoom
476 words · 3 min read6 sections · 11 concepts

Table of Contents

01

The Problem: Incomplete Predictive Models

86 words

Before this research, the expectations from language models were largely defined by existing . These laws suggested that merely increasing the size of models would enhance known capabilities. However, they did not account for the emergence of , which became evident with larger models. The limitations of these predictive models meant that researchers could not foresee the full range of abilities that scaling could unlock. This gap between expectation and reality highlighted a significant problem in understanding the true potential of large language models.

02

Key Insight: Emergent Abilities

85 words

The core insight of this paper is the discovery of that appear in large language models. These abilities are not simply enhanced versions of existing skills but are entirely new capabilities that arise from scaling. The presence of such abilities challenges the existing understanding and predictive models in the field, suggesting that there is more to model scaling than previously thought. This insight opens up new avenues for research and application, as it demonstrates the potential for language models to develop unforeseen skills.

03

Method: Model Scaling and Parameter Experiments

82 words

To explore the potential of large language models, the research involved scaling these models by increasing their parameters from millions to hundreds of billions. This process, known as , was central to the experiments conducted in the study. By varying the size of the models, the researchers were able to observe differences in performance and identify new abilities. The were designed to not only test known capabilities but also to uncover emergent abilities that only appeared in larger models.

04

Results: Linguistic and Reasoning Capabilities

74 words

The experiments revealed that large language models exhibit enhanced , including nuanced language understanding that smaller models could not achieve. Additionally, these models demonstrated that were surprising in their complexity and depth. One of the most striking findings was the models' ability to engage in creative problem-solving, which was not anticipated by existing scaling laws. These results underscore the idea that scaling introduces novel abilities rather than merely improving existing ones.

05

Impact: Enhanced Human-Computer Interaction

75 words

The discovery of in large language models has significant implications for the future of . By leveraging these newfound capabilities, companies like OpenAI and Google can create more natural and sophisticated AI-driven interfaces. These improvements could lead to better chatbots and virtual assistants, enhancing user experience by making interactions more seamless and intuitive. The potential for these models to redefine is substantial, offering new opportunities for innovation in the field.

06

Limitations & Open Questions: Managing Unpredictability

74 words

While the emergent abilities of large language models offer exciting opportunities, they also introduce . This poses challenges for developers aiming to ensure a consistent and reliable user experience. As these models continue to evolve, rigorous testing and careful management will be essential to harness their potential without compromising functionality. Open questions remain about how to best predict and control these emergent abilities, ensuring that advancements in AI are both beneficial and manageable.

Experience It

Live Experiment

Emergent Abilities

See Emergent Abilities in Action

You'll see how larger language models develop unexpected skills that smaller models lack, highlighting the impact of scaling.

Look for nuanced understanding and unexpected skills in the larger model's responses, as highlighted in the paper.

Try an example — see the difference instantly

⌘↵ to run

How grounded is this content?

Metrics are computed from available source text only — abstract, summary, and impact fields ingested into this system. Full paper PDF is not ingested; numerical claims that originate from within the paper body will not appear in these scores.

Source Richness88%

7 of 8 content fields populated. More fields = better-grounded generation.

Source Depth~225 words

Total source text analyzed by the model. Includes extended deep-dive summary — high confidence.

Methodology: Number grounding uses regex digit extraction against source text. Quote traceability uses token set intersection on content words stripped of stop-words. Neither metric validates semantic correctness or factual accuracy against the original paper. For full verification, cross-reference with the original paper via the arXiv link above.