Overparametrized LLM: COMPLEX Reasoning (Yale Univ)
Discover AI Discover AI
40.5K subscribers
4,700 views
206

 Published On Oct 5, 2024

Brand new AI research, published by Yale Univ et al, explores the emergence of intelligence in artificial systems, with a particular emphasis on overparameterized large language models (LLMs) trained on datasets derived from elementary cellular automata (ECA). It posits that exposure to complex yet structured datasets can facilitate the development of intelligence, even in models that are not inherently designed to process explicitly intelligent data. The authors employ ECA rules, specifically from Classes I-IV, to generate training data and evaluate LLM performance on downstream tasks.

The results indicate that models trained on rules operating near the "edge of chaos" (Class IV) demonstrate superior reasoning and chess move prediction capabilities compared to those trained on strictly ordered or purely chaotic data. These findings support the hypothesis that complexity—balanced between order and randomness—fosters the emergence of more sophisticated, generalized behavioral patterns in these models. Furthermore, training on such datasets appears to induce the development of intricate internal representations, as evidenced by attention mechanisms that effectively leverage historical context.

The methodology involves training modified GPT-2 models on ECA-generated datasets, with adaptations to handle binary inputs via linear projection layers. The study employs various complexity measures, including Lempel-Ziv complexity, Lyapunov exponents, and Krylov complexity, to characterize the ECA-generated data. Lempel-Ziv and compression complexities quantify the compressibility of the datasets, while Lyapunov exponents provide insights into the chaotic or stable dynamics of the generated sequences.

The findings suggest that models with overparameterized architectures can naturally explore non-trivial solutions, utilizing their excess capacity to form sophisticated representations of the input space, thereby elucidating their emergent reasoning capabilities. These results underscore that the emergence of intelligence in LLMs is not solely contingent on the nature of the data itself, but rather on the inherent complexity of the data, particularly when situated at the critical juncture between order and chaos.


All rights w/ authors:
INTELLIGENCE AT THE EDGE OF CHAOS
https://arxiv.org/pdf/2410.02536

00:00 Intelligence at the Edge of Chaos
02:47 Elementary Cellular Automata Complexity Data
03:43 GPT-4o canvas calculates Cellular Automata Complexities
07:32 Rule 30 vs Rule 90 vs Rule 108
10:12 GPT-4o codes Game of Life Automaton
12:30 Analyze the Findings (complexity and reasoning)
14:09 Overparametrization leads to non-trivial Solutions in LLMs
17:00 Complexity measures
18:00 Concept of Emergence of Intelligence
24:57 Edge of Chaos is crucial for Emergence of Intelligence in LLMs
26:32 How Intelligence Emerges in AI


#airesearch
#emergence
#aiagents
#intelligence

show more

Share/Embed