
Recent studies and discussions in the field of artificial intelligence have highlighted significant advancements in the capabilities of Large Language Models (LLMs). Researchers, including Beren Millidge, have pointed out that the parameter counts of human-level vision and language models are closer than previously thought, with current state-of-the-art models comprising 70 billion (70B SoTA) parameters, closely approximating the majority of human capabilities. Further investigations into the knowledge capacity of these models, as conducted by Z Allen-Zhu, Y Li of Meta & Mohamed bin Zayed University of AI (2024), have established that LLMs can store 2 bits of knowledge per parameter, even when quantized to int8, with around 1000 exposures. This finding was corroborated across multiple controlled datasets, indicating that even a 7 billion parameter model can effectively store 14 billion bits of knowledge. The studies also delve into how the size, training, architecture, and data quality of these models affect their capacity to store and utilize knowledge for downstream applications. Additionally, there is an observed trend where the persuasive capabilities of these models are increasing in correlation with their size, suggesting that they are becoming as persuasive as human writers without showing signs of plateauing.

We've reached the point where models are becoming as persuasive as human writers. There's a clear model size x persuasion scaling trend, and it doesn't look like it is plateauing. https://t.co/HWuvW2g43R
New study finds Large Language Models store 2 bits of knowledge per parameter, showing how size, training, architecture & data quality affect their capacity: https://t.co/uv1OrquLEk https://t.co/ncblHYO89C
The Physics of Language Models Investigates knowledge capacity scaling laws where it evaluates a model’s capability via loss or benchmarks, to estimate the number of knowledge bits a model stores. Quote from the paper: "Language models can and only can store 2 bits of knowledge… https://t.co/koFMZJPq4t