New Theory Suggests Chatbots Can Understand Text

Far from being “stochastic parrots,” the biggest large language models seem to learn enough skills to understand the words they’re processing.

[…]

A trained and tested LLM, when presented with a new text prompt, will generate the most likely next word, append it to the prompt, generate another next word, and continue in this manner, producing a seemingly coherent reply. Nothing in the training process suggests that bigger LLMs, built using more parameters and training data, should also improve at tasks that require reasoning to answer.

But they do. Big enough LLMs demonstrate abilities — from solving elementary math problems to answering questions about the goings-on in others’ minds — that smaller models don’t have, even though they are all trained in similar ways.

“Where did that [ability] emerge from?” Arora wondered. “And can that emerge from just next-word prediction?” —Quanta Magazine

Post was last modified on 26 Jan 2024 10:26 am

Share
Published by
Dennis G. Jerz
Tags: ai

Recent Posts

Another corner building. Designed and textured. Needs an interior. #blender3d #design #aesthetics #medievalyork #mysteryplay

Another corner building. Designed and textured. Needs an interior. #blender3d #design #aesthetics #medievalyork #mysteryplay

39 minutes ago

There’s No Longer Any Doubt That Hollywood Writing Is Powering AI

Two years after the release of ChatGPT, it may not be surprising that creative work…

2 days ago

The complex geometry on this wedge building took me all weekend.  #blender3d #medievalyork #mysteryplay #cgi #aesthetics #design

The complex geometry on this wedge building took me all weekend. The interior walls still…

4 days ago

Sesame Street had a big plot twist in November 1986

My older siblings say they remember our mother sitting them down to watch a new…

5 days ago