Discussion about this post

User's avatar
Gregoreite Roberts's avatar

this is one of your better posts. I don't fully agree with you on the "averaging" interpretation... that's a bit too simplistic, and is the same error I made when first judging MidJourney. And yet... I can say, it has challenges with anomalies and outliers. You repeat the misconception -- *intentionally*, you *know* this to be false -- that GPT has "memorized" the internet. Two clarifications:

a) The training dataset comprises significantly less than 1/3rd of the internet. And certainly (at this point) does not include video, which is a massive store of untapped information.

b) It isn't, as we now understand, memorization. Its fractal compression. Its pattern recognition. Its much much much more similar to the highly imperfect mechanism of human memory than it is like storing to a database or a hard drive with error-correction and fault-tolerance. From my understanding, GPT's method of "memory" is basically reconstructing context from pattern that was "burned in" to its neural net while digesting the training dataset and then re-re-inforced with months of RLHF. So it's much much more like reconstructive, symbolic human memory -- stories grown from "idea seeds," abstract relations of disparate concepts, strange triggers (smell) to expand massive sensory concepts (that day we met) -- than it is to literal bit-for-bit file storage.

Expand full comment
Daniel Nest's avatar

Another great read, Alberto!

The way ChatGPT appears to fill in people's deviating life paths reminds me of the fact that our own brains act in a similar way when it comes to how we percieve the world. There's the famous fact that our eyes have "blind spots" where they literally can't see, which the brain helpfully fills in with what it predicts should be there.

Then there's this relatively recent research showing that our brains tends to first spot the borders of objects and then fill in--or "color in"--the surface area (https://www.sciencedaily.com/releases/2007/08/070820135833.htm)

This quote by one of the professors is telling: "...a lot of what you perceive is actually a construction in your brain of border information plus surface information—in other words, a lot of what you see is not accurate."

I just find it curious how a large language model that's said to mimic our reasoning process ends up inadvertently acting like our brains in yet another way.

Expand full comment
22 more comments...

No posts