Discussion about this post

User's avatar
Jonathan Smith's avatar

Great article! There is a lot of emphasis these days on the probabilistic nature of trained systems, but that may be more a means of getting mechanisms to emerge than being fundamental to the mechanisms these systems end up developing.

If you reverse-engineer a trained neural (deep learning) network, you do see patterns that correspond to components that are carrying out specific operations on data (looking for edges, shapes etc). That is to say even though we don’t plan and build the mechanism, training may be developing mechanisms that we could reverse-engineer that would help us understand what is going on. Trained algorithms as a tool of science even if not a product of science.

For a while in my career I worked with people who worked in one branch of symbolic AI called Case-Based Reasoning. The basic idea is that humans store a collection of memories of situations and stories, and then have ways of combining and adapting those stories in new situations.

I suspect that LLMs (ChatGTP etc) are doing something similar, looking for patterns in past cases and adapting them to new situations. The fact that LLMs can get so far on pattern matching and adaptation supports many insights from the CBR community. Humans seem to do a lot of remembering and adapting past cases. But also it seems not to account for all of human intelligence.

LLMs also sometimes fail. They are not great at reasoning, and sometimes invent new stuff without testing their suppositions against available evidence and background knowledge. A little like a bright enthusiastic High School student who has new and creative ideas about science but has not yet developed the discipline to take generated ideas as hypotheses rather than as knowledge.

Expand full comment
Tanj's avatar

There is a related idea in analyzing human intelligence. Given the long delay that can be demonstrated in conscious reaction, where we react before we are conscious of the input, a hypothesis is that we're running a predictive model which allows us to react in real time. Perhaps humans are a generative ai with after-the-fact corrections. This may explain those videos where we watch a basketball game and don't notice the guy in the gorilla suit in the background, because how would we generate that - and if it does not interact with the events we are interested in we just eliminate it as noise.

Expand full comment
2 more comments...

No posts