“One year of research in neural networks is sufficient to believe in God.” The writing on the wall of John Hopfield’s lab at Caltech made no sense to me in 1992. Three decades later, and after years of building large language models, I see its sense if one replaces sufficiency with necessity: understanding neural networks as we teach them today requires believing in an immanent entity.
The article develops a theoretical framework contrasting traditional inductive learning (which emphasizes generalization over memorization) with transductive inference (which embraces memorization and reasoning). Here's a quote:
"What matters is that LLMs are inductively trained transductive-inference engines and can therefore support both forms of inference.[2] They are capable of performing inference by inductive learning, like any trained classifier, akin to Daniel Kahneman’s “system 1” behavior — the fast thinking of his book title Thinking Fast and Slow. But LLMs are also capable of rudimentary forms of transduction, such as in-context-learning and chain of thought, which we may call system 2 — slow-thinking — behavior. The more sophisticated among us have even taught LLMs to do deduction — the ultimate test for their emergent abilities."
Sadly, the opening quote is not elucidated.
The article is extremely technical and doesn't really explain the quote other than acknowledging that there are stuff we don't understand yet.
And really, a person will never grasp machine learning and AI as long as they keep drawing unbased parallels to humans and machines.
>> If the training data subtend latent logical structures, as do sensory data such as visual or acoustic data, models trained as optimal predictors are forced to capture their statistical structure.
There are a lot of red flags to pick from in this article, but this one stood out to me as the most absurd. AI doesn't get magical multimodal powers from reading secondhand accounts describing a sensation. You can say it in as fancy of a phrasing as you want, but the proof is in the pudding. The "statistical structure" of that text doesn't propagate a meaningful understanding of almost anything in the real world.
> And really, a person will never grasp machine learning and AI as long as they keep drawing unbased parallels to humans and machines.
I think you're right on the money with this one.
Maybe I should have just linked to the research paper:
[B'MOJO: Hybrid state space realizations of foundation models with eidetic and fading memory](https://www.arxiv.org/abs/2407.06324)