You have to wade through lots of hysterical videos about Illya’s recent talk at NeurIPS to find the actual video:
https://www.youtube.com/watch?v=WQQdd6qGxNs Illya’s a rare exceptions to the rule that if the ML hysteria turns you into an icon, you have a poor grasp of Kolmogorov Complexity. However even Illya fails to get the power of KC in so-called “reasoning” hence ASI. For example, when he talks about “in distribution” vs “out of distribution” he doesn’t think through the following chain of logic: In distribution is interpolation out of distribution is extrapolation. This is the same difference as that between statistics and dynamics. Statistics are easy compared to dynamics. Science almost always starts with statistical observations. It is only when scientists become serious about predictions that they are forced into dynamics. Dynamics inherently involve time. Some have meta-observed that there is a relationship between prediction and time. For example if you predict something after you have observed it… Well… Uh… So what are the “dynamics” of LLMs? This is where the folks get lost who are going on and on and on about “chain of thought” (they really mean “tree of thought”), tacked onto foundation models is the Royal Road To ASI. In the limit of KC (approximated by foundation models unadorned with the infamous lobotomy alignment layer let alone tree of thought), “next token prediction” (hence generation of the most likely token sequence after prompt tokens) AIXI’s Sequential Decision Theoretic decision tree is inevitable even within the Algorithmic Information (KC/Solomonoff Induced) half – prior to AIXI’s SDT use of KC in constructing its decision tree! This decision tree, internal to KC, is inevitable because the whole point of the KC model is to provide a probability distribution for the “next” token (aka “observation” in the Solomonoff sense) that takes the limit of all possibly infinite sequences of generated tokens. The first token in this possibly infinite most likely sequence is “the next token generated”. These sequences are TREES, as in “tree of thought”. That’s why, in the limit of KC, you already have “reasoning” which Illya seems to mistake as the “intuition” of the LLMs. His mistake is quite understandable given the audience. The hysteria has deracinated the field of ML from its foundational principles such as KC (which Illya gets more than the rest of the icons). When you’re trying to compose your thoughts for an ignorant audience, you must, to some extent, take on their ignorant world view. Even geniuses like Illya can lose it under that burden. ------------------------------------------ Artificial General Intelligence List: AGI Permalink: https://agi.topicbox.com/groups/agi/T82474edc49bd4568-M13b4973d42d44e7a91bb5086 Delivery options: https://agi.topicbox.com/groups/agi/subscription