No Free Lunch only holds over finite sets. The incomputability of prediction holds over infinite sets.
On Sun, Sep 27, 2020, 2:21 PM Danko Nikolic <[email protected]> wrote: > But this is the No Free Lunch right there. > > On Sun, 27 Sep 2020, 20:05 Matt Mahoney <[email protected]> wrote: > >> >> >> On Sun, Sep 27, 2020, 1:41 PM Danko Nikolic <[email protected]> >> wrote: >> >>> I see the no free lunch theorem striking every day. Every time we pick >>> one ML architecture for one type of problem and another architecture for >>> another type of problem, it is the No Free Lunch Theorem dictating the fact >>> that we have to make thos chices and are not able to have one the same >>> architecture for all kinds of problems. >>> >> >> There is no simple, universal prediction algorithm. Suppose you have one. >> Then I can create a simple sequence that you can't predict. My program runs >> a copy of your program and outputs the opposite of your prediction. >> >> The best compressors have lots of code to handle lots of rare, special >> cases. It's not because of the no free lunch theorem. It's because you can >> always find something that your program can't compress, and you have to add >> yet another special case. >> >>> *Artificial General Intelligence List <https://agi.topicbox.com/latest>* > / AGI / see discussions <https://agi.topicbox.com/groups/agi> + > participants <https://agi.topicbox.com/groups/agi/members> + delivery > options <https://agi.topicbox.com/groups/agi/subscription> Permalink > <https://agi.topicbox.com/groups/agi/Ta433301e9ac5fb42-M8b88a544d50382b263a7eb99> > ------------------------------------------ Artificial General Intelligence List: AGI Permalink: https://agi.topicbox.com/groups/agi/Ta433301e9ac5fb42-M61966cc9bd819064cf9ce4da Delivery options: https://agi.topicbox.com/groups/agi/subscription
