Good morning Waldek,

My response will be delving into some areas of philosophical and
metaphysical response.

On Thu, Jan 18, 2024 at 2:56 AM Waldek Hebisch <[email protected]> wrote:

> On Thu, Jan 18, 2024 at 01:16:55AM +1100, Hill Strong wrote:
> > On Wed, Jan 17, 2024 at 11:09 PM Tim Daly <[email protected]> wrote:
> >
> > They can raise the issue all they like. What they are not seeing is that
> > artificial stupidity (AI) systems are limited. As I said above. The only
> > intelligence you will find in these systems is the stuff generated by
> human
> > intelligence. No artificial stupidity (AI) system can ever exceed the
> > limitations of the programming entailed in them.
>
> Well, humans are at least as limited: your claim as true as claim
> that "humans can not ever exceed the limitations of the programming
> entailed in them".  In case of humans programming meaning both things
>

Though we are limited in every sort of way, we are capable of exceeding any
programming that is involved in our development. What is so not obvious is
that there is a category error when comparing humans to any sort of machine
that we build. Though we can build machines stronger, faster and able to
handle tedious tasks far better than we can do, due to our very distinctive
intelligence, we are able to do what no machine can do and that is solve
the problems that arise.


hardcoded in genome and chemical machinery of the body and "learned
> stuff".  Already at age 1 toys, customized environment and interactions
>

You speak here about [programming] hard-coded in the genome and chemical
machinery of the human body. But I somehow don't think you realise just how
far beyond us is that machinery. Within a single living cell, the automatic
adaptive capabilities are so far beyond our current levels of technology
that in comparison, we are no more advanced than some caveman when we
compare the entire planetary industrial and planetary capabilities that we
have already made.

The more we learn about cellular biology with its internal transportation,
control and manufacturing systems the more we should be recognising that we
are not capable of building any computerised system that will ever be
intelligent.

with other humans make significant difference to learning.  At later
> stages there are stories which were perfected for thousends of years,
> school curricula and books.  There were myths that people from
> non-western cultures are less "inteligent" than people from western
> culture.  More deeper research showed that part of our "inteligence"
> is really "programmed" (learned) and "programming" in differnet
> cultures were different.
>

Though part of our intelligence is learned there is something that is so
far beyond this, so that, even though this is an active area of research,
we haven't a clue as to what this is.


> In slightly different spirit, in fifties there were efforts to
> define inteligence and researcheres from that time postulated
> several abilities that every inteligent being should have.
> Based on that there were "proofs" that artificial inteligence
> is impossible.  One of such "proofs" goes as follows: people
> can prove math theorems.  But Goedel and Church proved that
> no machine can prove math theorem.  So no machine will match
> humans.  The fallacy of this argument is classic abuse of
> quantifiers: humans can prove same (easy) math theorems.
>

Here you are now arguing against the various incompleteness theorems and in
doing so, you are entering the realm of metaphysics and philosophy. These
incompleteness theorems highlight something quite interesting in that the
basis of our logic and mathematics (the fundamental axioms used) cannot be
proved true from within our logic and mathematics. We can argue for them
from metaphysics and philosophy but we are unable to prove them from within
the systems that are based on those fundamental axioms.

What you have missed here is that any theorems developed within the systems
are provable based on those unprovable axioms and previously proven
theorems. That is how we can build up the various systems of mathematics in
use. However, new mathematical systems require insights not possible to any
computerised system. These new insights are in the realm of human
intelligence which does exceed any programming that individual humans may
be exposed to.

No machine or human can prove _each_ math theorem.  Actually,
> we still do not know how hard is proving, but common belief
>

All theorems can be proved if they are based on the initial set of axioms
used. How hard it is to prove the theorem is another matter.


> is that complexity of proving is exponential in length of the
> proof.  What is proven is that that there is no computable
> bound on length of shortest proof.  Clearly this difficulty,
> that is large length of proofs affect humans as much as
> computers.
>

Maybe. The thing of interest here is that it is humans who can find
alternative ways of proving a theorem which is not a part of the exhaustive
processes previously used.

As you would well know, the Halting problem is not solvable by any Turing
machine or corresponding alternative. But we can see the problem
immediately. We make leaps and find solutions that are not possible within
a logical/mathematical system.


> To put is differently, if you put strong requirements on
> inteligence, like ability to prove each math theorem, then
> humans are not inteligent.  If you lower your requirements,
> so that humans are deemed inteligent, than appropriately
> programmed computer is likely to qualify.
>

Your comment above is false. It is false logically, rationally,
metaphysically and philosophically. The ability to prove a mathematics
theorem, if based on previous axioms and theorems, is not a sign of
intelligence. I would agree that we do this as a part of our intelligence,
but this is not a sign of intelligence. The axioms used at the fundamental
level which are designed and specified may well be a sign of intelligence,
but no computerised system does this. I have spent 40 years looking at
these kinds of things and every computerised system is based on the human
intellectual capacity (intelligence) for these systems to even exist.

>
> One more thing: early in history of AI there was Eliza.
> It was simple pattern matcher clearly having no inteligence,
> yet it was able to fool some humans to belive that they
> communicate with other human (ok, at least for some time).
>

The ability to fool people does not challenge our intelligence. Every human
is flawed and can be fooled. We see that every day in every society. This
does not lessen what humanity has in terms of being intelligent. We act
irrationally, we act in completely adverse ways. We make decisions that
make no sense. We even choose options that are not actually available to
us. We can see this in the recent Iowa caucus where the winner wasn't even
on the available options.

Machine programming doesn't allow this. Certainly, we can see all sorts of
errors and when fully studied, we see why those errors occurred. They are
logically defined.

Some people take this to consider all solved AI problems
> as kind of fake and show that the problem was not about
> inteligence.  But IMO there is different possiblity: that
> all our inteligence is "fake" in similar vein.  In other
>

If our intelligence is [fake] then the logical consequence of that argument
is that there is no mathematics, no scientific investigation, no
rationality, no intelligence, no progress in any form. This is a
contradiction. We do do mathematics, we do do scientific investigation, we
do do philosophy and metaphysics, we do do rational things.

words, we do not solve general problem but use tricks which
> happen to work in real life.  Or to put it differently,
>

Yes we do do tricks and this is one sign of intelligence here. Certainly we
are limited and I have been saying such things for a long time. However,
for all of our limitations, we are in a category that is very different to
all other things, living or otherwise and we often fail to recognise, let
alone understand that category difference.

I have seen too many people who dismiss philosophy and metaphysics as being
useless. Yet it is by these that we show forth our intelligence and by
which we can justify mathematics and scientific investigation as valid,
rational and proper intellectual pursuits.


we may be much more limited than we imagine.  Eliza clearly
> shows that we can be easily fooled into assuming that
> something has much more abilities than it really has
> (and "something" may be really "we").
>
> --
>                               Waldek Hebisch
>
> --
> You received this message because you are subscribed to the Google Groups
> "FriCAS - computer algebra system" group.
> To unsubscribe from this group and stop receiving emails from it, send an
> email to [email protected].
> To view this discussion on the web visit
> https://groups.google.com/d/msgid/fricas-devel/Zaf4t0bQZRQAsouA%40fricas.org
> .
>

-- 
You received this message because you are subscribed to the Google Groups 
"FriCAS - computer algebra system" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/fricas-devel/CAEnaMTHSQu_-0ewXP%3Dt-qsV2OU%3Dgs%3D5-OY8Q4kp8M4sd_j6rEg%40mail.gmail.com.

Reply via email to