Joe,

Yeah I read it, it’s a great book along with “The Martian”, couldn’t put
the Martian down.

Scott

On Mon, May 11, 2020 at 12:58 PM Joe Monk <joemon...@gmail.com> wrote:

> An even better story ...
>
> https://en.wikipedia.org/wiki/The_Adolescence_of_P-1
>
> Joe
>
> On Mon, May 11, 2020 at 11:31 AM Bob Bridges <robhbrid...@gmail.com>
> wrote:
>
> > I'll cheerfully leave political partisanship aside.  But if I may
> > attribute this equally to both sides (and thus avoid partisanship), I'm
> > with Joel ~and~ Lionel on this.  Most folks who misuse their power start
> > out, at least, in hopes of doing good.  What I'm saying is that although
> we
> > value altruism, I don't trust even altruists in the matter of exercising
> > power, especially when in pursuit of The Good of Humanity.
> >
> > Doesn't mean we won't keep building robots.  Doesn't even mean we
> > shouldn't.  But even altruists can be villains.  Ultron and Colossus both
> > wanted to save the world, after all.
> >
> > ---
> > Bob Bridges, robhbrid...@gmail.com, cell 336 382-7313
> >
> > /* The historian Macaulay famously said that the Puritans opposed
> > bearbaiting not because it gave pain to the bears but because it gave
> > pleasure to the spectators. The Puritans were right: Some pleasures are
> > contemptible because they are coarsening. They are not merely private
> > vices, they have public consequences in driving the culture's downward
> > spiral.  -George Will, "The challenge of thinking lower", 2001-06-22 */
> >
> > -----Original Message-----
> > From: IBM Mainframe Discussion List [mailto:IBM-MAIN@LISTSERV.UA.EDU] On
> > Behalf Of Lionel B Dyck
> > Sent: Monday, May 11, 2020 11:22
> >
> > Joel - can we please keep politics out of this listserv. Personally I
> > wouldn't trust anyone in power to act against their own self interests
> and
> > that applies to politicians and anyone else with power (as in money,
> > influence, etc.).
> >
> > There are altruistic individuals in the world and when it comes to the
> > development of an AI robot one prays/hopes that those are the software
> > developers who implement the code for the three laws.
> >
> > -----Original Message-----
> > From: IBM Mainframe Discussion List <IBM-MAIN@LISTSERV.UA.EDU> On Behalf
> > Of Joel C. Ewing
> > Sent: Monday, May 11, 2020 10:12 AM
> >
> > I've greatly enjoyed Asimov's vision of future possibilities, but when I
> > step back to reality it occurs to me that his perfect laws of robotics
> > would have to be implemented by fallible human programmers.  Even if
> > well-intentioned, how would they unambiguously convey to a robot the
> > concepts of "human", "humanity", "hurt", and "injure" when there have
> > always been minorities or "others" that are treated by one group of
> humans
> > as sub-human to justify injuring them in the name of "protecting"
> > them or protecting humanity?  And then there is the issue of who might
> > make the decision to build sentient robots:   For example, who in our
> > present White House would you trust to pay any heed to logic or
> scientific
> > recommendations or long-term consequences, if they were given the
> > opportunity to construct less-constrained AI robots that they perceived
> > offered some short-term political advantage?
> >
> > Humanity was also fortunate that when the hardware of Asimov's Daneel
> > began to fail, that he failed gracefully, rather than becoming a menace
> to
> > humanity.
> >
> > --- On 5/11/20 8:43 AM, scott Ford wrote:
> > > Well done Joel....I agree , But I can help to to be curious about the
> > > future of AI.
> > > a bit of Isaac Asimov ....
> > >
> > > --- On Mon, May 11, 2020 at 9:25 AM Joel C. Ewing <jcew...@acm.org>
> > wrote:
> > >>     And of course the whole point of Colossus, Dr Strangelove, War
> > >> Games, Terminator,  Forbidden Planet, Battlestar Galactica, etc. was
> > >> to try to make it clear to all the non-engineers and non-programmers
> > >> (all of whom greatly outnumber us) why putting lethal force in the
> > >> hands of any autonomous or even semi-autonomous machine is something
> > >> with incredible potential to go wrong.  We all know that even if the
> > >> hardware doesn't fail, which it inevitably will, that all software
> > >> above a certain level of complexity is guaranteed to have bugs with
> > >> unknown consequences.
> > >>     There is another equally cautionary genre in sci-fi about society
> > >> becoming so dependent on machines as to lose the knowledge to
> > >> understand and maintain the machines, resulting in total collapse
> > >> when the machines inevitably fail.  I still remember my oldest sister
> > reading E.M.
> > >> Forster, "The Machine Stops" (1909), to me  when I was very young.
> > >>     Various Star Trek episodes used both of these themes as plots.
> > >>     People can also break down with lethal  side effects, but the
> > >> potential  damage one person can create is more easily contained by
> > >> other people.   The  only effective way to defend again a berserk
> lethal
> > >> machine may be with another lethal machine, and Colossus-Guardian
> > >> suggests why that may be an even worse idea.
> > >>>
> > >>> -----Original Message-----
> > >>> From: Bob Bridges
> > >>> Sent: Sunday, May 10, 2020 10:21 PM
> > >>>
> > >>> I've always loved "Colossus: The Forbin Project".  Not many people
> > >>> have seen it, as far as I can tell.  The only problem I have with
> > >>> that movie - well, the main problem - is that no programmer in the
> > >>> world would make such a system and then throw away the Stop button.
> > >>> No engineer would do that with a machine he built, either.  Too many
> > >>> things can go wrong.  But a fun movie, if you can ignore that.
> > >>>
> > >>> -----Original Message-----
> > >>> From: scott Ford
> > >>> Sent: Sunday, May 10, 2020 11:38
> > >>>
> > >>> Like the 1970s flick , ‘Colossus , The Forbin Project’, Colossus and
> > >>> American computer and Guardian a Russian computer take over saying
> > >>> ‘Colossus and Guardian we are one’....
> >
> > ----------------------------------------------------------------------
> > For IBM-MAIN subscribe / signoff / archive access instructions,
> > send email to lists...@listserv.ua.edu with the message: INFO IBM-MAIN
> >
>
> ----------------------------------------------------------------------
> For IBM-MAIN subscribe / signoff / archive access instructions,
> send email to lists...@listserv.ua.edu with the message: INFO IBM-MAIN
>
-- 
Scott Ford
IDMWORKS
z/OS Development

----------------------------------------------------------------------
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@listserv.ua.edu with the message: INFO IBM-MAIN

Reply via email to