Are you willing to sign your name to this? Is anonymity important to presenting 
these ideas?

James Foster

> On Mar 22, 2023, at 5:34 AM, in_pharo_users--- via Pharo-users 
> <pharo-users@lists.pharo.org> wrote:
> 
> Offray,  and to all others,
> 
> you are missing the issue.
> 
> The problem we face is not to measure 'intelligence' of a system, but it's 
> ability to verbally act indistinguishable from a human.
> 
> This ability is allready given as chatbots are accepted by millions of users, 
> f.i. as user interfaces. (measurement = 'true', right?)
> 
> ChatGPT has the ability to follow a certain intention, f.i. to convince the 
> user to buy a certain product.  For this purpose, chat bots are getting  now 
> equipped with life like portrait pictures, speech input and output systems 
> with life like voices, phone numbers that they can use to make calls or being 
> called.  They are fed with all available data on the user, and we know that 
> ALL information about every single internet user in available and is being 
> consolidared on necessity.  The chat bots are able to use this information to 
> guide their conversational strategy, as the useful aspects of the users 
> mindset are extracted from his internet activity.
> 
> These chat bots are now operated on social network platforms with life like 
> names, 'pretending' to be human.
> 
> These bots act verbally indistinguishable from humans for most social media 
> users, as the most advanced psychotronic technology to manufacture consent.
> 
> The first goal of such a propaganda will naturally be to manufacture consent 
> about humans accepting being manipulated by AI chat bots, right?
> 
> How can this be achieved?  
> 
> Like allways in propaganda, the first attempt is to 
> - suppress awareness of the propaganda, then 
> - suppress the awareness of the problematic aspects of the propaganda 
> content, then 
> - reframe the propaganda content as acceptable, then as something to wish for,
> - achive collaboration of the propaganda victim with the goals of the 
> propaganda content.
> 
> Interestingly, this is exactly the schema that your post follows, Offray.
> 
> This often takes the form of domain framing, like we see in our conversation: 
>  the problem is shifted to the realm of academics - here informatics/computer 
> sciences - and thus delegated to experts exclusively.  We saw this in the 
> 9/11 aftermath coverup.
> 
> Then, Offray, you established yourself as an expert in color, discussing 
> aspects that have allready been introduced by others and including the groups 
> main focus 'Smalltalk', thus manufacturing consent and establishing yourself 
> as a reliable 'expert', and in reverse trying to hit at me, whom you have 
> identified as an adversary.
> 
> Then you offered a solution in color to the problem at hand with 'traceable 
> AI' and thus tried to open the possibility of collaboration with AI 
> proponents for the once critical reader.
> 
> I do not state, Offray, that you are knowingly an agent to promote the NWO AI 
> program.  I think you just 'learned' / have been programmed to be a 
> successful academic software developer, because to be successful in 
> academics, it is neccessary to learn to argue just like that since the 
> downfall of academic science in the tradition of, let's say, Humboldt.  So, I 
> grant that you may be a victim of propaganda yourself, instead of being a 
> secret service sposored agent. You took quite some time to formulate your 
> post, though.
> 
> You acted to contain the discussion about AI in this vital and important 
> informatics community to technical detail, when it is neccessary that 
> academics and community members look beyond the narrow borders of their 
> certifications and shift their thinking to the point of view where they can 
> see what technology does in the real world.
> 
> 
> 
> 
> 
> On 21.3.2023 at 7:21 PM, "Offray Vladimir Luna Cárdenas" 
> <offray.l...@mutabit.com> wrote:
>> 
>> I agree with Richard. The Turing test is not a good one to test 
>> intelligence and we have now just over glorified Eliza chatbots 
>> that 
>> appear to think and to understand but do none of them. ...

Reply via email to