pharo-users@lists.pharo.org

Any question about pharo is welcome

View all threads

Wow - Chat GPT understands Smalltalk

OV
Offray Vladimir Luna Cárdenas
Sun, Mar 26, 2023 6:33 PM

Hi,

Comments inlined below:

On 22/03/23 7:34, in_pharo_users--- via Pharo-users wrote:

Offray,  and to all others,

you are missing the issue.

The problem we face is not to measure 'intelligence' of a system, but it's ability to verbally act indistinguishable from a human.

This ability is allready given as chatbots are accepted by millions of users, f.i. as user interfaces. (measurement = 'true', right?)

ChatGPT has the ability to follow a certain intention, f.i. to convince the user to buy a certain product.  For this purpose, chat bots are getting  now equipped with life like portrait pictures, speech input and output systems with life like voices, phone numbers that they can use to make calls or being called.  They are fed with all available data on the user, and we know that ALL information about every single internet user in available and is being consolidared on necessity.  The chat bots are able to use this information to guide their conversational strategy, as the useful aspects of the users mindset are extracted from his internet activity.

These chat bots are now operated on social network platforms with life like names, 'pretending' to be human.

These bots act verbally indistinguishable from humans for most social media users, as the most advanced psychotronic technology to manufacture consent.

The first goal of such a propaganda will naturally be to manufacture consent about humans accepting being manipulated by AI chat bots, right?

I don't think I have missed the point, as we agreed (I think) on
chatbots not being intelligent, just having such appearance. That why
I'm calling "AI" #ApparentIntelligence (in the sense of look alike, but
not real). Of course, something looking like a real thing without being
the real thing can be used for manipulation since the first times of
gossip, printing press and now automatization, with the changes in
scale/danger that such medium changes imply.

I don't think that manufactured consent is so easy, as this very thread
shows. What is being automated is manufactured polarization (but humans
can do pretty well by our own on polarization).

How can this be achieved?

Like allways in propaganda, the first attempt is to

  • suppress awareness of the propaganda, then
  • suppress the awareness of the problematic aspects of the propaganda content, then
  • reframe the propaganda content as acceptable, then as something to wish for,
  • achive collaboration of the propaganda victim with the goals of the propaganda content.

Interestingly, this is exactly the schema that your post follows, Offray.

On the contrary, my post is advocating for a critical reading of
Apparent Intelligence, by reframing the terms and the acritical
technoutopic / technoapocalyptic readings/discourses that are spreading
rapidly on the wider web, as I think that this community has shown an
historical different position beyond/resisting hype and current trends.
So I don't see how any of the steps you mention are "blueprint followed"
in my post, and I think they will be difficult to locate without
specific examples.

This often takes the form of domain framing, like we see in our conversation:  the problem is shifted to the realm of academics - here informatics/computer sciences - and thus delegated to experts exclusively.  We saw this in the 9/11 aftermath coverup.

Then, Offray, you established yourself as an expert in color, discussing aspects that have allready been introduced by others and including the groups main focus 'Smalltalk', thus manufacturing consent and establishing yourself as a reliable 'expert', and in reverse trying to hit at me, whom you have identified as an adversary.

Then you offered a solution in color to the problem at hand with 'traceable AI' and thus tried to open the possibility of collaboration with AI proponents for the once critical reader.

Heh, heh. On the contrary seems that the one seeing a scheme and a
enemies location/confrontation with deep plots and tactics is you.
Providing external creditable sources beyond opinion, belonging to a
established discursive falsafiable tradition (i.e. one that you can
criticize instead of blindly accept) is a way to enrich
discourse/argumentation beyond conspiracy theories. You could also quote
your sources instead, which would allow the community to see where our
positions are hold/sustained, even if we use different domain frames,
which is better that claiming no domain or expertise in pursuit of
openness. So instead of this are my opinions without any external source
or reference to pretend no expertise or domain framing, we could
advocate for openness by welcoming different expertise and argumentation
and making our sources/bias as evident as possible.

I do not state, Offray, that you are knowingly an agent to promote the NWO AI program.  I think you just 'learned' / have been programmed to be a successful academic software developer, because to be successful in academics, it is neccessary to learn to argue just like that since the downfall of academic science in the tradition of, let's say, Humboldt.  So, I grant that you may be a victim of propaganda yourself, instead of being a secret service sposored agent. You took quite some time to formulate your post, though.

You acted to contain the discussion about AI in this vital and important informatics community to technical detail, when it is neccessary that academics and community members look beyond the narrow borders of their certifications and shift their thinking to the point of view where they can see what technology does in the real world.

I offered a view point with sources. In no way the discussion about
Apparent Intelligence was contained and, on the contrary, I tried to
offer arguments from cognition and philosophy, beyond technical details,
which have been pretty absent in the general discourse and rush gold of
technoutopism / technoapocalypse, conspiracy and paranoia. Other sources
could be added to enrich and inform de conversation and they would be
pretty welcomed (at least from me and several members I've known from
this community)

Best,

Offray

Hi, Comments inlined below: On 22/03/23 7:34, in_pharo_users--- via Pharo-users wrote: > Offray, and to all others, > > you are missing the issue. > > The problem we face is not to measure 'intelligence' of a system, but it's ability to verbally act indistinguishable from a human. > > This ability is allready given as chatbots are accepted by millions of users, f.i. as user interfaces. (measurement = 'true', right?) > > ChatGPT has the ability to follow a certain intention, f.i. to convince the user to buy a certain product. For this purpose, chat bots are getting now equipped with life like portrait pictures, speech input and output systems with life like voices, phone numbers that they can use to make calls or being called. They are fed with all available data on the user, and we know that ALL information about every single internet user in available and is being consolidared on necessity. The chat bots are able to use this information to guide their conversational strategy, as the useful aspects of the users mindset are extracted from his internet activity. > > These chat bots are now operated on social network platforms with life like names, 'pretending' to be human. > > These bots act verbally indistinguishable from humans for most social media users, as the most advanced psychotronic technology to manufacture consent. > > The first goal of such a propaganda will naturally be to manufacture consent about humans accepting being manipulated by AI chat bots, right? I don't think I have missed the point, as we agreed (I think) on chatbots not being intelligent, just having such appearance. That why I'm calling "AI" #ApparentIntelligence (in the sense of look alike, but not real). Of course, something looking like a real thing without being the real thing can be used for manipulation since the first times of gossip, printing press and now automatization, with the changes in scale/danger that such medium changes imply. I don't think that manufactured consent is so easy, as this very thread shows. What is being automated is manufactured polarization (but humans can do pretty well by our own on polarization). > How can this be achieved? > > Like allways in propaganda, the first attempt is to > - suppress awareness of the propaganda, then > - suppress the awareness of the problematic aspects of the propaganda content, then > - reframe the propaganda content as acceptable, then as something to wish for, > - achive collaboration of the propaganda victim with the goals of the propaganda content. > > Interestingly, this is exactly the schema that your post follows, Offray. On the contrary, my post is advocating for a critical reading of Apparent Intelligence, by reframing the terms and the acritical technoutopic / technoapocalyptic readings/discourses that are spreading rapidly on the wider web, as I think that this community has shown an historical different position beyond/resisting hype and current trends. So I don't see how any of the steps you mention are "blueprint followed" in my post, and I think they will be difficult to locate without specific examples. > > This often takes the form of domain framing, like we see in our conversation: the problem is shifted to the realm of academics - here informatics/computer sciences - and thus delegated to experts exclusively. We saw this in the 9/11 aftermath coverup. > > Then, Offray, you established yourself as an expert in color, discussing aspects that have allready been introduced by others and including the groups main focus 'Smalltalk', thus manufacturing consent and establishing yourself as a reliable 'expert', and in reverse trying to hit at me, whom you have identified as an adversary. > > Then you offered a solution in color to the problem at hand with 'traceable AI' and thus tried to open the possibility of collaboration with AI proponents for the once critical reader. Heh, heh. On the contrary seems that the one seeing a scheme and a enemies location/confrontation with deep plots and tactics is you. Providing external creditable sources beyond opinion, belonging to a established discursive falsafiable tradition (i.e. one that you can criticize instead of blindly accept) is a way to enrich discourse/argumentation beyond conspiracy theories. You could also quote your sources instead, which would allow the community to see where our positions are hold/sustained, even if we use different domain frames, which is better that claiming no domain or expertise in pursuit of openness. So instead of this are my opinions without any external source or reference to pretend no expertise or domain framing, we could advocate for openness by welcoming different expertise and argumentation and making our sources/bias as evident as possible. > > I do not state, Offray, that you are knowingly an agent to promote the NWO AI program. I think you just 'learned' / have been programmed to be a successful academic software developer, because to be successful in academics, it is neccessary to learn to argue just like that since the downfall of academic science in the tradition of, let's say, Humboldt. So, I grant that you may be a victim of propaganda yourself, instead of being a secret service sposored agent. You took quite some time to formulate your post, though. > > You acted to contain the discussion about AI in this vital and important informatics community to technical detail, when it is neccessary that academics and community members look beyond the narrow borders of their certifications and shift their thinking to the point of view where they can see what technology does in the real world. > > I offered a view point with sources. In no way the discussion about Apparent Intelligence was contained and, on the contrary, I tried to offer arguments from cognition and philosophy, beyond technical details, which have been pretty absent in the general discourse and rush gold of technoutopism / technoapocalypse, conspiracy and paranoia. Other sources could be added to enrich and inform de conversation and they would be pretty welcomed (at least from me and several members I've known from this community) Best, Offray
IP
in_pharo_users@nym.hush.com
Sun, Mar 26, 2023 7:16 PM

Dear Offray,

I have nothing to comment on this.


In general I have made the observation that  certain people who want to push an agenda to promote alterior motives tend to reiterate false prepositions and so false conclusions over and over.

If there is an apodictic statement that contradicts their alterior motives, this people just can't help themselves but to deny truth against better knowledge ad nauseam.

On 26.3.2023 at 8:34 PM, "Offray Vladimir Luna Cárdenas" offray.luna@mutabit.com wrote:

Hi,

Comments inlined below:

On 22/03/23 7:34, in_pharo_users--- via Pharo-users wrote:

Offray,  and to all others,

you are missing the issue.

The problem we face is not to measure 'intelligence' of a

system, but it's ability to verbally act indistinguishable from a
human.

This ability is allready given as chatbots are accepted by

millions of users, f.i. as user interfaces. (measurement = 'true',
right?)

ChatGPT has the ability to follow a certain intention, f.i. to

convince the user to buy a certain product.  For this purpose,
chat bots are getting  now equipped with life like portrait
pictures, speech input and output systems with life like voices,
phone numbers that they can use to make calls or being called.
They are fed with all available data on the user, and we know that
ALL information about every single internet user in available and
is being consolidared on necessity.  The chat bots are able to use
this information to guide their conversational strategy, as the
useful aspects of the users mindset are extracted from his
internet activity.

These chat bots are now operated on social network platforms

with life like names, 'pretending' to be human.

These bots act verbally indistinguishable from humans for most

social media users, as the most advanced psychotronic technology
to manufacture consent.

The first goal of such a propaganda will naturally be to

manufacture consent about humans accepting being manipulated by AI
chat bots, right?

I don't think I have missed the point, as we agreed (I think) on
chatbots not being intelligent, just having such appearance. That
why
I'm calling "AI" #ApparentIntelligence (in the sense of look
alike, but
not real). Of course, something looking like a real thing without
being
the real thing can be used for manipulation since the first times
of
gossip, printing press and now automatization, with the changes in
scale/danger that such medium changes imply.

I don't think that manufactured consent is so easy, as this very
thread
shows. What is being automated is manufactured polarization (but
humans
can do pretty well by our own on polarization).

How can this be achieved?

Like allways in propaganda, the first attempt is to

  • suppress awareness of the propaganda, then
  • suppress the awareness of the problematic aspects of the

propaganda content, then

  • reframe the propaganda content as acceptable, then as

something to wish for,

  • achive collaboration of the propaganda victim with the goals

of the propaganda content.

Interestingly, this is exactly the schema that your post

follows, Offray.

On the contrary, my post is advocating for a critical reading of
Apparent Intelligence, by reframing the terms and the acritical
technoutopic / technoapocalyptic readings/discourses that are
spreading
rapidly on the wider web, as I think that this community has shown
an
historical different position beyond/resisting hype and current
trends.
So I don't see how any of the steps you mention are "blueprint
followed"
in my post, and I think they will be difficult to locate without
specific examples.

This often takes the form of domain framing, like we see in our

conversation:  the problem is shifted to the realm of academics -
here informatics/computer sciences - and thus delegated to experts
exclusively.  We saw this in the 9/11 aftermath coverup.

Then, Offray, you established yourself as an expert in color,

discussing aspects that have allready been introduced by others
and including the groups main focus 'Smalltalk', thus
manufacturing consent and establishing yourself as a reliable
'expert', and in reverse trying to hit at me, whom you have
identified as an adversary.

Then you offered a solution in color to the problem at hand with

'traceable AI' and thus tried to open the possibility of
collaboration with AI proponents for the once critical reader.

Heh, heh. On the contrary seems that the one seeing a scheme and a
enemies location/confrontation with deep plots and tactics is you.
Providing external creditable sources beyond opinion, belonging to
a
established discursive falsafiable tradition (i.e. one that you
can
criticize instead of blindly accept) is a way to enrich
discourse/argumentation beyond conspiracy theories. You could also
quote
your sources instead, which would allow the community to see where
our
positions are hold/sustained, even if we use different domain
frames,
which is better that claiming no domain or expertise in pursuit of
openness. So instead of this are my opinions without any external
source
or reference to pretend no expertise or domain framing, we could
advocate for openness by welcoming different expertise and
argumentation
and making our sources/bias as evident as possible.

I do not state, Offray, that you are knowingly an agent to

promote the NWO AI program.  I think you just 'learned' / have
been programmed to be a successful academic software developer,
because to be successful in academics, it is neccessary to learn
to argue just like that since the downfall of academic science in
the tradition of, let's say, Humboldt.  So, I grant that you may
be a victim of propaganda yourself, instead of being a secret
service sposored agent. You took quite some time to formulate your
post, though.

You acted to contain the discussion about AI in this vital and

important informatics community to technical detail, when it is
neccessary that academics and community members look beyond the
narrow borders of their certifications and shift their thinking to
the point of view where they can see what technology does in the
real world.

I offered a view point with sources. In no way the discussion
about
Apparent Intelligence was contained and, on the contrary, I tried
to
offer arguments from cognition and philosophy, beyond technical
details,
which have been pretty absent in the general discourse and rush
gold of
technoutopism / technoapocalypse, conspiracy and paranoia. Other
sources
could be added to enrich and inform de conversation and they would
be
pretty welcomed (at least from me and several members I've known
from
this community)

Best,

Offray

Dear Offray, I have nothing to comment on this. --- In general I have made the observation that certain people who want to push an agenda to promote alterior motives tend to reiterate false prepositions and so false conclusions over and over. If there is an apodictic statement that contradicts their alterior motives, this people just can't help themselves but to deny truth against better knowledge ad nauseam. On 26.3.2023 at 8:34 PM, "Offray Vladimir Luna Cárdenas" <offray.luna@mutabit.com> wrote: > >Hi, > >Comments inlined below: > >On 22/03/23 7:34, in_pharo_users--- via Pharo-users wrote: >> Offray, and to all others, >> >> you are missing the issue. >> >> The problem we face is not to measure 'intelligence' of a >system, but it's ability to verbally act indistinguishable from a >human. >> >> This ability is allready given as chatbots are accepted by >millions of users, f.i. as user interfaces. (measurement = 'true', >right?) >> >> ChatGPT has the ability to follow a certain intention, f.i. to >convince the user to buy a certain product. For this purpose, >chat bots are getting now equipped with life like portrait >pictures, speech input and output systems with life like voices, >phone numbers that they can use to make calls or being called. >They are fed with all available data on the user, and we know that >ALL information about every single internet user in available and >is being consolidared on necessity. The chat bots are able to use >this information to guide their conversational strategy, as the >useful aspects of the users mindset are extracted from his >internet activity. >> >> These chat bots are now operated on social network platforms >with life like names, 'pretending' to be human. >> >> These bots act verbally indistinguishable from humans for most >social media users, as the most advanced psychotronic technology >to manufacture consent. >> >> The first goal of such a propaganda will naturally be to >manufacture consent about humans accepting being manipulated by AI >chat bots, right? > >I don't think I have missed the point, as we agreed (I think) on >chatbots not being intelligent, just having such appearance. That >why >I'm calling "AI" #ApparentIntelligence (in the sense of look >alike, but >not real). Of course, something looking like a real thing without >being >the real thing can be used for manipulation since the first times >of >gossip, printing press and now automatization, with the changes in >scale/danger that such medium changes imply. > >I don't think that manufactured consent is so easy, as this very >thread >shows. What is being automated is manufactured polarization (but >humans >can do pretty well by our own on polarization). > > >> How can this be achieved? >> >> Like allways in propaganda, the first attempt is to >> - suppress awareness of the propaganda, then >> - suppress the awareness of the problematic aspects of the >propaganda content, then >> - reframe the propaganda content as acceptable, then as >something to wish for, >> - achive collaboration of the propaganda victim with the goals >of the propaganda content. >> >> Interestingly, this is exactly the schema that your post >follows, Offray. > >On the contrary, my post is advocating for a critical reading of >Apparent Intelligence, by reframing the terms and the acritical >technoutopic / technoapocalyptic readings/discourses that are >spreading >rapidly on the wider web, as I think that this community has shown >an >historical different position beyond/resisting hype and current >trends. >So I don't see how any of the steps you mention are "blueprint >followed" >in my post, and I think they will be difficult to locate without >specific examples. > > >> >> This often takes the form of domain framing, like we see in our >conversation: the problem is shifted to the realm of academics - >here informatics/computer sciences - and thus delegated to experts >exclusively. We saw this in the 9/11 aftermath coverup. >> >> Then, Offray, you established yourself as an expert in color, >discussing aspects that have allready been introduced by others >and including the groups main focus 'Smalltalk', thus >manufacturing consent and establishing yourself as a reliable >'expert', and in reverse trying to hit at me, whom you have >identified as an adversary. >> >> Then you offered a solution in color to the problem at hand with >'traceable AI' and thus tried to open the possibility of >collaboration with AI proponents for the once critical reader. > >Heh, heh. On the contrary seems that the one seeing a scheme and a >enemies location/confrontation with deep plots and tactics is you. >Providing external creditable sources beyond opinion, belonging to >a >established discursive falsafiable tradition (i.e. one that you >can >criticize instead of blindly accept) is a way to enrich >discourse/argumentation beyond conspiracy theories. You could also >quote >your sources instead, which would allow the community to see where >our >positions are hold/sustained, even if we use different domain >frames, >which is better that claiming no domain or expertise in pursuit of >openness. So instead of this are my opinions without any external >source >or reference to pretend no expertise or domain framing, we could >advocate for openness by welcoming different expertise and >argumentation >and making our sources/bias as evident as possible. > > >> >> I do not state, Offray, that you are knowingly an agent to >promote the NWO AI program. I think you just 'learned' / have >been programmed to be a successful academic software developer, >because to be successful in academics, it is neccessary to learn >to argue just like that since the downfall of academic science in >the tradition of, let's say, Humboldt. So, I grant that you may >be a victim of propaganda yourself, instead of being a secret >service sposored agent. You took quite some time to formulate your >post, though. >> >> You acted to contain the discussion about AI in this vital and >important informatics community to technical detail, when it is >neccessary that academics and community members look beyond the >narrow borders of their certifications and shift their thinking to >the point of view where they can see what technology does in the >real world. >> >> > >I offered a view point with sources. In no way the discussion >about >Apparent Intelligence was contained and, on the contrary, I tried >to >offer arguments from cognition and philosophy, beyond technical >details, >which have been pretty absent in the general discourse and rush >gold of >technoutopism / technoapocalypse, conspiracy and paranoia. Other >sources >could be added to enrich and inform de conversation and they would >be >pretty welcomed (at least from me and several members I've known >from >this community) > >Best, > >Offray
OV
Offray Vladimir Luna Cárdenas
Sun, Mar 26, 2023 8:34 PM

Dear anonymous,

Me neither.

It is pretty difficult to make constructive discourse against hidden
agendas, ulterior motives, self evident truths, sources absence or
general affirmations without particular examples or detailed sustain.

Offray

On 26/03/23 14:16, in_pharo_users--- via Pharo-users wrote:

Dear Offray,

I have nothing to comment on this.


In general I have made the observation that  certain people who want to push an agenda to promote alterior motives tend to reiterate false prepositions and so false conclusions over and over.

If there is an apodictic statement that contradicts their alterior motives, this people just can't help themselves but to deny truth against better knowledge ad nauseam.

On 26.3.2023 at 8:34 PM, "Offray Vladimir Luna Cárdenas" offray.luna@mutabit.com wrote:

Hi,

Comments inlined below:

On 22/03/23 7:34, in_pharo_users--- via Pharo-users wrote:

Offray,  and to all others,

you are missing the issue.

The problem we face is not to measure 'intelligence' of a
system, but it's ability to verbally act indistinguishable from a
human.
This ability is allready given as chatbots are accepted by
millions of users, f.i. as user interfaces. (measurement = 'true',
right?)
ChatGPT has the ability to follow a certain intention, f.i. to
convince the user to buy a certain product.  For this purpose,
chat bots are getting  now equipped with life like portrait
pictures, speech input and output systems with life like voices,
phone numbers that they can use to make calls or being called.
They are fed with all available data on the user, and we know that
ALL information about every single internet user in available and
is being consolidared on necessity.  The chat bots are able to use
this information to guide their conversational strategy, as the
useful aspects of the users mindset are extracted from his
internet activity.
These chat bots are now operated on social network platforms
with life like names, 'pretending' to be human.
These bots act verbally indistinguishable from humans for most
social media users, as the most advanced psychotronic technology
to manufacture consent.
The first goal of such a propaganda will naturally be to
manufacture consent about humans accepting being manipulated by AI
chat bots, right?

I don't think I have missed the point, as we agreed (I think) on
chatbots not being intelligent, just having such appearance. That
why
I'm calling "AI" #ApparentIntelligence (in the sense of look
alike, but
not real). Of course, something looking like a real thing without
being
the real thing can be used for manipulation since the first times
of
gossip, printing press and now automatization, with the changes in
scale/danger that such medium changes imply.

I don't think that manufactured consent is so easy, as this very
thread
shows. What is being automated is manufactured polarization (but
humans
can do pretty well by our own on polarization).

How can this be achieved?

Like allways in propaganda, the first attempt is to

  • suppress awareness of the propaganda, then
  • suppress the awareness of the problematic aspects of the
    propaganda content, then
  • reframe the propaganda content as acceptable, then as
    something to wish for,
  • achive collaboration of the propaganda victim with the goals
    of the propaganda content.
    Interestingly, this is exactly the schema that your post
    follows, Offray.

On the contrary, my post is advocating for a critical reading of
Apparent Intelligence, by reframing the terms and the acritical
technoutopic / technoapocalyptic readings/discourses that are
spreading
rapidly on the wider web, as I think that this community has shown
an
historical different position beyond/resisting hype and current
trends.
So I don't see how any of the steps you mention are "blueprint
followed"
in my post, and I think they will be difficult to locate without
specific examples.

This often takes the form of domain framing, like we see in our
conversation:  the problem is shifted to the realm of academics -
here informatics/computer sciences - and thus delegated to experts
exclusively.  We saw this in the 9/11 aftermath coverup.
Then, Offray, you established yourself as an expert in color,
discussing aspects that have allready been introduced by others
and including the groups main focus 'Smalltalk', thus
manufacturing consent and establishing yourself as a reliable
'expert', and in reverse trying to hit at me, whom you have
identified as an adversary.
Then you offered a solution in color to the problem at hand with
'traceable AI' and thus tried to open the possibility of
collaboration with AI proponents for the once critical reader.

Heh, heh. On the contrary seems that the one seeing a scheme and a
enemies location/confrontation with deep plots and tactics is you.
Providing external creditable sources beyond opinion, belonging to
a
established discursive falsafiable tradition (i.e. one that you
can
criticize instead of blindly accept) is a way to enrich
discourse/argumentation beyond conspiracy theories. You could also
quote
your sources instead, which would allow the community to see where
our
positions are hold/sustained, even if we use different domain
frames,
which is better that claiming no domain or expertise in pursuit of
openness. So instead of this are my opinions without any external
source
or reference to pretend no expertise or domain framing, we could
advocate for openness by welcoming different expertise and
argumentation
and making our sources/bias as evident as possible.

I do not state, Offray, that you are knowingly an agent to
promote the NWO AI program.  I think you just 'learned' / have
been programmed to be a successful academic software developer,
because to be successful in academics, it is neccessary to learn
to argue just like that since the downfall of academic science in
the tradition of, let's say, Humboldt.  So, I grant that you may
be a victim of propaganda yourself, instead of being a secret
service sposored agent. You took quite some time to formulate your
post, though.
You acted to contain the discussion about AI in this vital and
important informatics community to technical detail, when it is
neccessary that academics and community members look beyond the
narrow borders of their certifications and shift their thinking to
the point of view where they can see what technology does in the
real world.

I offered a view point with sources. In no way the discussion
about
Apparent Intelligence was contained and, on the contrary, I tried
to
offer arguments from cognition and philosophy, beyond technical
details,
which have been pretty absent in the general discourse and rush
gold of
technoutopism / technoapocalypse, conspiracy and paranoia. Other
sources
could be added to enrich and inform de conversation and they would
be
pretty welcomed (at least from me and several members I've known
from
this community)

Best,

Offray

Dear anonymous, Me neither. It is pretty difficult to make constructive discourse against hidden agendas, ulterior motives, self evident truths, sources absence or general affirmations without particular examples or detailed sustain. Offray On 26/03/23 14:16, in_pharo_users--- via Pharo-users wrote: > Dear Offray, > > I have nothing to comment on this. > > --- > > In general I have made the observation that certain people who want to push an agenda to promote alterior motives tend to reiterate false prepositions and so false conclusions over and over. > > If there is an apodictic statement that contradicts their alterior motives, this people just can't help themselves but to deny truth against better knowledge ad nauseam. > > > On 26.3.2023 at 8:34 PM, "Offray Vladimir Luna Cárdenas" <offray.luna@mutabit.com> wrote: >> Hi, >> >> Comments inlined below: >> >> On 22/03/23 7:34, in_pharo_users--- via Pharo-users wrote: >>> Offray, and to all others, >>> >>> you are missing the issue. >>> >>> The problem we face is not to measure 'intelligence' of a >> system, but it's ability to verbally act indistinguishable from a >> human. >>> This ability is allready given as chatbots are accepted by >> millions of users, f.i. as user interfaces. (measurement = 'true', >> right?) >>> ChatGPT has the ability to follow a certain intention, f.i. to >> convince the user to buy a certain product. For this purpose, >> chat bots are getting now equipped with life like portrait >> pictures, speech input and output systems with life like voices, >> phone numbers that they can use to make calls or being called. >> They are fed with all available data on the user, and we know that >> ALL information about every single internet user in available and >> is being consolidared on necessity. The chat bots are able to use >> this information to guide their conversational strategy, as the >> useful aspects of the users mindset are extracted from his >> internet activity. >>> These chat bots are now operated on social network platforms >> with life like names, 'pretending' to be human. >>> These bots act verbally indistinguishable from humans for most >> social media users, as the most advanced psychotronic technology >> to manufacture consent. >>> The first goal of such a propaganda will naturally be to >> manufacture consent about humans accepting being manipulated by AI >> chat bots, right? >> >> I don't think I have missed the point, as we agreed (I think) on >> chatbots not being intelligent, just having such appearance. That >> why >> I'm calling "AI" #ApparentIntelligence (in the sense of look >> alike, but >> not real). Of course, something looking like a real thing without >> being >> the real thing can be used for manipulation since the first times >> of >> gossip, printing press and now automatization, with the changes in >> scale/danger that such medium changes imply. >> >> I don't think that manufactured consent is so easy, as this very >> thread >> shows. What is being automated is manufactured polarization (but >> humans >> can do pretty well by our own on polarization). >> >> >>> How can this be achieved? >>> >>> Like allways in propaganda, the first attempt is to >>> - suppress awareness of the propaganda, then >>> - suppress the awareness of the problematic aspects of the >> propaganda content, then >>> - reframe the propaganda content as acceptable, then as >> something to wish for, >>> - achive collaboration of the propaganda victim with the goals >> of the propaganda content. >>> Interestingly, this is exactly the schema that your post >> follows, Offray. >> >> On the contrary, my post is advocating for a critical reading of >> Apparent Intelligence, by reframing the terms and the acritical >> technoutopic / technoapocalyptic readings/discourses that are >> spreading >> rapidly on the wider web, as I think that this community has shown >> an >> historical different position beyond/resisting hype and current >> trends. >> So I don't see how any of the steps you mention are "blueprint >> followed" >> in my post, and I think they will be difficult to locate without >> specific examples. >> >> >>> This often takes the form of domain framing, like we see in our >> conversation: the problem is shifted to the realm of academics - >> here informatics/computer sciences - and thus delegated to experts >> exclusively. We saw this in the 9/11 aftermath coverup. >>> Then, Offray, you established yourself as an expert in color, >> discussing aspects that have allready been introduced by others >> and including the groups main focus 'Smalltalk', thus >> manufacturing consent and establishing yourself as a reliable >> 'expert', and in reverse trying to hit at me, whom you have >> identified as an adversary. >>> Then you offered a solution in color to the problem at hand with >> 'traceable AI' and thus tried to open the possibility of >> collaboration with AI proponents for the once critical reader. >> >> Heh, heh. On the contrary seems that the one seeing a scheme and a >> enemies location/confrontation with deep plots and tactics is you. >> Providing external creditable sources beyond opinion, belonging to >> a >> established discursive falsafiable tradition (i.e. one that you >> can >> criticize instead of blindly accept) is a way to enrich >> discourse/argumentation beyond conspiracy theories. You could also >> quote >> your sources instead, which would allow the community to see where >> our >> positions are hold/sustained, even if we use different domain >> frames, >> which is better that claiming no domain or expertise in pursuit of >> openness. So instead of this are my opinions without any external >> source >> or reference to pretend no expertise or domain framing, we could >> advocate for openness by welcoming different expertise and >> argumentation >> and making our sources/bias as evident as possible. >> >> >>> I do not state, Offray, that you are knowingly an agent to >> promote the NWO AI program. I think you just 'learned' / have >> been programmed to be a successful academic software developer, >> because to be successful in academics, it is neccessary to learn >> to argue just like that since the downfall of academic science in >> the tradition of, let's say, Humboldt. So, I grant that you may >> be a victim of propaganda yourself, instead of being a secret >> service sposored agent. You took quite some time to formulate your >> post, though. >>> You acted to contain the discussion about AI in this vital and >> important informatics community to technical detail, when it is >> neccessary that academics and community members look beyond the >> narrow borders of their certifications and shift their thinking to >> the point of view where they can see what technology does in the >> real world. >>> >> I offered a view point with sources. In no way the discussion >> about >> Apparent Intelligence was contained and, on the contrary, I tried >> to >> offer arguments from cognition and philosophy, beyond technical >> details, >> which have been pretty absent in the general discourse and rush >> gold of >> technoutopism / technoapocalypse, conspiracy and paranoia. Other >> sources >> could be added to enrich and inform de conversation and they would >> be >> pretty welcomed (at least from me and several members I've known > >from >> this community) >> >> Best, >> >> Offray
RO
Richard O'Keefe
Mon, Mar 27, 2023 5:31 AM

I tried it on three "dead" languages:

  • a bubble-sort in Pop-2
  • a system of linear equations solver in IMP-80
  • the array concatenation example in SNOBOL 4 from Rosetta Code.
    It got the first one right -- despite 'array' access
    being written 'subscr(index, array)' -- and the second
    one almost right -- mistaking an nxn array for a 2n
    array. It got the third one (which I had stripped of comments)
    right too.

It's impressive.  And it's impressive because it's clearly
NOT based on analysing a large corpus of code in the language
you give it (because there IS no corpus of Pop-2 and the
available corpus of IMP is both limited and very sparsely
commented.  Not much SNOBOL either.

On Wed, 15 Mar 2023 at 20:07, Tim Mackinnon tim@testit.works wrote:

Of course everyone is talking about ChatGPT, and I saw a post to
ExplainCode and didn’t expect it to understand Smalltalk, so was impressed
that it does! Presumably the Tonel Smalltalk code on GitHub has helped
along with numerous articles?

Try it out: https://whatdoesthiscodedo.com/

I gave this simple example: (1 to: 100 by: 3) reject: [:i | i isOdd]

And got this (neat eh):

The code above is written in the programming language Smalltalk.

It creates a collection of numbers from 1 to 100, incrementing by 3 at
each step. This is done using the "to:by:" construct where the starting
value is 1, the ending value is 100, and the increment is 3.

The resulting collection is then filtered using the "reject:" method,
which takes a block of code as an argument. In this case, the block checks
whether each element (denoted by the variable i) is odd by using the
"isOdd" method. If an element is odd, it will be rejected from the
collection.

So, the final result is a collection of all even numbers from 1 to 100,
incrementing by 3 at each step.
Share this explanation with colleagues: whatdoesthiscodedo.com/g/a0ec56e

I tried it on three "dead" languages: - a bubble-sort in Pop-2 - a system of linear equations solver in IMP-80 - the array concatenation example in SNOBOL 4 from Rosetta Code. It got the first one right -- despite 'array' access being written 'subscr(index, array)' -- and the second one almost right -- mistaking an nxn array for a 2n array. It got the third one (which I had stripped of comments) right too. It's impressive. And it's impressive because it's clearly NOT based on analysing a large corpus of code in the language you give it (because there IS no corpus of Pop-2 and the available corpus of IMP is both limited and very sparsely commented. Not much SNOBOL either. On Wed, 15 Mar 2023 at 20:07, Tim Mackinnon <tim@testit.works> wrote: > Of course everyone is talking about ChatGPT, and I saw a post to > ExplainCode and didn’t expect it to understand Smalltalk, so was impressed > that it does! Presumably the Tonel Smalltalk code on GitHub has helped > along with numerous articles? > > Try it out: https://whatdoesthiscodedo.com/ > > I gave this simple example: (1 to: 100 by: 3) reject: [:i | i isOdd] > > And got this (neat eh): > > The code above is written in the programming language Smalltalk. > > It creates a collection of numbers from 1 to 100, incrementing by 3 at > each step. This is done using the "to:by:" construct where the starting > value is 1, the ending value is 100, and the increment is 3. > > The resulting collection is then filtered using the "reject:" method, > which takes a block of code as an argument. In this case, the block checks > whether each element (denoted by the variable i) is odd by using the > "isOdd" method. If an element is odd, it will be rejected from the > collection. > > So, the final result is a collection of all even numbers from 1 to 100, > incrementing by 3 at each step. > Share this explanation with colleagues: whatdoesthiscodedo.com/g/a0ec56e >
D
danhunfeldz@mail.com
Mon, Mar 27, 2023 12:49 PM

I asked ChatGPT to write “Hello, World!” in different esoteric programming languages:

Kylix:

program HelloWorld;

begin WriteLn('Hello, World!');

end.

ActionScript (Flash 8):

trace("Hello, World!");

Limbo:

include "sys.m";

include "draw.m";

include "arg.m";

draw->init(nil, "hello");

sys->print("Hello, World!\n");

Alef:

include "sys.m";

include "draw.m";

init(nil, nil);

write("Hello, World!\n");

Alef was the predecessor of Limbo, which is the ancestor of Go. The results are not even close. ActionScript not even close. Kylix was wrong too. None of them are way out there. If anyone has worked on Plan 9 then they have heard of Alef and Limbo. Limbo is used for Inferno OS which was created from Plan 9 for set top boxes and came out about the same time as Java. What ChatGPT gave are not even starting points, except maybe Kylix and it derived the answer given from Delphi.

I asked ChatGPT to write “Hello, World!” in different esoteric programming languages: Kylix: `program HelloWorld; ` ` begin WriteLn('Hello, World!'); ` `end.` ActionScript (Flash 8): `trace("Hello, World!");` Limbo: `include "sys.m"; ` `include "draw.m"; ` `include "arg.m"; ` `draw->init(nil, "hello"); ` `sys->print("Hello, World!\n");` Alef: `include "sys.m"; ` `include "draw.m"; ` `init(nil, nil); ` `write("Hello, World!\n");` Alef was the predecessor of Limbo, which is the ancestor of Go. The results are not even close. ActionScript not even close. Kylix was wrong too. None of them are way out there. If anyone has worked on Plan 9 then they have heard of Alef and Limbo. Limbo is used for Inferno OS which was created from Plan 9 for set top boxes and came out about the same time as Java. What ChatGPT gave are not even starting points, except maybe Kylix and it derived the answer given from Delphi.