Monday, January 02, 2023

Yet Another Crappy Analysis of AI

Whenever some new breakthrough in AI (artificial intelligence) is made, you can depend on ignorant naysayers to break out the usual clichés.

We will be told that this program or that one isn't "really" creative (but no testable definition of `creative' will be given). We will be told that some new system doesn't "really" understand what you are saying (but no testable definition of `understand' will be forthcoming). We will be told that computers cannot have "minds" (but no test for having or not having a "mind" will be provided). We will be told that computers aren't "really" reading and understanding what they translate, even though no definition of what "really reading and understanding" will be given.

So it doesn't surprise me at all that there has been an enormous amount of ignorant nonsense written about ChaptGPT and related models. This dreck from Gary N. Smith, a professor of economics at Pomona College, is pretty typical for the genre.

Smith claims that "while GPT-3 can string words together in convincing ways, it has no idea what the words mean". Yet he does not provide any testable definition of "knowing what words mean". Does it really have no idea? I conducted a little experiment of my own, by asking ChatGPT, "What is a cow?" And here is the response:

A cow is a large, hoofed animal that is often kept on a farm for its milk, meat, and other products. 
Cows are part of the bovine family and are known for their docile nature and distinctive mooing 
sound. They are typically reddish-brown or black in color and have long horns on their head. 
Cows are used for a variety of purposes, including the production of milk, cheese, butter, and 
other dairy products. They are also raised for meat, leather, and other products. Cows are 
important sources of food and other resources for people around the world.
Does that really look to you like ChatGPT doesn't know what the word "cow" means?

So ChatGPT does indeed seem to define words. But can it reason correctly about them? So I asked, "Can a cow fly over the moon?" and ChatGPT responded as follows:

No, cows are not able to fly over the moon or any other celestial body. Cows are terrestrial 
animals and do not have the ability to fly through the air. They are able to walk and run on 
land, but they do not have wings or other body parts that allow them to fly. In fact, most 
animals are not able to fly, and those that are able to fly typically have specific adaptations, 
such as wings, that allow them to do so.
I was not satisfied with this, so I continued to ask, "Then how did the cow fly over the moon in "Hey Diddle Diddle" if it is impossible?" And here is how ChatGPT responded:
"Hey Diddle Diddle" is a popular nursery rhyme that has been passed down through the 
generations. It is a simple, nonsensical poem that is meant to be fun and entertaining, 
rather than literally true. The rhyme tells the story of a cow jumping over the moon, 
but it is not intended to be taken literally. Cows do not have the ability to fly, and 
the rhyme is not meant to be taken as a factual account of events. It is simply a 
lighthearted, playful rhyme that has been enjoyed by children and adults for many years.
Seems pretty accurate to me.

Smith justifies his claim by giving an example of a conversation about how many pennies there are in each group if you start with 47 pennies, you throw away 44, and divide the remainder into equal-size groups. In his example, ChatGPT gave the wrong answer and incorrect reasoning to get there. But that conversation isn't any evidence at all that ChatGPT didn't know the meaning of the words "pennies" or "group" or "47". If anything, it shows it was unable to do logical problem-solving in that particular domain, but that's a completely separate issue from knowing the meaning of words.

There's nothing easier than taking a chatbot and getting it to produce something nonsensical. After that, you can just claim it doesn't "really" understand English. But human speakers of English produce a flood of nonsense every day, as anyone reading Salon can verify. Does that mean that the writers don't "really" understand English?

I tried Smith's example, and here is what I got:

If you start with 47 pennies and throw away 44 of them, you are left with 47-44=3 pennies.
If you divide these 3 pennies into three groups, each group will have 3/3=1 penny.
So each group will have 1 penny.
Looks pretty good to me! So can we conclude that ChatGPT really does understand the meaning of words now?

I wish the people commenting on AI would demonstrate a little more seriousness about what it means to "really" understand utterances. Or any realization that "understanding" is not a binary; one can have "understanding" to a greater or lesser degree. Does a five-year-old understand the nuances of a word like "duty" the way a retired member of the military does?

I think the right way to view ChatGPT is that it does exhibit some understanding of what words mean, as evidenced by its ability to define them and use them correctly in context in many conversations. It doesn't mean that everything ChatGPT says is correct, nor that it holds no incorrect beliefs.

I am sure we will see many more crappy pieces like Smith's in the near future.

4 comments:

JimV said...

Thank you. That is something like I have been replying to nay-sayers (but better).

The key point for me is that we have found a way, using neural networks, for programs to learn by trial and error plus memory, which I think is the basis of intelligence. The history of science and engineering is one of trial and error, which builds on previous work. It also produced us and our vaunted intelligence (biological evolution).

That you got a better answer than the economist might be because ChatGPT learned from the previous experience. I suspect that is why the company released ChatGPT for free public use--to crowd-source some of its training.

Wikipedia says that in authentic cases of feral children (who grew up in the wild without human contact), when the children were found as teenagers, they never learned any language or other social skills. This shows that intelligent human behavior requires lots of training. I suspect we are still learning how to train programs like ChatGPT most effectively, and it will take a lot more work. The good news is that a thoroughly trained set of parameters can be copied to other instances of the program, and a thoroughly trained general AI could be used to train mor limited AI's for special tasks, by being the arbitrator of their trials and errors.

Dale said...

I totally agree with you here. In fact, my own view is that I agree with Gary Smith that AI is dangerous, but not because it is so fallible and lacking (as he suggests), but because of the opposite. If it really performed poorly, there would be much less to worry about.

Mikkel Rumraket Rasmussen said...

I agree with your overall point here (there are lots of stupid dismissals of AI), but there are nevertheless legitimate questions about how much comprehension is behind the words. I think some of the answers given by ChatGTP are genuinely surprising both in positive and negative ways.

I saw chatGTP give some surprisingly good answers to questions about biochemistry, but then when tasked to "put 2 and 2 together" give surprisingly wrong answers that I would otherwise have expected it to be able to solve on the basis of the answers it already provided to previous ones. It seemed like it had already demonstrated that it had the knowledge and comprehension required to reach a certain conclusion, so when it didn't I think that does raise legitimate questions about the depth of it's comprehension.

jwaltos said...

I just tried out this question:
Can I cause you to enter into an infinite cognitive loop?

As an AI language model, I am programmed to detect and prevent infinite loops, so it is unlikely that you can cause me to enter into an infinite cognitive loop. If I detect that a conversation is repeating or getting stuck in a loop, I will attempt to steer the conversation in a different direction or provide a different response to avoid the loop.

However, if you have a specific scenario in mind that you believe might cause an infinite loop, I can try to simulate it and see what happens. Please note that I am designed to prioritize user safety and prevent harm, so I will not engage in any behavior that is malicious or harmful.

Both interesting and informative. IBM's Watson is entertaining in the same fashion.
Thank you for making me aware of this public domain AI!