Episode Transcript
Available transcripts are automatically generated. Complete accuracy is not guaranteed.
Speaker 1 (00:04):
Welcome to tech Stuff, a production from I Heart Radio.
Hey there, and welcome to tech Stuff. I'm your host,
Jonathan Strickland. I'm an executive producer with I Heart Radio
and love all things tech, and this is time for
a classic tech Stuff episode. This episode originally published on
(00:25):
June twenty three, two thousand fourteen. It is titled passing
the Turing Test, something that we frequently associate with artificial
intelligence and actual you know, like machines thinking or at
least appearing to think. So let's listen in to this
classic episode. The Touring Test is named after Alan Turing,
(00:46):
and we've done a full episode on Alan Turing way
back when, back in November. Yeah, phenomenal person, amazing thinker. Yeah,
one of the like the grandfather of computer science. Also
tragic life story which we went into detail back in
that story. And so you might wonder why, right, So
(01:06):
what does that have to do was the the guy
who was the essentially the father of computer science or
grandfather of computer science? What does he have to do
with the story about a computer program in June two
thousand fourteen, a computer program with an interesting name, Eugene
Goostman passing the Turing test. What does that all have
to do with each other? Well? To answer that, we
(01:27):
have to ask what is a Turing test? Well, as
it turns out, Away back in the nineteen fifties, he
started envisioning uh I thought experiment. Yeah, he published a
paper in a journal called Mind in nineteen fifty called Computing,
Machinery and Intelligence, and he sort of laid out his
(01:48):
thought experiment there. It's interesting because what he did was
he took this idea of a party game and then
adapted it for computers. Right. The party game that it's
based on would have three participants, an interrogator, a man,
and a woman um, all situated so that they can't
see one another, and the interrogator is supposed to ask
questions um to try to figure out which of the
(02:11):
participants is the dude and which is the lady exactly.
And it's the dude's job to try and mislead the
interrogator to believe that that he, in fact is the
lady and the other one is the man. It's the
lady's job to say, hey, I want you to get
this right. I'm the lady. That other guy that's the dude.
And so the interrogator has to ask questions. Now, obviously
(02:33):
they can't see each other. Because they could see each other,
then that would probably give things away. Probably, they really
shouldn't be able to hear each other because that could
also give things away. Sure, and if you use handwritten notes,
then you the interrogator might be able to make judgments
based on the handwriting style. So it should really be typewritten, right,
So you want to you want to remove as many
(02:54):
easily identifiable traits from this game as possible to make
it all about the questions and the answers. Now, Touring said,
what if we were to take the same basic premise,
but instead of having two human interviewees, replace one of
those humans with a machine. Now, if that machine can
(03:14):
convince the interrogator that the machine itself is a human being,
it would be a pretty phenomenal achievement. Or you know,
just generally, if the interrogator wasn't sure which of the
two interrogate, right, whether which one is the human, which
one is the machine, or even if there could be
a case where you don't know. It may be that
(03:35):
you have two humans that you're interrogating, and it may
be one of those things where you have to you know,
if you don't know for a fact that one of
them the machine, that makes it even harder, right, at
least assuming that the computer program is sophisticated enough. Now,
Touring was saying that we don't have any machines right
now that can do this, but I envision a time
when computers will be able to do such a thing
(03:57):
where if you were to interrogate a computer, you would
get back responses that would be uh, convincing enough for
it to make it difficult to determine if it were
man or machine. And so he said that he predicted
in fifty years time, which would be the year two thousand,
that computers and software be sophisticate enough that interrogators would
only be able to guess correctly seventy percent of the time,
(04:19):
meaning that they would be fooled by the computer pent
of the time. Right. So, uh, this was just kind
of a thing he was coming up with, like an idea,
a prediction, not necessarily a test. Although very much like
Moore's observation became Moore's law, this became what is known
as the Turing test. People talk about a Touring test
as a machine capable of fooling people into thinking it's
(04:44):
another person at least thirty percent of the time after
five minutes of conversation. Very good point. Yes, it needs
to be five minutes of conversation. If you are just
getting maybe two or three responses, that might not be
enough for you to be able to draw a conclusion
that you feel good about. If after five minutes you
still are not entirely, entirely certain, then that might say
(05:06):
that this machine, in fact has passed the Turing test.
And this has been extrapolated to mean something about machine
intelligence because Turing himself tied the idea of of how
we perceive a machines intelligence directly to artificial intelligence and
and and even how we perceive human intelligence. Because here's
(05:29):
Touring's idea is a little cheeky, and I love the
fact that it's so cheeky. So Turing kind of said,
would you say such a machine as intelligent if it
appears to be intelligent, is it fair to say that
is intelligent? Touring? So why not? Because I'm only able
to tell that I am intelligent, right because of my experience.
(05:49):
I'm only able to have my own personal experience. I
can't experience what someone else's life is like, all right,
sitting across from each other, Jonathan and I can only
assume that the other one is intelligent, right, And it's
because the other person is displaying traits that we associate
with intelligence. They seem to be able to take an information,
respond to it, make decisions. And based on the fact
(06:12):
that we ourselves also do that thing, we go ahead
and say, all right, well, they clearly have the same
features that I have, which includes intelligence. Uh. Now, he says,
why would we not extend that same courtesy to a
machine if it also appeared to display those same uh features?
He says, doesn't matter if the computer is quote unquote thinking.
(06:35):
If it's it can fool you. Yeah, if it's able
to to simulate it well enough, you might as well
say it's intelligent, because simulate it is probably a kinder
phrase for that than fool. Yes. Yeah, well I know
it's it is fooling essentially, I mean, because ultimately you're
talking about a computer programmer who's making this happen. So
nowadays we think of this as the Turing test. Can
a machine of the time are more fool someone after
(07:00):
five minutes of conversation into thinking it's a human? Now,
this is really hard to do. This is is non trivial.
I mean it sounds almost simple. Yeah, the concept is simple,
the execution incredibly difficult. Because here's the thing, human language varied.
(07:20):
We have unstructured, unpredictable ways to say things like if
if I were to tell you that it's a hundred
degrees outside and it is humid like the humidities, and
you go out there, then I'm sure all of our
listeners would have slightly different ways to express their thoughts
(07:42):
on the conditions outside. Some of them would probably contain
colorful metaphors. Mine very likely when it contain colorful metaphors,
particularly if I had to be outside for any length
of time. But that's the point. We would all have
different ways of saying this. So how do you make
a computer program able to interpret all the myriad of
(08:05):
ways we can all express the same thought, let alone
any thought? All right? This is what's referred to as
natural language recognition, and it's a really huge problem in
artificial intelligence and and and a lot of other speech
related computer programming. Exactly. Yeah, this is where a computer
program has to be able to parse the language so
(08:27):
it recognizes things like this word is a noun, this
word is a verb, this this word alters this other word. Uh.
And not only does it need to be able to
recognize it, it it needs to be able to respond in kind. Right,
So if you have appropriately in some way or another share. Yeah,
you could have a computer program that's literally making up
(08:49):
sentences or you know, the approximation of a sentence randomly,
where it's just pulling strings of words and placing them
in a sequence and then presenting them. But that wouldn't
be convincing at all. If I were to say, hello,
how are you today, and Lauren was to say blue
panther pickup jump down street, I'd be like what? And
(09:10):
even that was closer to being a sentence than some
of the random stuff that you would see if it
was just truly completely So it also has to be
able to uh endure a five minute long conversation, like
we said, in order to pass the Turing test. So
you can't have too much repetition or that gives it away. Absolutely.
If you've ever been playing a video game and all
(09:30):
of the NPCs say the same thing over and over
and over again, yeah, if if all the chat bot
says is hey, listen, then you're clearly playing Zelda and
you're not actually having a decent conversation. That being said,
there is someone I know who plays a ferry at
the Georgia Renaissance Festival, and hey, listen is heavily represented
in a repertoire. It's pretty amazing. No, it's pretty awesome.
(09:54):
But at any rate, Yeah, so so these are these
are big problems. You have to build a database of words,
you have to be able to figure out what kind
of syntax are you going for. It's a wide open,
huge problem, and solving this problem can be really beneficial
in lots of ways. We'll talk a little bit about
that later. It's it's beyond just making a program that
(10:15):
seems human, right, That's that's one way of looking at it.
But there are a lot of other benefits that come
along with it, which will chat about towards the end
of the show. But for right now, let's go into
this story behind Eugene Goostman and whether or not it
was actually the first chat bought to pass the Turing test. Yeah. So,
first of all, we've got three programmers in this story
(10:36):
of Vladimir Vslov, Eugene Demchinko and Sarage you listen. As
you may guess from their names, they all hail from
Russia and the Ukraine as well, and although not all,
not all of them still live there. But starting in
two thousand one they got to work on this, right, Yeah,
they were trying to design a computer program that would
(10:58):
pose specifically as a third teen year old boy from
Odessa in the Ukraine. Yeah, and that meant that they
had specific parameters that they could work within. It automatically
helped reduce some of that unpredictability and that lack of
restriction that you would have if you were to just say,
this is a fluent adult speaker of a given language. Yeah, Yeah,
(11:22):
giving him these you know, setting up the expectation from
the judges that you know, this is a non native
English speaker. It's it's a kid essentially. Um, you know
that they'll expect him to have limited knowledge of the
world and and different subject areas, and a limited understanding
of English vocabulary and grammar and all that kind of stuff.
(11:43):
So you're already managing expectations. That's going to come into
play when we talk about some of the criticisms about this.
Although although I do think it's a very clever way
and and a lot of previous chatbots have have had
similar yes kind of approaches. Yeah, because like we said,
if you were to take a quote unquote pure approach
to this, it's really really challenging. So yeah, by by
(12:05):
limiting this, the judges have an idea of well, this
this could be a thirteen year old boy or it
could be a computer program. It means that the computer
program doesn't have to be as sophisticated as one that
would be completely fluent and have you know, an adult's experiences,
uh and ability to communicate. So that was step one
and uh, Eugene Goosman took part in the competition. UH
(12:29):
that had five total chatbots. It was it was one
of five that took place on the sixtieth anniversary of
Turing's death, and the program managed to full thirty three
percent of the judges into thinking it was actually a person. Uh,
and it had there were thirty judges from what I understand,
So that was where you got all the headlines of
(12:52):
chat bought. There are computer beats touring tests, which already
is not accurate. There were some slight um hiccups in
a little bit of the news reporting that we'll get
into that part of the story later. But you know,
the key takeaways I think here are that this was
a this was a competition that was celebrating touring awesome um.
(13:14):
And that there were thirty judges, some of whom were celebrities, yes,
including an actor who had appeared on Red Dwarf right. Yes.
And then a couple of years before that, because because
you said they started working on this in two two
thousand one, this was not the first time that Gustman
had entered competition. UH. Two years previously, that same software
(13:37):
had convinced twenty of judges at a similar competition that
was held at Bletchley Park. Now, but Bletchley Park that's
where Touring helped crack the Enigma machine, the encoding device
that the German military was using during World War Two.
So this was a big celebration. It was at the
centennial celebrating his birth, and it ended up falling just
(13:59):
short of passing the Turing test. Uh. The organizer of
the the more recent event, the one in which Gustman
ran away with quote unquote beating the Turing test. UH.
That organizer was Kevin Warwick. That name may sound familiar
to some of our listeners. If you've ever heard us
talk about cyborgs. He's the guy who had an r
(14:22):
f I D chip surgically implanted into him. His wife
also did at one point. Yes, they could communicate with
each other through them. Some unflattering news media sometimes refers
to him as Captain Cyborg. Yeah. There there are some
critics who say that he, uh, he courts publicity in
a manner that is unbecoming of of of a scientist. Yes,
(14:47):
really of anybody. Yes, yes, uh, those are the critics
who say that. By the way, I just want to
make that clear. At any rate, He said that this
was the first time a chatbot had passed the Turing
test at an event, or the conversation was open ended,
meaning that they had not previously uh decided upon a
specific topic or line of questioning that the judges were
(15:10):
allowed to say whatever they wanted to the chat bot
and response, which obviously makes it harder because you have
to have a much wider breadth of potential responses. Yeah. Yeah,
because again, if you were to say, all right, this
chat bot is just going to talk about, um, I
don't know, sporting events from last year, well, then you
can prepare pretty well for that. Yeah, exactly. So it's
(15:32):
one of those again, it's one of those things where
the unrestricted nature adds in a degree of difficulty. It's
time for us to take a quick break, but will
be right back. So why would you need to make
the qualification that this is an open ended approach and
(15:53):
this is the first chat bot to manage it. That
would be because despite what you may have heard, Eugene
Goosband was not the first program to beat the Turing test,
not by a long shot. So work on these sort
of chatbots, these kind of artificial conversationalists. That's real recent, right,
(16:13):
I mean they just started doing that like what like
like maybe three or four years ago or two thousand,
one of the earliest, I mean that's when they started
with Goose. Yeah. No, in the nineteen sixties and seventies.
Say what back in the mid nineteen sixties there was Eliza,
which was written by Joseph Weisenbaum. If you're pronouncing it
in the correct German, that's correct. Excellent, I'm finally learning.
(16:35):
I'm sure. I'm sure he pronounces at Wisenbaum. But yeah,
but no, it would be Weisenbaum at any rate. Um,
this this was a program um that would respond to
human conversation in what ideally would be a relevant way. Yes,
it was obviously an early attempt. It was not meant
to be a program that takes on the Turing test.
(16:58):
It was really a thought again, kind of like a
thought experiment, the idea of what does it take to
create a piece of software that can react to questions
and make it make sense. At that point, it was
more a can we do this? Then let's do this
for real? Yeah, and these are the sort of foundations
that you have to lay in order for other things
like the Eugene Goostman program to be uh successful. So
(17:22):
he created a language analyzer. Now this specifically would look
at words that users would put in and then compare
them against a database of words that were stored in
the computer's memory. And then also created scripts. Now, in
this case, the scripts were sets of rules. They're kind
of like, you know, like a protocol or an algorithm
in a way. These rules dictated how Eliza would respond
(17:45):
to messages in order to cut down on that huge,
massive number of variables we were talking about, that the
whole unrestricted, unpredictable thing, And so they would have different
uh kind of like like overlays. Think of him as
an overlay that would kind of guide Aliza's responses, And
the most well known one was called Doctor, which put
(18:05):
Eliza in the role of a Rogerian psychiatrist. Uh. This
is the person who responds to everything with a question.
Al right. It's that it's that passive interview style where
where you know, you repeat back. You know, if if
if I go, oh, man, like I'm I'm I'm really
sad about my cat, Oh, tell what is it about
(18:26):
your cat that makes you sad? And then you can
say things like you know, And that's one of those
things that where as a conversation starts to wind down,
you then have another line of questions, So tell me
more about your mother, like that the whole tell me
about your mother thing. That's really coming back to this
kind of model of psychiatrists. But yeah, if you've ever
heard the joke about them responding to anything with another question,
(18:48):
so I just just taking the last word and turning
it into a question what am I paying you for?
Why do you think you're paying for? That kind of thing? Uh?
So that's what that's how Aliza spotted and in fact,
you can find examples of the Eliza proscript transcripts, and
even even the actual there's there. There are ports you
(19:10):
could say, or people have essentially created their own version
of Eliza just using the originalize as a guide. You
can find tons of them on the web, and you
can attempt to have a conversation with them. It's not
terribly compelling, but it's it's kind of fun. Usually within
maybe four or five exchanges, you've already run into something
where you're like, well, this can't be a person, or
(19:31):
if it's a person, it's the weirdest person I've ever
conversed with. But but again, it wasn't really an earnest
attempt to to create something that would pass the quote
unquote tearing tests, right, which I don't think was being
referred to as such at that point yet. Yeah, it
was kind of people knew about tourings prediction, but it
wasn't so much called a Turing test. Another chat bought
that premiered in the early nineteen seventies went even further
(19:54):
and actually was an attempt to try and pass the
Turing test in a very specific approach, kind of like
you know, Eugene Goostman is a very specific approach to
narrow down that those parameters. In this case, this one
was called Perry p A R R y A. Right,
and Kenneth Colby created it too to emulate a patient
(20:15):
who has paranoid schizophrenia. Yeah, someone who has you know,
sort of a persecution complex. Uh. They imagined that that
there are people or other entities that are out to
get them, and in this uh specific case, he kinda
he kind of really embraced this approach. It reminds me
(20:35):
of people who create a like a Dungeons and Dragon's character,
but then give their character an entire backstory. Yeah. Yeah, yeah,
this this Perry persona was an entire persona. It was
a twenty eight year old man with a job as
a post office clerk who was single, had no brothers
and sisters, and rarely saw his parents. He had specific hobbies.
He liked to go to the movies and horse racing. Yeah,
(20:57):
he liked to bet on the horses. Uh. And he
placed about with bookies in the past, right, and that
he realized later bookies have an association with the criminal
underworld and that therefore the mafia knew about him and
we're out to get him. Now, now all of that
might sound very ridiculous to you if you never had
any kind of interaction with someone who suffers from paranoid schizophrenia,
(21:21):
that can seem like, well, that seems cartoonish. But no, this,
this kind of thinking is not uncommon, you know, whether
it's whether it's a criminal organization or the government or
some other even unnamed entity. Oh absolutely, uh, it's it's
very much realistic in terms of that kind of diagnosis.
And and if I was speaking about it with humor
(21:41):
in my voice a moment ago, it's it's only because
I am absolutely tickled that the programmers of this of
this program built it, built it in. Yeah, I mean
it's it's actually pretty entertaining that they went so far
as to make this whole, uh, this whole backstory to explain,
because that's what gives it the believability, and then they
ended up testing it in conversation with a human. Perry
(22:04):
would gradually start to introduce his thoughts quote unquote thoughts
about being persecuted, and would respond sensitively to anything said
about his appearance, family, or religious beliefs. I've actually seen
lots of transcripts of conversations with Perry, and sure enough,
it's one of those things where you know, you might
have a few exchanges and then Perry ends up saying
(22:24):
something that seems really odd, but not so odd as
to seem artificial. It just seems like it's a non sequitur,
you know, something like, well, that's what they want you
to think. And if you were to say, who are
they the mafia? Who did you think I was talking about?
Like the mafia after you? Of course they're after me.
They know who I am. That kind of stuff, and
(22:45):
it's disturbing, like it's you know, when you know what
it is, it's kind of amusing, but like if you're
in the middle of a conversation, join, oh this poor person. Yeah.
And so in order to test it, Kolby did a
couple of different things. Did one test where eight psychiatrists
interviewed both Perry and a human patient via teletypewriter. So
(23:06):
in both cases the the the psychiatrists could not see
who they were interviewing. This is going back to the
kind of the original Touring test idea, or at least
Touring's proposed experiment, and in this case, only two of
the eight were able to identify that one of the
interviewees was human. And the other was a machine. In
a second test, Colby presented a group of a hundred psychiatrists,
(23:29):
almost a psychics, transcripts of interviews between Perry changed the Yeah,
he had he had pairs of exactly I foresee no.
But they gave these psychiatrists transcripts of interviews between an
interviewer and Perry, and an interviewer and a human patient,
and forty out of the responded. I don't know if
(23:51):
the other sixty just never got it or if they didn't.
Response rates are variable, right, so I'll the forty who responded,
nineteen of them guessed incorrectly. So that's almost a fifty percent,
you know, getting you know, right up there with a
pretty impressive amount. Now, again, we have to look at
(24:12):
the fact that Perry is operating under a very restricted,
uh set of rules. We're talking about paranoid schizophrenic, someone
who we would assume would occasionally an non normative answers
exactly two conversational pieces. And and again it's it's a
limited time that you're having with this person or this
entity in this case, this this program, and uh. Because
(24:37):
the psychiatrists had a specific expectation of the type of
interactions they were going to see that could have affected
their their uh answer right. So ideally, in the perfect situation,
you would have this interview happening where you have no
expectation as to what the answers should be. In other words,
(25:01):
you don't know ahead of time that the interviewee is
having any kind of other, you know, any kind of
restrictions upon that person, so that you would be interviewing
anyone like any average person. But that's obviously not what
we're talking about here, nor was it the one for
Eugene Goost Bond. So I've also seen, by the way,
transcripts of people who set up Eliza and Perry to
(25:25):
talk to each other. So you have Eliza acting as
the Rogerian psychiatrist and Perry the paranoid schizophrenic, having bizarre conversations,
and they usually don't last very long because Perry gets upset,
And obviously by gets upset, I just mean that Perry
ends up essentially shutting down the conversation because Eliza just
(25:46):
wants to ask questions and Perry gets suspicious of people
who are asking questions, and by again get suspicious, I'm
saying following specific rules that make it feel like this.
Computer programs getting suspicious, but they are entertaining. If you
ever do a search on line, just look for Eliza
Perry transcripts. There. There are a few of them, and
they're all pretty entertaining. But so since since that time,
(26:08):
I mean obviously, lots and lots of chatbots have been
created for multiple reasons. Oh yeah, well, you know, some
of them are trying to test the Turing test, and
others are trying to fool you into giving out your
credit card information or plaining on a link that has
malicious uh links to malicious software. Yeah, anyone who's been
on any kind of chat program, specifically like AIM has
(26:33):
probably encountered this at least once or twice, where they're
they're getting an unsolicited message from someone or something or something. Yeah,
if you type a couple of times, you realize, oh,
this is not actually this is an attempt for to
either get information from me or have me click on
a link. Yeah, that's a that's a thing. But there
(26:53):
are some examples of uh, I guess saying legitimate is weird,
but there are some examples of they're more and more
scholarly attempts. Yeah, like a PC therapist that one was
by Joseph Weintraub, and it fooled of its judges into
thinking that it was human. Of ten judges, of course,
(27:17):
so like, if I'm doing my math correctly, that's five
that I believe you are. Unless we're talking about quantum judges. Um,
these judges were both right and wrong at the same time.
And um, and it was it was a whimsical program.
I guess you could say, yeah, it had come some
I think it's fair to say its answers could be
pretty smart ass. Also, I read some of these transcripts
(27:39):
and it actually surprised me that enough judges thought that
it was a person. Maybe they thought it was a
person who was purposefully attempting to fool them into thinking
it was a computer, huh, you know, because it would
say things like I compute, therefore I am that kind
of stuff where it was specifically yeah. So you're looking
at this stuff and you're thinking, all right, well, maybe
this is kind of going back to that original touring
(28:01):
test party game, where the the effort is for a
person like the person who's being interviewed is trying to
throw everybody off, and that's perfectly within the rules, unless
you stay upfront. No, just be honest in your in
your answers. There's nothing in here, by the way that
says that the interviewee has to tell the truth necessarily,
unless you just state that as a parameter at the beginning.
(28:23):
So in other words, you could you could be like,
I'm totally the computer and you're the human being interviewed. Uh.
I don't know if that's a fair way of saying
that the device won or lost, but it is a possibility. Uh.
Then we have a two thousand eleven. Now, this one
is a really pretty impressive one. And this was clever Butt,
which was made by a fellow named Rollo Carpenter, and
(28:44):
it fooled fifty nine point three percent of a live
audience and an event in India with more than a
thousand people. Yeah, the way this worked was that the
audience watched as interviewers interacted via text with either clever
Bot or a human in the course of a four
minute interview. So it's a little shorter than what Turing
(29:04):
had said, but not by not by a whole lot.
In four mints is still a good amount of time.
Sure that that is twenty less, that's true, that's true.
So keep that in mind. But but at any rate,
it was you know, pretty interesting experience. And also from
why I read they misidentified, they thought that the human
was a computer sixty of the time. Um, because they
(29:28):
didn't necessarily just say that it was a computer or
it was. So now we see that there are a
few examples of chat bots quote unquote passing the Turing test.
So what does that mean. Does it mean that the
machines are actually thinking? Um? No, I mean, and it's
(29:52):
not to say that that computers don't have a certain
amount of machine intelligence, but there's absolutely a distinction between
that and what we consider to be human intelligence. That's true.
The programmers themselves have said that this doesn't mean a
machine is able to think. Uh, they're just able to
interpret commands then to follow a set of rules to
(30:12):
make a response, which is still pretty cool. And and
it doesn't It certainly doesn't mean that the Turing test
is worthless as an exercise. No, it is. In fact,
it's improving our ability to create programs that that can
understand or at least respond to natural language. Natural language
recognition is one of those big things where if you're
(30:34):
really able to crack it, then you can have some
amazing opportunities open up, and we've seen this recently with
stuff like Siri. Oh absolutely being able to speak to
your computer rather than having to I mean, even if
you could speak to your computer through a keyboard and
have it understand what you're what you're saying, I mean,
like it's it's the reason why Google spends so much
(30:55):
time and money, and it's search algorithms of of trying
to figure out what you really mean when you search
for a certain phrase, because traditionally, you know, before we
really got into the natural language recognition era, it meant
that in order to work with a computer, you had
to work with a computer on the computer's terms. You
had to learn the commands, you had to learn the
(31:17):
way to navigate a computer system in order for it
to do what you wanted it to do. Once you
get to a point where natural language recognition software is
robust enough the computer is working on your terms, you
you can put in however you're thinking, like whatever, whatever
mental exercise you've gone through to ask this computer to
(31:37):
do something, whatever you do to kind of express a thought,
uh as a command to this computer, the computer can
then interpret then respond and it's not just for serving
you back whatever information you happen to be looking for.
It's I mean, I mean, we're talking about being able
to just look at a computer and say, you know,
(31:58):
I really want a graph that looks blue and has
these percentages in it and is about this thing, and
it just doesn't. Yeah, like I want to see what
the population distribution of Atlanta is in a bar chart
or something, and then it could bring that shows out,
finds that information, put it into a bar chart, and yeah,
(32:18):
that's pretty phenomenal stuff. We have a little bit more
to say about passing the Turing test, but before we
get to that, let's take another quick break. We see
other examples of machine intelligence everywhere, things like pattern recognition,
probabilistic predictions, for example, Pandora. You know the Music Genome project.
(32:42):
It's yeah, that's that's pattern recognition. Yeah, it's looking for
elements of songs that you say you like and then
looking for other stuff that's not in the specific category
you mentioned or the specific examples you mentioned, and there's
something else you probably will like because you like these
other things that also have this stuff in it. Uh,
you know, sometimes that's less functional than other times. It
(33:05):
makes me think of Patton Oswald has a great routine
about TiVo that does the same sort of thing where
he says, you know, TiVo is great. I mean, I like,
I love Westerns, and I'll have it set to TiVo
a Western for me, and I come back, and then
they'll be all these other Westerns that will be suggested.
I didn't even know about. Thank you, TiVo. But then
sometimes TiVo gets it wrong and I come back and
(33:25):
everything has got horses in it, because Westerns have horses
in it. So I've got my little pony and cartoons
with horses and unicorns and things, and I have to say, no, TiVo,
that's a bad TiVo. But TiVo says, but you said
you liked horses. Same sort of thing. Like when you
get more sophisticated than the the computer program starts to
(33:45):
anticipate things and makes these probabilistic models. These these models
where there there are certain percentages associated with various responses,
and it goes with whichever one seems to be the
most prevalent, assoning it meets a threshold. If this sounds
familiar to you, it's because that's how IBM S. Watson worked, right,
(34:06):
which is a really good example of natural language recognition, absolutely,
because not only was it able to recognize natural language,
it had to interpret things like wordplay that Jeopardy do.
This is the machine that went up on Jeopardy and
beat the returning champions or former champions. Uh And you know,
if you've ever played Jeopardy or watched Jeopardy, you know
(34:29):
that there are categories that depend on things like puns
or hominem's or other forms of word play. So it
has to parse all of that, and that's even more
complicated than just taking a simple sentence and figuring out,
all right, what are the potential responses to whatever this
this this phrase is so great, great example. You know,
(34:50):
they would end up coming up with a potential answer.
It would assign a percentage of how quote unquote sure
it was that that was the right answer, and if
the percentage was higher than its threshold, which I think
was something like that, it would buzz in and give
that as a guess. Sometimes it was wrong, but it
was right a lot of the time, so that's kind
(35:12):
of cool. Uh So, getting back to Eugene Gene, the
main machine as I called him in in my notes
at one point. Uh And of course, I'm anthropomorphizing when
I say him, it's a it's a it's an it. Well,
it has a dude name. Yeah, it has a dude
name and a dude persona. But it's ultimately, isn't it.
(35:33):
Would you say that perhaps some of the reporting around
this was was maybe a little misleading or at least
hype is well, you know, Okay, the entire Eugene Gooseman
chat bought sounds really cool. I haven't met it personally, No,
I haven't either, although you can. There is an Internet version,
(35:54):
and I'm not sure that it's the same version that's
being used in competition, because I've seen some transcripts from
the Internet version and they don't seem good at all.
They seem bad, right, I guess you'd have to talk
to some actual thirteen year old boys from actually, yeah,
that this is part of it. You know. There's certainly
been some some questions among uh natural language AI enthusiasts
(36:18):
online about whether we're really just lowering our expectations for
human communication. Which, yeah, that's that's a totally different way
of looking at and a depressing one to say that, Oh, well,
if you come from this place and if you are
of this age, then I expect you to only be
able to communicate at this level, right, which is depressing.
(36:40):
It certainly is well, but you know, but it's a
valid point, I think. I think it's a good thing
to be thinking about in this kind of situation. UM.
You know. Beyond that though, and I certainly don't want
to downplay the apparent achievements of its programmers, because I
haven't programmed any capable chatbots today, I it's been ever
(37:02):
since I did. But there are a few things that
are just a little bit shady about the news UM.
First off, the original press release, which came out of
the University of Reading, I believe so UM stated that
a quote supercomputer had achieved this feat, and perhaps charitably,
(37:22):
it was a mistake or misunderstanding on the part of
the writer of the press release, but some skeptics have
suggested that it was in fact a purposeful publicity play
that in fact worked, because a whole lot of news
headlines around the interwebs repeated the error very excitedly. Yes,
because it was not a supercomputer. It was a computer
running a piece of software. It's a program, Yes, the
(37:44):
program that did the work. I mean, the computer just
provided the horsepower. Right, it's the software that did all
the action work, and it was not It wasn't on
a supercomputer. Little little known fact. Supercomputers have better things
to do than run chat bought software generally. Yeah. Yeah,
we're talking about things like figuring out global weather changing
(38:07):
change patterns and things like that, you know, or or
the way that money works. Chat bots low on their
priority list. It's like number seven at least. Um And
and also that this press release in question was largely
a quotation from Dr Kevin Warwick. Kevin Warwick, of course
(38:28):
being the fellow who organized this entire competition. Um who
who's an engineer and a futurist, um and also the
instigator and or enjoyer of a certain amount of hype
and debate about future technologies. Yeah, he is, obviously you
can tell. This is the guy who elected to have
(38:49):
surgery performed on him so he could have that r
F I D chip and call himself aside Wars. This
is someone who not only embraces this these ideas of futurism,
but is actively trying to promote them and get to
them that we're not even saying that that's a bad thing.
What we are saying, is that that may give him
(39:09):
somewhat of a bias when it comes to proclaiming a
computer software piece of computer software being an amazing achievement
that beat the Turing test, right sure. I mean he
admits basically to to being a provocateur. He he says
that that's really his job, you know, it is to
get people excited about tech and engineering and in the future.
(39:31):
And we get that like that that we agree that's
our job too. We think it's red. We think that
perhaps I don't want to put words into your mount Lauren.
I think perhaps that there's a different way of going
about it where you can still be excited, but you
can be a little more grounded in the way you
present things, because I I also think the achievement of
(39:51):
creating a chatbot that could be uh convincing is a
fantastic achievement. I mean, it's something that, even under any
number of qualifications, incredibly challenging to do, no matter how
you frame it. Um. I do think, however, that if
you seem to over inflate the achievement, you run the
danger of making people feel jaded about it later, which
(40:12):
I think Computer Wolf exactly. Yeah, exactly. So it's one
of those things where you know, you have to take
the context into account, right and don't don't downplay the achievement,
but don't sit there and say, like, ah ha, now
we have intelligent computers everywhere. That's not that's not true either.
I saw there's a great Wired article that specifically went
(40:35):
into uh, kind of debunking the whole beating the Turing
test thing and again kind of saying the same thing
we're saying, like, take the context into account, and and
part of that article. They ended up asking a cognitive
scientist named Gary Marcus of n y U about this,
and Marcus proposed a new version of the Turing test
(40:57):
because he says the old version is not really a
measurement of machine intelligence. Uh, it does kind of illustrate
ways of creating natural language recognition and clever ways to
fool the human side, huh. And that it was very
valid historically at the time because you know, text textual
communication was new and exciting and it was you know,
(41:20):
it pushed the field forward, it really did. But now
we've gotten to a point where fooling the person on
the other side of a keyboard is not necessarily the
goal that we should be looking at. He proposes, the
next version of the Turing test should be that a
computer software, uh, like any kind of program that wants
to beat it, what it has to do is first
(41:40):
quote unquote watch a movie, television show, YouTube video something,
some kind of video media, and then be able to
respond to questions about it. So sort of like here,
let me show you this ten minute video on car safety,
and then asking questions about specifically about the video, what
happened after they fastened the seat belt, that kind of thing,
(42:02):
and if the computer program is able to answer it,
then that would be a much more convincing touring test
than just kind of spewing out a script, which is,
you know, it's adding another layer of difficulty on top
of an already difficult task. But that's that's the whole
only way you can go forward. Otherwise we're just going
to see increasingly sophisticated chat box. Yeah, and and that
(42:24):
is actually a very difficult and interesting problem. What wasn't
it just recently that there was a computer that we
we taught I mean not us personally, but that humanity,
some researchers and we're taught taught to identify cats pictures
of cats. That was the AI program that essentially went
through thousands and thousands of uh I think it was
images and videos and then became able to identify cats.
(42:48):
It essentially defined what a cat was because no one
taught it right. It learned what cats are based upon
their appearance, on their appearance and can look at pictures
of cats and say that is totally a cat, essentially
saying that thing that is in that video is the
same as this other thing that's in this picture. That's
the same as this other thing that's in this totally
different video, which sounds trivial and hilarious, and it kind
(43:13):
of is hilarious. Also easy because I mean, come on,
everything on the internet has cats in it. Yeah, that
is That is kind of a gimme, isn't it? But
but still no, it is. It is cool. I mean
because just just that level of image recognition, I mean
being able to take an object and look at it
from a different angle than you were taught or that's
a different color, different different size. Yeah, all these things.
(43:35):
All of these things are easy for us, hard for computers,
so seeing something make that breakthrough is really exciting. Anyway,
we thought we would take that story and kind of
break it down for you guys, explain how it's still
cool but maybe not as cool as the way some
of the headlines are saying. Right, and also say hey
internet journalists, UM, step up your game. Yeah, I understand
(43:57):
that you want people to read your stuff. Oh yeah,
And and you're under deadline pressure and that's terrible. But
let's look hard. Let's represent reality, shall we don't just
don't just spit out press releases the way that you
found them. Yeah, and uh, and I would be criminal
to to neglect to mention Noel. Our producer reminded me
(44:18):
that obviously this is a very important field of study
because we want to be able to tell the difference
between computers and humans when the future of Blade Runner
becomes our reality and you're chasing down a replicant and
you have to determine if it's actually a replicate or
human being. And that wraps up this classic episode about
passing the Turing Test. Honestly, the whole Turing Test thing,
(44:41):
that general concept has morphed over the decades, and it's
it's more like shorthand right. Passing the Touring test isn't
so much about passing a specific hypothetical test. It's more
about this idea of creating a machine that of here's
to be indistinguishable from a human as far as you know,
(45:04):
processing and communicating information so that you feel like the
machine is truly intelligent. I think the turning test is
one of those concepts that changes over time, and and
it's more like you start to find out what it's
not rather than what it is. It's an interesting things,
very much like our artificial intelligence itself, or even machine consciousness.
(45:26):
These are all concepts that are are kind of wibbly wobbly.
As dr who might say, well, if you have suggestions
for topics I should cover in future episodes of tech Stuff,
reach out to me. The best way to do that
is over on Twitter. The handle for the show is
text Stuff HSW and I'll talk to you again really soon.
(45:52):
Text Stuff is an I Heart Radio production. For more
podcasts from my Heart Radio, visit the i Heart Radio app,
Apple Podcasts, or wherever you listen to your favorite shows.