Transcript
ImKkaeUx1MU • Melanie Mitchell: Concepts, Analogies, Common Sense & Future of AI | Lex Fridman Podcast #61
/home/itcorpmy/itcorp.my.id/harry/yt_channel/out/lexfridman/.shards/text-0001.zst#text/0271_ImKkaeUx1MU.txt
Kind: captions
Language: en
the following is a conversation with
Melanie Mitchell she's the professor of
computer science at Portland State
University and an external professor at
Santa Fe Institute she has worked on and
written about artificial intelligence
from fascinating perspectives including
adaptive complex systems genetic
algorithms and the copycat cognitive
architecture which places the process of
analogy making at the core of human
cognition from her doctoral work with
her advisers Douglas Hofstadter and John
Holland - today she has contributed a
lot of important ideas to the field of
AI including her recent book simply
called artificial intelligence a guide
for thinking humans this is the
artificial intelligence podcast if you
enjoy it subscribe on YouTube give it
five stars on Apple podcast supported on
patreon or simply connect with me on
Twitter at Lex Friedman spelled Fri D ma
n I recently started doing ads at the
end of the introduction I'll do one or
two minutes after introducing the
episode and never any ads in the middle
that can break the flow of the
conversation I hope that works for you
it doesn't hurt the listening experience
I provide time stamps for the start of
the conversation but it helps if you
listen to the ad and support this
podcast by trying out the product the
service being advertised
this show is presented by cash app the
number one finance app in the App Store
I personally use cash app to send money
to friends but you can also use it to
buy sell and deposit Bitcoin in just
seconds cash app also has a new
investing feature you can buy fractions
of a stock say $1 worth no matter what
the stock price is brokerage services
are provided by cash app investing a
subsidiary of square and member s IBC
I'm excited to be working with cash app
to support one of my favorite
organizations called first best known
for their first robotics and Lego
competitions they educate and inspire
hundreds of thousands of students in
over 110 countries and have a perfect
rating and charity navigator which means
that donated money is used to maximum
effectiveness when you get cash app from
the App Store or Google Play and use
code Lex podcast you'll get ten dollars
in cash up will also donate ten dollars
the first which again is an organization
that I've personally seen inspire girls
and boys to dream of engineering a
better world and now here's my
conversation with Melanie Mitchell
the name of your new book is artificial
intelligence subtitle a guide for
thinking humans the name of this podcast
is artificial intelligence so let me
take a step back and ask the old
Shakespeare question about roses and
what do you think of the term artificial
intelligence for our big and complicated
and interesting field I'm not crazy
about the term I think it has a few
problems because it it's means so many
different things to different people and
intelligence is one of those words that
isn't very clearly defined either
there's so many different kinds of
intelligence degrees of intelligence
approaches to intelligence John McCarthy
was the one who came up with the term
artificial intelligence and what from
what I read he called it that to
differentiate it from cybernetics which
was another related movement at the time
and he later regretted calling it
artificial intelligence Herbert Simon
was pushing for calling it complex
information processing which got nixed
but you know probably is equally vague I
guess is it the intelligence or the
artificial in terms of words that it's
the most problematic you would you say
yeah I think it's a little of both but
you know it has some good size because I
personally was attracted to the field
because I was interested in phenom
phenomenons of intelligence and if it
was called complex information
processing maybe I'd be doing something
wholly different now what do you think
of I've heard the term used cognitive
systems for example so using cognitive
yeah I mean cognitive has certain
associations with it and people like to
separate things like cognition and
perception which I don't actually think
are separate but often people talk about
cognition is being different from sort
of other aspects of intelligence it's
sort of higher level so to you cognition
is this broad beautiful mess of things
that's in calm
the whole thing memory yeah I I think
it's hard to draw lines like that when I
was coming out of grad school in the
night in 1990 which is when I graduated
that was during one of the AI winters
and I was advised to not put AI
artificial intelligence on my CV but
instead call it intelligent systems so
that was kind of a euphemism I guess
what about the stick briefly on on terms
and words the idea of artificial general
intelligence or or like beyond Laocoon
prefers human level intelligence sort of
starting to talk about ideas that that
achieve higher and higher levels of
intelligence and somehow artificial
intelligence seems to be a term used
more for the narrow very specific
applications of AI and sort of the
there's the what set of terms appeal to
you to describe the thing that perhaps
would strive to create people have been
struggling with this for the whole
history of the field and defining
exactly what it is that we're talking
about you know John Searle had this
distinction between strong AI and weak
AI and weak AI could be generally AI but
his idea was strong AI was the view that
a machine is actually thinking that as
opposed to simulating thinking or
carrying out intelligent processes that
we would call intelligent
high level if you look at the founding
of the field of McCarthy in sterlin and
so on are we closer to having a better
sense of that line between narrow weak
AI and strong AI yes I think we're
closer to having a better idea of what
that line is early on for example a lot
of people thought that playing chess
would be you couldn't play chess if you
didn't have sort of general human level
intelligence and of course once
computers were able to play chess better
than humans that revised that view and
people said ok well maybe now we have to
revise what we think of intelligence as
or and and so that's kind of been a
theme throughout the history of the
field is that once a machine can do some
task we then have to look back and say
oh well that changes my understanding of
what intelligence is because I don't
think that machine is intelligent at
least that's not what I want to call
intelligence do you think that line
moves forever or will we eventually
really feel as a civilization like we
cross the line if it's possible it's
hard to predict but I don't see any
reason why we couldn't in principle
create something that we would consider
intelligent I don't know how we will
know for sure maybe our own view of what
intelligence is will be refined more and
more until we finally figure out what we
mean when we talk about it but I I think
eventually we will create machines in a
sense that have intelligence they may
not be the kinds of machines we have now
and one of the things that that's going
to produce is is making us sort of
understand our own machine like
qualities that we in a sense are
mechanical in the sense that like an
eles cells are kind of mechanical they
part they have algorithms they process
information by and somehow out of this
mass of cells we get this emergent
property that we call intelligence but
underlying it is really just cellular
processing and and lots and lots and
lots of it do you think we'll be able to
do you think it's possible to create
intelligence without understanding our
own mind you said sort of in that
process we'll understand more and more
but do you think it's possible to sort
of create without really fully
understanding from a mechanistic
perspective sort of from a functional
perspective how our mysterious mind
works if I had to bet on it I would say
no we we we do have to understand our
own minds at least to some significant
extent but it I think that's a really
big open question I've been very
surprised at how far kind of brute force
approaches based on say big data and
huge networks can can take us I wouldn't
have expected that and they have nothing
to do with the way our minds work so
that's been surprising to me so it could
be wrong to explore the psychological
and the philosophical do you think we're
okay as a species with something that's
more intelligent than us do you think
perhaps the reason we're pushing that
line farther and farther is we're afraid
of acknowledging that there's something
stronger better smarter than us humans
well I'm not sure we can define
intelligence that way because you know
smarter then is with with respect to
what what you know computers are already
smarter than us in some areas they could
multiply much better than we can they
they can figure out driving routes to
take much faster and better than we can
they have a lot more information to draw
on they know about you know traffic
conditions and all that stuff so
for any given particular task sometimes
computers are much better than we are
and we're totally happy with that right
I'm totally happy with that I don't
doesn't bother me at all I guess the
question is you know what which things
about our intelligence would we feel
very sad or or upset that machine's had
been able to recreate so in the book I
talk about my former PhD advisor Douglas
Hofstadter who encountered a music
generation program and that was really
the line for him that if a machine could
create beautiful music that would be
terrifying for him because that is
something he feels is really at the core
of what it is to be human creating
beautiful music art literature I you
know I don't think he doesn't like the
fact that machines can recognize spoken
language really well like he doesn't he
personally doesn't like using speech
recognition I don't think it bothers him
to his core because it's like okay
that's not at the core of humanity but
it may be different for every person
what what really they feel would usurp
their humanity and I think maybe it's a
generational thing also maybe our
children or our children's children will
be adapted they'll adapt to these new
devices that can do all these tasks and
and say yes this thing is smarter than
me in all these areas but that's great
because it helps me looking at the broad
history of our species why do you think
so many humans have dreamed of creating
artificial life and artificial
intelligence throughout the history of
our civilization so not just this
century or the 20th century but really
many throughout many centuries that
preceded it that's a really good
question and I have wondered about that
because I'm I myself
you know was driven by curiosity about
my own thought processes and thought it
would be fantastic to be able to get a
computer to mimic some of my thought
process season I'm not sure why we're so
driven I think we want to understand
ourselves better and we also want
machines to do things for us but I don't
know there's something more to it
because it's so deep in in the kind of
Mythology or the dose of our species and
I don't think other species have this
drive so I don't know if you were to
sort of psychoanalyze yourself and
you're in your own interest in AI are
you what excites you about creating
intelligence you said understanding our
own selves
yeah I think that's what drives me
particularly I'm really interested in
human intelligence but I'm all I'm also
interested in the sort of the phenomenon
of intelligence more generally and I
don't think humans are the only thing
with intelligence
you know I or even animals that I think
intelligence is a concept that
encompasses a lot of complex systems and
if you think of things like insect
colonies or cellular processes or the
immune system or all kinds of different
biological or even societal processes
have as an emergent property some
aspects of what we would call
intelligence you know they have memory
they do in process information they have
goals they accomplish their goals etc
and to me that the question of what is
this thing we're talking about here was
really fascinating to me and and
exploring it using computers seem to be
a good way to approach the question so
do you think kind of
intelligence do you think of our
universes a kind of hierarchy of complex
systems and then intelligence is just
the property of any you can look at any
level and every level has some aspect of
intelligence so we're just like one
little speck in that giant hierarchy of
complex systems I don't know if I would
say any system like that has
intelligence but I guess what I want to
I don't have a good enough definition of
intelligence to say that so let me let
me do sort of multiple choice I guess
though
so you said ant colonies so our ant
colonies intelligent are the bacteria in
our body in intelligent and then look
going to the physics world molecules and
the behavior at the quantum level of of
electrons and so on is are those kinds
of systems do they possess intelligence
like words where's the line that feels
compelling to you I don't know I mean I
think intelligence is a continuum and I
think that the ability to in some sense
have intention have a goal have a some
kind of self-awareness is part of it
so I'm not sure if you know it's hard to
know where to draw that line I think
that's kind of a mystery but I wouldn't
say that say that you know this the
planets orbiting the Sun her is an
intelligent system I mean I would find
that maybe not the right term to
describe that and this is you know
there's all this debate in the field of
like what's what's the right way to
define intelligence what's the right way
to model intelligence should we think
about computation should we think about
dynamics and should we think about you
know free energy and all of that stuff
and I think that it's it's a fantastic
time to be in the field because there's
so many questions and so much we don't
understand there's so much work to do so
are we are we the most special kind of
intelligence
this kind of you said there's a bunch of
different elements and characteristics
of intelligent systems and colonies are
his human intelligence the thing in our
brain is that the most interesting kind
of intelligence in this continuum
well it's interesting to us because
because it is us I mean interesting to
me yes and because I'm part of the you
know human but to understanding the
fundamentals of intelligence what I'm
yeah yeah Jerry is studying the human is
sort of if everything we've talked about
will you talk about in your book what
just the AI field this notion yes it's
hard to define but it's usually talking
about something that's very akin to
human intelligence to me it is the most
interesting because it's the most
complex I think it's the most self-aware
it's the only system at least that I
know of that reflects on its own
intelligence and you talk about the
history of AI and us in terms of
creating artificial intelligence being
terrible at predicting the future or the
Iowa tech in general so why do you think
we're so bad at predicting the future
are we hopelessly bad so no matter what
well there's this decade or the next few
decades every time I make a prediction
there's just no way of doing it well or
as the field matures we'll be better and
better at it I believe as the field
matures we will be better and I think
the reason that we've had so much
trouble is that we have so little
understanding of our own intelligence so
there's the famous story about Marvin
Minsky assigning computer vision as a
summer project to his undergrad students
and I believe that's actually a true
story ya know there's a there's a
write-up on it everyone should read it's
like a I think it's like a proposal
this describes everything done in that
project is hilarious because that I mean
you can explain it but for my sort of
recollection it described
is basically all the fundamental
problems of computer vision many of
which they still haven't been solved
yeah and and I don't know how far they
really expected to get but I think that
and and they're really you know Marvin
Minsky is super smart guy and very
sophisticated thinker but I think that
no one really understands or understood
still doesn't understand how complicated
how complex the things that we do are
because they're so invisible to us you
know to us vision being able to look out
at the world and describe what we see
that's just immediate it feels like it's
no work at all so it didn't seem like it
would be that hard but there's so much
going on
unconsciously sort of invisible to us
that I think we overestimate how easy it
will be to get computers to do it and
sort of for me to ask an unfair question
you've done research you've thought
about many different branches of AI and
through this book widespread looking at
where AI has been where it is today what
if you were to make a prediction how
many years from now would we as a
society create something that you would
say achieved human level intelligence or
superhuman level intelligence that is an
unfair question a prediction that will
most likely be wrong so but it's just
your notion because okay I'll say I'll
say more than a hundred years more than
a hundred years and there I quoted
somebody in my book who said that human
level intelligence is a hundred Nobel
Prizes away which I like because it's a
it's a nice way to to sort of it's a
nice unit for prediction and it's like
that many fantastic discoveries have to
be made and of course there's no Nobel
Prize in
if we look at that hundred years your
senses
really the journey to intelligence has
to go through something something more
complicated as again to our own
cognitive systems understanding them
being able to create them in in the
artificial systems as opposed to sort of
taking the machine learning approaches
of today and really scaling them and
scaling them and scaling them
exponentially with both computing
hardware and and data that would be my
that would be my guess
you know I think that in in the the sort
of going along in the narrow AI that
these current the current approaches
will get better you know I think there's
some fundamental limits to how far
they're gonna get I might be wrong but
that's what I think but and there's some
fundamental weaknesses that they have
that I talked about in the book that
that just comes from this approach of
supervised learning we require requiring
sort of feed-forward networks and so on
it it's just I don't think it's a
sustainable approach to understanding
the world yeah I'm I'm personally torn
on it sort of I've everything read about
in the book and sort of we're talking
about now I agreed I agree with you but
I'm more and more depending on the day
first of all I'm deeply surprised by the
successful machine learning and deep
learning in general and from the very
beginning that when I was it's really
been many focus of work I'm just
surprised how far it gets
and I'm also think we're really early on
in these efforts of these narrow AI so I
think there will be a lot of surprise
off how far it gets
I think will be extremely impressed like
my senses everything I've seen so far
and we'll talk about autonomous driving
and so on I think we can get really far
but I also have a sense that we will
discover just like you said is that even
though we'll get really far in order to
create something like our own
intelligence is actually much farther
than we realized
right I think these methods are a lot
more powerful than people give them
credit for actually so that of course
there's the media hype but I think
there's a lot of researchers in the
community especially like not undergrads
right but like people who've been in AI
they're skeptical about how far deep
learning yet and I'm more and more
thinking that it can actually get
farther than I realize it's certainly
possible one thing that surprised me
when I was writing the book is how far
apart different people are in the field
are artisan their opinion of how how far
the field has come and what is
accomplished and what's what's gonna
happen next what's your sense of the
different who are the different people
groups mindsets thoughts in the
community about where AI is today yeah
they're all over the place so so there's
there's kind of the the singularity
transhumanism group I don't know exactly
how to characterize that approach which
is there as well yeah the sort of
exponential exponential progress we're
on the sort of almost at the the hugely
accelerating part of the exponential and
by in the next 30 years we're going to
see super intelligent AI and all that
and we'll be able to upload our brains
and that so there's that kind of extreme
view that most I think most people who
work in AI don't have they disagree with
that but there are people who who are
maybe don't aren't you know singularity
people but but they're they do think
that the current approach of deep
learning is going to scale and is going
to kind of go all the way basically and
take us to
ái or human-level AI or whatever you
want to call it and there's quite a few
of them and a lot of them like a lot of
the people I've met who work at big tech
companies in AI groups kind of have this
view that we're really not that far you
know just to linger on that point sort
of if I can take as an example like
Yannick kun I don't know if you know
about his work and so a few points
unless I do he believes that there's a
bunch of breakthroughs like fundamental
like Nobel Prizes there's yeah he did
still write but I think he thinks those
breakthroughs will be built on top of
deep learning right and then there's
some people who think we need to kind of
put deep learning to the side a little
bit as just one module that's helpful in
the bigger cognitive framework right so
so I think some what I understand yan
laocoön is rightly saying supervised
learning is not sustainable we have to
figure out how to do unsupervised
learning that that's going to be the key
and you know I think that's probably
true
I think unsupervised learning is going
to be harder than people think I mean
the way that we humans do it then
there's the opposing view you know that
there's a the the Gary Marcus kind of
hybrid view or where deep learning is
one part but we need to bring back kind
of these symbolic approaches and combine
them of course no one knows how to do
that very well which is the more
important part right to emphasize and
how do they how do they fit together
what's what's the foundation what's the
thing that's on top yeah the cake was
the icing right yeah then there's people
pushing different different things
there's the people the causality people
who say you know deep learning as its
formulated a completely lacks any notion
of causality and that's dooms it and
therefore we have to somehow give it
some kind of notion of cause
there's a lot of push from the more
cognitive science crowd saying we have
to look at developmental learning we
have to look at how babies learn we have
to look at intuitive physics all these
things we know about physics and it's
somebody kind of quipped we also have to
teach machines intuitive metaphysics
which means like objects exist causality
exists you know these things that maybe
were born with I don't know that that
they don't have the machines don't have
any of that you know they look at a
group of pixels and they maybe they get
10 million examples but they they can't
necessarily learn that there are objects
in the world so there's just a lot of
pieces of the puzzle that people are
promoting and with different opinions of
like how how how important they are and
how close we are to the you know we'll
put them all together to create general
intelligence looking at this broad field
what do you take away from it who is the
most impressive is that the cognitive
folks Gary Marcus camp the yawn camp son
supervising their self supervise there's
the supervisor and then there's the
engineers who are actually building
systems you have sort of the Andrey
Carpathia Tesla building actual you know
it's not philosophy it's real writing
systems that operate in the real world
what yeah what do you take away from all
all this beautiful yeah I don't know if
you know these these different views are
not necessarily mutually exclusive and I
think people like Jung McCune agrees
with the developmental psychology
causality intuitive physics etc but he
still thinks that it's learning like
end-to-end learning is the way to go
we'll take us perhaps all the way yeah
and that we don't need there's no sort
of innate
stuff that has to get built in this is
you know it's because no it's a hard
problem
I personally you know I'm very
sympathetic to the cognitive science
side because that's kind of where I came
in to the field I've become more and
more sort of an embodiment adherent
saying that you know without having a
body it's gonna be very hard to learn
what we need to learn about the world
that's definitely something like I'd
love to talk about in a little bit to
step into the cognitive world then if
you don't mind because you've done so
many interesting things if you look to
copycat taking a couple of decades step
back
you'd Douglas Hofstadter and others have
created and developed copycat more than
thirty years ago
ah that's painful here what is it what
is what is copycat it's a program that
makes analogies in an idealized domain
idealized world of letter strings so as
you say thirty years ago Wow
so I started working on it when I
started grad school in 1984 Wow and it's
based on Doug Hofstadter's ideas that
about that analogy is really a core
aspect of thinking I remember he has a
really nice quote in in in the book by
by himself and Emmanuel Sanders called
surfaces and essences I don't know if
you've seen that book but it's it's
about analogy he says without concepts
there can be no thought and without
analogies there can be no concepts so
the view is that analogy is not just
this kind of reasoning technique where
we go you know shoe is to foot as glove
as to what you know these kinds of
things that we have on IQ tests or
whatever that but that it's much deeper
much more pervasive in everything we do
in everything our language our thinking
our perception so we so he had a view
that was a very active perception idea
so the idea was that instead of having
kind of what a passive network in which
you have input that's being processed
through these feed-forward layers and
then there's an output at the end that
perception is really a dynamic process
you know we're like our eyes are moving
around and they're getting information
and that information is feeding back to
what we look at next influences what we
look at next and how we look at it and
so copycat was trying to do that kind of
simulate that kind of idea where you
have these agents it's kind of an agent
based system and you have these agents
that are picking things to look at and
deciding whether they were interesting
or not whether they should be looked at
more and and that would influence other
agents how do they interact so they
interacted through this global kind of
what we call the workspace so this
actually inspired by the old blackboard
systems where you'd have agents that
post information on a blackboard a
common blackboard this is like old very
old fashioned a set is that we're
talking about like in physical space is
a computer program computer programs
agents posting concepts on a blackboard
yeah we called it a workspace and it
it's the workspace is a data structure
the agents are little pieces of code
that you can think of them as detect
little detectors or little filters then
say I'm gonna pick this place to look
and I'm gonna look for a certain thing
and it's just the thing I I think is
important is it there so it's almost
like you know a convolution in way
except a little bit more general and
saying and then highlighting it on the
on the work in the workspace wasn't once
it's in the workspace how do the things
they're highlighted relate to each other
like what
so there's different kinds of agents
that can build connections between
different things so just to give you a
concrete example what copycat did was it
made analogies between strings of
letters so here's an example ABC changes
to a BD what does ijk change to and the
program had some prior knowledge about
the alphabet new the sequence of the
alphabet it you know had a concept of
letter successor of letter it had
concepts of sameness so it has some
innate things programmed in but then it
could do things like say discover that
ABC is a group of letters in succession
hmm and then it an agent can mark that
so the idea that there could be a
sequence of letters is that a new
concept that's formed or if that's a
concept that's a concept that's innate
sort of can you form new concepts or all
so in this program all the concepts of
the program were innate so cuz because
we weren't I mean obviously that limits
it quite up quite a bit but what we were
trying to do is say suppose you have
some innate concepts how do you flexibly
apply them to new situations right and
how do you make analogies let's step
back for a second so I really like that
quote that he said without concepts
there can be no thought and without
analogies that can be no concepts you
know in a Santa Fe presentation you said
that it should be one of the mantras of
AI yes and that you all see yourself
said how to form and fluidly use concept
is the most important open problem in AI
yes how to form and fluidly use concepts
is the most important open problem in AI
so let's what is the concept and what is
an analogy a concept is in some sense a
fundamental unit of thought so say we
have a concept
of a dog okay and a concept is embedded
in a whole space of concepts so that
there's certain concepts that are closer
to it or farther away from it are these
concepts are they really like
fundamental like we mention innate look
almost like XE o matic like very basic
and then there's other stuff built on
top of it or just include everything is
are they're complicated like you can
certainly have form new concepts right I
guess that's the question I'm asked yeah
can you form new concepts that our
company complex combinations of other
ago yes absolutely and that's kind of
what we we do you know learning and then
what's the role of analogies in that so
analogy is when you recognize that one
situation is essentially the same as
another situation and essentially is
kind of the key word there and because
it's not the same so if I say last week
I did a podcast interview in actually
like three days ago in Washington DC and
that situation was very similar to this
situation although it wasn't exactly the
same you know it was a different person
sitting across from me we had different
kinds of microphones the questions were
different
the building was different there's all
kinds of different things but really it
was analogous or I can say so by doing a
podcast interview that's kind of a
constant it's a new concept you know I
never had that concept before I mean and
I can make an analogy with it like being
interviewed for a news article in a
newspaper and I can say well you kind of
play the same role that the the
newspaper the reporter played it's not
exactly the same because maybe they
actually emailed me some written
questions rather than
and the writing the written questions
play the you know are analogous to your
spoken questions you know there's just
all kinds of this somehow probably
connects to conversations you have over
Thanksgiving dinner just general
conversations you could there's like a
thread you can probably take that just
stretches out in all aspects of life
that connect to this podcast I mean sure
conversations between humans sure and
and if I go and tell a friend of mine
about this podcast interview my friend
might say oh the same thing happened to
me you know let's say you know you ask
me some really hard question and I have
trouble answering it my friend could say
the same thing happened to me but it was
like it wasn't a podcast interview it
wasn't it was a completely different
situation and yet my friend is seen
essentially this the same thing you know
we say that very fluidly the same thing
happened to me essentially the same
thing we don't even say that right
things they imply it yes yeah and the
view that kind of what went into say
coffee cat that that whole thing is that
that that that act of saying the same
thing happened to me is making an
analogy and in some sense that's what's
underlies all of our concepts why do you
think analogy making that you're
describing is so fundamental to
cognition like it seems like it's the
main element action of what we think of
us cognition yeah so it can be argued
that all of this generalization we do
concepts and recognizing concepts in
different situations is done by analogy
that that's every time I'm recognizing
that say you're a person that's by
analogy because I have this concept of
what person is and I'm applying it to
you and every
time I recognize a new situation like
one of the things I talked about it in
the book was the the concept of walking
a dog that that's actually making an
analogy because all that you know the
details are very different so it's so
now--so reasoning could be reduced on to
sense your analogy making so all the
things we think of as like yeah like you
said perception so what's perception is
taking raw sensory input and it's
somehow integrating into our our
understanding of the world updating the
understanding and all of that has just
this giant mess of analogies that are
being made I think so yeah if you just
linger on it a little bit like what what
do you think it takes to engineer a
process like that for us in our
artificial systems we need to understand
better I think how how we do it how
humans do it and it comes down to
internal models I think you know people
talk a lot about mental models that
concepts are mental models that I can in
my head I can do a simulation of a
situation like walking a dog and that
there there's some work in psychology
that promotes this idea that all of
concepts are really mental simulations
that whenever you encounter a concept or
situation in the world or you read about
it or whatever you do some kind of
mental simulation that allows you to
predict what's going to happen to
develop expectations of what's going to
happen mm-hm so that's the kind of
structure I think we need is that kind
of mental model that and the in our
brain somehow these mental models are
very much inter connected again so a lot
of stuff we're talking about it they're
essentially open problems right so if I
ask a question I don't mean that you
would know the answer already just
hypothesizing but how big do you think
is the the network graph data structure
of concepts that's in our head like if
we're trying to build that ourselves
like it's we take it and that's one of
the things we take for granted we think
I mean that's why we take common sense
for granted within common sense is
trivial but how big of a thing of
concepts is on that underlies what we
think of as common sense for example
yeah I don't know and I'm not I don't
even know what units to measure it in
beautifully put right but but you know
we have you know it's really hard to
know we have what a hundred billion
neurons or something I don't know and
they're connected via trillions of
synapses and there's all this chemical
processing going on there's just a lot
of capacity for the stuff and their
informations encoded in different ways
in the brain it's encoded in chemical
interactions it's encoded and electric
like firing and firing rates and and
nobody really knows how it's encoded but
it just seems like there's a huge amount
of capacity so I think it's it's huge
it's just enormous and it's amazing how
much stuff we know yeah and but we know
and not just know like facts but it's
all integrated into this thing that we
can make analogies with yes there's a
dream of semantic web and there's
there's a lot of Dreams from expert
systems of building giant knowledge
bases or do you see a hope for these
kinds of approaches of building of
converting Wikipedia into something that
could be used in analogy making sure and
I think people have have made some
progress along those lines I mean people
have been working on this for a long
time but the problem is and this I think
was is is the problem of common sense
like people have been trying to get
these common sense networks here at MIT
there's this concept net project right
but the problem is that as I said most
of the knowledge that we have is
invisible to us it's not in Wikipedia
it's very basic things about you know
intuitive physics intuitive psychology
to ative metaphysics all that stuff if
you were to create a website that
described intuitive physics intuitive
psychology would it be bigger or smaller
than Wikipedia what do you think
I guess describe to whom no that's very
really good right yeah that's a hard
question because you know how do you
represent that knowledge is the question
right I can certainly write down F
equals MA and Newton's laws and a lot of
physics can be deduced from that but
that's probably not the best
representation of that knowledge for for
doing the kinds of reasoning we want a
machine to do so so I don't know it's
it's it's impossible to say and you know
the projects like there's a famous the
famous psych project right that Doug
Douglass Lynott did that was trying
still going I think it's still going and
if the the idea was to try and encode
all of common-sense knowledge including
all this invisible knowledge in some
kind of logical representation and it
just never I think could do any of the
things that he was hoping it could do
because that's just the wrong approach
of course that's what they always say
you know and then the history books will
say well the psych project finally found
a breakthrough in 2058 or something and
it did you know we're so much progress
has been made in just a few decades that
yeah okay knows what the next
breakthroughs will be it could be a
certainly a compelling notion what the
psych project stands for
I think Lenin was one of the early
people do say common sense is what we
need and that's what we need all this
like expert system stuff that is not
going to get you to AI you need common
sense and he basically gave up his whole
academic career to to go pursue that I
told my er that but I think that the
approach itself will not what do you
think is wrong with approach what kind
of approach would might be successful
well again he knows the answer right I
knew that you know one of my talks one
of the people in the audience's a
published lecture one of the people in
the audience said what AI companies are
you investing in advice I'm a college
professor extra funds to invest but also
like no one knows what's gonna work in
AI right that's the problem let me ask
another impossible question in case you
have a sense in terms of data structures
that will store this kind of information
do you think they've been invented yet
both in hardware and software or is
something else needs to be are we
totally you know I think something else
has to be invented I that's my guess is
the breakthroughs that's most promising
would that be in hardware and software
do you think we can get far with the
current computers or do we need to do
something you're saying I don't know if
Turing computation is gonna be
sufficient probably I would guess it
will I don't I don't see any reason why
we need anything else but so so in that
sense we have invented the hardware we
need but we just need to make it faster
and bigger and we need to figure out the
right algorithms and and the right sort
of architecture touring that's a very
mathematical notion when we try to have
to build intelligence it's not an
engineering notion where you throw all
that stuff
I guess I guess it is a it is a question
that their people have brought up this
question you know and when you asked
about like is our current Hardware will
our current Hardware work well turing
computation says that like our current
hardware is in principle a Turing
machine right so all we have to do is
make it faster and bigger but there have
been people like Roger Penrose if you
might remember that he said Turing
machines cannot produce intelligence
because intelligence requires continuous
valued numbers I mean that was sort of
my reading of his argument and quantum
mechanics and what else whatever you
know but I don't see any evidence for
that that we need new computation
paradigms but I don't know if we're you
know I don't think we're going to be
able to scale up our current approaches
to programming these computers what is
your hope for approaches like copycat or
other cognitive architectures I've
talked to the creator of sore for
example I've used that arm myself I
don't know if you're familiar with yeah
woody what do you think is what's your
hope of approaches like that in helping
develop systems of greater and greater
intelligence in the coming decades well
that's what I'm working on now is trying
to take some of those ideas and
extending it so I think there are some
really promising approaches that are
going on now that have to do with more
active generative models so this is the
idea of this simulation in your head a
concept when you if you want to when
you're perceiving a new a new situation
you have some simulations in your head
those are generative models they're
generating your expectations they're
generating predictions that's part of a
perception you haven't met the model
that generates a prediction then you
come
parrot with ya and then the difference
and you also that that generative model
is telling you where to look and what to
look at and what to pay attention to and
it I think it affects your perception
it's not that just you compare it with
your perception it it becomes your
perception in a way it is kind of a
mixture of that bottom-up information
coming from the world and your top-down
model being opposed in the world is what
becomes your perception so your hope is
something like that can improve
perception systems and that they can
understand things better yes understand
things yes what's the what's the step
was the analogy making step there well
there the the the idea is that you have
this pretty complicated conceptual space
you know you can talk about a semantic
network or something like that
with these different kinds of concept
models in your brain that are connected
so so let's let's take the example of
walking a dog we were talking about that
okay let's see I say see someone out on
the street walking a cat some people
walk their cats I guess this seems like
a bad idea but yeah so my model of my
you know there's connections between my
model of a dog and model of a cat and I
can immediately see the analogy of that
those are analogous situations but I can
also see the differences and that tells
me what to expect so also you know I
have a new situation so another example
with the walking the dog thing is
sometimes people I see people riding
their bikes with Elise holding a leash
and the dogs running alongside okay so I
know that the I recognize that as kind
of a dog walking situation even though
the person's not walking right and the
dogs not walking because I I have the
these these models that say okay
riding a bike
is sort of similar to walking or it's
connected it's a means of transportation
but I because they have their dog there
I assume they're not going to work but
they're going out for exercise and you
know these analogies help me to figure
out kind of what's going on what's
likely but sort of these analogies are
very human interpreter Bowl mm-hmm so
that's that kind of space and then you
look at something like the current deep
learning approaches they kind of help
you to take raw sensory information and
just to automatically build up
hierarchies of role you can even call
them concepts they're just not human
interpretive or concepts
what's your what's the link here do you
hope it's sort of the hybrid system
question how do you think that two can
start to meet each other what's the
value of learning in this systems of
forming of analogy making the the goal
of I you know the original goal of deep
learning in at least visual perception
was that you would get the system to
learn to extract features that at these
different levels of complexities may be
edge detection and that would lead into
learning you know simple combinations of
edges and then more complex shapes and
then whole objects or faces and this was
based on that the ideas of the
neuroscientists Hubel and Wiesel who had
seen laid out this kind of structure and
brain and I think that is that's right
to some extent of course people have
come found that the whole story is a
little more complex than that and the
brain of course always is and there's a
lot of feedback and so I see that
as absolutely a good brain inspired
approach to some aspects of perception
but one thing that it's lacking for
example is all of that feedback which is
extremely important the interactive
element do you mentioned the expectation
the sexual level go back and forth with
the the expectation the perception and
yes going back and forth so right so
that is extremely important and you know
one thing about deep neural networks is
that in a given situation like you know
they they're trained right they get
these weights everything but then now I
give them a new a new image let's say
yes they treat every part of the image
in the same way you know they apply the
same filters at each layer to all parts
of the image mm-hmm there's no feedback
to say like oh this part of the image is
irrelevant right I shouldn't care about
this part of the image or this part of
the image is the most important part and
that's kind of what we humans are able
to do because we have these conceptual
expectations there's a little bit work
in that there's certainly a lot more in
a tent what's under the called attention
in natural language processing knowledge
ease it's a that's exceptionally
powerful and it's a very just as you say
it's really powerful idea but again in
sort of machine learning it all kind of
operates in an automated way that's not
human it's not it's not also okay so
that yeah right it's not dynamic I mean
in the sense that as a perception of a
new example is being processed those
attentions weights don't change right so
I mean there's a this
kind of notion that there's not a memory
so you're not aggregating the idea of
the this mental model yes yeah he that
seems to be a fundamental idea there's
not a really powerful I mean there's
some stuff with memory but there's not a
powerful way to represent the world in
some sort of way that's deeper than and
it's it's so difficult because uh you
know neural networks do represent the
world they do have a mental model right
but it just seems to be shallow I like
it it's it's hard to it's it's hard to
criticize them at the fundamental level
to me at least it's easy to it's it's
easy to criticize and we'll look like
exactly you're saying mental models sort
of almost from a sec I'll put a
psychology head on say look these
networks are clearly not able to achieve
what we humans do with forming mental
models but analogy making so on but that
doesn't mean that they fundamentally
cannot do that like you can it's very
difficult to say that I mean I used to
me do you have a notion that the
learning approaches really I mean
they're going to not not only are they
limited today but they will forever be
limited in being able to construct such
mental models I think the idea of the
dynamic perception is key here the idea
that moving your eyes around and getting
feedback and that's something that you
know there's been some models like that
there's certainly recurrent neural
networks that operate over several time
steps and but the problem is that it
that the actual the recurrence is you
know basically the the feedback is to
the next time step is the entire hidden
state yes the network which which is it
that it that's that doesn't work very
well does he hit the the thing I'm
saying is mathematically speaking it has
the information in that recurrence to
capture everything it just doesn't seem
to work yeah so like my you know it's
like it's the same touring machine
question right
yeah maybe theoretically it computers
and anything that's throwing a universal
Turing machine can can be intelligent
but practically the architecture might
be very specific kind of architecture to
be able to create it so just I guess
it's sort of ask almost the same
question again is how big of a role do
you think deep learning needs will play
or needs to play in this in perception I
think deep learning as it's currently as
it currently exists you know will place
that kind of thing will play some role
and but I think that there's a lot more
going on in perception but who knows you
know that the definition of deep
learning I mean it this it's pretty
broad it's kind of an umbrella so what I
mean is purely sort of neural networks
yeah and a feed-forward neural networks
essentially or there could be recurrence
but yeah sometimes it feels like for us
I'll talk to Gary Marcus it feels like
the criticism of deep learning is kind
of like us birds criticizing airplanes
for not flying well or that they're not
really flying do you think deep learning
do you think it could go all the way
like you're looking things do you think
that yeah the brute force learning
approach can go all the way I don't
think so no I mean I think it's an open
question but I I tend to be on the
innate Ness side that there has that
there's some things that we've been
evolved to be able to learn and
that learning just can't happen without
them so so one example here's an example
I had in the book that that I think is
useful to me at least in thinking about
this so this has to do with the
deepmind's atari game playing program
okay and learned to play these Atari
video games just by getting input from
the pixels of the screen and it learned
to play the game break out thousand
percent better than humans okay that was
one of the results and it was great and
and it learned this thing where it
tunneled through the side of the the
bricks in the breakout game and the ball
could bounce off the ceiling and then
just wipe out bricks okay so there was a
group who did an experiment where they
took the paddle you know that you move
with the joystick and moved it up to
pixels or something like that and then
they they looked at a deep Q learning
system that had been trained on breakout
and said could it now transfer its
learning to this new version of the game
of course a human could but and it
couldn't maybe that's not surprising but
I guess the point is it hadn't learned
the concept of a paddle it hadn't
learned that it hadn't learned the
concept of a ball or the concept of
tunneling it was learning something you
know we caught we looking at it kind of
anthropomorphised it and said oh it
here's what it's doing and the way we
describe it but it actually didn't learn
those concepts and so because it didn't
learn those concepts it couldn't make
this transfer yes so that's a beautiful
statement but at the same time by moving
the paddle we also anthropomorphize
flaws to inject into the system that
will then flip out how impressed we are
by it what I mean by that is to me the
Atari games were to me deeply impressive
that that was possible at all so that
guy first pause on that and people
should look at that just like the game
of Go
which is fundamentally different to me
then then what deep blue did even though
there's still mighty calls distillate
research it's just everything in deep
mind is done in terms of learning
however limited it is still deeply
surprising to me yeah i i'm not i'm not
trying to say that what they did wasn't
impressive i think it was incredibly
impressive to me is interesting is
moving the path aboard just another love
another thing that needs to be learned
so like we've been able to maybe maybe
been able to through the current neural
networks learn very basic concepts that
are not enough to do this general
reasoning and it may be with more data i
mean the data that you know the
interesting thing about the examples
that you talk about and beautifully is
they it's often flaws of the data well
that's the question i mean i i think
that is the key question it whether it's
a flaw of the data or not or the mexico
the reason I brought up this example was
because you were asking do I think that
you know learning from data could go all
the way yes and that this was why I
brought up the example because I think
and this was is not at all to to take
away from the impressive work that they
did but it's to say that when we look at
what these systems learn do they learn
the human the things that we humans
consider to be the relevant concepts and
in that example
it didn't sure if you train it on a
movie you know the pat paddle being in
different places maybe it could deal
with maybe it would learn that concept
I'm not totally sure but the question is
you know scaling that up to more
complicated worlds to what extent could
a machine that only gets this very raw
data learn to divide up the world into
relevant concepts and I don't know the
answer but I would bet that that
without some innate notion that it can't
do it
yeah ten years ago a hundred percent
agree with you as the deal most experts
in a system but now I have a one but
like I have a glimmer of hope okay
have you no that's very nice and I think
I think that's what deep learning did in
the community is no no I still if I had
to bet all my money it's a hundred
percent deep learning will not takes all
the way but there's still other it still
I was so personally sort of surprised
mm-hmm why the Thar games by go by by
the power of self play of just yeah I'm
playing against you that I was like many
other times just humbled of how little I
know about what's possible you know yeah
I think fair enough self play is
amazingly powerful and you know that's
that goes way back to Arthur Samuel
Wright with his checker playing program
and that which was brilliant and
surprising that it did so well so just
for fun let me ask you a topic of
autonomous vehicles it's the area that
that I work at least these days most
closely on and it's also area that I
think is a good example that you use a
sort of an example of things we as
humans don't always realize how hard it
is to do it's like the the constant
trend AI but the different problems that
we think are easy when we first try them
and then realize how hard it is okay so
why you've talked about this autonomous
driving being a difficult problem more
difficult than we realize you must give
it credit for why is it so difficult one
of the most difficult parts in your view
I think it's difficult because of the
world is so open-ended as to what what
kinds of things can happen so you have
sort of what normally happens which is
just you drive along and nothing nothing
surprising happens and autonomous
vehicles can do the ones we have now
evidently can do really well on most
normal situations as long
as long as you know the weather is
reasonably good and everything but if
some we have this notion of edge cases
or or you know things in the tail of the
distribution you call it the long tail
problem which says that there's so many
possible things that can happen that was
not in the training data of the machine
that it won't be able to handle it
because it doesn't have common sense
right it's the old the paddle moved yeah
it's the paddle moved problem right and
so my understanding and you probably are
more of an expert than I am on this is
that current self driving car vision
systems have problems with obstacles
meaning that they don't know which
obstacles which quote unquote obstacles
they should stop for and which ones they
shouldn't stop for and so a lot of times
I read that they tend to slam on the
brakes quite a bit and the most common
accidents with self-driving cars are
people rear-ending them because they
were surprised they've warned expecting
the machine the car to stop yeah so
there's there's a lot of interesting
questions there whether because because
you mentioned kind of two things so one
is the the problem of perception of
understanding of interpreting the
objects that are detected right
correctly and the other one is more like
the policy the action that you take how
you respond to it so a lot of the cars
braking is a kind of notion of to
clarify there's a lot of different kind
of things that are people calling
autonomous vehicles but a lot the L for
vehicles with a safety driver are the
ones like way moe and cruise and those
companies they tend to be very
conservative and cautious so they tend
to be very very afraid of hurting
anything or anyone and getting in any
kind of accidents so their policy is
very kind of that it that results in
being exceptionally responsive to
anything that could possibly be an
obstacle right
right which which which the human
drivers around it it's unpredictably
yeah that's not a very human thing to do
caution that's not the thing we're good
at specially in driving we're in a hurry
often angry and etc especially in Boston
so and then there's of another and a lot
of times that's machine learning is not
a huge part of that it's becoming more
and more unclear to me how much you you
know sort of speaking to public
information because a lot of companies
say they're doing deep learning and
machine learning just attract good
candidates the reality is in many cases
it's still not a huge part of the the
perception this is this lidar there's
other sensors that are much more
reliable for obstacle detection and then
there's Tesla approach which is vision
only and there's I think a few companies
doing that protest the most sort of
famously pushing that forward and that's
because the lidar is too expensive right
well I mean yes but I would say if you
were to for free give to every test
vehicle I mean Elon Musk fundamentally
believes that lidar is a crutch right
fantasy said that that if you want to
solve the problem of machine learning
lidar is not should not be the primary
sensor is the belief okay the camera
contains a lot more information mm-hmm
so if you want to learn you want that
information but if you want to not to
hit obstacles you want like are it's
sort of it's this weird trade-off
because yeah it's sort of what Tesla
vehicles have a lot of which is really
the thing the price of the fallback the
primary fallback sensor is radar which
is a very crude version of lighter it's
a good detector of obstacles except when
those things are standing right the
stopped vehicle right that's why it had
problems with crashing into stop fire
trucks stop fire trucks right so the
hope there is that the vision sensor
would somehow catch that and infer
there's a lot of problems of perception
I they are doing actually some
incredible stuff in the almost like an
active learning space where it's
constantly taking edge cases and pulling
back in there's a state data pipeline
another aspect that is really important
that people are studying now is called
multitask learning which is sort of
breaking apart this problem whatever the
problem is in this case driving into
dozens or hundreds of little problems
that you can turn into learning problems
so this giant pipeline the you know it's
kind of interesting I've been skeptical
from the very beginning we've become
less and less skeptical over time how
much of driving can be learned I'm still
think it's much farther than then the
CEO of that particular company thinks it
will be but it it is costly surprising
that through good engineering and data
collection and active selection of data
how you can attack that long tail and
it's an interesting open question that
you're absolutely right there's a much
longer tail and all these edge cases
that we don't think about but it's this
it's a fascinating question that applies
to natural language in all spaces how
big how how big is that long tail right
and I mean not to linger on the point
but what's your sense in driving in
these practical problems of the human
experience can it be learned so the
current what are your thoughts are sort
of Elon Musk thought let's forget the
thing that he says it'd be solved in a
year but can it be solved in in a
reasonable timeline or do fundamentally
other methods need to be invented so I I
don't I think that ultimately driving so
so it's a trade-off in a way I you know
being able to drive and deal with any
situation that comes up does require
kind of full human
telogen sand even in humans aren't
intelligent enough to do it because
humans I mean most human accidents are
because the human wasn't paying
attention or the humans drunk or
whatever and not because they weren't
intelligent but not because they weren't
intelligent enough right whereas the
accidents with autonomous vehicles is
because they weren't intelligent enough
they're always paying attention so it's
a it's a trade off you know and I think
that it's a very fair thing to say that
autonomous vehicles will be ultimately
safer than humans because humans are
very unsafe it's kind of a low bar but
just like you said
the III I think he was get a bad rap
right cuz we're really good at the
common-sense thing yeah we're great at
the common-sense thing we're bad at the
paying atten thing being attached a
thing especially moral you know driving
is kind of boring and we have these
phones to play with and everything but I
think what what's gonna happen is that
for many reasons not just AI reasons but
also like legal and other reasons that
the the definition of self-driving is
going to change or autonomous is going
to change it's not going to be just I'm
gonna go to sleep in the back and you
just drive me anywhere
it's gonna be more certain areas are
going to be instrumented to have the
sensors and the mapping and all the
stuff you need for that that the
autonomous cars won't have to have full
common sense and they'll do just fine in
those areas as long as pedestrians don't
mess with them too much that's another
question I don't think we will have
fully autonomous self-driving in the way
that like most the average person thinks
of it for a very long time and just to
reiterate this is the interesting open
question that I think I agree with you
on is to solve fully
Thomas driving you have to be able to
engineer in common sense yes I think
it's an important thing to hear and
think about I hope that's wrong but I
currently I could agree with you that
unfortunately you do have to have to be
more specific sort of these deep
understandings of physics and yeah of
the way this world works and also the
human dynamics like you mentioned
pedestrians and cyclists actually that's
whatever that nonverbal communication is
some people call it there's that dynamic
that is also part of this common sense
right and we're pretty we humans are
pretty good at predicting what other
humans are gonna do and how are our
actions impacts the behaviors of yes
this is weird game theoretic dance that
we're good at somehow and work well the
funny thing is is because I've watched
countless hours of pedestrian video and
talked to people
we humans are also really bad at
articulating the knowledge we have right
which is a been a huge challenge yes so
you've mentioned embodied intelligence
what do you think it takes to build a
system of human level intelligence does
he need to have a body I'm not sure but
I I'm coming around to that more and
more and what does it mean to be I don't
mean to keep breaking on up yeah Laocoon
he looms very large yeah well he
certainly has a large personality yes he
thinks that the system needs to be
grounded meaning he needs to sort of be
able to interact with reality but it
doesn't think it necessarily need to
have a body
so when you think of what's the
difference I guess I want to ask when
you mean body do you mean you have to be
able to play with the world or do you
also mean like there's a body that you
that you have to preserve oh that's a
good question I haven't really thought
about that but I think both I would
guess because it's because I think you I
think intelligence it's so hard to
separate it from
self our desire for self-preservation
our emotions are all that non rational
stuff that kind of gets in the way of
logical thinking because we the way you
know if we're talking about human
intelligence or human level intelligence
whatever that means
a huge part of it is social that you
know we were evolved to be social and to
deal with other people and that's just
so ingrained in us that it's hard to
separate intelligence from that I I
think you know AI for the last 70 years
or however long has been around it it
has largely been separated there's this
idea that there's like it's kind of very
Cartesian there's this you know thinking
thing that we're trying to create but we
don't care about all this other stuff
and I think the other stuff is very
fundamental so there's idea that things
like emotion get in the way of
intelligence as opposed to being an
integral part and part of it so I mean
I'm Russian so romanticize the notions
of emotion and suffering and all that
kind of fear of mortality those kinds of
things so I I especially sort of by the
way did you see that there was this
recent thing going around the internet
of this so some I think he's a Russian
or some Slavic head had written this
thing a sort of anti the idea of super
intelligence mmm-hmm I forgot maybes
polish anyway so at all these arguments
and one one was the argument from Slavic
pessimism do you remember what the
argument is it's like nothing ever works
so what what do you think is the role
like that's such a fascinating idea that
the what we perceive as serve the limits
of human
of the human mind which is emotion and
fear and all those kinds of things are
integral to intelligence could could you
elaborate on that like what why is that
important
do you think for human level
intelligence at least the way the humans
work it's a big part of how it affects
how we perceive the world it affects how
we make decisions about the world it
affects how we interact with other
people it affects our understanding of
other people you know for me to
understand your what you're going what
you're likely to do I need to have kind
of a theory of mine and that's very much
a theory of emotions and motivations and
goals and and to understand that I you
know we have the this whole system of
you know mirror neurons you know I sort
of understand your motivations through
sort of simulating it myself so you know
it's not something that I can prove
that's necessary but it seems very
likely so ok you've written the op-ed in
New York Times titled we shouldn't be
scared by super intelligent AI and it
criticized a little bit just to rustle
in the boss room can you try to
summarize that articles key ideas so it
was spurred by a earlier New York Times
op-ed by Stewart Russell which was
summarizing his book called human
compatible and the article was saying
you know if we if we have super
intelligent AI we need to have its
values align with our values and it has
to learn about what we really want and
he gave this example what if we have a
super intelligent AI and we give it the
prob
of solving climate change and it decides
that the best way to lower the carbon in
the atmosphere is to kill all the humans
okay so to me that just made no sense at
all because a super intelligent AI first
of all thinking what trying to figure
out what what super intelligence means
and it doesn't it seems that something
that super intelligent can't just be
intelligent along this one dimension of
okay I'm gonna figure out all the steps
the best optimal path to solving climate
change and not be intelligent enough to
figure out that humans don't want to be
killed that you could get to one without
having the other and you know
boström in his book talks about the
orthogonality hypothesis where he says
he thinks that systems I can't remember
exactly what it is but it like a systems
goals and it's uh values don't have to
be aligned there's some orthogonal 'ti
there which didn't make any sense to me
so you're saying it in any system that's
sufficiently not even super intelligent
but is it approach greater greater
intelligence there's a holistic nature
that will sort of attention that will
naturally emerge
yes events it from sort of any one
dimension running away yeah yeah exactly
so so you know
boström had this example of the the
super intelligent AI that that makes
that turns the world into paperclips
because its job is to make paper clips
or something and that just as a thought
experiment didn't make any sense to me
well as a thought experiment or the
thing that could possibly be realized
either so so I think that you know what
my op ed was trying to do was say that
that intelligence is more complex than
these people are presenting it that it's
not like it's not so separable the
rationality the the values the emotions
all of that that it's the the view that
you could separate all these dimensions
and build the machine that has one of
these dimensions and it's super
intelligent in one dimension but it
doesn't have any of the other dimensions
that's what I was trying to criticize
that that that I don't believe that
so can I read a few sentences from
yoshua bengio who is always super
eloquent so he writes I have the same
impression as Melanie that our cognitive
biases are linked with our ability to
learn to solve many problems they may
also be a limiting factor for AI however
this is a may in quotes things may also
turn out differently and there's a lot
of uncertainty about the capabilities of
future machines but more importantly for
me the value alignment problem is a
problem well before we reached some
hypothetical super intelligence it is
already posing a problem in the form of
super powerful companies whose objective
function may not be sufficiently aligned
with humanity's general well-being
creating all kinds of harmful side
effects so he goes on to argue that at
you know the orthogonality and those
kinds of things the concerns of just
aligning values with the capabilities of
the system is something that might come
long before we reach anything like in
super intelligence so your criticism
it's kind of really nice as saying this
idea of super intelligence systems seem
to be dismissing fundamental parts of
what intelligence would take and then
you know kind of says yes but if we look
at systems that are much less
intelligent there might be these same
kinds of problems that emerge sure but I
guess the example that he gives there of
these corporations that's people right
those are people's values I mean we're
talking about people the corporations
are their value
are the values of the people who run
those corporations but the idea is the
algorithm that's right so does the
fundamental person that the fundamental
element of what does the bad thing as a
human being
yeah but the the algorithm kind of
controls the behavior this mass of human
beings which help whatever for a company
that's the outs of for example if it's
advertisement driving company that
recommends certain things and encourages
engagement so it gets money by
encouraging engagement and therefore the
company more and more it's like the
cycle that builds an algorithm that
enforces more engagement and made
perhaps more division in the culture and
so on so on again I guess the question
here is sort of who has the agency so
you might say for instance we don't want
our algorithms to be racist right and
facial recognition you know some people
have criticized some facial recognition
systems as being racist because they're
not as good on darker skin and lighter
skin okay but the agency there the the
the the actual algal recognition
algorithm isn't what has the agency it's
it's not the racist thing right it's
it's the that the I don't know the the
combination of the training data the
cameras being used I whatever but my
understanding of and I'll say I told
agree with Benjy oh there that he you
know I think there are these value
issues with our use of algorithms but my
understanding of what Russell's argument
was is more that the algorithm itself
has the agency now it's the thing that's
making the decisions and it's the thing
that has what we would call values yes
so whether that's just a matter of
degree you know it's hard it's hard to
say right because but I would say that's
sort of qualitatively different than a
face recognition neural network and to
broadly linger on that point if you look
at Elon Musk goes to a rustle or boström
people who are worried about existential
risks of AI however far into the future
the argument goes is it eventually
happens we don't know how far but it
eventually happens
do you share any of those concerns and
what kind of concerns in general do you
have a body I that approach anything
like existential threat to humanity so I
would say yes it's possible but I think
there's a lot more closer in existential
threats you had as you said like a
hundred years for so your times more
more than a hundred more than a hundred
years and so that maybe even more than
500 years I don't I don't know I mean
it's so the existential threats are so
far out that the future is the immune
there'll be a million different
technologies that we can't even predict
now that will fundamentally change the
nature of our behavior reality society
and so on before then I think so I think
so and you know we have so many other
pressing existential threats going on
new hangouts even their nuclear weapons
climate problems you know
poverty possible pandemics that you can
go on and on and I think though you know
worrying about existential threat from
AI is it's not the best priority for
what we should be worried about that
that's kind of my view because we're so
far away but I you know I I'm not I'm
not necessarily criticizing Russell or
boström or whoever for worrying about
that and I'm I think it's some some
people should be worried about it it's
it's certainly fine but I I was more
sort of getting at their their view of
intelligible intelligence is mmm-hmm so
I was more focusing on like their view
of the super intelligence then uh just
the fact of them worrying and the title
of the article was written by the the
New York Times editors I wouldn't have
called it that we shouldn't be scared by
super intelligent and no if you wrote it
be like we should redefine what you mean
by super in I actually said it said you
know something like super intelligence
is not is is not a sort of coherent idea
that's not like it's only New York Times
would put in and the follow-up argument
that Yoshio makes also not argument but
a statement and I've heard him say it
before and I think I agree he's kind of
has a very friendly way of phrasing it
is it's good for a lot of people to
believe different things yeah well no
but he's it's also practically speaking
like we shouldn't be like while your
article stands like Stuart Russell does
amazing work boström does amazing work
you do amazing work and even when you
disagree about the definition of super
intelligence or the usefulness of even
the term it's still useful to have
people that like use that term all right
and then argue it sir I
I absolutely agree with video there and
I think it's great that you know and
it's great that New York Times will
publish all this stuff that's right it's
an exciting time to be here what what do
you think is a good test of intelligence
IQ is is natural language ultimately a
test that you find the most compelling
like the the original or the what you
know the higher levels of the Turing
test kind of yeah yeah I still think the
original idea of the Turing test is a
good test for intelligence I mean I
can't think of anything better
you know the Turing tests the way that
it's been carried out so far has been
very impoverished if you will but I
think a real Turing test that really
goes into depth like the one that I
mentioned I talk about in the book I
talk about Ray Kurzweil and Mitchell
Kapoor have this bet right that that in
2029 I think is the date there a machine
will pass the Turing test and turn says
and they have a very specific like how
many hours many expert judges and all of
that and you know Kurzweil says yes
Kapoor says no we can't we only have
like nine more years to go to see I you
know if something a machine could pass
that I would be willing to call it
intelligent of course nobody will they
will say that's just a language model if
it does so you would be comfortable it's
a language a long conversation that well
yeah here I mean you're right because I
think probably to carry out that long
conversation you would literally need to
have deep common-sense understanding of
the world I think so and the
conversation is enough to reveal that so
another super fun topic of complexity
that you have worked on written about
let me ask the basic question what is
complexity so complexity is another one
of those terms like intelligence
it's perhaps overused but my book about
complexity was about this wide area of
complex systems studying different
systems in nature in technology in
society in which you have emergence kind
of like I was talking about with
intelligence you know we have the brain
which has billions of neurons and each
neuron individually could be said to be
not very complex compared to the system
as a whole but the system the the
interactions of those neurons and the
dynamics creates these phenomena that we
call we call intelligence or
consciousness you know that are we
consider to be very complex so the field
of complexity is trying to find general
principles that underlie all these
systems that have these kinds of
emergent properties and the the
emergence occurs from like underlying
the complex system is usually simple
fundamental interactions yes and the
emergence happens when there's just a
lot of these things interacting yes sort
of what and then most of science to date
can you talk about what what is
reductionism
well reductionism is when you try and
take a system and divide it up into its
elements whether those be cells or atoms
or subatomic particles whatever your
field is and then try and understand
those elements and then try and build up
an understanding of the whole system by
looking at sort of the sum of all the
elements so what's your sense whether
we're talking about intelligence or
these kinds of interesting complex
systems is it possible to understand
them in in a reductionist way it's just
probably the approach of most of science
today right
I don't think it's always possible to
understand the things we want to
understand the most so I don't think
it's possible to look at single neurons
and understand what we call intelligence
you know just look at sort of summing up
and the sort of the summing up is the
issue here that were you know that one
example is that the human genome alright
so there was a lot of work on excitement
about sequencing the human genome
because the idea would be that we'd be
able to find genes that underlies
diseases but it turns out that and I was
a very reductionist idea you know we
figure out what all the the parts are
and then we would be able to figure out
which parts cause which things but it
turns out that the parts don't cause the
things that we're interested in it's
like the interactions it's the networks
of these parts and so that kind of
reductionist approach didn't yield the
the explanation that we wanted would he
would use the most beautiful complex
system that you've encountered most
beautiful that you've been captivated by
is it sort of I mean for me that is the
simplest to be cellular automata oh yeah
so I was very captivated by cellular
automata and worked on cellular automata
for several years do you find it amazing
or is it surprising that such simple
systems such simple rules and cellular
Domino can create sort of seemingly
unlimited complexity yeah that was very
surprising to me I didn't make sense of
it how does that make you feel this is
just ultimately humbling or is there
hope to somehow leverage this into a
deeper understanding and even able to
engineer things like intelligence
it's definitely humbling how humbling in
that also kind of awe-inspiring that
it's that inspiring like part of
mathematics that these credible
simple rules can produce this very
beautiful complex hard to understand
behavior and that that's it's mysterious
you know and and surprising still but
exciting because it does give you kind
of the hope that you might be able to
engineer complexity just from from these
can you briefly say what is the Santa Fe
Institute its history its culture its
ideas its future stuff I've never
semester G I've never been but so has
been this in my - mystical place where
brilliant people study the edge of chaos
exactly so the Santa Fe Institute was
started in 1984 and it was created by a
group of scientists a lot of them from
Los Alamos National Lab which is about a
40-minute drive from the Santa Fe
Institute
they were mostly physicists and chemists
but they were frustrated in their field
because they felt so that their field
wasn't approaching kind of big
interdisciplinary questions like the
kinds we've been talking about and they
wanted to have a place where people from
different disciplines could work on
these big questions without sort of
being siloed into physics chemistry
biology whatever so they started this
Institute and this was people like
George Cowan who is a chemist in the
Manhattan Project and Nicholas
Metropolis who mathematician physicist
Murray gell-mann physicist nism so some
really big names here ken arrow an
economist Nobel prize-winning economist
and they started having these workshops
and this whole enterprise kind of grew
into this Research Institute that's
itself has been
kind of on the edge of chaos its whole
life because it doesn't have any it
doesn't have a significant endowment and
it's just been kind of living on
whatever funding it can raise through
donations and grants and however it can
you know business business associates
and so on but it's a great place it's a
really fun place to go think about ideas
from that you wouldn't normally
encounter I saw Sean Carroll so
physicists yeah yeah external faculty
and you mentioned that there's so
there's some external faculty and
there's people there's a very small
group of resident faculty maybe maybe
about ten who are there for five year
terms that can sometimes get renewed and
then they have some postdocs and then
they have this much larger on the order
of a hundred external faculty or people
come like me who come and visit for
various periods of time so what do you
think this is the future of the Santa Fe
Institute like what and if people are
interested like what what's there in
terms of the public interaction or
students or so on that's that could be a
possible interaction on the Santa Fe
Institute or its ideas yeah so there's a
there's a few different things they do
they have a complex system summer school
for graduate students and postdocs and
sometimes faculty attend to and that's a
four week very intensive residential
program where you go and you listen to
lectures and you do projects and people
people really like that I mean it's a
lot of fun they also have some specialty
summer schools there's one on
computational social science there's one
on
climate and sustainability I think it's
called there's a few and then they have
short courses where just a few days on
different topics they also have an
online education platform that offers a
lot of different courses and tutorials
from SFI faculty
including an introduction to complexity
course that I talk and there's a bunch
of talks to online from there's guest
speakers and so on they they host a lot
of yeah they have sort of technical
seminars and colloquia they all and they
have a community lecture series like
public lectures and they put everything
on their YouTube channel so you can see
it all watching douglas hofstadter
author of get olestra bach was your PhD
adviser he mentioned a couple times and
collaborator do you have any favorite
lessons or memories from your time
working with him that continues to this
day yes but just even looking back
through throughout your time working
with him so one of the things he taught
me was that when you're looking at a
complex problem to to idealize it as
much as possible to try and figure out
what are really what is the essence of
this problem and this is how like the
copycat program came into being was by
taking an analogy making and saying how
can we make this as idealized as
possible but still retain really the
important things we want to study and
that's really kept you know been a core
theme of my research I think and I
continue to try and do that and it's
really very much kind of physics
inspired Hofstadter was a PhD in physics
that was his background it's like first
principles kind of thinking like you
reduced to the the most fundamental
aspect of the problem yeah so there you
can focus on solving that fun than I
thought yeah and in AI you know that was
people used to work in these micro
worlds right like the blocks world was
very early important area in AI and then
that got criticized because they said oh
you know you can't scale that to the
real world and so people started working
on much like more real world like
problems but now there's been kind of a
return even to the blocks world itself
you know we've seen a lot of people who
are trying to work on
more of these very idealized problems or
things like natural language and common
sense so that's an interesting evolution
of those ideas so the perhaps the
block's world's represents the
fundamental challenges of the problem of
intelligence more than people realized
it might yeah is there sort of when you
look back at your body of work and your
life you've worked in so many different
fields is there something that you're
just really proud of in terms of ideas
that you've gotten chance to explore
create yourself so I am really proud of
my work on the copycat project I think
it's really different from what almost
everyone is done in AI I think there's a
lot of ideas there to be explored and I
guess one of the happiest days of my
life you know aside from like the births
of my children was the birth of copycat
when it actually started to be able to
make really interesting analogies and I
remember that very clearly you know it
was very exciting time well you kind of
gave life yes artificial so that's right
what in terms of what people can
interact I saw there's like a I think
it's called meta copy kinetic hat mad
cat and there's a Python three
implementation at if people actually
want to play around with it and actually
get into it and study it maybe integrate
into whether it's with deep learning or
any other kind of work they're doing
what what would you suggest they do to
learn more about it and to take it
forward in different kinds of directions
yeah so that there's a Douglas
Hofstadter's book called fluid concepts
and creative analogies talks in great
detail about copycat I have a book
called analogy making as perception
which is a version of my PhD thesis on
it
there's also code that's available that
you can get it to run I have some links
on my web page to where people can get
the code for it and I think that that
would really be the best way I get into
it yeah play with it well Melanie is a
honor talking to you I really enjoyed it
thank you so much for your time today
has been really great
thanks for listening to this
conversation with Melanie Mitchell and
thank you to our presenting sponsor cash
app downloaded use code Lex podcast
you'll get ten dollars and ten dollars
will go to first a stem education
nonprofit that inspires hundreds of
thousands of young minds to learn and to
dream of engineering our future if you
enjoyed this podcast subscribe on
youtube give it five stars an apple
podcast supported on patreon or connect
with me on Twitter and now let me leave
you some words of wisdom from Douglas
Hofstadter and Melanie Mitchell without
concepts there can be no thought and
without analogies there can be no
concepts and Melanie adds how to form
and fluidly use concepts is the most
important open problem in AI
thank you for listening and hope to see
you next time
you