Transcript
Whtt2H5_isM • David Ferrucci: IBM Watson, Jeopardy & Deep Conversations with AI | Lex Fridman Podcast #44
/home/itcorpmy/itcorp.my.id/harry/yt_channel/out/lexfridman/.shards/text-0001.zst#text/0146_Whtt2H5_isM.txt
Kind: captions
Language: en
following is a conversation with David
Ferrucci he led the team that built
Watson the IBM question-answering system
that beat the top humans in the world at
the game of Jeopardy for spending a
couple hours of David I saw a genuine
passion not only for abstract
understanding of intelligence but for
engineering it to solve real-world
problems under real-world deadlines and
resource constraints where science meets
engineering is where brilliant simple
ingenuity emerges people who work
adjoining it to have a lot of wisdom
earned two failures and eventual success
David is also the founder CEO and chief
scientist of elemental cognition a
company working to engineer AI systems
that understand the world the way people
do this is the artificial intelligence
podcast if you enjoy it subscribe on
YouTube give it five stars and iTunes
support it on patreon or simply connect
with me on Twitter Alex Friedman spelled
Fri D M a.m. and now here's my
conversation with David Ferrucci
your undergrad was in biology with a
with an eye toward medical school before
you went on for the PhD in computer
science so let me ask you an easy
question what is the difference between
biological systems and computer systems
in your when you sit back look at the
Stars and think philosophically I often
wonder I often wonder whether or not
there is a substantive difference and I
think the thing that got me into
computer science and artificial
intelligence was exactly this
presupposition that if we can get
machines to think or I should say this
question this philosophical question if
we can get machines to think to
understand to process information the
way do we do so if we can describe a
procedure or describe a process even if
that process where the intelligence
process itself then what would be the
difference
so from philosophical standpoint I'm not
trying to convince that there are there
is I mean you can go in the direction of
spirituality you can go in the direction
of a soul but in terms of you know what
we can what we can experience from an
intellectual and physical perspective
I'm not sure there is clearly there
implement there are different
implementations but if you were to say
as a biological information processing
system fundamentally more capable than
one we might be able to build out of
silicon or or some other substrate I
don't I don't know that there is how
distant do you think is the biological
implementation so fundamentally they may
have the same capabilities but is it
really a far mystery where a huge number
of breakthroughs are needed to be able
to understand it or is that something
that for the most part in the important
aspects echoes are the same kind of
characteristics yeah that's interesting
I mean I so you know your question
presupposes that there's this goal to
recreate you know what we perceive is
biological intelligence I'm not I'm not
sure that's the I'm not sure that that's
how I would state the goal I mean I
think that studying the goal good so I
think there are a few goals I think that
understanding the human brain and how it
works is important for us to be able to
diagnose and treat issues for us to
understand our own strengths and
weaknesses both intellectual
psychological and physical so
neuroscience and on sending the brain
from that perspective has a there's a
clear clear goal there from the
perspective of saying I want to I want
to I want to mimic human intelligence
that one's a little bit more interesting
human intelligence certainly has a lot
of things we Envy it's also got a lot of
problems too so I think we're capable of
sort of stepping back and saying what do
we want out of it what do we want out of
an intelligence how do we want to
communicate with that intelligence how
do we want to behave how do we want it
to perform now of course it's it's
somewhat of an interesting argument
because I'm sitting here as a human with
a biological brain and I'm critiquing
this trends and weaknesses of human
intelligence and saying that we have the
capacity just the capacity to step back
and say gee what what is intelligence is
what do we really want out of it and
that even in and of itself suggests that
human intelligence is something quite
amiable that it could you know it can it
can it can introspect that it could
introspect that way and the flaws you
mentioned the flaws the human self yeah
but I think I think that flaws that
humans wholeness house is extremely
prejudicial and bias and the way it
draws many inferences do you think those
are sorry to interrupt you think those
are features or are those bugs do you
think the the prejudice the
forgetfulness the fear what other flaws
list them all what love maybe that's a
flaw you think those are all things that
can be get gotten getting in the way of
intelligence or the essential components
of and
well again if you go back and you define
intelligence as being able to sort of
accuracy accurately precisely rigorously
reason develop answers and justify those
answers in an objective way
yeah then human intelligence has these
flaws and that it tends to be more
influenced by some of the things you
said and it's and it's largely an
inductive process meaning it takes past
data uses that to predict the future
very advantageous in some cases but
fundamentally biased and prejudicial in
other cases because it's gonna be
strongly influenced by its priors
whether they're whether they're right or
wrong from some you know objective
reasoning perspective you're gonna favor
them because that's those are the
decisions or those are the paths that
succeeded in the past and I think that
mode of intelligence makes a lot of
sense for when your primary goal is to
act quickly and and and survive and make
fast decisions and I think those create
problems when you want to think more
deeply and make more objective and
reasons that decisions of course humans
capable of doing both they do sort of
one more naturally than they do the
other but they're capable of doing both
you're saying they do the one that
responds quickly in it more naturally
right because that's the thing you kind
of need to not be eaten by the Predators
in the world for example but I mean
better than we've we've learned to
reason through logic we've developed
science we train people to do that
I think that's harder for the individual
to do I think it requires training and
you know and and and teaching I think we
are human - certainly is capable of it
but we find more difficult and then
there are other weaknesses if you will
as you mentioned earlier it's just
memory capacity and how many chains of
inference can you actually go through
without like losing your way so just
focus and so the way you think about
intelligence and we're really sort of
floating this philosophical slightly
but I think you're like the perfect
person to talk about this because we'll
get to jeopardy and beyond that's like
an incredible one of the most incredible
accomplishments in AI in the history of
AI but hence the philosophical
discussion so let me ask you've kind of
alluded to it but let me ask again
what is intelligence underlying the
discussions we'll have with with
jeopardy and beyond how do you think
about intelligence is it a sufficiently
complicated problem being able to reason
your way through solving that problem is
that kind of how you think about what it
means to be intelligent so I think of
intelligence to primarily two ways one
is the ability to predict so in other
words if I have a problem what's gonna
can I predict what's going to happen
next whether it's to you know predict
the answer of a question or to say look
I'm looking at all the market dynamics
and I'm going to tell you what's going
to happen next or you're in a in a room
and somebody walks in and you're going
to predict what they're going to do next
or what they're going to say next doing
that in a highly dynamic environment
full of uncertainty be able to lots of
lockdown the more the more variables the
more complex the more possibilities the
more complex but can I take a small
amount of prior data and learn the
pattern and then predict what's going to
happen next accurately and consistently
that's a that's certainly a form of
intelligence what do you need for that
by the way you need to have an
understanding of the way the world works
in order to be able to unroll it into
the future all right thank you one thing
is needed to predict depends what you
mean by understanding IIIi need to be
able to find that function and this is
very much like what function deep
learning does machine learning does is
if you give me enough prior data and you
tell me what the output variable is that
matters I'm going to sit there and be
able to predict it and if I can predict
you predict it accurately so that I can
get it right more often than not I'm
smart if I do that with less data and
less training time I'm even smarter
if I can figure out what's even worth
predicting I'm smarter meaning I'm
figuring out what path is gonna get me
toward a goal
what about picking a goal so again well
that's interesting about picking our
goal sort of an interesting thing I
think that's where you bring in what do
you pre-programmed to do we talked about
humans and humans a pre-programmed to
survive so sort of their primary you
know driving goal what do they have to
do to do that and that that could be
very complex right so it's not just it's
not just figuring out that you need to
run away from their ferocious tiger but
we survive in social context as an
example so understanding the subtleties
of social dynamics becomes something
that's important for surviving finding a
mate reproducing right so we're
continually challenged with complex sets
of variables complex constraints rules
if you will that we we or patterns and
we learn how to find the functions and
predict the things in other words
represent those patterns efficiently and
be able to predict what's going to
happen that's a form of intelligence
that doesn't really record that doesn't
really require anything specific other
than ability to find that function and
and predict that right answer it's
certainly a form of intelligence but
then when we when we say well do we
understand each other in other words do
would you perceive me as as intelligent
beyond that ability to predict so now I
can predict but I can't really
articulate how I'm going to that process
what my underlying theory is for
predicting and I can't get you to
understand what I'm doing so that you
can follow you can figure out how to do
this yourself if you hadn't if you did
not have for example the right pattern
matching machinery that I did and now we
have potentially have this breakdown
where in effect I'm intelligent but I'm
sort of an alien intelligence relative
to you you're intelligent but nobody
knows about it or I can see the I can
see the output knowing so so you're
saying let's
to separate the two things one is you
explaining why you were able to predict
the future and and the second is me
being able to like impressing me that
you're intelligent me being able to know
that you successfully predicted the
future do you think that's well it's not
a pressing you item intelligent in other
words you may be convinced that I'm
intelligent in some form so high well
because of my ability to predict so I
would imagine that wow wow you're right
all here you're you're right more times
than I am you're doing something
interesting that's a form that's a form
of intelligence but then what happens is
if I say how are you doing that and you
can't communicate with me and you can't
describe that to me now I'm a label you
a savant I mean I may say well you're
doing something weird and it's and it's
just not very interesting to me because
you and I can't really communicate and
and so now this is interesting right
because now this is you're in this weird
place where for you to be recognized as
intelligent the way I'm intelligent then
you and I sort of have to be able to
communicate and then my we start to
understand each other and then my
respect and my my appreciation my
ability to relate to you starts to
change so now you're not an alien
intelligence anymore yours you're our
human intelligence now because you and I
can communicate and so I think when we
look at when we look at when we look at
animals for example animals can do
things we can't quite comprehend we
don't quite know how they do them but
they can't really communicate with us
they can't put what they're going
through in our terms and so we think of
them in sort of low there are these
alien intelligences and they're not
really worthless so what we're worth we
don't treat them the same way as a
result of that but it's it's hard
because who knows what you know what's
going on so just a quick elaboration on
that the explaining that you're
intelligent the explaining the the
reasoning the one end to the prediction
is not some kind of mathematical proof
if we look at humans look at political
debates and discourse on Twitter it's
mostly just telling stories so you
usually your task is sorry that your
task is not to tell an accurate
depiction of how you reason but to tell
a story real or not that convinces me
that there was a mechanism by which you
ultimately that's what a proof is I mean
even a mathematical proof is is that
because ultimately the other
mathematicians have to be convinced by
your proof otherwise in fact they're
been that the measurement success yeah
yeah there have been several proofs out
there where mathematicians would study
for a long time before they were
convinced that it actually proved
anything right you never know if it
proved anything until the community of
mathematicians decided that it did so I
mean so it's but it's it's a real thing
yeah and and that's sort of the point
right is that ultimately on you know
this notion of understanding us
understanding something there's
ultimately a social concept in other
words you I have to convince enough
people that I I did this in a reasonable
way I did this in a way that other
people can understand and and replicate
and that make sense to them so we're
very human Houghton's is bound together
in that way we're bound up in that sense
we sort of never really get away with it
until we can consider convince others
that our thinking process you know make
sense did you think the general question
of intelligence is then also social
constructs so if we task asked questions
of an artificial intelligence system is
this system intelligent the answer will
ultimately be a socially constructed I
think I think so I so I think you're
making to be a mess I'm saying we can
try to define intelligence in this super
objective way that says here here's this
data I want to predict this type of
thing learn this function and then if
you get it right often enough we
consider you intelligent but that's more
like a stepfather that I think it I
think it is it doesn't mean it's
useful if it could be incredible useful
it could be solving a problem we can't
otherwise solve and can solve it more
reliably than we can but then there's
this notion of can humans take
responsibility for the decision that
you're that you're making can we make
those decisions ourselves can we relate
to the process that you're going through
and now you as an agent whether you're a
machine or another human frankly are now
obliged to make me understand how it is
that you're arriving at that answer and
allow me I mean me or the obviously a
community or a judge of people to decide
whether or not whether or not that makes
sense and by the way that happens with
the humans as well you're sitting down
with your staff for example and you ask
for suggestions about what to do next
and someone says well I think you should
buy and I think you should buy this much
or would have or sell or whatever it is
or I think you should launch the product
today or tomorrow or launch this product
versus that product whatever decision
may be and you ask why and the person so
I just have a good feeling about it and
it's not you're not very satisfied now
that person could be you know you might
say well you've been right you know
before but I'm gonna put the company on
the line can you explain to me why I
should believe this and that explanation
may have nothing to do with the truth
just them and all them convinced the
wrong yes they'll be wrong she's got to
be convincing but it's ultimately got to
be convinced and that's why I'm saying
it's we're bound together right our
intelligences are bound together in that
sense we have to understand each other
and and if for example you're giving me
an explanation I mean this is a very
important point right you're giving me
an explanation and I'm and I and I and I
have iton I'm not good and then I'm not
good at reasoning well and being
objective and following logical paths
and consistent paths and I'm not good at
measuring and sort of computing
probabilities across those paths what
happens is collectively we're not going
to do we're not going to do well
how hard is that problem the second one
so we I think will talk quite a bit
about the the first on a specific
objective metric benchmark performing
well but being able to explain the steps
the reasoning how hard is that probably
that's I think that's very hard I mean I
think that that's um well it's hard for
humans the thing that's hard for humans
as you know may not necessarily be hard
for computers and vice-versa so sorry so
how hard is that problem for computers I
think it's hard for computers and the
reason why are related to or saying that
it's also hard for humans is because I
think when we step back and we say we
want to design computers to do that one
of the things we have to recognize is
we're not sure how to do it well I'm not
sure we have a recipe for that and even
if you wanted to learn it it's not clear
exactly what data we use and what
judgments we use to learn that well and
so what I mean by that is if you look at
the entire enterprise of science science
is supposed to be at a bad objective
reason and reason right so we think
about who's the most intelligent person
or group of people in the world do we
think about the savants who can close
their eyes and give you a number we'd
think about the think tanks or the
scientists of the philosophers who kind
of work through the details and write
the papers and come up with the
thoughtful logical proves and use the
scientific method and I think it's the
latter and my point is that how do you
train someone to do that and that's what
I mean by it's hard how do you what's
the process of training people to do
that well that's a hard process we work
as a society we work pretty hard to get
other people to understand our thinking
and to convince them of things now we
could for so
weighed them obviously talked about this
like human flaws or weaknesses we can
persuade through persuade then through
emotional means but to but to get them
to understand and connect to and follow
a logical argument is difficult we try
it we do it we do it as scientists we
try to do it as journalists we know we
try to do it as you know even artists in
many forms as writers as teachers we go
to a fairly significant training process
to do that and then we could ask what
why is that so hard
but it's hard and for humans it takes a
lot of work and when we step back and
say well step back and say well how do
we get a machine - how do we get a
machine to do that it's a vexing
question how would you begin to try to
solve that and maybe just a quick pause
because there's an optimistic notion in
the things you're describing which is
being able to explain something through
reason but if you look at algorithms
that recommend things that we look at
next
well there's Facebook Google advertising
based companies you know their goal is
to convince you to buy things based on
anything so that could be reason because
the best of advertisement is showing you
things that you really do need and
explain why you need it but it could
also be through emotional manipulation
the algorithm that describes why a
certain reason a certain decision was
was made how hard is it to do it through
emotional manipulation and why is that a
good or a bad thing so you've kind of
focused on reason logic really showing
in a clear way why something is good one
is that even a thing that us humans do
and and and - how do you think of the
differences in the reasoning aspect and
the emotional manipulation
well they you know so you call it
emotional manipulation but more
objectively is essentially saying you
know thing you know there are certain
features of things that seem to attract
your attention I'm gonna kind of give
you more of that stuff
manipulation is a bad word yeah I mean
I'm not saying it's good right or wrong
is it it works to get your attention and
it works to get you to buy stuff and
when you think about algorithms that
look at the patterns of the you know
patterns of features that you seem to be
spending your money on and is there
going to give you something with a
similar pattern so I'm going to learn
that function because the objective is
to get you to click on and/or get you to
buy and or whatever it is I don't know I
mean that it is like it is what it is I
mean that's what the algorithm does
you can argue whether it's good or bad
it depends what your you know what your
what your goal is
I guess this seems to very useful for
convincing telling us the thing for
convincing humans yeah it's good because
you gives again this goes back to how
does a human you know what is the human
behavior like how does a human you know
brain respond to things I think there's
a more optimistic view of that too which
is that if you're searching for certain
kinds of things you've already reasoned
that you need them and these these
algorithms are saying look that's up to
you
the reason whether you need something or
not that's your job you know you you met
you may have an unhealthy addiction to
this stuff or you may have a reasoned
and thoughtful
explanation for why it's important to
you and the algorithms are saying hey
that's like whatever like that's your
problem all I know is you're buying
stuff like that you're interested in
stuff like that could be a bad reason
could be a good reason that's up to you
I'm gonna show you more of that stuff
and so and I and I and I think that
that's it's not good or bad it it's not
reason or not reason the algorithm is
doing what it does which is saying you
seems to be interested in this I'm going
to show you more that stuff and I think
we're seeing it's not just in buying
stuff but even in social media you're
reading this kind of stuff I'm not
judging on whether it's good or bad I'm
not reasoning at all I'm just saying I'm
gonna show you other stuff with similar
features and you know and like and
that's it and I wash my hands from it
and I say that's all you know that's all
what's going on you know there is you
know people are so harsh on AI systems
so one the bar of performance is
extremely high and yet we also asked
them to in the case of social media to
help find the better angels of our
nature and help make a better society so
what do you think about the role of it
that so that agrees you that's that's
the interesting dichotomy right because
on one hand we're sitting there and
we're sort of doing the easy part which
is finding the patterns we're not
building the systems not building a
theory that it's consumable and
understandable other humans that could
being explained and justified and and so
on one hand to say oh you know AI is
doing this why isn't doing this other
thing well those other things a lot
harder and it's interesting to think
about why why why it's harder and
because you're interpreting you're
interpreting the data in the context of
prior models in other words
understandings of what's important in
the world what's not important what are
all the other abstract features that
drive our decision-making
what's sensible what's not sensible
what's good what's bad what's moral
what's valuable what is it where is that
stuff no one's applying the
interpretation so when I when I see you
clicking on a bunch of stuff and I look
at these simple features the raw
features the features that are there in
a data like what words are being used
or how long the material is more other
very superficial features what colors
are being used in the material like I
don't know why you're clicking on the
stuff you're looking or if it's products
what the price of what the price is or
what the categories or stuff like that
and I just feed you more of the same
stuff that's very different than kind of
getting in there and saying what does
this mean what the stuff you're reading
like why are you reading it what
assumptions are you bringing to the
table are those assumptions sensible is
the miss the material make any sense
does it does it lead you to thoughtful
good conclusions again there's judgment
this interpretation judgment involved in
that process that isn't really happening
in in in the AI today that's harder
right because you have to start getting
at the meaning of this of the of the
stop of the content you have to get at
how humans interpret the content
relative to their value system and
deeper thought processes so that's what
meaning means is not just some kind of
deep timeless semantic thing that the
statement represents but also how a
large number of people are likely to
interpret so that's again even meaning
is a social construct it's so you have
to try to predict how most people would
understand this kind of statement yeah
meaning is often relative but meaning
implies that the connections go beneath
the surface of the artifact so if I show
you a painting it's a bunch of colors in
a canvas what does it mean to you and it
may mean different things at different
people because of their different
experiences it may mean something even
different to the artist to who painted
it as we try to get more rigorous with
our communication we try to really nail
down that meaning so we go from abstract
art to precise mathematics precise
engineering drawings and things like
that we're really trying to say I want
to narrow that that space of possible
interpretations
because the precision of the
communication ends up becoming more and
more important and so that means that I
have to specify and I think that's why
this becomes really hard because if I'm
just showing you an artifact and you're
looking at it superficially whether it's
a bunch of words on a page or whether
it's you know brushstrokes on a canvas
or pixels on a photograph you can sit
there and you can interpret lots of
different ways at many many different
levels but when I want to when I want to
align our understanding of that I have
to specify a lot more stuff that's
actually not in it not directly in the
artifact now I have to say well how you
were how are you interpreting this image
and that image and what about the colors
and what do they mean to you what's what
perspective are you bringing to the
table
what are your prior experiences with
those artifacts
what are your fundamental assumptions
and values what what is your ability to
kind of reason to chain together logical
implication as you're sitting there and
saying well if this is the case then I
would conclude this and if that's the
case then I would conclude that and it
so your reasoning processes and how they
work your prior models and what they are
your values and your assumptions all
those things now come together into the
interpretation getting in sick of that
is hard and yet humans able to intuit
some of that without any pre because
they have the shared experience me and
we're not talking about shared two
people have any shares know me as a
society that's correct we have this
shared experience and we have similar
brains so we tend to Institute in other
words part of our shared experiences are
shared local experience like we may live
in the same culture we may live in the
same society and therefore we have
similar education we have similar what
we like to call prior models about the
world prior experiences and we use that
as a think of it as a wide collection of
interrelated variables and they're all
bound to similar things and so we take
that as our background and we start
interpreting things similarly but as
humans we have it we have
a lot of shared experience we do have
similar brains similar goals similar
emotions under similar circumstances
because we're both humans so now one of
the early questions you ask well how is
biological and you know computer
information systems fundamentally
different well one is you know one is
come you means come with a lot of
pre-programmed stuff yeah a ton of
program stuff and they were able to
communicate because they have a lot of
it because they share that stuff do you
think that shared knowledge if it can
maybe escape the hardware question how
much is encoded in the hardware just the
shared knowledge in the software the the
history the many centuries of wars and
so on that came to today that shared
knowledge how hard is it to encode and
did you have a hope can you speak to how
hard is it to encode that knowledge
systematically in a way that could be
used by a computer so I think it is
possible to learn to form machine to
program machine to acquire that
knowledge with a similar foundation in
other words in a similar interpretive
interpretive foundation for processing
that knowledge but what do you mean by
that so in other in other words
foundation we view the world in a
particular way and so in other words we
we have i if you will as humans we have
a frame reference for bringing the world
around us so we have multiple frameworks
for interpreting the world around us but
if you're interpreting for example
social political interactions you're
thinking about what there's people
there's collections and groups of people
they have goals the goals largely built
around survival and quality of life that
are their fundamental economics around
scarcity of resources and when when
humans come and start interpreting a
situation like that because you've
brought you've grown up like historical
events they start interpreting
situations like that they apply a lot of
this a lot of this this fundamental
framework for interpreting that well who
are the people
what were their goals what
users did they have how much power
influence that they have over the other
like this fundamental substrate if you
will for interpreting and reasoning
about that so I think it is possible to
in view a computer with that that stuff
that humans like take for granted when
they go and sit down and try to
interpret things and then and then with
that with that foundation they acquire
they start acquiring the details the
specifics in any given situation are
then able to interpret it with regard to
that framework and then given that
interpretation they can do what they can
predict but not only can they predict
they can predict now with an explanation
that can be given in those terms in the
terms of that underlying framework that
most humans share now you could find
humans that come in interpret events
very differently than other humans
because they're like using a different
different framework you know movie
matrix comes to mind where you know they
decided the humans were really just
batteries and that's how they
interpreted the value of humans as a
source of electrical energy so but um
but I think that you know for the most
part we we have a way of interpreting
the events or do social events around us
because we have to share at framework it
comes from again the fact that we're
we're similar beings that have similar
goals similar emotions and we is we can
make sense out of these these frameworks
make sense to us so how much knowledge
is there do you think so it's you said
it's possible well there's all its
tremendous amount of detailed knowledge
in the world there you know you can
imagine you know effectively infinite
number of unique situations and unique
configurations of these things but the
the knowledge that you need what I refer
to as like the frameworks for you for
interpreting them I don't think I think
that's those are finite you think the
frameworks I'm more important than the
bulk of them now so it's like framing
yeah because the frameworks do is they
give you now the ability to interpret
and reason and to interpret and
reasoning to interpret and reason over
the specific
in ways that other humans would
understand what about the specifics you
know who acquired the specifics by
reading and by talking to other people
and so mostly actually just even if we
can focus on even the beginning the
common-sense stuff the stuff that
doesn't even require reading or
animalistic requires playing around with
the world or something just being able
to sort of manipulate objects drink
water and so on all does that every time
we try to do that kind of thing in
robotics or AI it seems to be like an
onion you seem to realize how much
knowledge is really required to perform
you in some of these basic tasks do you
have that sense as well and if so how do
we get all those details are they
written down somewhere idea they have to
be learned through experience so I think
when like if you're talking about sort
of the physics the basic physics around
us for example acquiring information
about for acquiring how that works yeah
I think that I think there's a
combination of things going I think
there's a combination of things going on
I think there is like fundamental
pattern matching like what were you
talking about before where you see
enough examples enough data about
something you start assuming that and
with similar input I'm going to predict
similar outputs you don't can't
necessarily explain it at all you may
learn very quickly that when you let
something go it falls to the ground
that's a that's a sickness is horribly
explained that but that's such a deep
idea if you let something go like they
do gravity I mean people were letting
things go and counting on them falling
well before they understood gravity but
that seems to be a that's exactly what I
mean is before you take a physics class
or the or study anything about Newton
just the idea that stuff falls to the
ground and they be able to generalize
that other all kinds of stuff falls to
the ground it just seems like a non if
without encoding it like hard coding it
in it seems like a difficult thing to
pick up it seemed like gift of Allah
of different knowledge to be able to
integrate that into the framework sort
of into everything else so both know
that stuff falls to the ground and start
to reason about social political
discourse so both like the very basic
and the high-level reasoning
decision-making I guess my question is
how hard is this problem and sorry to
linger on it because again and we'll get
to it for sure as well Watson with
jeopardy did its take on a problem
that's much more constrained but has the
same hugeness of scale at least from the
outsider's perspective so I'm asking the
general life question of to be able to
be an intelligent being and reason in
the in the world about both gravity and
politics how hard is that problem
so I think it's solvable okay now
beautiful so what about what about time
travel okay convinced not as convinced
yet okay no I said I I think it is I
mean I I took it as solvable I mean I
think that it's alert it's versatile
it's about getting machines to learn
learning is fundamental and I think
we're already in a place that we
understand for example how machines can
learn in various ways right now our
learning our learning stuff is sort of
primitive in that we haven't sort of
taught machines to learn the frameworks
we don't communicate our frameworks
because of our shared in some cases we
do but we don't annotate if you will all
the data in the world with the
frameworks that are inherent or
underlying our understanding instead we
just operate with the data so if we want
to be able to reason over the data in
similar terms in the common frameworks
we need to be able to teach the computer
or at least we need to program the
computer to require to have access to
and acquire
learn the frameworks as well and connect
the frameworks to the data I think this
I think this can be done I think we can
start
I think machine learnings for example
with enough examples can start to learn
these basic dynamics will they relate
the necessary to gravity not unless they
can also acquire those theories as well
and put the experiential knowledge and
connected back to the theoretical
knowledge I think if we think in terms
of these class of architectures that are
are designed to both learn the specifics
find the patterns but also acquire the
frameworks and connect the data to the
frameworks if we think in terms of
robust architectures like this I think
there is a path toward getting there
jeez in terms of encoding architectures
like that do you think systems they were
able to do this will look like and you
know that works or representing if you
look back to the eighties and nineties
of the expert systems so more like
graphs the systems that are based in
logic able to contain a large amount of
knowledge where the challenge was the
automated acquisition of that knowledge
the I guess the question is when you
collect both the frameworks and the
knowledge from the data what do you
think that thing will look like yeah so
I mean I think think is asking a
question they look like neural networks
is a bit of a red herring I mean I think
that they they will they will certainly
do inductive or pattern match based
reasoning and I've already experimented
with architectures that combine both
that use machine learning and neural
networks to learn certain classes of
knowledge in other words to find
repeated patterns in order or in order
for it to make good inductive guesses
but then ultimately to try to take those
learnings and and marry them in other
words connect them to frameworks so that
it can then reason over that in terms of
their humans understand so for example
at elemental cognition we do both we
have architectures that that do both but
both those things but also have a
learning method for acquiring the
frameworks themselves and saying look
ultimately I need to take this data I
need to interpret it in the form of
these frameworks so they can reason over
it so there is a fundamental knowledge
representation like what you saying like
these graphs of logic if you will there
are also neural networks that acquire
certain class of information they then
they they and align them with these
frameworks but there's also a mechanism
to acquire the frameworks themselves yes
so it seems like the idea of framework
requires some kind of collaboration with
humans absolutely so do you think of
that collaboration as well and unless to
be clear let's be clear only for the for
the express purpose that you're
designing you you're designing machine
designing and intelligence that can
ultimately communicate with humans in
terms of frameworks that help them
understand things right so so now to be
really clear you can create you can
independently create an a machine
learning system and an intelligent
intelligence that I might call an
alien's elegans that does a better job
than you with some things but can't
explain the framework to you that
doesn't mean is it might be better than
you at the thing it might be that you
cannot comprehend the framework that it
may have created for itself that is
inexplicable to you that's a reality but
you're more interested in a case where
you can I I am yeah I per might sort of
approach to AI is because I've set the
goal for myself I want machines to be
able to ultimately communicate
understanding with human I want to meet
would acquire and communicate acquire
knowledge from humans and communicate
knowledge to humans they should be using
what you know inductive machine learning
techniques are good at which is to
observe patterns of data whether it be
in language or whether it be in images
or videos or whatever to acquire these
patterns to induce the generalizations
from those patterns but then ultimately
work with humans to connect them to
frameworks interpretations if you will
that ultimately make sense to humans of
course the machine is gonna have the
strength egg it has the richer or longer
memory but that you know it has the more
rigorous reasoning abilities the deeper
reasoning abilities so be it interesting
you know complementary relationship
between the human and the machine do you
think that ultimately needs explained
ability like a machine so if we look we
study for example Tesla autopilot a lot
or humans I don't know if you've driven
the vehicle or are aware of what is it
so you basically
the human and machine are working
together there and the human is
responsible for their own life to
monitor the system and you know the
system fails every few miles and so
there's there's hundreds of there's
millions of those failures a day and so
that's like a moment of interaction DC
yeah that's exactly right that's a
moment of interaction where you know the
the the machine has learned some stuff
it has a failure
somehow the failures communicated the
human is now filling in the mistake if
you will or maybe correcting or doing
something that is more successful in
that case the computer takes that
learning so I believe that the
collaboration between human and machine
I mean that's sort of a permanent
example of sort of a more another
example is where the machine is
literally talking to you and saying look
I'm I'm reading this thing I know I know
that like the next word might be this or
that but I don't really understand why I
have my gas can you help me understand
the framework that supports this and
then can kind of take acquire that take
that and reason about it and reuse it
the next time it's reading to try to
understand something not on not unlike a
human student might do I mean I remember
like when my daughter was the first
great in she was had a reading
assignment about electricity and you
know somewhere in in the text it says
and electricity is produced by water
flowing over turbines or something like
that and then there's a question that
says well how was electricity created
and so my daughter comes to me and says
I mean I could you know created and
produced or kind of synonyms in this
case so I can go back to the text and I
can copy by water flowing over turbines
but I have no idea what that means like
I don't know how to interpret water
flowing over turbines and what
electricity even is I mean I can get the
answer right by matching the text but I
don't have any framework for
understanding what this means at all and
framework really I mean it's a set of
not to be mathematical but axioms of
ideas that you bring to the table and
interpreting stuff and then you build
those up somehow you build them up with
the expert
that there's a shared understanding of
what they are Sheriff it's the social
network that us humans do you have a
sense that humans on earth in general
share a set of like how many frameworks
are there I mean it depends on how you
bound them right so in other words how
big or small like their their individual
scope but there's lots and there are new
ones I think they're I think the way I
think about is kind of an a layer I
think that the architectures are being
layered in that there's there's a small
set of primitives that allow you the
foundation to build frameworks and then
there may be you know many frameworks
but you have the ability to acquire them
and then you have the ability to reuse
them I mean one of the most compelling
ways of thinking about this is or
reasoning by analogy where I could say
oh wow I've learned something very
similar you know I never heard of this I
never heard of this game soccer but if
it's like basketball in the sense that
the goals like the hoop and I have to
get the ball in the hoop and I have
guards and I have this and I have that
like we're weird is the where where are
the similarities and where the
difference is and I have a foundation
now for interpreting this new
information and then the different
groups like the Millennials will have a
framework and then and then well that
you never you know yeah well Kratz and
Republicans
well I Neal's nobody wants that
framework well I mean I think
understands it right I mean you're
talking about political and social ways
of interpreting the world around them
and I think these frameworks are still
largely largely similar I think they
differ in maybe what some fundamental
assumptions and values are now from a
reasoning perspective like the ability
to process the framework of Magna might
not be that different the implications
of different fundamental values or
fundamental assumptions in those
framework frameworks may reach very
different conclusions so from so from a
social perspective that conclusions may
be very different from an intelligence
perspective I you know I just followed
where my assumptions took me yeah the
product the process itself would look
similar but that's a fascinating idea
that frameworks really helped carve how
a statement will be interpreted
I mean having a Democrat and the
Republican framework and read the exact
same statement and the conclusions that
you derive would be totally different
from an ad respective is fascinating
what we would want out of the AI is to
be able to tell you that this
perspective one perspective one set of
assumptions is going to lead you here in
other setups as luncheons is gonna leave
you there and to and in fact you know to
help people reason and say oh I see
where I see where our differences lie
yeah you know I have this fundamental
belief about that I have this
fundamental belief about that yeah
that's quite brilliant from my
perspective and NLP there's this idea
that there's one way to really
understand a statement but there
probably isn't there's probably an
infinite number of ways then just as
well well there's a lot finding on
there's lots of different
interpretations and the you know the the
broader you know the broader to the the
contents the richer it is and so you
know you you and I can have very
different experiences with the same text
obviously and if we're committed to
understanding each other we start and
that's the other important point like if
we're committed to understanding each
other we start decomposing and breaking
down our interpretation towards more and
more primitive components until we get
to that point where we say oh I see why
we disagree and we try to understand how
fundamental that disagreement really is
but that requires a commitment to
breaking down that interpretation in
terms of that framework in a logical way
otherwise you know and this is why I
like I think of a eyes is really
complementing and helping human
intelligence to overcome some of its
biases and its predisposition to be
persuaded by you know buys but more
shallow reasoning in the sense that like
we get over this idea well I you know
you know I'm right because I'm a
Republican or I'm right because I'm
democratic and someone labeled this is
democratic point of view or it has the
following keywords in it and and if the
machine can help us break that argument
down and say wait a second you know what
do you really think about this right so
essentially holding us accountable to
doing more critical thinking
to sit and think about that as fast
that's I love that I think that's really
empowering use of AI for the public
discourse it's completely disintegrating
currently I don't know as we learn how
to do it on social medias right so one
of the greatest accomplishments in the
history of AI is Watson competing
against in a game of Jeopardy against
humans and you were a lead in that
accrue at a critical part of that let's
start the very basics what is the game
of Jeopardy the game for us humans human
versus human right so it's to take a
question and answer it actually no but
it's not right it's really not it's
really it's really to get a question and
answer but it's what we call a factoid
questions so this notion of like it's it
really relates to some fact that
everything few people would argue
whether the facts are true or not in
fact most people what and jeopardy kind
of counts on the idea that these these
statements have factual answers and and
the idea is to first of all determine
whether or not you know the answer which
is sort of an interesting twist so first
of all understand the question you have
to understand the question what is it
asking and that's a good point because
the questions are not asked directly
right they're all like the way the
questions are asked is nonlinear it's
like it's a little bit witty it's a
little bit playful sometimes it's a it's
a little bit tricky yeah they're asked
and exactly in numerous witty tricky
ways exactly what they're asking is not
obvious it takes it takes an experienced
humans a while to go what is it even
asking right and it's sort of an
interesting realization that you have
was a missus Oh what's the Jeopardy is a
question answering Shou and there's a go
like I know a lot and then you read it
and you're you're still trying to
process the question and the champions
have answered and moved on there's like
there's three questions ahead at the
time you figured out what the question
even met so there's there's definitely
an ability there to just parse out what
the question even is so that was
certainly challenging it's
interesting historically though if you
look back at the jeopardy games much
earlier
you know 63 yeah and I think the
questions were much more direct it
weren't quite like that they got sort of
more and more interesting the way they
asked them that sort of got more and
more interesting and subtle and nuanced
and humorous and witty over time which
really required the human to kind of
make the right connections and figuring
out what the question was even asking so
yeah you have to figure out the
questions even asking then you have to
determine whether or not you think you
know the answer and because you have to
buzz in really quickly you sort of have
to make that determination as quickly as
you possibly can otherwise you lose the
opportunity buzz in you've been going
before you really know if you know the
answer I think well I think a lot of
humans will will assume they'll they'll
look at the look at their process of
very superficially in other words what's
the topic what are some key words and
just say do I know this area or not
before they actually know the answer
then they'll buzz in and then I'll buzz
in and think about it it's interesting
what humans do now some people who know
all things like Ken Jennings or
something or the more recent big
jeopardy player that knows about that
though just assume they know although
jeopardy and I'll just pose it you know
Watson interestingly didn't even come
close to knowing all of Jeopardy right
Watson even at the peak even at that's
been yeah so for example I mean we had
this thing called recall which is like
how many of all the Jeopardy questions
you know how many did could we even find
like find the right answer for like
anywhere like could we come up with if
we look you know we had up a big body of
knowledge some of the order of several
terabytes I mean from from a web scale
was actually very small but from like a
book scales talking about millions in
bucks right so the equivalent millions
of books and cyclopædia is dictionaries
books it's a ton of information and you
know for I think was 80 only 85% was the
answer anywhere to be found hmm so
you're ready down you're ready down at
that level just to get just to get
started right so and so was important to
get a very quick sense of do you think
you know the right answer to this
question so we have to compute that
confidence as quickly as we possibly
could so it's in effect
to answer it and at least you know spend
some time essentially answering it and
then judging the confidence that we you
know that that our answer was right and
in deciding whether or not we were
confident enough to buzz in and that
would depend on what else was going on
in the game it could because it was a
risk so like if you're really in a
situation where I have to take a gas I
have very little to lose
then you'll buzz in with less confidence
so that was the counter for the the
financial standings of the different
competitors cracks yeah how much of the
game was laughs how much time was left
and where were you were in the standings
things like that what how many hundreds
of milliseconds that we're talking about
here do you have a sense of what is we
targets because we yeah was the targeted
so I mean we targeted answering and
under three seconds and buzzing it so
the decision to buzz in and then the
actual answering are those two yes there
were two there were two different things
in fact we had multiple stages whereas
like we would say let's estimate our
confidence which which is sort of a
shallow answering process and then
ultimate and then ultimately decide to
buzz in and then we may take another
second or something it's kind of go in
there and and do that but by and large
we're saying like we can't play the game
we can't even compete if we can't on
average answer these questions and
around three seconds or less
so you stepped in so there's this
there's these three humans playing a
game and you stepped in with the idea
that IBM Watson would be one of replaced
one of the humans and compete against
two can you tell the story of Watson
taking on this game sure seems
exceptionally difficult yeah so the
story was that um it was or it was
coming up I think the 10-year
anniversary of a big blue an optical
deep blues IBM wanted to do sort of
another kind of really you know fun
challenge public challenge that can
bring attention to IBM research and the
kind of cool stuff that we were doing
I had been working in an AI at IBM for
some time I had a team doing what's
called open domain factoids
question-answering which is you know
we're not gonna tell you what the
questions are we're not even gonna tell
you what they're about
can you go off and get accurate answers
to these questions and it was an area of
AI research that I was involved in and
so it was a big Pat it was a very
specific passion of mine language
understanding and always always been a
passion of mine one sort of narrow slice
on whether or not you could do anything
was language was this notion of open
domain and meaning I could ask anything
about anything
factoids meaning it essentially had an
answer and and you know being able to do
that accurately and quickly so that was
a research area that might even already
been in and so completely independently
several you know IBM exactly there's
like what are we gonna do what's the
next cool thing to do and Ken Jennings
was on his winning streak this was like
whatever was 2004 I think was on his win
winning streak when someone thought hey
that'd be really cool um if the computer
can play jeopardy and so this was like
in 2004 they were shopping this thing
around and everyone who's telling the
the research execs no way like this is
crazy
and we had some pretty you know senior
people know if you'll understand the
others crazy and he'll come across my
desk and I was like but that's kind of
what what I'm really interested in doing
and but there was such this prevailing
sense of this is nots we're not going to
risk IBM's reputation on this we're just
not doing it and this happened in 2004
it happened in 2005 at the end of 2006
it was coming around again and I was
coming off of a I was doing that the
open domain question-answering stuff but
I was coming off a couple other projects
I had a lot more time to put into this
and I argued that it could be done and I
argue it would be crazy not to do this
can I
you could be honest at this point so
even though you argued for it what's the
confidence that you had yourself
privately that this could be done it was
we just totally told the story of how
you tell stories to convince others how
confident were you what was your
estimation of the problem
at that time so I thought it was
possible and a lot of people thought it
was impossible I thought it was possible
a reason why I thought it was possible
is because I did some brief
experimentation I knew a lot about how
we were approaching on open domain
factoids question asked me we have been
doing it for some years I looked at the
Japanese stuff I said this is going to
be hard for a lot of the points that you
mentioned earlier hard to interpret the
question hard to do it quickly enough
hard to compute an accurate confidence
none of this stuff had been done well
enough before but a lot of the
technologies were building with the
kinds of technologies that should work
but more to the point what was driving
me was I was an IBM research I was a
senior leader in IBM Research and this
is the kind of stuff we were supposed to
do we were basically supposed to the
moonshot this is I mean we were supposed
to take things and say this is an active
research area it's our obligation to
kind of if we have the opportunity to
push it to the limits and if it doesn't
work to understand more deeply why we
can't do it and so I was very committed
to that notion saying folks this is what
we do it's crazy not not to do this is
an active research area we've been in
this for years why wouldn't we take this
Grand Challenge and and push it as hard
as we can at the very least we'd be able
to come out and say here's why this
problem is is way hard here's what we've
tried and here's how we failed so I was
very driven as a scientist from that
perspective and then I also argued based
on what we did a feasibility study oh
why I thought it was hard but possible
and I showed examples of you know where
it succeeded where it failed why it
failed and sort of a high level
architecture approach for why we should
do it but for the most part that at that
point the execs really were just looking
for someone crazy enough to say yes
because for several years at that point
everyone has said no I'm not willing to
risk my reputation and my career you
know on this thing clearly you did not
have such fears okay I did not say you
died right in and yet for what I
understand it was
performing very poorly in the beginning
so what were the initial approaches and
why did they fail well there were lots
of hard aspects to it I mean one of the
reasons why prior approaches that we had
worked on in the past
um failed was because of because the
questions were difficult difficult to
interpret like what are you even asking
for right very often like if if the
question was very direct like what city
you know or what you know even then it
could be tricky but but you know what
city or what person was often when it
would name it very clearly you would
know that and and if there was just a
small set of them in other words we're
gonna ask about these five types like
it's gonna be an answer and the answer
will be a city in this state or a city
in this country the answer will be a
person of this type right like an actor
or whatever it is but turns out that in
jeopardy there were like tens of
thousands of these things and it was a
very very long tale meaning you know
that it just went on and on and and so
even if you focused on trying to encode
the types at the very top like there's
five that were the most let's say five
of the most frequent you still cover a
very small percentage of the data so you
couldn't take that approach of saying
I'm just going to try to collect facts
about these five or ten types or twenty
types or fifty types or whatever so that
was like one of the first things like
what do you do about that and so we came
up with a an approach toward that and
the approach to look promising and we we
continue to improve our abilities to
handle that problem throughout the
project the other issue was that right
from the outside I said we're not going
to I committed to doing this in three
five years so we did in four so I got
lucky
um but one of the things that that
putting that like stake in the ground
was I and I knew how hard the language
of the standard problem was I said we're
not going to actually understand
language to solve this problem we are
not going to
interpret the question and the domain of
knowledge the question refers to in
reason over that to answer these
questions were obviously we're not going
to be doing that at the same time simple
search wasn't good enough to confidently
answer with this you know a single
correct answer first others like
brilliant that's such a great mix of
innovation in practical engineering
three three four eight so you're not
you're not trying to solve the general
NLU problem you're saying let's solve
this in any way possible oh yeah no I
was committed to saying look we're gonna
solving the open the main question
answering problem we're using jeopardy
as a driver for that management hard
enough big benchmark exactly and now
we're how do we do it we're just like
whatever like just figure out what works
because I want to be able to go back to
the académica scientific community and
say here's what we tried here's what
work here's what didn't work I don't
want to go in and say oh I only have one
technology hammer and only gonna use
this I'm gonna do whatever it takes I'm
like I'm gonna think out of the box do
whatever it takes one um and I also
Baloo's another thing I believed I
believe that the fundamental NLP
technologies and machine learning
technologies would be would be adequate
and this was an issue of how do we
enhance them how do we integrate them
how do we advance them so I had one
researcher and came to me who had been
working on question answering with me
for a very long time
who had said we're gonna need Maxwell's
equations for question-answering and I
said if we if we need some fundamental
formula that breaks new ground and how
we understand language we're screwed
yeah we're not gonna get there from here
like we I am not counting I am that my
assumption is I'm not counting on some
brand new invention what I'm counting on
is the ability to take everything that
has done before to figure out a an
architecture on how to integrate it well
and then see where it breaks and make
the necessary advances we need to make
and sold this thing works
yeah push it hard to see where it breaks
and then patch it up I mean that's how
people change the world and that's the
you know mosque approaches Rockets
SpaceX that's the Henry Ford and so on a
lot and and I happen to be and in this
case I happen to be right but but like
we didn't know right but you kind of
have to put a second or so how you gonna
run the project so yep and backtracking
to search so if you were to do what's
the brute force solution what what would
you search over so you have a question
how would you search the possible space
of answers look web search has come a
long way even since then but at the time
like you know you first of all I mean
there are a couple of other constraints
around the problems interesting so you
couldn't go out to the web you couldn't
search the Internet in other words the
AI experiment was we want a
self-contained device device if devices
as big as a room fine it's as big as a
room but we want a self-contained advice
contained device you're not going out
the internet you don't have a life
lifeline to anything so it had to kind
of fit in a shoebox if you will or at
least the size of a few refrigerators
whatever it might be
see but also you couldn't just get out
there you couldn't go off Network right
to kind of go so there was that
limitation but then we did it but the
basic thing was go go do what go do a
web search the problem was even when we
went and did a web search I don't
remember exactly the numbers but someone
the order of 65% at a time the answer
would be somewhere you know in the top
10 or 20 documents so first of all
that's not even good enough to play Jack
pretty you know the words even if you
could pull the avian if you could
perfectly pull the answer out of the top
20 documents top 10 documents whatever
was which we didn't know how to do but
even if you could that do that your
you'd be at and you knew it was Ryan
Lizza we've had enough confidence in it
right so you have to pull out the right
answer you have you depth of confidence
it was the right answer and and then
you'd have to do that fast enough to now
go buzz in and you'd still only get 65%
of them right with nine doesn't even put
you in the winner's circle winner's
circle you have to be up over 70 and you
have to do it really quick and you do
really quickly but now the problem is
well even if I had somewhere in the top
10 documents how do I figure out where
in the top 10 documents that answer is
and how do i compute a confidence of all
the possible candidates so it's not like
I go in knowing the right answer and I
have to pick it I don't know the right
answer I have a bunch of documents
somewhere in there's the right answer
how do i as a machine go out and figure
out which ones right and then how do I
score it so and now how do I deal with
the fact that I can't actually go out to
the web first of all if you pause and
then just think about it if you could go
to the web do you think that problem is
solvable if you just pause on it just
thinking even beyond jeopardy do you
think the problem of reading text
defined where the answer is but we saw
we solved that and some definition of
solves given the Jeopardy challenge how
did you do it forever so how did you
take a body of work and a particular
topic and extract the key pieces of
information so what so now forgetting
about the the huge volumes that are on
the web right so now we have to figure
out we did a lot of source research in
other words what body of knowledge is
gonna be small enough but broad enough
to answer Jeffrey and we ultimately did
find the body of knowledge that did that
I mean it included Wikipedia and a bunch
of other stuff so like encyclopedia type
of stuff I don't know if you use Mary's
different types of semantic resources
unlike wordnet and other types of Mantic
resources like that as well as like some
web crawls in other words where we went
out and took that content and then
expanded it based on producing
statistical see you know statistically
producing sees using those sees for
other searchers searches and then
expanding that so using these like
expansion techniques we went out and had
found enough content and we're like okay
this is good and we even up and totally
and you know we had a threat of
resources always trying to figure out
what content could we efficiently
include I mean there's a lot of popular
cut like what is the church lady well I
think was one of the end hey yeah what
we ready I guess that's probably an
encyclopedia so it's a pepino is that
but then we would but then we would take
that stuff when we would go out and we
would expand in other words we go find
other content that wasn't in the core
resources and expanded you know the
amount of content will grew it by an
order of magnitude but still so again
from a web scale perspective this is
very small amount of content it's very
select we then we then took all that
content so we we pre analyzed the crap
out of it meaning we we we parsed it you
know broke it down into all this
individual words and then we did
semantic static and semantic parses on
it you know had computer algorithms that
annotated it and we in that we indexed
that in a very rich and very fast index
so we have a relatively huge amount of
you know let's say the equivalent of for
the sake of argument two to five million
bucks
we've now analyzed all that blowing up
at size even more because now with all
this metadata and we then we richly
indexed all of that and in by way in a
giant in-memory cache so Watson did not
go to disk so the infrastructure
component there if you just speak to it
how tough it I mean I know mm maybe this
is 2089
you know that that's kind of a long time
ago right
how hard is it to use multiple machines
Olivia how hard is the infrastructure
part of the hardware component we used
IBM we so we used IBM hardware we had
something like I figured exactly but
2,000 to 3,000 cores completely
connected so had a switch were you know
every CPU was connected to every other
scene they were sharing memory in some
kind of way Lauren up close shared
memory right and all this data was pre
analyzed and put into a very fast
indexing structure that was all all all
in all in memory and then
we took that question we would analyze
the question so all the content was now
pre analyzed so if I so if I went and
tried to find a piece of content it
would come back with all the metadata
that we had pre computed how do you
shove that question how do you connect
the the big stuff with the meta the the
big knowledgebase of the metadata and
that's indexed to the simple little
witty confusing question right so
therein lies
you know the Watson architects right so
we would take the question we would
analyze the question so which means that
we would parse it and interpret it a
bunch of different ways we try to figure
out what is it asking about so we would
come we had multiple strategies to kind
of determine what was it asking for that
might be represented as a simple string
and character string or was something we
would connect back to different semantic
types that were from existing resources
so anyway the bottom line is we would do
a bunch of analysis and the question and
question analysis had to finish and had
to finish fast so we do the question
analysis because then from the question
analysis we would now produce searches
so we would and we had built using open
source search engines we modified them
we had a number of different search
engines we would use that had different
characteristics we went in there and
engineered and modified those search
engines ultimately to now take our
question analysis produce multiple
queries based on different
interpretations of the question and fire
out a whole bunch of searches in
parallel and they would produce combate
with passages so this is these are
passive search algorithms they would
come back with passages and so now you
let's say you had a thousand passages
now for each passage you you parallel
eyes again so you went out and you
paralyze those paralyze the search each
search would now come back with a whole
bunch of passages maybe you had a total
of a thousand or five thousand different
passages for each passage now you don't
figure out whether or not there was a
candidate it would call it candidate
answer in there so you had a whole bunch
of other a whole bunch of other
algorithms
that would find candidate answers
possible answers to the question and so
you had candidate answers jet cold
candidate answers generators a whole
bunch of those so for every one of these
components the team was constantly doing
research coming up better ways to
generate search queries from the
questions better ways to analyze the
question better ways to generate
candidates and speed so better is
accuracy and speed cracked so right and
speed and accuracy for the most part
we're separated we handle that sort of
in separate ways like I focus purely on
accuracy and to an accuracy are we
ultimately getting more questions and
producing more accurate confidences and
they had a whole nother team that was
constantly analyzing the workflow to
find the bottlenecks and then if you're
getting out of both parallel eyes and
drive the algorithm speed but anyway so
so now think of it like you have this
big fan out now right because you have
you had multiple queries now you have
now you have thousands of candidate
answers for each candidate answer you're
gonna score it so you're gonna use all
the data that built up you're gonna use
the question analysis you can use how
the query was generated you're going to
use the passage itself and you're going
to use the candidate answer that was
generated and you're gonna score that so
now we have a group of researchers
coming up with scores there are hundreds
of different scores so now you're
getting a fan at it again from however
many candidate answers you have to all
the different scorers so if you have a
200 different scores and you never a
thousand candidates now you have two
thousand scores and and so now you got
to figure out you know how do I now rank
these rank these answers based on the
scores that came back and I want to rank
them based on the likelihood that there
are correct answer to the question so
every score was its own research project
what do you mean by score so is that the
annotation process of basically human
being saying that this this answer do
you think you think of if you want to
think of it what you're doing you know
if you want to think about what a human
would be doing human would be looking at
a possible answer they'd be reading the
you know Emily Dixon Dickinson they've
been reading the passage in which that
occurred they'd be looking at the
question
they'd be making a decision of how
likely it is that Emily Dixon Dickinson
given this evidence in this passage is
the right answer to that quad got it
so that that's the annotation task that
Stan Johnson scoring task so but scoring
implies zero to one kind of trite
continuance is not a binary no give it a
score give it a zero yeah exactly so
it's what humans did give different
scores so that you have to somehow
normalize and all that kind of stuff
that deal with all that depends on what
your strategy is we both we could be
relative to it could be we actually
looked at the raw scores as well
standardized scores because humans are
not involved in this
humans are not involved sorry so I mean
I'm misunderstanding the the the process
here this is passages where is the
ground truth coming from grass root
there's only there were answers to the
questions so it's end to end
it's end to end so we also I was always
driving and and performance a very
interesting a very interesting you know
engineering approach and ultimately
scientific and researcher personal
always driving in 10 now that's not to
say we wouldn't make hypotheses that
individual component performance was
related in some way to n10 performance
of course we would because people would
have to build individual components but
ultimately to get your component
integrates with the system you had to
show impact on end-to-end performance
question-answering performance as
there's many very smart people work on
this and they're basically trying to
sell their ideas as a component that
should be part of the system that's
right and and they would do research on
their component and they would say
things like you know I'm going to
improve this as a candidate generator
I'm going to improve this as a question
score or as a passive scorer I'm going
to proved as or as a parser and I can
improve it by two percent on its
component metric like a better parse or
better candidate or a better type
estimation or whatever it is and then I
would say I need to understand how the
improvement on that computer metric is
going to affect the end-to-end
performance if you can't estimate
that and can't do experiments to
demonstrate that it doesn't get in
that's like the best run AI project I've
ever heard that's awesome okay what
breakthrough would you say like I'm sure
there's a lot of day to day break this
but it was there like a breakthrough
that really helped improve performance
like wait what people began to believe
or is it just a gradual process well I
think it was a gradual process but one
of the things that I think gave people
confidence that we can get there was
that as we fouled as as we follow this
procedure of different ideas build
different components plug them into the
architecture run the system see how we
do do the error analysis start off new
research projects to improve things and
the and and and the very important idea
that the individual component work did
not have to deeply understand everything
that was going on with every other
component and this is where we we
leverage machine learning in a very
important way so while individual
components could be statistically driven
machine learning components some of them
were your wrist ik some of them were
machine learning components the system
has a whole combined all the scores
using machine learning this was critical
because that way you can divide and
conquer so you can say okay you work on
your candidate generator or you work on
this approach to answer scoring you work
on this approach to type scoring you
work on this approach to passage search
or the passive selection and so forth
but when we you just plug it in and we
had enough training data to say now we
can we can train and figure out how do
we weigh all the scores relative to each
other based on the predicting the
outcome which is right right or wrong on
jeopardy and we had enough training data
to do that so this enabled people to
work independently and to let the
machine learning do the integration
beautiful so that yeah the machine
learning is doing the fusion and then
it's a human orchestrated ensemble
that's right friend approaches as a
great still impressive they were able to
get it done a few years that not obvious
to me that it's doable if I just put
myself in that mindset but when you look
back at the Jeopardy challenge again
when you're looking up at the stars what
are you most proud of looking back at
those days I'm most proud of my um my
commitment and my team's commitment to
be true to the science to not be afraid
to fail that's beautiful because there's
so much pressure because it is a public
event this is a public show that you
were dedicated to the idea that's right
do you think it was a success in the
eyes of the world it was a success by
your I'm sure exceptionally high
standards is there something you regret
you would do differently
it was a success it was a success for
our goal our goal was to build the most
advanced open domain question-answering
system we went back to the old problems
that we used to try to solve and we did
dramatically better on all of them as
well as we beat jeopardy so we wanted
jeopardy so it was it was a success it
was I worried that the world would not
understand that has success because it
came down to only one game and I knew
statistically speaking this can be a
huge technical success and we could
still lose that one game and that's a
whole nother theme of this of the
journey but it was a success it was not
a success in natural language
understanding but that was not the goal
yeah that was but I would argue I
understand what you're saying in terms
of the science but I would argue that
the inspiration of it right the they not
a success in terms of solving natural
language understanding there was a
success of being an inspiration to
future challenges absolutely drive
future efforts what's the difference
between how human being compete in
jeopardy and how Watson does it that's
important in terms of intelligence yeah
so thats that actually came out very
early on in the project also in fact I
had people who wanted to be on the
project who were
early on who has sort of approached me
once I committed to do it had wanted to
think about how humans do it and they
were you know from a cognition
perspective like human cognition and how
that should play and I would not take
them on the project because another
assumption or another stake I put in the
ground was I don't really care are you
into this at least in the context of
this prior need to build in the context
to this project in NLU and in building
an AI that understands how it needs to
alter that communicate with humans I
very much care yeah so wasn't that I
didn't care in general in fact as an AI
scientist I care a lot about that but
I'm also a practical engineer and I
committed to getting this thing done and
I wasn't gonna get distracted I had to
kind of say look if I'm gonna get this
done and when it charts this path and
this path says we're gonna engineer a
machine that's gonna get this thing done
and we know what search and NLP can do
we have to build on that foundation if I
come in and take a different approach
and start wondering about how the human
mind might or might not do this I'm not
going to get there from here in the time
and you know in the timeframe I think
that's a great way to lead the team but
now there's done and then one when you
look back right so analyse what's the
difference sexy right so so I was a
little bit surprised actually to
discover over time as this would come up
from time to time and would reflect on
it that and and talking to Ken Jennings
a little bit and hearing Ken Jennings
talk about it about how he answered
questions that it might have been closer
to the way humans answer questions than
I might have imagined previously because
humans are probably in the game of
Jeopardy at the level of Ken Jennings
probably also cheating their weight into
winning right now one else is shallow
they're doing that fast as possible
they're doing shallow analysis so they
are very quickly analyzing the question
and coming up with some you know key you
know key vectors or cues if you will and
they're taking those cue
they're very quickly going through like
their library of stuff not deeply
reasoning about what's going on and then
sort of like a lots of different like
what we call these these scores which
kind of score that in a very shallow way
and then say oh boom you know that's
what it is and and so it's interesting
as we reflected on that so we may be
doing something that's not too far off
from the way humans do it but we certain
certainly didn't approach it by saying
you know how would you even do this now
in an elemental cognition like the
project I'm leading now we asked those
questions all the time because
ultimately we're trying to do something
that is to make the the the intelligence
in the machine and the intelligence of
the human very compatible
well compatible in the sense they can
communicate with one another and they
can reason with this shared
understanding so how they think about
things and how they build answers how
they build explanations becomes a very
important question to consider so what's
the difference between this open domain
but cold constructed question answering
or jeopardy and more something that
requires understanding for shared
communication with humans and machines
yeah well this goes back to the
interpretation of what we were talking
about before anyway jeopardy the systems
on trying to interpret the question and
that's not interpreting the content
that's reasoning and with regard to any
particular framework I mean it's it is
parsing it and like parsing the contents
and using grammatical cues and stuff
like that so if you think of grammar as
a human framework in some sense and as
that but when you get into the richer
semantic frameworks what are people how
do they think what motivates them what
are the events that are occurring and
why are they occurring and what causes
what else to happen and and and when it
where are things in time and space and
it's like when you started thinking
about how humans formulate and structure
the knowledge that they acquire in their
head and wasn't doing any of that what
do you think are the essential
challenges of like free flowing
communication free flowing
log versus question-answering even with
the framework of the interpretation
dialogue yep
do you see free-flowing dialogue as a
fundamentally more difficult than
question answering even with shared so
dialogue is as important in number of
different ways I mean it's a challenge
so first of all when I think about the
machine that when I think about a
machine that understands language and
ultimately can reason in an objective
way that can take the information that
it perceives through language or other
means and connects it back to these
frameworks reason and explain itself
that system ultimately needs to be able
to talk to humans or I needs to be able
to interact with humans so in some
sentence to dialogue that doesn't mean
that it it that like sometimes people
talk about dialogue and they think you
know how do humans talk how do you
montork talk to each other in a casual
conversation then you could mimic casual
conversations we're not trying to mimic
casual conversations we're really trying
to produce the machine as goal is it is
to help you think and help you reason
about your answers and explain why so
instead of like talking to your friend
down the street about having a smoke
having a small talk conversation with
your friend down the street this is more
about like you would be communicating to
the commuter computer on Star Trek we're
like what do you want to think about
like what do you want to reason about
I'm going to tell you the information I
have I'm gonna have to summarize it I'm
gonna ask you questions you're gonna
answer those questions I'm gonna go back
and forth with you I'm gonna figure out
what your mental model is I'm gonna I'm
gonna now relate that to the information
I have and present it to you in a way
that you can understand it and we could
ask follow-up questions so it's that
type of dialogue that you want to
construct it's more structured it's more
goal oriented but it needs to be fluid
in other words it can't it can't it has
to be engaging and fluid it has to be
productive and not distracting so there
has to be a model of the words the
machine has to have a model of how
humans think through things and discuss
them so basically a productive rich
conversation unlike this part yes but
what I'd like to think it's more similar
to this pocket as in joking I'll ask you
about humor as well actually but what's
the hardest part of that because it
seems we were quite far away as a
community from thats though to be able
to so one is having a shared
understanding as i think a lot of the
stuff you said with frameworks is quite
brilliant
but just creating a smooth discourse
yeah it feels clunky right now well
which aspects of this whole problem you
just specified all having a productive
conversation is the hardest and that
were or maybe maybe any aspect of it you
can comment on because it's so shrouded
in mystery so I think do this you kind
of have to be creative in the following
sense if I were to do this is purely a
machine learning approach and someone
said learn how to have a good flue in
structured knowledge acquisition
conversation I'd go out and say okay I
have to collect a bunch of data of
people doing that people reasoning well
having a good structured conversation
that both acquires knowledge efficiently
as well as produces answers and
explanations as part of the process and
you struggle I don't know
elect a day to collect the data because
I don't know how much data is like that
I think okay okay so this one there's a
human but also even if it's out there
say was out there how do you like
alright so I think I think like an
accessible right so I think any like any
problem like this where you don't have
enough data to represent the phenomenon
you want to learn in other words you
want you if you have enough data you
could potentially learn the pattern in
an example like this it's hard to do it
this is the you know Susie sort of a
human sort of thing to do what you
recently came out IBM was the debate or
projects and surest thing right because
now you had you do have these structured
dialogues these debate things where they
did use machine learning techniques to
generate the you know generate these
debates dialogues are a little bit
tougher in my opinion than generating a
a structured argument where you have
lots of other structural arguments like
this you could potentially annotate that
data and you could say this is a good
response a bad response in a particular
domain here
I have to be responsive and I have to be
opportunistic with regard to what is the
human saying what so I'm goal-oriented
and saying I want to solve the problem I
want to acquire the knowledge necessary
but I also have to be opportunistic and
responsive to what the human is saying
so I think that it's not clear that we
could just train on a body of data to do
this but we could bootstrap it in other
words we can be creative and we could
say what do we think what do we think
the structure of a good dialogue is that
does this well and we can start to
create that if we can if we can create
that more programmatic programmatically
at least to get this process started and
I can create a tool that now engages
humans effectively I could start both I
could start generating data I could
start with the human learning process
and I can update my machine but I can
also start the automatic learning
process as well but I have to understand
what features to even learn over so I
have to bootstrap the process a little
bit first and that's a creative design
task that I could then use as input into
a more automatic learning task this is
some creativity and bootstrapping all
right what elements of conversation do
you think you would like to see so one
of the benchmarks for me is humor right
that seems to be one of the hardest if
you end to me the biggest contrast is
from Watson so one of the greatest
sketches of comedy sketches of all time
right is the SNL celebrity jeopardy with
uh with with Alex Trebek and Sean
Connery and Burt Reynolds and so on with
uh with the Sean Connery commentating on
Alex Trebek smile there a lot so and I
think all of them are in the negative
point what's why so they're clearly all
losing in terms of the game of Jeopardy
but they're winning in terms of comedy
so what do you think about humor in this
whole interaction in the dialogue that's
productive or even just whatever what
human represents to me is it the same
idea that you're saying about a
framework because humor only exists
within a particular human framework
so what do you think about humor what do
you think about things like humor that
connect to the kind of creativity you
mentioned that's needed I think there's
a couple things going on there so I I I
sort of feel like and I might be too
optimistic this way but I think that
there are we did a little bit about with
with this and with puns and in jeopardy
we literally sat down and said well you
know how do puns work and you know it's
like wordplay and you could formalize
these things so I think there's a lot
aspects of humor that you could
formalize you could also learn new Murr
you could just say what do people laugh
at and if you have enough again if you
have enough data to represent the
phenomenon you know might be able to you
know weigh the features and figure out
you know what humans find funny and what
they don't find funny
you might the Machine might not be able
to explain why the my buddy unless we
unless we sit back and think about that
more formally I think again I think you
do a combination of both and I'm always
a big proponent that I think you know
robust architectures and approaches are
always a little bit combination of us
reflecting and being creative about how
things are structured and how to
formalize them and then taking advantage
of large data and doing learning and
figuring how to combine these two
approaches I think there's another
aspect of human to human though which
goes to the idea that I feel like I can
relate to the person telling the story
telling the person telling the story and
I think that's that's a interesting
theme in the whole AI theme which is do
I feel differently when I know it's a
robot and when I know when I imagine
there's a row but is not conscious the
way I'm conscious when they imagine the
robot does not actually have the
experiences that I experience do I find
it you know funny or do because it's not
as related I don't imagine that the
person is relating it to it the way I
relate to it I think this also you see
this in in the arts and in entertainment
where like you know sometimes you have
savants who are remarkable at a thing
whether it's sculpture it's music or
whatever but the people who get the most
attention are the people who can't who
can evoke a similar emotional response
who can get you to emote right about the
way they
in other words who can basically make
the connection from the artifact from
the music of the painting of the
sculpture to the to the emotion and get
you to share that emotion with them and
then and that's when it becomes
compelling so they're communicating at a
whole different level they're just not
communicating the artifact they're
communicating their emotional response
to the artifact and then you feel like
oh wow I can relate to that person I can
connect to that I can connect to that
person so I think humor has that has
that aspect as well so the idea that you
can connect to that person person being
the critical thing but we're also able
to anthropomorphize objects pretty
robots and AI systems pretty well
so we're almost looking to make them
human there may be from your experience
with Watson maybe you can comment on did
you consider that as part well obviously
the problem of Jeopardy doesn't require
int the promotoras ation but
nevertheless well there was some
interest in doing that and I've that's
an that's another thing I didn't want to
do so I didn't want to distract from the
from the actual scientific test nights
so you're absolutely right I mean humans
do anthropomorphize and and without
necessarily a lot of work I mean just
put some eyes in a couple of eyebrow
movements and you're getting humans to
react emotionally and I and I think you
can do that so I didn't mean to suggest
that that that connection can't cannot
be mimicked I think that connection can
be mimicked and can get you to can
produce that emotional response I just
wonder though if you're told what's
really going on if you know that the
machine is not conscious not having the
same richness of emotional reactions and
understanding that doesn't really share
the understanding but is essentially
just moving inside brow or drooping its
eyes or making them big or whatever it's
doing that's getting the emotional
response will you still feel it
interesting I think you probably would
for a while and then when it becomes
more important that there's a deeper
under depreciate understanding it may
run flat but I don't know I'm pretty I'm
pretty confident that it will
the majority of the world even if you
tell them how no matter well it will not
matter especially if the Machine herself
says that she is cautious that's very
possible so you the scientists that made
the machine is saying that this is how
the algorithm works everybody will just
assume you're lying and that there's a
conscious being there so you're deep
into the science fiction shop you're on
right now but yeah I think it's actually
psychology I think it's not science
fiction
I think it's reality I think it's a
really powerful one that will have to be
exploring in the next few decades it's a
very interesting element of intelligence
so what do you think we've talked about
social constructs of intelligences and
and frameworks and the way humans kind
of interpret information what do you
think is a good test of intelligence in
your view so there's the Alan Turing
with the Turing test
Watson accomplished something very
impressive with Jeopardy what do you
think is a test that would impress the
heck out of you that you saw that a
computer could do they say this is
crossing a kind of threshold that's that
gives me pause in a good way
expectations for a are generally high
what does high look like by the way so
not the threshold test as a threshold
what do you think is the destination
what do you think is the ceiling
I think machines will in many measures
will be better than us will become more
effective in other words better
predictors about a lot of things and
then then then ultimately we can do I
think where they're gonna struggle is
what we talked about before which is
relating to communicating with and
understanding humans in deeper ways and
and so I think that's a key point like
we can create the super parrot what I
mean by the super parrot is given enough
data a machine can mimic your emotional
response can even generate language that
will sound smart and what someone else
might say under similar circumstances
look how its paws on that like that's a
super parrot right so given similar
circumstances moves its face its faces
in similar ways changes its tone of
voice in similar ways produce the
strings of language that you know would
similar that a human might say not
necessarily being able to produce a
logical interpretation or understanding
that would ultimately satisfy a critical
interrogation or a critical
understanding I think you guys describe
me in a nutshell
I think I think philosophically speaking
you could argue that that's all we're
doing as human beings to war so I was
gonna say it's very possible you know
humans do behave that way too and so
upon deeper probing and deeper
interrogation you may find out that
there isn't a shared understanding
because I think humans do both like
humans are statistical language model
machines and and and they are capable
reasoner's you know they're they're both
and you don't know which is going on
right so and I think it's I think it's
an interesting problem we talked earlier
about like where we are in our social
and political landscape can you
distinguish some
who can string words together and sound
like they know what they're talking
about from someone who actually does can
you do that without dialogue without
integrity of a programming dialogue so
it's interesting because humans are
really good at in their own mind
justifying or explaining what they hear
because they project their understanding
on onto yours so you could say you could
put together a string of words and and
someone will sit there and interpret in
a way that's extremely biased this is
the way they want to interpret it they
want to assuming you're an idiot and
they'll true put it one way they've all
seen you're a genius and interpreted
another way that suits their needs so
this is tricky business so I think the
answer your question as AI gets better
and better at better and better mimic
you we create the super parrots we're
challenged just as we are with we're
challenged with humans do you really
know what you're talking about do you
have a meaningful interpretation a
powerful framework that you could reason
over and justify your answers justify
your predictions and your beliefs why
you think they make sense can you
convince me what the implications are
you know can you
so can you reason intelligently and make
me believe that those um the
implications of your prediction and so
forth so what happens is it becomes
reflective my standard for judging your
intelligence depends a lot on mine but
you're saying that there should be a
large group of people with a certain
standard of intelligence that would be
convinced by this particular AI system
then there should be by I think one of
the depending on the content one of the
problems we have there is that if that
large community of people are not
judgment judging it with regard to a
rigorous standard of objective logic and
reason you still have a problem like
masses of people can be persuaded the
Millennials yeah to turn them turn their
brains off
right okay sorry I have nothing against
the warning I just so you you're a part
of one of the great benchmarks
challenges of AI history what do you
think about alpha zero open AI five
alpha star accomplishments on video
games recently which are also I think at
least in the case of go without fagala
now for zero playing go was a monumental
accomplishment as well what are your
thoughts about that challenge I think it
was a giant lamare I I think it was
phenomenal I mean as one of those other
things nobody thought like solving go
was gonna be easy particularly because
it's again it's hard for particularly
hard for humans our team is to learn how
for humans to excel at and so it was up
another measure a measure of
intelligence it's very cool I mean it's
very interesting you know what they did
I mean and I loved how they solved like
the data problem which again they
bootstrapped it and got the machine to
play itself to generate enough data to
learn from I think that was brilliant I
think that was great and and and of
course the result speaks for itself I
think it makes us think about again it
is okay what's intelligence what aspects
of intelligence are important can the
can the go machine help me make me a
better go player is it an alien
intelligence it was is am I even capable
of like again if we if we put in very
simple terms it found the function we
found the go function can I even
comprehend the go function can I talk
about the go function can i
conceptualize the go function like
whatever it might be so one of the
interesting ideas of that system is it
plays against itself right yeah but
there's no human in the loop there so
like you're saying it could have by
itself created an alien intelligence how
torta torta gorrik imagine you're
sentencing you're judging you're
sentencing people or you're setting
policy or you're you know you're making
medical decisions and you can't explain
you can't get anybody to understand what
you're doing or why so it's it's it's an
interesting dilemma
for the applications of AI do we hold AI
to this accountability that says you
know humans have to be humans have to be
able to take responsibility you know for
for the decision in other words can you
explain why you would do the thing well
you will use get up and speak to other
humans and convince them that this was a
smart decision is the AI enabling you to
do that can you get behind the logic
that was made there do you think sorry
to link on this point because it's a
there's a fascinating one that's a great
goal for AI do you think it's achievable
in many cases or do you okay there's two
possible worlds that we have in the
future one is where AI systems do like
medical diagnosis or things like that
would drive a car without ever
explaining to you why it fails when it
does that's one possible world then
we're okay with it or the other where we
are not okay with it and we really hold
back the technology from getting to good
before it gets able to explain which of
those worlds are more likely do you
think and which are concerning to you or
not I think the reality is it's gonna be
a mix you know I'm not trying a problem
with that I mean I think there are tasks
that perfectly fine with machines show a
certain level of performance and that
level of performance is already better
it is already better than humans so for
example I don't know that I get tape
driverless cars if driverless cars learn
how to be more effective drivers than
humans but can't explain what they're
doing but bottom line statistically
speaking there you know ten times safer
than humans I I don't know that I care I
think when we we have these edge cases
when something bad happens and we want
to decide who's liable for that thing
and who made that mistake in what we do
about that and I think in those those
educators are interesting cases and now
do we go to designers of the AI and the
I says I don't know if that's what it
learned to do and it says well you
didn't train it properly you know you
you were you were negligent in the
training data that you gave that machine
like how do we drive down and realize oh
so I think those are I think those are
interesting questions so the
optimization problem there sorry
is to create a system that's able to
explain the lawyers away there you go um
I think that uh uh I think it's gonna be
interesting I mean I think this is where
technology and social discourse are
gonna get like deeply intertwined and
how we start thinking about problems
decisions and problems like that I think
in other cases it becomes more obvious
where you know it's I got like why did
you decide to give that person you know
a longer sentence or or to deny them
parole again policy decisions or why did
you pick that treatment like that
treatment up killing that guy like why
was that a reasonable choice to make so
so and people are gonna demand
explanations now there's a reality
though here and the reality is that it's
not I'm not sure humans are making
reasonable choices when they do these
things they are using statistical
hunches biases or even systematically
using statistical averages to make
Osmonds is what happened my dad if you
saw that target gave about that but you
know I mean they decided that my father
was brain dead he had went into cardiac
arrest and it took a long time for the
ambulance to get there and wasn't not
resuscitated right away and so forth and
they came they told me he was brain dead
and why was he brain dead because
essentially they gave me a purely
statistical argument under these
conditions with these four features 98%
chance he's brain dead innocent but can
you just tell me not inductively but
deductively go there and tell me his
brain stopped functioning is the way for
you to do that and they and and their
the the protocol in response was no this
is how we make this decision I said this
is adequate for me I understand the
statistics and I don't have you know
there's a two percent chance he's so
like I just don't know the specifics I
need the specifics of this case and I
want the deductive logical argument
about why you actually know he's brained
it so I wouldn't sign that do not
resuscitate and I don't know it was like
they went through lots of procedures as
a big long story but the bottom was a
fascinating story by the way but how I
reasoned and how the doctors reasoned
through this whole process but I don't
know somewhere around 24 hours later or
something he was sitting up
that would zero bushido brain damage any
what lessons do you draw from that story
that experience that the data that
they're you that the data that's being
used to make sophistical inferences
doesn't adequately reflect the
phenomenon so in other words you're
getting shit Ramsar you're getting stuff
wrong because you're your model is not
robust enough and you might be better
off not using statistical inference and
statistical averages in certain cases
when you know the models insufficient
and that you should be reasoning at
about the specific case more logically
and more deductively and hold yourself
responsible to hold yourself accountable
to doing that and perhaps AI has a role
to say the exact thing we just said
which is perhaps this is a case you
should think for yourself you should
reason deductively so it's hard it's
it's so it's hard because it's hard to
know that you know you'd have to go back
and you'd have to have enough data to
essentially say and this goes back to
how do we this goes back to the case of
how do we decide whether the AI is good
enough to do a particular task and
regardless of whether or not it produces
an explanation
so um and and what standards do we hold
right for that so um you know if you
look at you you look more broadly for
example as my father as a metal kick
medical case the medical system
ultimately helped him a lot throughout
his life without it he probably would
have died much sooner
so overall sort of you know work for him
and sort of a net in that kind of way
actually I don't know that's fair um but
it maybe not in that particular case but
overall like oh the medical system
overall that's more given a system
overall you know was doing more more
good than bad now is another argument
that suggests that wasn't the case but
for the for the sake of argument let's
say like that's let's say a net positive
and I think you have to sit there and
there and take
take that into consideration now you
look at a particular use case like for
example making this this decision have
you done enough studies to know how good
that prediction really is right and how
you have you done enough studies to
compare it to say well what if we what
if we dug in and in a more direct you
know let's get the evidence let's let's
do the deductive thing and not use the
statistics here how often would that
have done better right you just so you
have to do this studies to know how good
the AI actually is and it's complicated
because depends how fast you have to
make decision so if you have to make the
decision superfast do you have no choice
right if you have more time right but if
you're ready to pull the plug and this
is a lot of the argument that I had was
a doctor I said what's he gonna do if
you do it what's gonna happen to him in
that room if you do it my way you know
if you do well he's gonna die anyway so
let's do it my way though I mean it
raises questions for our society to
struggle with as was the case with your
father but also when things like race
and gender start coming into play when
when certain when when judgments are
made made based on things that are
complicated in our society at least in
this course and it starts you know I
think I think I'm safe to say that most
of the violent crimes committed by males
so if you discriminate based you know as
a male versus female saying that if it's
a male more likely to commit the crime
so this is one of my my very positive
and optimistic view views of why the
study of artificial intelligence the
process of thinking and reasoning
logically and statistically and how to
combine them is so important for the
discourse today because it's causing a
regardless of what what state AI device
devices are or not
it's causing this dialogue to happen
this is one of the most important
dialogues that in my view the human
species can have right now which is how
to think well yeah how to reason well
how to understand our own cognitive
biases and what to do about them that
has got to be one of the most important
things we as as as a species can be
doing honestly we are reached we've
created an incredibly complex society
we've created amazing abilities to
amplify noise faster than we can play
amplifies signal we are challenged
we are deeply deeply challenged we have
you know big segments of the population
getting hit with enormous amounts of
information do they know how to do
critical thinking do they know how to
objectively objectively reason do they
understand what they are doing nevermind
with their AI is doing this is such an
important dialogue you know to be having
and and and you know we are
fundamentally are thinking can be and
easily becomes fundamentally bias and
there are statistics and we shouldn't
blind our so we shouldn't discard
statistical inference but we should
understand the nature of such this
conference as us as a society as you
know we decided to reject statistical
inference to favor individual
understanding and and deciding on the
individual yes we we consciously make
that choice so even if the statistics
said even if the Cystic said males are
more likely to have you know to be
violent criminals we still take each
person as an individual and we treat
them based on the logic and the
knowledge of that situation we
purposefully and intentionally reject
the statistical
once we do that at a respect for the
individual for the individual yeah and
then that requires reasoning and
cracking looking forward what Grand
Challenges would you like to see in the
future because the the Jeopardy
challenge you know captivated the world
alpha go alpha zero cap day of the world
deep blue certainly beating Kasparov
Gary's bitterness aside and captivated
the world what do you think do you have
ideas for next grand challenges for
future challenges of that oh you know I
look I mean I think there are lots of
really great ideas for Grand Challenges
I'm particularly focused on one right
now which is Kent you know can you
demonstrate that they understand that
they could read and understand that they
can they can acquire these frameworks
and communicate you know reason and
communicate with humans so it is kind of
like the Turing test but it's a little
bit more demanding than the Turing test
it's not enough it's not enough to
convince me that you might be human
because you could you know you can
parrot a conversation I think you know
the the this standard is a little bit
higher is for example can you you know
the santa is higher and I think one of
the challenges of devising this grand
challenge is that we're not sure what
intelligence is we're not sure how to
determine whether or not two people
actually understand each other and then
what depth they understand it they you
know and what to what depth they
understand each other so the challenge
becomes something along the lines of can
you satisfy me that we have a shared
understanding so if I were to probe and
probe and you probe me can can can can
machines really act like thought
partners where they could satisfy me
that they that we have a share our
understanding is shared enough that we
can collaborate and produce the answers
together and that you know they they can
help me explain and justify those
answers so maybe here's an idea so we'll
have a Isis
run for president and convinced that's
too easy
from sorry oh no you have to convince
the voters that they should vote for it
so they s what I would again again I
that's why I think this is such a
challenge because we go back to the
emotional persuasion we go back to you
know now we're checking off an aspect of
human cognition that is in many ways
weak or flawed right we're so easily
manipulated our minds are drawn for
often the wrong reasons right not the
reasons that ultimately matter to us but
the reasons that can easily persuade us
I think we can be persuaded to believe
one thing or another
for reasons that ultimately don't serve
us well in the long term and a good
benchmark should not play with those
elements of emotional manipulation I
don't think so I think that's where we
have to set the set the higher standard
for ourselves of what you know what does
it mean this goes back to rationality
and it goes back to objective thinking
and can you produce can you acquire
information and produce reasoned
arguments and to those reasons arguments
pass a certain amount of muster and is
it and can you acquire new knowledge you
know can you can you under can you
reason oh I have acquired new knowledge
can you identify where it's consistent
or contradictory with other things
you've learned and can you explain that
to me and get me to understand that so I
think another way to think about it
perhaps is kind of machine teach you can
the hell
really nice less than that's where to
put it can you understand something that
you didn't really understand before
where's where is you know it's taking it
so you're not you know again it's almost
like can it can it teach you can it help
you learn and and in an arbitrary space
so it can open those domain space so can
you tell the Machine and again this
borrows from some science fiction's abut
can you go off and learn about this
topic that I'd like to understand better
and then work with me to help me
understand it that's quite brilliant
what the machine that passes that kind
of test do you think it would need to
have self-awareness or even
consciousness what do you think about
consciousness and the importance of it
maybe in relation to having a body
having a presence an entity do you think
that's important you know people used to
ask if Watson was conscious and I used
to think and he said he's the conscious
of what exactly I mean I think you know
main cell it depends what it is that
you're conscious I mean like so you know
did it if you you know it's certainly
easy for it to answer questions about it
would be trivial to program it so the
answer questions about whether or not it
was playing jeopardy I mean it could
certainly answer questions that will
imply that it was aware of things
exactly what does it mean to be aware
and what does it mean to conscious and
it's sort of interesting I mean I think
that we differ from one another based on
what we're conscious of but wait wait
for sure there's degrees of
consciousness in there so it well in
those areas like it's not just agrees
what do you what do you what are you
aware of like what are you not aware but
nevertheless there's a very subjective
element to our experience let me even
not talk about consciousness let me talk
about another to me really interesting
topic immortality fear or mortality
Watson as far as I could tell did not
have a fear of death certainly not most
most humans do wasn't conscious of death
it wasn't that so there's an element of
finiteness to our existence that I think
like we like I mentioned survival that
adds to the whole thing that I mean
consciousness is tied up with that that
we are us thing it's a subjective thing
that ends and that seems to add a color
and flavor to our motivations in a way
that seems to be fundamentally important
for intelligence or at least the kind of
human intelligence well I take for
generating goals again I think you could
have you could have an intelligence
capability and a capability to learn I
capability to predict but I think
without I mean again you get a fear but
essentially without the goal to survive
so you think you can just encode that
without having to million code I mean
can you create a robot now and you could
say you know and plug it in and say
protect your power source you know and
give it some capabilities and we'll sit
there and operate to try to protect this
power source and survive I mean I so I
don't know that that's false awfully a
hard thing to demonstrate it sounds like
a fairly easy thing to demonstrate that
you can give it that goal we'll come up
with that goal by itself as you have to
program that goal in but there's
something because I think as we touched
on intelligence is kind of like a social
construct the the fact that a robot will
be protecting its power source
would would add depth and grounding to
its intelligence in terms of us being
able to respect I mean ultimately it
boils down to us acknowledging that it's
intelligent and the fact that it can die
I think is an important part of that the
interesting thing to reflect on is how
trivial that would be and and I don't
think if you knew how trivial that was
you would associate that with being
intelligence I mean I literally put in a
statement of code that says you know you
have the following actions you can take
you give it a bunch of actions like you
mount a laser gun on her or you may do
you the ability to scream a screech or
whatever and you know and you you say
you know if you see your power source
then you could program that in and you
know you're gonna print it you're gonna
take these actions to protect it you
know you teach it checking it on a bunch
of things so and and now you're gonna
look at that and you say well you know
that's intelligence because it's
protecting power source maybe but that's
again at this human bias that says the
thing I had then I identify my
intelligence and my conscious so
fundamentally with the desire or at
least the behaviors associated with the
desire to survive that if I see another
thing doing that I'm going to assume
it's intelligent
what timeline year will society have a
something that would that you would be
comfortable calling an artificial
general intelligence system well what's
your intuition nobody can predict the
future
certainly not next few months or twenty
years away but what's your intuition how
far away are we
I the ideas hearts make these
predictions and I would be you know I
would be guessing and there's so many
different variables including just how
much we want to invest in it and how
important it you know and how important
we think it is
what kind of investment are willing to
make in it what kind of talent we end up
bringing to the table all you know the
incentive structure all these things so
I think it is possible to do this sort
of thing I think it's I think trying to
sort of ignore many of the variables and
things like that
is it a ten-year thing as a 23 it's
probably closer to a 20-year thing I
guess but not as little no I don't think
it's several hundred years I don't think
it's several hundred years but again so
much depends on how committed we are to
investing and incentivizing this type of
work this type of work and it's sort of
interesting like I don't think it's
obvious
how incentivize we are I think from a
task perspective you know if we see
business opportunities to take this
technique is a technique to solve that
problem I think that's the main driver
for many from any of these things from a
from a general Tosta seems kind of an
interesting question are we really
motivated to do that and and like we
just struggled ourselves right now to
even define what it is so it's hard to
incentivize when we don't even know what
it is we're incentivized to create and
if you said mimic a human intelligence
I just think there are so many
challenges with the the significance and
meaning of that there's not a clear
directive there's no clear directive to
do precisely that thing so assistance in
a larger and larger number of tasks so
being able to a system that's
particularly able to operate my
microwave and making a grilled cheese
sandwich I don't even know how to make
one of those and then the same system
would be doing the vacuum cleaning and
then the same system would be teaching
my kids that I don't have math I think
that when when when you get into a
general intelligence for learning
physical tasks and again yeah I want to
go back to your body questions it's on
your body question was interesting but
you want to go back to you know learning
abilities do physical tasks you might
have we might get Majan in that
timeframe we will get better and better
at learning these kinds of tasks whether
it's mowing your lawn or driving a car
or whatever it is I think we will get
better and better at that where it's
learning how to make predictions over
large bodies of data as if we're going
to continue to get better and better at
that and machines will out you know
outpace humans and and a variety of
those things the underlying mechanisms
for doing that may be the same meaning
that you know maybe these are deep Nats
there's infrastructure to train them
reusable components to get them to
different classes of tasks and we get
better and better at building these
kinds of machines you could see argue
that the general learning infrastructure
in there is a form of a general type of
intelligence I think what starts getting
harder is this notion of you know can we
can we effectively communicate and
understand and build that shared
understanding because of the layers of
interpretation that are required to do
that and the need for the machine to be
engaged with humans at that level at a
continuous basis so how do you get in
how do you get the machine in the game
how do you get the machine in the
intellectual game yeah and to solve AGI
you probably have to solve that problem
you have to get the machine so it's a
little bit of a bootstrapping can we get
the machine engaged and
you know in the intellectual calling a
game but in the intellectual dialogue
with the humans are the humans
sufficiently an intellectual dialogue
with each other to generate enough to
generate enough data in this context and
how do you bootstrap that because every
one of those conversations every one of
those conversations those intelligent
interactions require so much prior
knowledge that is a challenge to
bootstrap it so that's so as so the
question is and how committed so I think
that's possible but when I go back to
are we incentivized to do that I know
we're incentivized to do the former are
we incentivize to do the latter
significantly enough to people
understand what the latter really is
well enough part of the elemental
cognition mission is to try to
articulate that better and better you
know through demonstrations and to
trying to craft these grand challenges
and get people to say look this is a
class of intelligence this is a class of
AI do we do we want this what what is
the potential of this what are the
business what's the business potential
what's the societal potential to that
and so you know and to build up that
incentive system around that yeah I
think if people don't understand yet I
think they will and is a huge business
potential here so it's exciting that
you're working on it you've kind of
skipped over but I'm a huge fan of
physical presence of things do you think
you know Watson head of body do you
think having a body as to the
interactive element between the AI
system and a human or just in general to
intelligence so I think I think going
back to that shared understanding bit
humans are very connected to their
bodies I mean is one of the reasons one
of the challenges in getting an AI to
kind of be a compatible human
intelligence is that our physical bodies
are generating a lot of features that
make up the input so in other words
where our bodies are are the the tool we
use to affect output but they're also
but they also generate a lot of input
for our brains so we generate emotion we
generate all these feelings we
generate all these signals that machines
don't have so missions that have this as
the input data and they don't have the
the feedback that says okay I've gotten
this I've gotten this emotion or I've
gotten this idea I now want to process
that and then I can it then affects me
as a physical being and then I and I and
I can play that out in other words I
could realize the implications of tax
implications again on my bond mind body
complex I then process that and the
implications again are internal features
are generated I learned from them they
have an effect on my mind body complex
so it's interesting when we think do we
want a human intelligence well if we
want a human compatible intelligence
probably the best thing to do is to
embed it embedded in a human body
just to clarify and both concepts
beautiful is humanoid robots so robots
that look like humans is one or did you
mean actually sort of what Hamas was
working with neural link really
embedding intelligence systems that the
ride-alongs human bodies know I was
riding along is different I meant like
if you want to create an intelligence
that is human compatible meaning that it
can learn and develop a shared
understanding of the world around it you
have to give it a lot of the same
substrate part of that substrate you
know is the idea that it generates these
kinds of internal features like sort of
emotional stuff it has similar senses it
has to do a lot of the same things with
those same sentences um right so I think
if you want that again I don't know that
you want that like man like that's not
my specific goal I think that's a
fascinating scientific goal I think it
has all kinds of other implications
that's sort of not to go like I want it
I want to create I think of it as I
create intellectual thought martyrs for
humans so that kind of that kind of
intelligence
I know other companies that are creating
physical thought partners the fiscal
partners to figure out for you but
that's kind of not where we're you know
I'm at but but but the the important
point is that a big part of how of what
we process is that physical experience
of the world around us on the point of
thought partners
what role does an emotional connection
or forgive me love have to play in that
thought partnership is that something
you're interested in put another way
sort of having a deep connection beyond
intellectual with the AI yeah with the a
between human and ass
is that something that gets in the way
of the the rational discourse is there
something that's useful I worry about
biases you know obviously so in other
words if you develop an emotional
relationship with the machines do all of
a sudden you start are more likely to
believe what it's saying even if it
doesn't make any sense
so I you know I worry about that but at
the same time I think the opportunity to
use machines to provide human
companionship is actually not crazy and
it's again the intellectual and social
companionship is not crazy the idea do
you have concerns as a few people do you
know Musk Sam Harris about long-term
existential threats of AI and perhaps
short-term threats of AI we talked about
bias we talked about different misuses
but do you have concerns about thought
partners systems that are able to help
us make decisions together humans
somehow having a significant negative
impact on society in the long term I
think there aren't things to worry about
I think the giving machines too much
leverage is a problem and what I mean by
leverage is is too much control for
things that can hurt us whether it's
socially psychological intellectually or
physically and if you give them machines
too much control I think that's a
concern you forget about the AI just
when you give them too much control
human bad actors can hack them and
produce havoc so um you know that's a
problem and you imagine hackers taking
over the driverless car Network and you
know creating all kinds of havoc but you
could also imagine given given the ease
at which humans could be persuaded one
way or the other and now we have
algorithms that can easily take control
over over that over that and amplify
noise and move people one direction or
another I mean humans do that to other
humans all the time and we have
marketing campaigns we have political
campaigns that take it to image of our
our emotions or our fears and this is
done all the time when but with machines
machines are like giant mecha phones
right we can amplify this and orders of
magnitude and can fine-tune its control
so we can tailor the message we can now
very rapidly and efficiently tailor the
message to the audience taking taking
advantage of you know of their biases
and amplifying them and using them to
pursue a them in one direction or
another in ways that are not fair not
logical not objective not meaningful and
humans the machines and power that so so
that's what I mean by leverage like it's
not new but wow it's powerful because
machines can do it more effectively more
more you know more quickly and we see
that already going on and and and social
media not the plays and other places
that's scary and and that's why like I'm
I'm that's why I go back to saying one
of the most important public dialogues
we could be having is about the nature
of intelligence and the nature of
inference and logic and reason and
rationality and us understanding our own
biases us understanding our own
cognitive biases and how they work and
then how machines work and how do we use
them to complement and sit basically so
that in the end we have a stronger
overall system that's just incredibly
important I don't
most people understand that so so like
telling telling your kids or telling
your students this goes back to the
cognition here's how your brain works
here's how easy it is to trick your
brain right there are fundamental
cognitive but you should appreciate the
different the different types of
thinking and how they work and what
you're prone to and you know and what
and what do you prefer
and under what conditions does this make
sense versus that makes sense and then
say here's what AI can do here's how it
can make this worse and here's how it
can make this better and then that's
where the as a role is to reveal that
then the that trade-off so if you
imagine a system that is able to beyond
any definition of the Turing test of the
benchmark really an AGI system as a
thought partner that you one day will
create what question what topic of
discussion if you get to pick one would
you have with that system what would you
ask and you get to find out the truth
together so you threw me a little bit
with finding the truth at the end but
this is a whole nother topic but the I
think the beauty of it I think what
excites me is the beauty of it is if I
really have that system I don't have to
pick so in other words I can you know I
can go to and say this is where I care
about today and and and that's what we
mean by like this general capability go
out read this stuff in the next three
milliseconds and I want to talk to you
about it I want to draw analogies I want
to understand how this affects this
decision or that decision what if this
were true what if that were true what
what knowledge should I be aware of that
could impact my decision
here's what I'm thinking is the main
implication can you find can you prove
that out can you give me the evidence
that supports that can you give me
evidence supports this oh there's a boy
that would that be incredible you
would that be just incredible just a
long discourse just to be part of
whether it's a medical diagnosis or
whether it's you know the various
treatment options or whether it's a
legal case or whether it's a social
problem that people are discussing like
be part of the dialogue one that holds
itself and us accountable to reasons an
objective dialogue you know I just I get
goosebumps talking about it right so
when when you create it please come back
on the podcast well the discussion
together and make it even longer this is
a record for the longest conversation
now there's an honor it was a pleasure
David thank you so much for thanks so
much a lot of fun
you