Episode Transcript
Transcripts are displayed as originally observed. Some content, including advertisements may have changed.
Use Ctrl + F to search
0:02
Hello, Hello, Welcome to Smart Talks with
0:04
IBM, a podcast from Pushkin
0:06
Industries, iHeartRadio and
0:09
IBM. I'm Malcolm Gabwell. This
0:11
season, we're continuing our conversation with
0:14
new creators visionaries
0:16
who are creatively applying technology
0:18
in business to drive change, but
0:20
with a focus on the transformative
0:22
power of artificial intelligence and
0:25
what it means to leverage AI
0:27
as a game changing multiplier for your
0:29
business. Our guest today
0:32
is doctor David Cox, VP
0:34
of AI Models at IBM
0:36
Research and IBM Director
0:38
of the MIT IBM Watson
0:41
AI Lab, a first of its kind
0:43
industry academic collaboration between
0:46
IBM and MIT focused
0:48
on the fundamental research of artificial
0:51
intelligence. Over the course
0:53
of decades, David Cox watched
0:56
as the AI revolution steadily
0:58
grew from the sim ideas of
1:00
a few academics and technologists
1:02
into the industrial boom we are experiencing
1:05
today. Having dedicated
1:07
his life to pushing the field of AI
1:09
towards new horizons, David has
1:11
both contributed to and presided
1:14
over many of the major breakthroughs
1:17
in artificial intelligence. In
1:19
today's episode, you'll hear David
1:21
explain some of the conceptual
1:23
underpinnings of the current AI
1:26
landscape, things like foundation models
1:28
in surprisingly comprehensible terms,
1:30
I might add, we'll also get into some
1:33
of the amazing practical applications
1:35
for AI in business, as well as what implications
1:37
AI will have for the future of work
1:40
and design. David spoke with Jacob
1:42
Goldstein, host of the Pushkin podcast
1:45
What's Your Problem. A veteran
1:47
business journalist, Jacob has reported
1:49
for The Wall Street Journal, the Miami Herald,
1:52
and was a longtime host of the NPR
1:54
program Planet Money.
1:57
Okay, let's get to the interview.
2:05
Tell me about your job at IBM.
2:08
SO. I wear two hats at IBM. SO
2:10
one, I'm the IBM Doctor of the MI t IBM
2:13
Watson AI Lab. So that's a joint
2:16
lab between IBM and MIT where we
2:18
try and invent what's next in AI. It's been
2:20
running for about five years, and then
2:22
more recently I started as the vice president
2:24
for AI Models, and I'm in charge
2:27
of building IBM's foundation
2:29
models, you know, building these these
2:31
big models, generative models that allow us to have all
2:33
kinds of new exciting capabilities in AI.
2:36
So, so I want to talk to you a lot about
2:38
foundation models, about genitive AI.
2:40
But before we get to that. Let's just spend a minute on
2:43
the on the IBM MI
2:45
T collaboration. Where
2:47
where did that partnership start? How did it originate?
2:51
Yeah, so, actually it turns out that MI T
2:53
and IBM have been collaborating for
2:56
a very long time in the area of AI. In
2:58
fact, the term artificial
3:00
intelligence was coined in a nineteen
3:03
fifty six workshop that was held
3:05
at Dartmouth. It was actually organized by an IBM
3:07
or Nathaniel Rochester, who led
3:09
the development of the IBM seven and one. So
3:12
we've really been together in AIS since
3:14
the beginning, and as
3:16
AI kept accelerating more and
3:18
more and more, I think
3:20
there was a really interesting decision to say,
3:23
let's make this a formal partnership. So IBM
3:25
in twenty seventeen and now, so it'd be committing close to a
3:27
quarter billion dollars over ten years to
3:30
have this joint lab with MIT, and
3:32
we located ourselves right on the campus
3:35
and we've been developing very very deep relationships
3:37
where we can really get to know each other, work
3:39
shoulder to shoulder, conceiving
3:41
what we should work on next, and then executing the projects.
3:44
And it's really very
3:46
few entities like this exist
3:49
between academia industry. It's been really
3:51
fun the last five years to be a part of
3:53
it.
3:53
And what do you think are some of the most important
3:55
outcomes of this collaboration between
3:58
IBM and MIT.
4:00
Yeah, so we're really kind
4:02
of the tip of the sphere for for IBM's
4:05
AI strategy. So we're we're really
4:07
looking what, you know, what's coming ahead, and
4:10
you know, in areas like Foundation models, you know, as
4:12
the field changes and I T
4:15
people are interested in working on you know, faculty,
4:17
students and staff are interested in working on what's the latest
4:19
thing, what's the next thing. We at IBM Research very
4:22
much interested in the same so we can kind
4:24
of put out feelers, you know, interesting things
4:27
that we're seeing in our research, interesting
4:29
things we're hearing in the field. We can go and chase those opportunities.
4:32
So when something big comes, like the big
4:34
change that's been happening lately with Foundation
4:36
Models, we're ready to jump on it. That's
4:38
really the purpose, that's that's the lab functioning
4:41
the way it should. We're also really interested
4:43
in how do we advance you
4:45
know, AI that can help with climate change
4:48
or you know, build better materials
4:50
and all these kinds of things. That are you know, a broader
4:52
aperture sometimes than than what we might
4:55
consider just looking at the product portfolio
4:57
of IBM, and that that gives us again a
4:59
breadth where we can connections that we might
5:01
not have seen otherwise. We can, you
5:03
know, think things that help out society and
5:05
also help out our customers.
5:08
So the last whatever
5:10
six months, say, there has been this
5:14
wild rise in the
5:16
public's interest in AI, right clearly
5:18
coming out of these generative
5:20
AI models that are really accessible, you know,
5:23
certainly chat GPT language
5:25
models like that, as well as models that generate images
5:28
like mid Journey. I mean, can
5:30
you just sort of briefly talk about
5:32
the breakthroughs in AI
5:34
that have made this moment feel so exciting,
5:37
so revolutionary for artificial intelligence.
5:41
Yeah. You know, I've been studying
5:44
AI basically my entire adult
5:46
life. Before I came to IABM, I was a professor
5:48
at Harvard. I've been doing this a long time,
5:51
and I've gotten used to being surprised. It sounds
5:53
like a joke, but it's serious, Like I'm
5:55
getting used to being surprised at the acceleration
5:58
of the pace. Again. It tracks
6:00
actually a long way back. You know, there's
6:03
lots of things where there was an idea that
6:05
just simmered for a really
6:07
long time. Some of the key
6:09
math behind the
6:12
stuff that we have today, which is amazing. There's
6:14
an algorithm called backpropagation, which
6:17
is sort of key to training neural networks that's
6:19
been around, you know, since the eighties in
6:21
wide use. And really
6:23
what happened was it simmered for a
6:25
long time, and then enough
6:28
data and enough compute came. So
6:30
we had enough data because you
6:33
know, we all started carrying multiple
6:35
cameras around with us. Our mobile phones have
6:37
all, you know, all these cameras and this we
6:39
put everything on the Internet, and there's all this data
6:42
out there. We called a lucky break that there
6:44
was something called a graphics processing unit, which
6:46
you know, turns out to be really useful for doing these kinds
6:48
of algorithms, maybe even more useful than
6:50
it is for doing graphics. They're great graphics too.
6:53
And things just kept kind
6:55
of adding to the snowball. So we had
6:57
deep learning, which is sort of a
7:00
a rebrand of neural networks
7:02
that I mentioned from the eighties, and that was
7:04
enabled again by data because we digitalized
7:07
the world and compute because because we
7:09
kept building faster and faster and more powerful computers,
7:12
and then that allowed us to make this big
7:14
breakthrough. And then, you know, more
7:16
recently, using the same building
7:19
blocks, that inexorable rise
7:21
of more and more and more data met
7:24
the technology called self supervised
7:26
learning, where the key
7:29
difference there in traditional
7:31
deep learning, you know, for classifying images,
7:33
you know, like is this a cat or is this a dog? And
7:35
a picture those technologies
7:38
require supervision, so you have to take
7:41
what you have and then you have to label it. So you have to take
7:43
a picture of a cat, and then you label it as a cat,
7:46
and it turns out that you know, that's very
7:48
powerful, but it takes a lot of time to label
7:51
gats and to label dogs, and there's
7:53
only so many labels that us in the world. So
7:55
what really changed more recently is
7:58
that we have self supervised learning where you don't
8:00
have to have the labels. We can just take unannotated
8:02
data. And what that does is it lots you use
8:05
even more data. And that's
8:07
really what drove this latest
8:10
sort of rage. And then and then all of a sudden
8:12
we start getting these these really powerful
8:14
models. And then really this
8:16
has been simmering technologies,
8:19
right, this has been happening
8:21
for a while and progressively
8:23
getting more and more powerful. One of
8:26
the things that really happened with
8:28
CHATGBT and technologies like
8:30
stable Diffusion and mid Journey was
8:33
that they made it visible to the public.
8:36
You know, you put it out there. The public can touch
8:38
and feel and they're like, wow, not only is there
8:40
palpable change and wow this you
8:43
know, I can talk to this thing. Wow, this thing can generate
8:45
an image. Not only that, but everyone
8:47
can touch and feel and try. My
8:49
kids can use some
8:51
of these AI art generation technologies.
8:54
And that's really just launched.
8:57
You know. It's like a propelled slingshot
8:59
at a into a different regime.
9:01
In terms of the public awareness of these technologies.
9:04
You mentioned earlier in the conversation foundation
9:07
models, and I want to talk a little bit about that.
9:09
I mean, can you just tell me, you
9:11
know, what are foundation models
9:13
for AI and why are they a big
9:15
deal?
9:17
Yeah, So this term foundation model
9:19
was coined by a group at Stanford,
9:22
and I think it's actually a really
9:24
apt term because remember
9:26
I said, you know, one of the big things
9:28
that unlocked this latest excitement was
9:31
the fact that we could use large amounts of unannotated
9:34
data. We could train a model. We don't have
9:37
to go through the painful effort of labeling
9:39
each and every example. You still
9:41
need to have your model do something you wanted to
9:43
do. You still need to tell it what you want
9:46
to do. You can't just have a model that doesn't
9:48
have any purpose. But what a foundation models
9:50
that provides a foundation, Like
9:52
a literal foundation, you can sort of stand
9:54
on the shoulders of giants. You can have them these massively
9:57
trained models, and then do a little bit
9:59
on top. You know, you could use just a few
10:01
examples of what you're looking for and
10:04
you can get what you want from the model. So
10:07
just a little bit on top. Now it gets to the
10:09
results that a huge amount of effort used to have
10:11
to put in, you know, to get from the ground
10:13
up to that level.
10:15
I was trying to think of
10:17
of an analogy for sort
10:19
of foundation models versus what came
10:21
before, and I don't know that I came up with a
10:24
good one, But the best I could do was this. I
10:26
want you to tell me if it's plausible. It's
10:29
like before foundation models, it
10:31
was like you had these sort of single
10:33
use kitchen appliances. You could make a
10:35
waffle iron if you wanted waffles, or you could
10:38
make a toaster if you wanted to make toast.
10:40
But a foundation model is like like an oven
10:43
with a range on top. So it's like this machine
10:45
and you could just cook anything with this
10:47
machine.
10:48
Yeah, that's a great analogy. They're
10:51
very versatile. The other
10:53
piece of it, too, is that they dramatically
10:55
lower the effort that it takes
10:57
to do something that you want to do. And
11:00
I used to say about the old world
11:02
of AI, would say, you know, the problem with automation
11:05
is that it's too labor intensive, which
11:07
sounds like I'm making a joke.
11:09
Indeed, famously, if automation does
11:11
one thing, it substitutes machines
11:14
or computing power for labor. Right,
11:16
So what does that mean to say AI
11:18
is or automation is too labor
11:21
intensive.
11:22
It sounds like I'm making a joke, but I'm actually serious. And
11:24
what I mean is that the effort it took
11:27
the old regime to automate something was
11:29
very, very high. So if
11:31
I need to go and curate
11:33
all this data, collect all this data, and then carefully
11:36
label all these examples, that labeling
11:39
itself might be incredibly expensive
11:41
and time, and we estimate anywhere between
11:43
eighty to ninety percent of the effort it
11:46
takes to feel an AI solution actually
11:48
is just spent on data, so
11:50
that that has some consequences, which
11:52
is the threshold for
11:55
bothering. You know, if you're going to
11:57
only get a little bit of value back from
12:00
something, are you going to go through this huge effort
12:02
to curate all this data? And then
12:05
when it comes time to train the model, you need highly
12:07
skilled people defensive
12:09
or hard to find in the labor market.
12:12
You know, are you really going to do something that's just a tiny, little
12:14
incremental thing. Now you're going to do the only
12:16
the highest value things that warrn't right
12:19
level because you.
12:20
Have to essentially build the whole machine
12:23
from scratch, and there aren't many
12:25
things where it's worth that much work to build
12:27
a machine that's only going to do one narrow
12:29
thing that's.
12:31
Right, and then you tackle the next
12:33
problem and you basically have to start over.
12:35
And you know, there are some nuances here, like
12:37
for images, you can pre train a model
12:39
on some other task and change it around. So
12:41
there are some examples of this, like non
12:44
recurring cost that we have in
12:46
the old world too, but by and large, it's just
12:48
a lot of effort. It's hard. It
12:50
takes you know, a large level of
12:52
skill to implement. One
12:55
analogy that I like is, you know,
12:57
think about it as you know, you have a river of
12:59
data, you know, running through your company or
13:01
your institution. Traditional AI
13:04
solutions are kind of like building a dam on
13:06
that river. You know, Dams are very
13:08
expensive things to build. They require
13:10
highly specialized skills and
13:12
lots of planning. And you know, you're
13:15
only going to put a dam on a river
13:17
that's big enough that you're gonna get
13:19
enough energy out of it that it was worth your trouble.
13:21
You're gonna get a lot of value out of that dam. If you have
13:23
a river like that, you know, a river of data, but
13:26
it's actually the vast majority
13:28
of the water you know, in your kingdom
13:30
actually isn't in that river. It's in
13:33
puddles and greeks and vallet bricks.
13:35
And you know, there's a lot of
13:38
value left on the table because it's like, well,
13:40
I can't there's nothing you can do about it. It's just
13:42
that that's too low value,
13:45
so it takes too much effort, so
13:47
I'm just not going to do it. The return on investment just
13:49
isn't there so you just end up not automating
13:51
things. It's too much of a pain. Now
13:54
what foundation models do is they say, well,
13:56
actually, no, we can train a
13:58
base model a foundation they can work on, don't
14:01
We don't care. We not specify what the task is ahead
14:03
of time. We just need to learn about the domain
14:05
of data. So if we want to build something
14:08
that can understand English language,
14:10
there's a ton of English language text available
14:13
out in the world. We can now train models
14:15
on huge quantities of it,
14:18
and then it learned the structure, learned
14:21
how language you know, good part of how
14:23
language works on all that unlabeled data,
14:25
and then when you roll up with your task, you
14:27
know, I want to solve this particular
14:29
problem. You don't have to start
14:31
from scratch. You're starting from a very, very
14:34
very high place. So that just
14:36
gives you the ability to just you know, now,
14:38
all of a sudden, everything is accessible.
14:40
All the puddles and greeks and babbling books
14:42
and gelipons, you know, those are all accessible
14:46
now. And that's that's very exciting.
14:48
But it just changes the equation on what kinds of
14:50
problems you could use AI to solve.
14:52
And so foundation models basically
14:55
mean that automating
14:57
some new task is much less labor
14:59
and intensive. The sort of marginal effort
15:02
to do some new automation thing is much
15:04
lower because you're building on top of the foundation
15:06
model rather than starting from scratch.
15:09
Absolutely, So that is that
15:12
is like the exciting good
15:14
news. I do feel like there's a
15:16
little bit of a countervailing idea that's worth
15:19
talking about here, and that is the idea that even
15:21
though there are these foundation models
15:24
that are really powerful, that are relatively
15:26
easy to build on top of, it's still
15:28
the case right that there is not some one
15:31
size fits all foundation model. So
15:34
you know, what does that mean and why is that
15:37
important to think about.
15:38
In this context? Yeah, so
15:41
we believe very strongly that there isn't
15:43
just one model to rule
15:45
them all. There's a number of reasons why that could
15:47
be true. One which I think is
15:49
important and very relevant today is
15:52
how much energy these
15:54
models can consume. So these
15:56
models, you know, can get
15:59
very very large. So one
16:01
thing that we're starting
16:04
to see or starting to believe, is that you
16:06
probably shouldn't use one giant
16:08
sledgehammer model to solve every
16:10
single problem, you know, like we should pick
16:13
the right size model to solve the problem. We shouldn't
16:15
necessarily assume that we need the biggest,
16:18
baddest model for every little
16:21
use case. And we're also seeing that, you
16:23
know, small models that are trained like to
16:25
specialize on particular domains can
16:28
actually outperform much bigger models. So bigger
16:30
isn't always even better.
16:32
So they're more efficient and they do
16:34
the thing you want them to do better as well.
16:37
That's right. So Stanford, for
16:39
instance, a group of Stanford trained a model. It
16:42
is a two point seven billion parameter model,
16:44
which isn't terribly big by today's standards.
16:46
They trained it just on the biomedical literature,
16:48
you know, this is the kind of thing that universities do.
16:51
And what they showed was that this model
16:54
was better at answering questions about the biomedical
16:56
literature than some models that are one
16:58
hundred billion parameters you any times
17:00
larger. So it's a little bit like
17:03
you know, asking an expert for help
17:05
on something versus asking the smartest
17:07
person you know, the smartest person you
17:09
know, maybe very smart, but they're not
17:11
going to be expertise. And then
17:14
as an added bonus, you know, this is now a much smaller
17:16
model, it's much more efficient to run. We are
17:18
you know, you know, it's cheaper, so
17:21
there's lots of different advantages there. So I
17:23
think we're going to see attention in
17:26
the industry between vendors
17:29
that say, hey, this is the one, you know, big model,
17:31
and then others that say, well, actually, you know,
17:33
there's there's you know, lots of different
17:35
tools we can use that all have this nice quality
17:37
that we outligned at the beginning, and
17:39
then we should really pick the one that makes the most sense
17:41
for the task at hand.
17:44
So there's sustainability basically
17:46
efficiency. Another kind of set
17:48
of issues that come up a lot with ai A
17:50
are bias hallucination. Can
17:53
you talk a little bit about bias and hallucination,
17:56
what they are and how you're working to mitigate
17:58
those problems.
17:59
Yeah, so there are lots of issues
18:01
still as amazing as these technologies are, and
18:03
they are amazing, let's be very clear,
18:06
lots of great things we're going to enable with these
18:08
kinds of technologies. Bias isn't
18:10
a new problem, so you know,
18:13
basically we've seen this
18:15
since the beginning of AI. If you train a model
18:18
on data that has a bias in it,
18:21
the model is going to recapitulate that bias
18:23
when it provides its answers. So every
18:26
time, you know, if all the text you have says,
18:29
you know, it's more likely to refer to female nurses
18:31
and male scientists, then you're going to
18:33
get models that you know. For instance, there was
18:35
an example where a machine learning
18:37
based translation system translated from Hungarian
18:40
to English. Hungarian doesn't
18:42
have gendered pronouns. English does,
18:44
and when you ask them to translate, it would translate
18:46
they are a nurse to she is a nurse,
18:49
would translate they are a scientist to he
18:51
is a scientist. And that's not because the people
18:54
who wrote the algorithm were building in bias
18:56
and coding in like oh, it's got to be this way. It's
18:58
because the data was like that. You know, we
19:01
have biases in our society and
19:03
they're reflected in our data
19:05
and our text and our images everywhere.
19:08
And then the models they're just mapping
19:11
from what they've what they've seen in their training data to
19:13
to the result that you're trying to get them to do and
19:16
to give, and then these biases
19:18
come out. So there's a very
19:20
active program of research
19:23
in you know, we we do quite a bit at IBM
19:25
research and I, but also
19:28
all over the community and industry and academia
19:30
trying to figure out how do we explicitly
19:33
remove these biases, how do we identify them,
19:35
how do you know, how do we build tools that allow
19:37
people to audit their systems to make sure they aren't
19:40
biased. So this is a really important
19:42
thing. And you know, again this was here since
19:45
the beginning, you know, of machine
19:47
learning and AI, but foundation
19:49
models and large language models and generative AI
19:53
just bring it into sharper even sharper focus
19:55
because there's just so much data and it's sort
19:57
of building in baking and all
19:59
these different biases we have. So
20:01
that's that's absolutely a
20:03
problem that these models have. Another
20:06
one that you mentioned was hallucinations. So
20:08
even the most impressive of our models
20:11
will often just make
20:13
stuff up. You know, the technical term
20:15
that the heels chosen is hallucination.
20:18
To give you an example, I asked chat
20:20
tbt to create a biography
20:22
of David Cox IBM, and
20:25
you know, it started off really well, you
20:27
know, the identifying that I was the director of the mt
20:29
IBM Watsonday and said a few words about
20:31
that, and then it proceeded to create
20:33
an authoritative but completely fake
20:36
biography of me where I was British.
20:39
I was born in the UK. I
20:41
went to British university, you know, universities
20:44
in the UK. I was professor in the authority.
20:46
Right, it's the certainty that that is
20:48
weird about it, Right, it's it's dead certain
20:51
that you're from the UK, et cetera.
20:53
Absolutely, yeah, it has all kinds of flourishes,
20:56
like I want awards in the UK. So
20:58
yeah, it's it's problematic
21:01
because it kind of pokes at a lot of
21:03
weak spots in our human psychology,
21:06
where if something sounds coherent,
21:09
we're likely to assume it's true. We're
21:11
not used to interacting with people who eloquently
21:13
and authoritatively, you know, admit
21:16
complete nonsense like yeah, you
21:18
know, you know we can debate about that, but yeah, we.
21:20
Can debate about that. But yes, the
21:23
sort of blive confidence throws
21:25
you off when you realize it's completely wrong.
21:28
Right, that's right. And we do
21:30
have a little bit of like a great and powerful
21:32
aws sort of vibe going
21:34
sometimes where we're like, well, you know, the AI is
21:36
all knowing and therefore whatever
21:38
it says must be true. But but these
21:41
things will make up stuff, you know, very
21:44
aggressively, and you know, you
21:46
everyone can try asking it for their their bio. You'll
21:49
you'll get something that You'll always get something
21:51
that's of the right form, that has the right
21:54
tone. But you know, the facts just aren't necessarily
21:56
there. So that's obviously a problem.
21:58
We need to figure out how to close those gaps, fix
22:00
those problems. There's lots of ways
22:03
we could use them more easily.
22:05
I'd just like to say, faced with the awesome
22:07
potential of what these technologies might do,
22:10
it's a bit encouraging to hear that even
22:12
chat GPT has a weakness
22:14
for inventing flamboyant, if
22:16
fictional versions of people's lives.
22:19
And while entertaining ourselves with chat GPT
22:22
and mid journey is important, the
22:24
way lay people use consumer facing chatbots
22:27
and generative AI is just
22:29
fundamentally different from the
22:31
way an enterprise business uses AI.
22:34
How can we harness the abilities of artificial
22:36
intelligence to help us solve the problems
22:39
we face in business and technology.
22:41
Let's listen on as David and Jacob
22:43
continue their conversation.
22:45
We've been talking in a somewhat abstract way
22:48
about AI in the ways it can
22:50
be used. Let's talk in a little bit
22:52
more of a specific way. Can
22:54
you just talk about some examples
22:57
of business challenges that can be solved
22:59
with automation. With this kind of automation
23:02
we're talking about.
23:03
Yeah, so there really really this guy's
23:06
the limit. There's a whole set
23:08
of different applications that these
23:10
models are really good at. And basically
23:12
it's a super set of everything we used to use
23:14
ALI for in business. So, you
23:17
know, the simple kinds of things are like,
23:19
hey, if I have text and i'm you know, I have product
23:21
reviews and I want to be able to tell if these
23:23
are positive or negative. You know, like let's look
23:25
at all the negative reviews so we can have a human look through
23:27
them and see what was up. Very
23:30
common business use case. You can
23:32
do it with traditional deep learning based
23:34
AI. So so there's things like
23:36
that that are you know, it's very prosaic sort
23:38
that we were already doing it. We've been doing it for a long time.
23:42
Then you get situations that are
23:44
that were harder for the old day. I like, if
23:46
i'm I want to compress
23:48
something like I want to I have like they have
23:51
a chat transcript, like a customer called in
23:53
and they had a complaint, they called
23:56
back. Okay, Now a new you
23:58
know, person on the line needs to go read
24:00
the old transcript to catch up. Wouldn't
24:03
it be better if we could just summarize that,
24:05
just condense it all down a quick little paragraph,
24:07
you know, customer call they were upset about this, rather than
24:09
having to read the blow by blow. There's just
24:11
lots of settings like that where summarization
24:14
is really helpful. Hey, you have a meeting and
24:17
I'd like to just automatically, you
24:19
know, have have that meeting or that email or
24:21
whatever. I'd like to just have a condensed down so I can
24:23
really quickly get to the heart of the matter. These
24:26
models are are really good at doing that. They're
24:28
also really good at question answering. So if
24:30
I want to find out what's how many vacation days
24:33
do I have? I can now interact
24:35
in natural language with a system
24:38
that can go and that has access
24:40
to our HR policies, and I can actually
24:42
have a you know, multi turn conversation
24:44
where I can, you know, like I would have with you
24:46
know, somebody, you know, actual HR
24:49
professional or customer service representative.
24:52
So a big part, you
24:54
know, of what this is doing is it's
24:56
it's putting an interface. You know, when
24:58
we think of computer interfaces, usually thinking about
25:01
UI user interface elements where I
25:03
click on menus and there's buttons and all
25:05
this stuff. Increasingly, now we
25:07
can just talk, you know, you just
25:10
in words. You can describe what you want, you
25:12
want to answer, ask a question, you
25:14
want to sort of command the system to do something,
25:17
rather than having to learn how to do that clicking buttons,
25:19
which might be inefficient. Now we can just sort of spell
25:21
it out.
25:22
Interesting, right, the graphical user interface
25:24
that we all sort of default to, that's
25:27
not like the state of nature, right, That's
25:29
a thing that was invented and just came
25:31
to be the standard way that we interact with computers.
25:33
And so you could imagine, as you're saying,
25:36
like chat essentially chatting
25:39
with the machine could could become
25:41
a sort of standard user interface, just like
25:43
the graphical user interface, did you know
25:45
over the past several decades.
25:47
Absolutely, And I think those kinds of
25:49
conversational interfaces are going to be hugely
25:52
important for increasing our productivity.
25:54
It's just a lot easier if I if I have to
25:56
learn how to use a tool, or I don't have to kind
25:59
of have awkward, you know, interactions
26:01
for the computer. I can just tell it what I want, and I can understand
26:03
it, could you know, potentially even ask questions
26:06
back to clarify and have those kinds of conversations
26:09
that can be extremely powerful.
26:11
And in fact, one area where that's going to I think be absolutely
26:15
game changing is in code. When we write
26:17
code. You know, programming
26:19
languages are a way
26:21
for us to sort of match between
26:24
our very sloppy way of talking and
26:27
the very exact way that you need to command a computer
26:29
to do what you wanted to do. They're cumbersome
26:32
to learn, they can you know, create very complex
26:34
systems that are very hard to reason about. And
26:37
we're already starting to see the ability to just
26:39
write down what you want and AI will
26:41
generate the code for you. And I think we're
26:43
just going to see a huge revolution of like we just
26:45
converse you and we can have a conversation to
26:47
say what we want, and then the computer can
26:50
actually not only do fixed
26:52
actions and do things for us, but it can actually
26:54
even write code to do new things, you know, and
26:57
generate software itself. Given how much
26:59
software we have, of how much craving we
27:01
have for software, like we'll never have enough
27:03
software in our world, uh,
27:05
you know, the ability to have a systems
27:07
as a helper in that, I
27:09
think we're going to see a lot of a lot of value
27:12
there.
27:13
So if you if you think about the different
27:15
ways AI might be applied
27:17
to business, I mean you've talked about a number of the sort
27:19
of classic use cases. What
27:22
are some of the more out
27:24
there use cases. What are some you know, unique
27:27
ways you could imagine AI being applied
27:29
to business.
27:31
Yeah, there's really disguised the limit.
27:33
I mean, we have one project that I'm kind of a fan
27:36
of where we actually were
27:38
working with a mechanical engineering professor
27:40
at MIT working on a classic
27:42
problem, how do you build linkage systems
27:45
which like you imagine bars and joints
27:47
and ogres, you know, the things
27:49
that are.
27:50
Building a thing, building a physical
27:52
machine of some.
27:53
Kind of like real like metal
27:55
and you know nineteenth
27:58
century just old school industrial
28:00
revolution. Yeah yeah, yeah, but you know the little
28:03
arm that's that's holding up my microphone in front
28:05
of me. Cranes get bold, your buildings,
28:07
you know, parts of your engines. This is like classical
28:09
stuff. It turns out that you know humans,
28:11
if you want to build an advanced system, you
28:14
decide what like curve you want to create,
28:16
and then a human together with a computer
28:18
program can build a five or six bar
28:21
linkage, and then that's kind of where you top out. It is
28:23
because it gets too complicated to work more
28:26
than that. We built a generative AI
28:28
system that can build twenty bar linkages,
28:30
like arbitrarily complex. So these are
28:32
machines that are beyond the capability of
28:34
a human to design themselves.
28:38
Another example, we have an AI system
28:40
that can generate electronic circuits. You know,
28:42
we had a project where we're working where we were building
28:44
better power converters which allow our
28:47
computers and our devices to be more efficient,
28:50
save energy, you know, less
28:52
less carbon ote. But I think the world
28:54
around us has always been shaped by
28:56
technology. If you look around, you know, just think
28:58
about how many steps and how people and
29:00
how many designs went into the table
29:02
and the chair and the lamp. It's
29:05
it's really just astonishing. And that's
29:07
already you know, the fruit of automation
29:10
and computers and those kinds of tools. But we're going to see
29:12
that increasingly be product also
29:15
of AI. It's just going to be everywhere around
29:17
us. Everything we touch is going to have to you
29:19
know, helped in some way to get get
29:22
to you by a.
29:23
You know, that is a pretty profound transformation
29:26
that you're talking about in business. How
29:28
do you think about the implications of that, both
29:30
for the sort of you know, business
29:33
itself and also for for employees.
29:37
Yeah, so I think for businesses
29:39
this is gonna cut costs, make
29:42
new opportunities to like customers,
29:44
you know, like there's just you
29:46
know, it's sort of all upside right, like for
29:49
the for the workers, I think the story is mostly
29:52
good too. You know, like how many things
29:54
do you do in your day that you'd
29:57
really rather not right? You know, and we're
29:59
used to have I think, things we don't like automated
30:01
away, you know, we didn't
30:04
you know, if you didn't like walking many miles
30:06
to work, then you know, like you have a car and
30:08
you can drive there. Or we used to have a
30:10
huge fraction over ninety percent of the US
30:12
population engaged in agriculture, and then we
30:15
mechanized it how very few people work
30:17
in agriculture. A small number of people can do the work
30:19
of a large number of people. And then you
30:21
know, things like email, and you know, they've
30:23
led to huge productivity enhancements because
30:25
I don't need to be writing letters and sending them
30:28
in the mail. I can just instantly communicate with
30:30
people. We just become more
30:32
effective. Like our jobs have transformed,
30:36
whether it's a physical job like agriculture
30:38
or whether it's a knowledge worker job where
30:40
you're sending emails and communicating
30:42
with people and coordinating teams, we've
30:44
just gotten better. And you know, the technology
30:46
has just made us more productive. And this is
30:48
just another example. Now, you know,
30:51
there are people who worry that, you know, we'll
30:53
be so good at that that maybe jobs
30:55
will be displaced, and
30:57
that's that's a legitimate concern. But
30:59
just like how in agriculture,
31:02
you know, it's not like suddenly we had ninety percent of
31:04
the population unemployed. You know, people
31:06
transitioned to other jobs.
31:09
And the other thing that we've found, too, is that
31:12
our appetite for doing more
31:14
things is as humans
31:16
is sort of insatiable. So even if we
31:19
can dramatically increase how much you know,
31:21
one human can do, that doesn't
31:23
necessarily mean we're going to do a fixed amount of
31:25
stuff. There's an appetite to have even more, so
31:27
we're going to you can continue to grow the pie.
31:30
So I think at least certainly in the near
31:32
term, you know, we're going to see a lot of drudgery go away
31:34
from work. We're going to see people
31:37
be able to be more effective at their
31:39
jobs. You know, we will see some transformation
31:42
in jobs and what look like. But we've
31:44
seen that before and
31:47
the technology a least has the potential to make our
31:49
lives a lot easier.
31:52
So IBM recently launched
31:54
Watson X, which includes Watson
31:56
x dot AI. Tell me about
31:59
that, Tell me about you know what it is and the new
32:01
possibilities that it opens up.
32:03
Yeah, So Watson next is
32:05
obviously a bit of a
32:08
new branding on the Watson
32:10
brand. TJ. Watson that was the
32:12
founder of IBM and
32:15
our EI technologies have had the Watson
32:17
brand. Watson X is
32:19
a recognition that there's
32:21
something new, there's something that actually has changed
32:23
the game. We've gone from this
32:26
old world of automation is
32:28
to labor intensive to this new world of possibilities
32:31
where it's much easier to use AI. And
32:34
what Watson X does
32:36
it brings together tools for
32:38
businesses to harness that power. So
32:41
whattsonex dot AI foundation
32:44
models that our customers can use. It includes
32:47
tools that make it easy to run, easy
32:49
to deploy, easy to experiment.
32:52
There's a watsonex dot Data component
32:54
which allows you to sort of organize
32:57
and access to your data. So what we're really
32:59
trying to do is give our customers a
33:01
cohesive set of tools
33:03
to harness the value of
33:06
these technologies and at the same time be
33:08
able to manage the risks and other
33:10
things that you have to keep an eye on in
33:12
an enterprise context.
33:15
So we talk about the guests on this
33:17
show as new creators,
33:20
by which we mean people who are creatively
33:22
applying technology in business
33:25
to drive change. And I'm
33:27
curious how creativity
33:30
plays a role in the research that you do.
33:33
I honestly, I think the creative
33:36
aspects of this job,
33:38
this is what makes this work exciting.
33:41
You know, I should say, you know, the folks who
33:43
work in my organization are
33:45
doing the creating, and I.
33:47
Guess you're doing the
33:49
managing so that they could do the creator.
33:52
I'm helping them be their best and
33:55
I still get to get involved in the
33:57
weeds of the research as much as I can.
33:59
But you know, there's something really exciting
34:01
about inventing, you know,
34:04
Like one of the nice things about doing
34:06
invention and doing research on AI
34:08
in industry is it's usually grounded
34:10
and a real problem that somebody's having. You
34:12
know, a customer wants to solve this problem
34:15
it's losing money or there
34:17
there would be a new opportunity. You identify
34:19
that problem and then you build
34:22
something that's never been built before
34:24
to do that. And I think that's honestly
34:26
the adrenaline rush that keeps
34:28
all of us in this field. How
34:30
do you do something that nobody else on
34:33
earth has done before or
34:35
tried before, So that that kind of
34:37
creativity, and there's also creativity
34:39
as well, and identifying what those problems are, being
34:42
able to understand the places
34:45
where you know the technology
34:47
is close enough to solving a problem,
34:49
and doing that matchmaking between problems
34:53
that are now solvable, you know, and
34:55
in AI, where the field is moving so fast, this
34:57
is constantly growing horizon
35:00
of things that we might be able to solve. So
35:02
that matchmaking, I think is also a really
35:04
interesting creative problem. So
35:07
I think I think that's that's that's why it's
35:09
so much fun. And it's a fun environment
35:11
we have here too. It's you know, people drawing
35:14
on whiteboards and writing on
35:16
pages of math and.
35:18
You know, like in a movie, like in a movie.
35:21
Yeah, straight from special casting.
35:23
The drawing on the window, writing on the window in sharp.
35:26
Absolutely so,
35:29
so let's close with the really long
35:31
view. How
35:33
do you imagine AI and
35:35
people working together twenty.
35:38
Years from now? Yeah,
35:42
it's really hard to make predictions. The
35:45
vision that I like,
35:49
actually this came from an MIT
35:51
economist named David Ottur, which
35:54
was imagine AI almost
35:57
as a natural resource. You
35:59
know, we know how natural resources
36:02
work, right, Like there's an ore we can dig up
36:04
out of the earth that comes from kind of springs
36:06
from the earth, or we usually think of
36:08
that in terms of physical stuff. With
36:10
AI, you can almost think of it as like there's a new kind
36:13
of abundance potentially twenty years
36:15
from now where not only can we have
36:17
things we can build or eat or use or burn
36:19
or whatever, now we have, you know, this
36:21
ability to do things and understand things
36:23
and do intellectual work. And I
36:26
think we can get to a world where
36:28
automating things is just seamless. We're
36:31
surrounded by capability
36:33
to augment ourselves to get
36:36
things done. And you
36:38
could think of that in terms of like, well, that's
36:40
going to displace our jobs, because eventually the AI system
36:42
is going to do everything we can do. But you
36:44
could also think of it in terms of like, wow,
36:46
that's just so much abundance that we now have,
36:49
and really how we use that abundance
36:51
is sort of up to us, you know,
36:53
like when you can writing software is super
36:55
easy and fast and anybody can do it. Just
36:58
think about all the things you can do now, think
37:00
about all the new activities, and go out all the
37:02
ways we could use that to enrich our lives.
37:05
That's where I'd like to see us in
37:07
twenty years. You know, we can we
37:09
can do just so much more than
37:11
we were able to do before abundance.
37:14
Great, thank you so much
37:17
for your time.
37:18
Yeah, it's been a pleasure. Thanks for inviting me.
37:22
What a far ranging, deep conversation.
37:25
I'm mesmerized by the vision David just described.
37:27
A world where natural conversation between
37:30
mankind and machine can generate
37:32
creative solutions to our most
37:34
complex problems. A world where
37:36
we view AI not as our
37:38
replacements, but as a powerful
37:41
resource we can tap into and
37:43
exponentially boost our innovation
37:46
and productivity. Thanks so much
37:48
to doctor David Cox for joining us
37:50
on smart Talks. We deeply appreciate
37:53
him sharing his huge breadth
37:55
of AI knowledge with us and for explaining
37:58
the transformative potential of foundation
38:00
models in a way that even I can
38:02
understand. We eagerly await his
38:05
next great breakthrough. Smart
38:08
Talks with IBM is produced by Matt Romano,
38:10
David jaw nishe Venkat
38:13
and Royston Preserve with Jacob
38:15
Goldstein. We're edited by Lydia
38:17
Jane Kott. Our engineers are Jason
38:19
Gambrel, Sarah Buguier and
38:21
Ben Holliday. Theme song by
38:24
Gramosco. Special thanks
38:26
to Carli Megliori, Andy Kelly,
38:28
Kathy Callahan and the eight Bar
38:30
and IBM teams, as well as
38:33
the Pushkin marketing team. Smart
38:35
Talks with IBM is a production of Pushkin
38:38
Industries and iHeartMedia. To
38:40
find more Pushkin podcasts, listen
38:42
on the iHeartRadio app, Apple Podcasts,
38:45
or wherever you listen to podcasts.
38:48
Him Malcolm Gladwell. This is a
38:50
paid advertisement from IBM.
Podchaser is the ultimate destination for podcast data, search, and discovery. Learn More