跟TED演讲学英文:AI isn‘t as smart as you think -- but it could be by Jeff Dean

2024-04-29 19:36

本文主要是介绍跟TED演讲学英文:AI isn‘t as smart as you think -- but it could be by Jeff Dean,希望对大家解决编程问题提供一定的参考价值,需要的开发者们随着小编来一起学习吧!

AI isn’t as smart as you think – but it could be

在这里插入图片描述

Link: https://www.ted.com/talks/jeff_dean_ai_isn_t_as_smart_as_you_think_but_it_could_be

Speaker: Jeff Dean

Jeffrey Adgate “Jeff” Dean (born July 23, 1968) is an American computer scientist and software engineer. Since 2018, he has been the lead of Google AI.[1] He was appointed Alphabet’s chief scientist in 2023 after a reorganization of Alphabet’s AI focused groups.[2]

Date: August 2021

文章目录

  • AI isn't as smart as you think -- but it could be
    • Introduction
    • Vocabulary
    • Transcript
    • Q&A with Chris Anderson
    • Summary
    • 后记

Introduction

What is AI, really? Jeff Dean, the head of Google’s AI efforts, explains the underlying technology that enables artificial intelligence to do all sorts of things, from understanding language to diagnosing disease – and presents a roadmap for building better, more responsible systems that have a deeper understanding of the world. (Followed by a Q&A with head of TED Chris Anderson)

Vocabulary

wedge into:挤进

when we were all wedged into a tiny office space 当我们都挤在一个狭小的办公室里

emulate:美 [ˈemjuleɪt] 模仿,仿效

a series of interconnected artificial neurons that loosely emulate the properties of your real neurons. 一系列相互连接的人造神经元,松散地模拟了真实神经元的特性。

tailored to:合身的,适合的,裁剪的

we also got excited about how we could build hardware that was better tailored to the kinds of computations neural networks wanted to do. 我们还对如何构建更适合神经网络想要进行的计算类型的硬件感到兴奋。

hone:磨砺,磨练

hydroponic:美 [ˌhaɪdrə’pɒnɪk] 水耕法的, 水栽法的

I’ve been honing my gardening skills, experimenting with vertical hydroponic gardening. 我一直在磨练我的园艺技能,尝试垂直水培园艺。

Transcript

Hi, I’m Jeff.

I lead AI Research and Health at Google.

I joined Google more than 20 years ago,

when we were all wedged
into a tiny office space,

above what’s now a T-Mobile store
in downtown Palo Alto.

I’ve seen a lot of computing
transformations in that time,

and in the last decade, we’ve seen AI
be able to do tremendous things.

But we’re still doing it
all wrong in many ways.

That’s what I want
to talk to you about today.

But first, let’s talk
about what AI can do.

So in the last decade,
we’ve seen tremendous progress

in how AI can help computers see,
understand language,

understand speech better than ever before.

Things that we couldn’t do
before, now we can do.

If you think about computer vision alone,

just in the last 10 years,

computers have effectively
developed the ability to see;

10 years ago, they couldn’t see,
now they can see.

You can imagine this has had
a transformative effect

on what we can do with computers.

So let’s look at a couple
of the great applications

enabled by these capabilities.

We can better predict flooding,
keep everyone safe,

using machine learning.

We can translate over 100 languages
so we all can communicate better,

and better predict and diagnose disease,

where everyone gets
the treatment that they need.

So let’s look at two key components

that underlie the progress
in AI systems today.

The first is neural networks,

a breakthrough approach to solving
some of these difficult problems

that has really shone
in the last 15 years.

But they’re not a new idea.

And the second is computational power.

It actually takes a lot
of computational power

to make neural networks
able to really sing,

and in the last 15 years,
we’ve been able to have that,

and that’s partly what’s enabled
all this progress.

But at the same time,
I think we’re doing several things wrong,

and that’s what I want
to talk to you about

at the end of the talk.

First, a bit of a history lesson.

So for decades,

almost since the very
beginning of computing,

people have wanted
to be able to build computers

that could see, understand language,
understand speech.

The earliest approaches
to this, generally,

people were trying to hand-code
all the algorithms

that you need to accomplish
those difficult tasks,

and it just turned out
to not work very well.

But in the last 15 years,
a single approach

unexpectedly advanced all these different
problem spaces all at once:

neural networks.

So neural networks are not a new idea.

They’re kind of loosely based

on some of the properties
that are in real neural systems.

And many of the ideas
behind neural networks

have been around since the 1960s and 70s.

A neural network is what it sounds like,

a series of interconnected
artificial neurons

that loosely emulate the properties
of your real neurons.

An individual neuron
in one of these systems

has a set of inputs,

each with an associated weight,

and the output of a neuron

is a function of those inputs
multiplied by those weights.

So pretty simple,

and lots and lots of these work together
to learn complicated things.

So how do we actually learn
in a neural network?

It turns out the learning process

consists of repeatedly making
tiny little adjustments

to the weight values,

strengthening the influence
of some things,

weakening the influence of others.

By driving the overall system
towards desired behaviors,

these systems can be trained
to do really complicated things,

like translate
from one language to another,

detect what kind
of objects are in a photo,

all kinds of complicated things.

I first got interested in neural networks

when I took a class on them
as an undergraduate in 1990.

At that time,

neural networks showed
impressive results on tiny problems,

but they really couldn’t scale to do
real-world important tasks.

But I was super excited.

(Laughter)

I felt maybe we just needed
more compute power.

And the University of Minnesota
had a 32-processor machine.

I thought, "With more compute power,

boy, we could really make
neural networks really sing."

So I decided to do a senior thesis
on parallel training of neural networks,

the idea of using processors in a computer
or in a computer system

to all work toward the same task,

that of training neural networks.

32 processors, wow,

we’ve got to be able
to do great things with this.

But I was wrong.

Turns out we needed about a million times
as much computational power

as we had in 1990

before we could actually get
neural networks to do impressive things.

But starting around 2005,

thanks to the computing progress
of Moore’s law,

we actually started to have
that much computing power,

and researchers in a few universities
around the world started to see success

in using neural networks for a wide
variety of different kinds of tasks.

I and a few others at Google
heard about some of these successes,

and we decided to start a project
to train very large neural networks.

One system that we trained,

we trained with 10 million
randomly selected frames

from YouTube videos.

The system developed the capability

to recognize all kinds
of different objects.

And it being YouTube, of course,

it developed the ability
to recognize cats.

YouTube is full of cats.

(Laughter)

But what made that so remarkable

is that the system was never told
what a cat was.

So using just patterns in data,

the system honed in on the concept
of a cat all on its own.

All of this occurred at the beginning
of a decade-long string of successes,

of using neural networks
for a huge variety of tasks,

at Google and elsewhere.

Many of the things you use every day,

things like better speech
recognition for your phone,

improved understanding
of queries and documents

for better search quality,

better understanding of geographic
information to improve maps,

and so on.

Around that time,

we also got excited about how we could
build hardware that was better tailored

to the kinds of computations
neural networks wanted to do.

Neural network computations
have two special properties.

The first is they’re very tolerant
of reduced precision.

Couple of significant digits,
you don’t need six or seven.

And the second is that all the
algorithms are generally composed

of different sequences
of matrix and vector operations.

So if you can build a computer

that is really good at low-precision
matrix and vector operations

but can’t do much else,

that’s going to be great
for neural-network computation,

even though you can’t use it
for a lot of other things.

And if you build such things,
people will find amazing uses for them.

This is the first one we built, TPU v1.

“TPU” stands for Tensor Processing Unit.

These have been used for many years
behind every Google search,

for translation,

in the DeepMind AlphaGo matches,

so Lee Sedol and Ke Jie
maybe didn’t realize,

but they were competing
against racks of TPU cards.

And we’ve built a bunch
of subsequent versions of TPUs

that are even better and more exciting.

But despite all these successes,

I think we’re still doing
many things wrong,

and I’ll tell you about three
key things we’re doing wrong,

and how we’ll fix them.

The first is that most
neural networks today

are trained to do one thing,
and one thing only.

You train it for a particular task
that you might care deeply about,

but it’s a pretty heavyweight activity.

You need to curate a data set,

you need to decide
what network architecture you’ll use

for this problem,

you need to initialize the weights
with random values,

apply lots of computation
to make adjustments to the weights.

And at the end, if you’re lucky,
you end up with a model

that is really good
at that task you care about.

But if you do this over and over,

you end up with thousands
of separate models,

each perhaps very capable,

but separate for all the different
tasks you care about.

But think about how people learn.

In the last year, many of us
have picked up a bunch of new skills.

I’ve been honing my gardening skills,

experimenting with vertical
hydroponic gardening.

To do that, I didn’t need to relearn
everything I already knew about plants.

I was able to know
how to put a plant in a hole,

how to pour water, that plants need sun,

and leverage that
in learning this new skill.

Computers can work
the same way, but they don’t today.

If you train a neural
network from scratch,

it’s effectively like forgetting
your entire education

every time you try to do something new.

That’s crazy, right?

So instead, I think we can
and should be training

multitask models that can do
thousands or millions of different tasks.

Each part of that model would specialize
in different kinds of things.

And then, if we have a model
that can do a thousand things,

and the thousand and first
thing comes along,

we can leverage
the expertise we already have

in the related kinds of things

so that we can more quickly be able
to do this new task,

just like you, if you’re confronted
with some new problem,

you quickly identify
the 17 things you already know

that are going to be helpful
in solving that problem.

Second problem is that most
of our models today

deal with only a single
modality of data –

with images, or text or speech,

but not all of these all at once.

But think about how you
go about the world.

You’re continuously using all your senses

to learn from, react to,

figure out what actions
you want to take in the world.

Makes a lot more sense to do that,

and we can build models in the same way.

We can build models that take in
these different modalities of input data,

text, images, speech,

but then fuse them together,

so that regardless of whether the model
sees the word “leopard,”

sees a video of a leopard
or hears someone say the word “leopard,”

the same response
is triggered inside the model:

the concept of a leopard

can deal with different
kinds of input data,

even nonhuman inputs,
like genetic sequences,

3D clouds of points,
as well as images, text and video.

The third problem
is that today’s models are dense.

There’s a single model,

the model is fully activated
for every task,

for every example
that we want to accomplish,

whether that’s a really simple
or a really complicated thing.

This, too, is unlike
how our own brains work.

Different parts of our brains
are good at different things,

and we’re continuously calling
upon the pieces of them

that are relevant for the task at hand.

For example, nervously watching
a garbage truck

back up towards your car,

the part of your brain that thinks
about Shakespearean sonnets

is probably inactive.

(Laughter)

AI models can work the same way.

Instead of a dense model,

we can have one
that is sparsely activated.

So for particular different tasks,
we call upon different parts of the model.

During training, the model can also learn
which parts are good at which things,

to continuously identify what parts
it wants to call upon

in order to accomplish a new task.

The advantage of this is we can have
a very high-capacity model,

but it’s very efficient,

because we’re only calling
upon the parts that we need

for any given task.

So fixing these three things, I think,

will lead to a more powerful AI system:

instead of thousands of separate models,

train a handful of general-purpose models

that can do thousands
or millions of things.

Instead of dealing with single modalities,

deal with all modalities,

and be able to fuse them together.

And instead of dense models,
use sparse, high-capacity models,

where we call upon the relevant
bits as we need them.

We’ve been building a system
that enables these kinds of approaches,

and we’ve been calling
the system “Pathways.”

So the idea is this model
will be able to do

thousands or millions of different tasks,

and then, we can incrementally
add new tasks,

and it can deal
with all modalities at once,

and then incrementally learn
new tasks as needed

and call upon the relevant
bits of the model

for different examples or tasks.

And we’re pretty excited about this,

we think this is going
to be a step forward

in how we build AI systems.

But I also wanted
to touch on responsible AI.

We clearly need to make sure
that this vision of powerful AI systems

benefits everyone.

These kinds of models raise
important new questions

about how do we build them with fairness,

interpretability, privacy and security,

for all users in mind.

For example, if we’re going
to train these models

on thousands or millions of tasks,

we’ll need to be able to train them
on large amounts of data.

And we need to make sure that data
is thoughtfully collected

and is representative of different
communities and situations

all around the world.

And data concerns are only
one aspect of responsible AI.

We have a lot of work to do here.

So in 2018, Google published
this set of AI principles

by which we think about developing
these kinds of technology.

And these have helped guide us
in how we do research in this space,

how we use AI in our products.

And I think it’s a really helpful
and important framing

for how to think about these deep
and complex questions

about how we should
be using AI in society.

We continue to update these
as we learn more.

Many of these kinds of principles
are active areas of research –

super important area.

Moving from single-purpose systems
that kind of recognize patterns in data

to these kinds of general-purpose
intelligent systems

that have a deeper
understanding of the world

will really enable us to tackle

some of the greatest problems
humanity faces.

For example,

we’ll be able to diagnose more disease;

we’ll be able to engineer better medicines

by infusing these models
with knowledge of chemistry and physics;

we’ll be able to advance
educational systems

by providing more individualized tutoring

to help people learn
in new and better ways;

we’ll be able to tackle
really complicated issues,

like climate change,

and perhaps engineering
of clean energy solutions.

So really, all of these kinds of systems

are going to be requiring
the multidisciplinary expertise

of people all over the world.

So connecting AI
with whatever field you are in,

in order to make progress.

So I’ve seen a lot
of advances in computing,

and how computing, over the past decades,

has really helped millions of people
better understand the world around them.

And AI today has the potential
to help billions of people.

We truly live in exciting times.

Thank you.

(Applause)

Q&A with Chris Anderson

Chris Anderson: Thank you so much.

I want to follow up on a couple things.

This is what I heard.

Most people’s traditional picture of AI

is that computers recognize
a pattern of information,

and with a bit of machine learning,

they can get really good at that,
better than humans.

What you’re saying is those patterns

are no longer the atoms
that AI is working with,

that it’s much richer-layered concepts

that can include all manners
of types of things

that go to make up a leopard, for example.

So what could that lead to?

Give me an example
of when that AI is working,

what do you picture happening in the world

in the next five or 10 years
that excites you?

Jeff Dean: I think
the grand challenge in AI

is how do you generalize
from a set of tasks

you already know how to do

to new tasks,

as easily and effortlessly as possible.

And the current approach of training
separate models for everything

means you need lots of data
about that particular problem,

because you’re effectively trying
to learn everything

about the world
and that problem, from nothing.

But if you can build these systems

that already are infused with how to do
thousands and millions of tasks,

then you can effectively
teach them to do a new thing

with relatively few examples.

So I think that’s the real hope,

that you could then have a system
where you just give it five examples

of something you care about,

and it learns to do that new task.

CA: You can do a form
of self-supervised learning

that is based on remarkably
little seeding.

JD: Yeah, as opposed to needing
10,000 or 100,000 examples

to figure everything in the world out.

CA: Aren’t there kind of terrifying
unintended consequences

possible, from that?

JD: I think it depends
on how you apply these systems.

It’s very clear that AI
can be a powerful system for good,

or if you apply it in ways
that are not so great,

it can be a negative consequence.

So I think that’s why it’s important
to have a set of principles

by which you look at potential uses of AI

and really are careful and thoughtful
about how you consider applications.

CA: One of the things
people worry most about

is that, if AI is so good at learning
from the world as it is,

it’s going to carry forward
into the future

aspects of the world as it is
that actually aren’t right, right now.

And there’s obviously been
a huge controversy about that

recently at Google.

Some of those principles
of AI development,

you’ve been challenged that you’re not
actually holding to them.

Not really interested to hear
about comments on a specific case,

but … are you really committed?

How do we know that you are
committed to these principles?

Is that just PR, or is that real,
at the heart of your day-to-day?

JD: No, that is absolutely real.

Like, we have literally hundreds of people

working on many of these
related research issues,

because many of those
things are research topics

in their own right.

How do you take data from the real world,

that is the world as it is,
not as we would like it to be,

and how do you then use that
to train a machine-learning model

and adapt the data bit of the scene

or augment the data with additional data

so that it can better reflect
the values we want the system to have,

not the values that it sees in the world?

CA: But you work for Google,

Google is funding the research.

How do we know that the main values
that this AI will build

are for the world,

and not, for example, to maximize
the profitability of an ad model?

When you know everything
there is to know about human attention,

you’re going to know so much

about the little wriggly,
weird, dark parts of us.

In your group, are there rules
about how you hold off,

church-state wall
between a sort of commercial push,

“You must do it for this purpose,”

so that you can inspire
your engineers and so forth,

to do this for the world, for all of us.

JD: Yeah, our research group
does collaborate

with a number of groups across Google,

including the Ads group,
the Search group, the Maps group,

so we do have some collaboration,
but also a lot of basic research

that we publish openly.

We’ve published more
than 1,000 papers last year

in different topics,
including the ones you discussed,

about fairness, interpretability
of the machine-learning models,

things that are super important,

and we need to advance
the state of the art in this

in order to continue to make progress

to make sure these models
are developed safely and responsibly.

CA: It feels like we’re at a time
when people are concerned

about the power of the big tech companies,

and it’s almost, if there was ever
a moment to really show the world

that this is being done
to make a better future,

that is actually key to Google’s future,

as well as all of ours.

JD: Indeed.

CA: It’s very good to hear you
come and say that, Jeff.

Thank you so much for coming here to TED.

JD: Thank you.

(Applause)

Summary

Jeff Dean’s speech focuses on the transformative potential of AI, highlighting the significant progress made in the last decade. He discusses the capabilities of AI in various domains, such as computer vision, language understanding, and medical diagnosis. Dean attributes this progress to advancements in neural networks and computational power. However, he acknowledges several shortcomings in current AI systems, including their single-task nature, limited modalities, and dense architecture.

Dean proposes solutions to address these shortcomings, advocating for the development of multitask models capable of handling diverse tasks and modalities. He emphasizes the importance of sparse, high-capacity models that can efficiently activate relevant components for specific tasks. Dean introduces the concept of “Pathways,” a system designed to enable these approaches, which he believes will represent a significant step forward in AI research and development.

In addition to discussing technical advancements, Dean emphasizes the importance of responsible AI, highlighting the need for fairness, interpretability, privacy, and security in AI systems. He outlines Google’s AI principles as a guiding framework for responsible AI development and acknowledges the ongoing research and challenges in this area. Dean concludes his speech by expressing optimism about the potential of AI to address complex societal challenges and the importance of multidisciplinary collaboration in realizing this potential.

后记

2024年4月29日18点23分完成这篇演讲的学习。

这篇关于跟TED演讲学英文:AI isn‘t as smart as you think -- but it could be by Jeff Dean的文章就介绍到这儿,希望我们推荐的文章对编程师们有所帮助!



http://www.chinasem.cn/article/946974

相关文章

Ilya-AI分享的他在OpenAI学习到的15个提示工程技巧

Ilya(不是本人,claude AI)在社交媒体上分享了他在OpenAI学习到的15个Prompt撰写技巧。 以下是详细的内容: 提示精确化:在编写提示时,力求表达清晰准确。清楚地阐述任务需求和概念定义至关重要。例:不用"分析文本",而用"判断这段话的情感倾向:积极、消极还是中性"。 快速迭代:善于快速连续调整提示。熟练的提示工程师能够灵活地进行多轮优化。例:从"总结文章"到"用

AI绘图怎么变现?想做点副业的小白必看!

在科技飞速发展的今天,AI绘图作为一种新兴技术,不仅改变了艺术创作的方式,也为创作者提供了多种变现途径。本文将详细探讨几种常见的AI绘图变现方式,帮助创作者更好地利用这一技术实现经济收益。 更多实操教程和AI绘画工具,可以扫描下方,免费获取 定制服务:个性化的创意商机 个性化定制 AI绘图技术能够根据用户需求生成个性化的头像、壁纸、插画等作品。例如,姓氏头像在电商平台上非常受欢迎,

从去中心化到智能化:Web3如何与AI共同塑造数字生态

在数字时代的演进中,Web3和人工智能(AI)正成为塑造未来互联网的两大核心力量。Web3的去中心化理念与AI的智能化技术,正相互交织,共同推动数字生态的变革。本文将探讨Web3与AI的融合如何改变数字世界,并展望这一新兴组合如何重塑我们的在线体验。 Web3的去中心化愿景 Web3代表了互联网的第三代发展,它基于去中心化的区块链技术,旨在创建一个开放、透明且用户主导的数字生态。不同于传统

AI一键生成 PPT

AI一键生成 PPT 操作步骤 作为一名打工人,是不是经常需要制作各种PPT来分享我的生活和想法。但是,你们知道,有时候灵感来了,时间却不够用了!😩直到我发现了Kimi AI——一个能够自动生成PPT的神奇助手!🌟 什么是Kimi? 一款月之暗面科技有限公司开发的AI办公工具,帮助用户快速生成高质量的演示文稿。 无论你是职场人士、学生还是教师,Kimi都能够为你的办公文

Andrej Karpathy最新采访:认知核心模型10亿参数就够了,AI会打破教育不公的僵局

夕小瑶科技说 原创  作者 | 海野 AI圈子的红人,AI大神Andrej Karpathy,曾是OpenAI联合创始人之一,特斯拉AI总监。上一次的动态是官宣创办一家名为 Eureka Labs 的人工智能+教育公司 ,宣布将长期致力于AI原生教育。 近日,Andrej Karpathy接受了No Priors(投资博客)的采访,与硅谷知名投资人 Sara Guo 和 Elad G

AI hospital 论文Idea

一、Benchmarking Large Language Models on Communicative Medical Coaching: A Dataset and a Novel System论文地址含代码 大多数现有模型和工具主要迎合以患者为中心的服务。这项工作深入探讨了LLMs在提高医疗专业人员的沟通能力。目标是构建一个模拟实践环境,人类医生(即医学学习者)可以在其中与患者代理进行医学

AI行业应用(不定期更新)

ChatPDF 可以让你上传一个 PDF 文件,然后针对这个 PDF 进行小结和提问。你可以把各种各样你要研究的分析报告交给它,快速获取到想要知道的信息。https://www.chatpdf.com/

【北交大信息所AI-Max2】使用方法

BJTU信息所集群AI_MAX2使用方法 使用的前提是预约到相应的算力卡,拥有登录权限的账号密码,一般为导师组共用一个。 有浏览器、ssh工具就可以。 1.新建集群Terminal 浏览器登陆10.126.62.75 (如果是1集群把75改成66) 交互式开发 执行器选Terminal 密码随便设一个(需记住) 工作空间:私有数据、全部文件 加速器选GeForce_RTX_2080_Ti

AI Toolkit + H100 GPU,一小时内微调最新热门文生图模型 FLUX

上个月,FLUX 席卷了互联网,这并非没有原因。他们声称优于 DALLE 3、Ideogram 和 Stable Diffusion 3 等模型,而这一点已被证明是有依据的。随着越来越多的流行图像生成工具(如 Stable Diffusion Web UI Forge 和 ComyUI)开始支持这些模型,FLUX 在 Stable Diffusion 领域的扩展将会持续下去。 自 FLU

AI基础 L9 Local Search II 局部搜索

Local Beam search 对于当前的所有k个状态,生成它们的所有可能后继状态。 检查生成的后继状态中是否有任何状态是解决方案。 如果所有后继状态都不是解决方案,则从所有后继状态中选择k个最佳状态。 当达到预设的迭代次数或满足某个终止条件时,算法停止。 — Choose k successors randomly, biased towards good ones — Close