The Promise of Deep Learning

By Yoshua Bengio

Humans have long dreamed of creating machines that think. More than 100 years before the first programmable computer was built, inventors wondered whether devices made of rods and gears might become intelligent. And when Alan Turing, one of the pioneers of computing in the 1940s, set a goal for computer science, he described a test, later dubbed the Turing Test, which measured a computer’s performance against the behavior of humans.

In the early days of my academic field, artificial intelligence, scientists tackled problems that were difficult for humans but relatively easy for computers–such as large-scale mathematical calculations. In more recent years, we’re taking on tasks that are easy for people to perform but hard to describe to a machine–tasks humans solve “without thinking,” such as recognizing spoken words or faces in a crowd.

That more difficult quest gave rise to the domain of machine learning, the ability of machines to learn. This is what interests me. It’s not really my goal to make machines that think like humans do. My aim is to understand the fundamental principles that may enable an entity, machine or living being, to be intelligent. I have long ago made the bet that this would happen thanks to the ability of such an entity to learn, and my focus is on building machines that can learn and understand the world by themselves, i.e., learn to make sense of it.

The reason I’m laying out this chronology is that I believe we’re at a turning point in the history of artificial intelligence–and, indeed, computing itself. Thanks to more powerful computers, the availability of large and varied datasets, and advances in algorithms, we’re able to cross a threshold that has long held back computer science. Machine learning is shifting from a highly manual process where humans have had to design good representations for each task of interest into an automated process where machines learn more like babies do — through experience –  building internal representations that help to make sense of the world. This is the field of deep learning.

Deep learning isn’t brand new. Indeed, when I was a student in the 1980s, it was the concept of neural networks, the precursor of deep learning, that got me interested in pursuing an academic career in computer science. What’s new is that the accumulation of many scientific and technical advances has yielded breakthroughs in AI applications such as speech recognition, computer vision, and natural language processing.  This has brought into the field a large group of researchers, mostly graduate students, and we’re now making progress in deep learning at a gallop.

We’re able to do that because of advances in creating hierarchies of concepts and representations that computers discover by themselves. The hierarchies allow a computer to learn complicated concepts by building them out of simpler ones.  This is also how humans learn and build their understanding of the world; they gradually refine their model of the world to better fit what they observe and discover new ideas from the composition of older ones, new ideas that help them to better fit the evidence, the data.

For example, a deep learning system can represent the concept of an image of a cat by combining simpler concepts, such as corners and contours, which are in turn defined in terms of edges. But we don’t have to teach it explicitly about these intermediate concepts, it learns them on its own. We don’t have to show the system pictures of all the possible cat colors, shapes, and behaviors for such object recognition systems to correctly identify that it is a Siamese cat that’s somersaulting in a photograph. When it “sees” a cat, it “knows” it is one.

I’m privileged to be part of a troika of computer scientists who are widely credited with spearheading advances in this field–along with Geoffrey Hinton and Yann LeCun. We co-authored a paper, Deep Learning,which was published in the journal Nature in May, where we laid out the promise of our branch of A.I. But this isn’t a field where a few “media stars” are doing all that needs to be done. To produce the advances that are possible and to find applications for them will require thousands of scientists and engineers–in academia and in industry.

That’s why I’ve been dedicated to rallying people to our exciting project. I’m co-authoring a book, Deep Learning, with Ian Goodfellow and Aaron Courville. Our core audiences are university students studying machine learning and software engineers working in a wide variety of industries that are likely to find important uses for it. This book-in-progress is posted on the Web, and we welcome people to read, learn and give us feedback.

Which brings me to another key point: I’m an advocate of open science. Like open source developers, participants in the open science movement believe that we should share knowledge as soon as we gain it to increase the pace at which the boundaries of science are pushed, and for the benefit of all. Many of my research colleagues and I contribute all of our deep learning inventions to the Theano project and its derivatives on GitHub. There, anybody who is building deep learning systems can use the algorithms and programming tools, and we urge them to contribute back to the project: hundreds already do so.

Just as sharing is essential to open science, so is collaboration–the kind that’s done transparently. The whole enterprise of science is a giant brainstorm. The Montreal Institute for Learning Algorithms (MILA), with its 60 researchers — including 5 professors, contributes to it via numerous collaborative research projects with scientists in universities and industry.

The newest of our collaborative research partners is IBM. We look forward to working with scientists and engineers in IBM Research and the Watson Group on a very ambitious research agenda, including deep learning for language, speech and vision. We believe that, together, we’ll be able to scale up and extend deep learning methods by using powerful computers to take on very large datasets. It will help machines learn more, across broader domains, faster and from a larger set of data sources, including the vast amounts of unlabeled data – that have not been curated by humans.

I’m tremendously excited about the future of deep learning. We’ve made rapid progress, and while we’re far from solving the great riddle of what it will take to enable machines to truly understand the world, I’m very hopeful that we’ll crack it.

And then the floodgates will open. Once computers truly understand text, speech, images and sounds, they will become our indispensible assistants. This will revolutionize the way we interact with computers, helping us live more conveniently in our day-to-day lives and perform more effectively at work. It will enable society to take on some of the grand challenges that matter to us–such as curing deadly diseases and spreading knowledge and wealth more broadly. As importantly, it will help us understand who we are and that part of who we are that has always fascinated me, i.e., how intelligence arises. This has been my dream for more than 30 years, and it’s fast becoming our reality.

The Promise of Deep Learning的更多相关文章

  1. (zhuan) Where can I start with Deep Learning?

    Where can I start with Deep Learning? By Rotek Song, Deep Reinforcement Learning/Robotics/Computer V ...

  2. (转) The major advancements in Deep Learning in 2016

    The major advancements in Deep Learning in 2016 Pablo Tue, Dec 6, 2016 in MACHINE LEARNING DEEP LEAR ...

  3. (转) Deep Learning in a Nutshell: Reinforcement Learning

    Deep Learning in a Nutshell: Reinforcement Learning   Share: Posted on September 8, 2016by Tim Dettm ...

  4. Rolling in the Deep (Learning)

    Rolling in the Deep (Learning) Deep Learning has been getting a lot of press lately, and is one of t ...

  5. 深度学习材料:从感知机到深度网络A Deep Learning Tutorial: From Perceptrons to Deep Networks

    In recent years, there’s been a resurgence in the field of Artificial Intelligence. It’s spread beyo ...

  6. [C3] Andrew Ng - Neural Networks and Deep Learning

    About this Course If you want to break into cutting-edge AI, this course will help you do so. Deep l ...

  7. Deep learning:五十一(CNN的反向求导及练习)

    前言: CNN作为DL中最成功的模型之一,有必要对其更进一步研究它.虽然在前面的博文Stacked CNN简单介绍中有大概介绍过CNN的使用,不过那是有个前提的:CNN中的参数必须已提前学习好.而本文 ...

  8. 【深度学习Deep Learning】资料大全

    最近在学深度学习相关的东西,在网上搜集到了一些不错的资料,现在汇总一下: Free Online Books  by Yoshua Bengio, Ian Goodfellow and Aaron C ...

  9. 《Neural Network and Deep Learning》_chapter4

    <Neural Network and Deep Learning>_chapter4: A visual proof that neural nets can compute any f ...

随机推荐

  1. javascript笔记08:javascript的if...else语句

    案例代码如下: <!DOCTYPE html> <html> <body> <p>点击这个按钮,获得基于时间的问候.</p> <but ...

  2. Java RMI详解

    RMI:远程方法调用(Remote Method Invocation).能够让在某个java虚拟机上的对象像调用本地对象一样调用另一个java 虚拟机中的对象上的方法. RMI远程调用步骤: 1,客 ...

  3. 简单的实现QQ通信功能(四)

    第四部分:主界面的设计及代码 一:效果图及界面设计 1. 效果图: 2. 界面设计: (1)上面显示自己信息用一个PictureBox和两个Label,用来显示自己的头像和昵称备注名. (2)下面用了 ...

  4. LoadRunner安装包(性能测试工具分享)

    今天在测试课指导老师朱香元的指导下,开始了测试软件的安装,下面我分享一下整个安装流程,最后我会附带安装包的 第一步:链接:http://pan.baidu.com/s/1pXqk2 密码:csjk , ...

  5. UpdateProgress使用

    UpdateProgress是一个进度显示条,加在AJAX里能显得更加的人性化(个人认为).现在我们就开始吧: 第一.新建一个AJAX项目.在页面上加上ScriptManager,UpdatePane ...

  6. 2.redis.3.2 下载,安装、配置、使用 - 2

    上篇简单介绍了 下载,安装,测试,现在直接使用了,看结果 使用的redis服务便是,上篇临时搭建的简易服务,,注意,说的是简易,因为它只是一个单点的“玩具”: 临时在项目登录的时候模拟了一下,这里使用 ...

  7. 企业级搜索引擎Solr使用入门指南

    由于搜索引擎功能在门户社区中对提高用户体验有着重在门户社区中涉及大量需要搜索引擎的功能需求,目前在实现搜索引擎的方案上有集中方案可供选择: 基于Lucene自己进行封装实现站内搜索. 工作量及扩展性都 ...

  8. SQL的update from 理解

    学习了sql的语句都有快3年,工作上使用都一年半的,最近突然想起update from语句,感觉好像很模糊,虽然语法上使用一直正确,一直都是这样使用,但是就好像不是很明白里面的深处意思. 今天特意测试 ...

  9. 控制器的跳转-modal与push

    一.modal与pushmodal从下面往上盖住原来的控制器,一般上一个控制器和下一个控制器没有什么关联时用modal,比如联系人的加号跳转页面,任何控制器都可以用modal push一般是上下文有关 ...

  10. redis基础-前篇

    设置键值 #设置值 set key value #获取值 get key 设置自增 #自增1 incr num #指定增长跨度 incrby num 10 设置自减 #自增1 decr num #指定 ...