全部版块 我的主页
论坛 计量经济学与统计论坛 五区 计量经济学与统计软件 LATEX论坛
1090 2
2016-08-12

Last week, deep learning research leader Yann LeCun took part in a Quora Session, during which he answered questions from community members on a wide variety of (mostly machine/deep learning) topics.


But... what does Yann LeCun think he does?

During the session, this question was posed:

When will we see a theoretical background and mathematical foundation for deep learning?

The answer turned into a very eloquent overview of three particular thoughts on why deep learning works so well. Here is a quick overview.

LeCun's first point of explanation, which maps to a good reason why deep learning works so well, is as follows:

One theoretical puzzle is why the type of non-convex optimization that needs to be done when training deep neural nets seems to work reliably.

The main idea here is that local minima do not arise in very high dimensional space, so greedy-search gradient optimization is not trapped in a "box." As LeCun states:

It’s hard to build a box in 100 million dimensions.

Moving on, LeCun introduces his next point as:

Another interesting theoretical question is why multiple layers help.

The point here, beyond LeCun stating that there is not a complete understanding as to why, is that multiple layers help to implement complex functions more concisely. While he points out that computer scientists are accustomed to the idea of sequential steps and multiple layers of computation, this doesn't quite cover the reasons why multiple layers in deep neural networks work as they do.

For his last point, he turns to a specific neural network architecture.

A third interesting question is why ConvNets work so well.

Interesting question indeed. This article gets cited as reading for why ConvNet architectures are right for analyzing certain types of signals, touching on the fact that ConvNets actually work very well for some type of signals, like spatial. Sorting out why is this is so is one thing, but noting that it is true places squarely into perspective just how well ConvNets work when they do.

While Yann LeCun was answering the question posed to him on when we could expect a mathematical foundation for deep leaning to emerge, in doing so he provided valuable insight into why deep learning functions as well as it does.


二维码

扫码加我 拉你入群

请注明:姓名-公司-职位

以便审核进群资格,未注明则拒绝

全部回复
2016-8-12 11:18:10
谢谢分享
二维码

扫码加我 拉你入群

请注明:姓名-公司-职位

以便审核进群资格,未注明则拒绝

2016-8-12 13:41:47
谢谢分享
二维码

扫码加我 拉你入群

请注明:姓名-公司-职位

以便审核进群资格,未注明则拒绝

相关推荐
栏目导航
热门文章
推荐文章

说点什么

分享

扫码加好友,拉您进群
各岗位、行业、专业交流群