Paper Reading: Deep Networks with Stochastic Depth

Written on 2016/08/31

Application:

This simple approach complements the recent success of **residual network ** to reduce training time and improve the test error.

Challenge:

  1. Very deep models become worse at function approximation (called ** degradation ** problem) is not caused by overfitting, but caused by training signals vanishing.
  2. Effective and efficient training methods for very deep models need to be found.

Problem:

Motivated by ** ResNets ** which simplifies ** Highway Networks **, authors proposed a method new called Stochastic Depth to go a step further to reduce ResNet's test error and training time.

Solution:

  1. Shrink the depth of a network during training, while keeping it unchanged during testing.
  2. By a survival probability, randomly dropping entire ResBlocks during training and by bypassing their transformations through skip connections.
  3. Survival probabilities can adopt uniform distribution or linear decay (better)

Insights:

  1. This method(Stochastic depth) is designed for ResNet. Therefore, other networks without ResBlocks is not compatible with this method.
  2. This method can be regarded as an implicit model ensemble.
  3. A new more competitive method has been proposed (http://arxiv.org/pdf/1603.05027.pdf), which can be employed on deeper model and acquire lower test error.

One word to summarize:

This paper proposes a deep network with stochastic depth, a procedure to train very deep neural networks effectively and efficiently.

最后编辑于
©著作权归作者所有,转载或内容合作请联系作者
平台声明:文章内容(如有图片或视频亦包括在内)由作者上传并发布,文章内容仅代表作者本人观点,简书系信息发布平台,仅提供信息存储服务。

推荐阅读更多精彩内容

  • Jimintt阅读 2,880评论 0 0
  • 孩子,今天对簿公堂,并非我所愿,但凡还能有条活路,我都不会为了苟活而选择伤害你。 你是我们最小的孩子,人们常说,八...
    丽子_c15b阅读 2,720评论 0 0
  • 健康是我们最亲密的伙伴。没有健康做伴,啥事都将成为泡沫,某33岁男子,在一家企业做计算机工程,每天精神都要处在高度...
    熏莉阅读 1,573评论 0 0
  • 愚人节的那天,微信收到一条消息。 我喜欢你。 时间显示是在凌晨一点多。 依我对他的了解,应该是鼓足了勇气才说出的这...
    小韩老师a阅读 3,308评论 0 0