Artikel in einem Konferenzbericht,

Deep Learning is not a Matter of Depth but of Good Training

, und .
International Conference on Pattern Recognition and Artificial Intelligence (ICPRAI), Seite 683--687. CENPARMI, Concordia University, (2018)

Zusammenfassung

In the past few years, deep neural networks have often been claimed to provide greater representational power than shallow networks. In this work, we propose a wide, shallow, and strictly sequential network architecture without any residual connections. When trained with cyclical learning rate schedules, this simple network achieves a classification accuracy on CIFAR-100 competitive to a 10 times deeper residual network, while it can be trained 4 times faster. This provides evidence that neither depth nor residual connections are crucial for deep learning. Instead, residual connections just seem to facilitate training using plain SGD by avoiding bad local minima. We believe that our work can hence point the research community to the actual bottleneck of contemporary deep learning: the optimization algorithms.

Tags

Nutzer

  • @bjoern.barz
  • @s364315

Kommentare und Rezensionenanzeigen / verbergen

  • @s364315
    @s364315 vor 3 Jahren
    Reference for Plain11 neural network architecture
Bitte melden Sie sich an um selbst Rezensionen oder Kommentare zu erstellen.