Input Fast-Forwarding for Better Deep Learning
Ahmed Ibrahim, A. Lynn Abbott
Abstract
This paper introduces a new architectural framework, known as input fast-forwarding, that can enhance the performance of deep networks. The main idea is to incorporate a parallel path that sends representations of input values forward to deeper network layers. This scheme is substantially different from “deep supervision”, in which the loss layer is re-introduced to earlier layers. The parallel path provided by fast-forwarding enhances the training process in two ways. First, it enables the individual layers to combine higher-level information (from the standard processing path) with lower-level information (from the fast-forward path). Second, this new architecture reduces the problem of vanishing gradients substantially because the fast-forwarding path provides a shorter route for gradient backpropagation. In order to evaluate the utility of the proposed technique, a Fast-Forward Network (FFNet), with 20 convolutional layers along with parallel fast-forward paths, has been created and tested. The paper presents empirical results that demonstrate improved learning capacity of FFNet due to fast-forwarding, as compared to GoogLeNet (with deep supervision) and CaffeNet, which are 4× and 18x
larger in size, respectively.
Ahmed Ibrahim, A. Lynn Abbott, Mohamed E. Hussein: Input Fast-Forwarding for Better Deep Learning. ICIAR 2017: 363-370
People
Publication Details
- Date of publication:
- June 2, 2017
- Conference:
- International Conference on Image Analysis and Recognition
- Page number(s):
- 363-370