Skip to main content

Research Repository

Advanced Search

Train Your Classifier First: Cascade Neural Networks Training from Upper Layers to Lower Layers

Zhang, Shucong; Do, Cong-Thanh; Doddipatla, Rama; Loweimi, Erfan; Bell, Peter; Renals, Steve

Authors

Shucong Zhang

Cong-Thanh Do

Rama Doddipatla

Erfan Loweimi

Peter Bell

Steve Renals



Abstract

Although the lower layers of a deep neural network learn features which are transferable across datasets, these layers are not transferable within the same dataset. That is, in general, freezing the trained feature extractor (the lower layers) and retraining the classifier (the upper layers) on the same dataset leads to worse performance. In this paper, for the first time, we show that the frozen classifier is transferable within the same dataset. We develop a novel top-down training method which can be viewed as an algorithm for searching for high-quality classifiers. We tested this method on automatic speech recognition (ASR) tasks and language modelling tasks. The proposed method consistently improves recurrent neural network ASR models on Wall Street Journal, self-attention ASR models on Switchboard, and AWD-LSTM language models on WikiText-2.

Presentation Conference Type Conference Paper (Published)
Conference Name ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)
Start Date Jun 6, 2021
End Date Jun 11, 2021
Online Publication Date May 13, 2021
Publication Date 2021
Deposit Date Apr 3, 2024
Publisher Institute of Electrical and Electronics Engineers
Series ISSN 2379-190X
Book Title ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)
DOI https://doi.org/10.1109/icassp39728.2021.9413565
Public URL http://researchrepository.napier.ac.uk/Output/3585855