Graduate Thesis Or Dissertation


State Denoised Recurrent Neural Networks Public Deposited

Downloadable Content

Download PDF
  • We investigate the use of attractor neural networks for denoising the internal states of another neural network, thereby boosting its generalization performance. Denoising is most promising for recurrent sequence-processing networks (i.e. recurrent neural networks), in which noise can accumulate in the hidden states over the elements of a sequence. We call our architecture state- denoised recurrent neural network (SD-RNN). We conduct a series of experiments to demonstrate the benefit of internal denoising, from small experiments like detecting parity of a binary sequence to larger natural language processing data sets. We characterize the behavior of the network using an information theoretic analysis, and we show that internal denoising causes the network to learn better on less data.
Date Issued
  • 2018
Academic Affiliation
Committee Member
Degree Grantor
Commencement Year
Last Modified
  • 2019-11-17
Resource Type
Rights Statement