Graduate Thesis Or Dissertation

 

State Denoised Recurrent Neural Networks Public Deposited

Downloadable Content

Download PDF
https://scholar.colorado.edu/concern/graduate_thesis_or_dissertations/rr171x54d
Abstract
  • We investigate the use of attractor neural networks for denoising the internal states of another neural network, thereby boosting its generalization performance. Denoising is most promising for recurrent sequence-processing networks (i.e. recurrent neural networks), in which noise can accumulate in the hidden states over the elements of a sequence. We call our architecture state- denoised recurrent neural network (SD-RNN). We conduct a series of experiments to demonstrate the benefit of internal denoising, from small experiments like detecting parity of a binary sequence to larger natural language processing data sets. We characterize the behavior of the network using an information theoretic analysis, and we show that internal denoising causes the network to learn better on less data.
Creator
Date Issued
  • 2018
Academic Affiliation
Advisor
Committee Member
Degree Grantor
Commencement Year
Subject
Last Modified
  • 2019-11-17
Resource Type
Rights Statement
Language

Relationships

Items