Graduate Thesis Or Dissertation
State Denoised Recurrent Neural Networks Public Deposited
Downloadable Content
Download PDF
https://scholar.colorado.edu/concern/graduate_thesis_or_dissertations/rr171x54d
- Abstract
- We investigate the use of attractor neural networks for denoising the internal states of another neural network, thereby boosting its generalization performance. Denoising is most promising for recurrent sequence-processing networks (i.e. recurrent neural networks), in which noise can accumulate in the hidden states over the elements of a sequence. We call our architecture state- denoised recurrent neural network (SD-RNN). We conduct a series of experiments to demonstrate the benefit of internal denoising, from small experiments like detecting parity of a binary sequence to larger natural language processing data sets. We characterize the behavior of the network using an information theoretic analysis, and we show that internal denoising causes the network to learn better on less data.
- Creator
- Date Issued
- 2018
- Academic Affiliation
- Advisor
- Committee Member
- Degree Grantor
- Commencement Year
- Subject
- Last Modified
- 2019-11-17
- Resource Type
- Rights Statement
- Language
Relationships
Items
Thumbnail | Title | Date Uploaded | Visibility | Actions |
---|---|---|---|---|
stateDenoisedRecurrentNeuralNetworks.pdf | 2019-11-17 | Public | Download |