C. Wang, and M. Niepert. (2019)cite arxiv:1901.08817Comment: to appear at ICML2019, 20 pages.
Abstract
Recurrent neural networks are a widely used class of neural architectures.
They have, however, two shortcomings. First, it is difficult to understand what
exactly they learn. Second, they tend to work poorly on sequences requiring
long-term memorization, despite having this capacity in principle. We aim to
address both shortcomings with a class of recurrent networks that use a
stochastic state transition mechanism between cell applications. This
mechanism, which we term state-regularization, makes RNNs transition between a
finite set of learnable states. We evaluate state-regularized RNNs on (1)
regular languages for the purpose of automata extraction; (2) nonregular
languages such as balanced parentheses, palindromes, and the copy task where
external memory is required; and (3) real-word sequence learning tasks for
sentiment analysis, visual object recognition, and language modeling. We show
that state-regularization (a) simplifies the extraction of finite state
automata modeling an RNN's state transition dynamics; (b) forces RNNs to
operate more like automata with external memory and less like finite state
machines; (c) makes RNNs have better interpretability and explainability.
%0 Journal Article
%1 wang2019stateregularized
%A Wang, Cheng
%A Niepert, Mathias
%D 2019
%K duality fsm readings recurrent
%T State-Regularized Recurrent Neural Networks
%U http://arxiv.org/abs/1901.08817
%X Recurrent neural networks are a widely used class of neural architectures.
They have, however, two shortcomings. First, it is difficult to understand what
exactly they learn. Second, they tend to work poorly on sequences requiring
long-term memorization, despite having this capacity in principle. We aim to
address both shortcomings with a class of recurrent networks that use a
stochastic state transition mechanism between cell applications. This
mechanism, which we term state-regularization, makes RNNs transition between a
finite set of learnable states. We evaluate state-regularized RNNs on (1)
regular languages for the purpose of automata extraction; (2) nonregular
languages such as balanced parentheses, palindromes, and the copy task where
external memory is required; and (3) real-word sequence learning tasks for
sentiment analysis, visual object recognition, and language modeling. We show
that state-regularization (a) simplifies the extraction of finite state
automata modeling an RNN's state transition dynamics; (b) forces RNNs to
operate more like automata with external memory and less like finite state
machines; (c) makes RNNs have better interpretability and explainability.
@article{wang2019stateregularized,
abstract = {Recurrent neural networks are a widely used class of neural architectures.
They have, however, two shortcomings. First, it is difficult to understand what
exactly they learn. Second, they tend to work poorly on sequences requiring
long-term memorization, despite having this capacity in principle. We aim to
address both shortcomings with a class of recurrent networks that use a
stochastic state transition mechanism between cell applications. This
mechanism, which we term state-regularization, makes RNNs transition between a
finite set of learnable states. We evaluate state-regularized RNNs on (1)
regular languages for the purpose of automata extraction; (2) nonregular
languages such as balanced parentheses, palindromes, and the copy task where
external memory is required; and (3) real-word sequence learning tasks for
sentiment analysis, visual object recognition, and language modeling. We show
that state-regularization (a) simplifies the extraction of finite state
automata modeling an RNN's state transition dynamics; (b) forces RNNs to
operate more like automata with external memory and less like finite state
machines; (c) makes RNNs have better interpretability and explainability.},
added-at = {2019-10-27T19:40:07.000+0100},
author = {Wang, Cheng and Niepert, Mathias},
biburl = {https://www.bibsonomy.org/bibtex/29cf3821264815fd834e08c7078f0857e/kirk86},
description = {[1901.08817] State-Regularized Recurrent Neural Networks},
interhash = {d27a843eae5a8bd1cae21c813e8ddeb3},
intrahash = {9cf3821264815fd834e08c7078f0857e},
keywords = {duality fsm readings recurrent},
note = {cite arxiv:1901.08817Comment: to appear at ICML2019, 20 pages},
timestamp = {2019-10-27T19:40:07.000+0100},
title = {State-Regularized Recurrent Neural Networks},
url = {http://arxiv.org/abs/1901.08817},
year = 2019
}