2018
DOI: 10.1109/tnnls.2017.2733544
|View full text |Cite
|
Sign up to set email alerts
|

Using Directional Fibers to Locate Fixed Points of Recurrent Neural Networks

Abstract: We introduce mathematical objects that we call "directional fibers," and show how they enable a new strategy for systematically locating fixed points in recurrent neural networks. We analyze this approach mathematically and use computer experiments to show that it consistently locates many fixed points in many networks with arbitrary sizes and unconstrained connection weights. Comparison with a traditional method shows that our strategy is competitive and complementary, often finding larger and distinct sets o… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
3
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 9 publications
(3 citation statements)
references
References 18 publications
0
3
0
Order By: Relevance
“…These fixed points, along with local linearizations of the RNN dynamics about those fixed points, provide a mechanistic description of how the network implements a computation. Identifying the fixed points of an RNN requires optimization tools that have previously been tailored for specific, "vanilla" or Hopfield-like RNN architectures (Sussillo & Barak, 2013, Katz & Reggia (2018). These approaches rely on hard-coded analytic derivatives (e.g., of the hyperbolic tangent function) and thus can be cumbersome for complicated models (e.g., gated architectures), for studies that involve multiple different models, or when frequent architectural changes are required during the development of a model.…”
mentioning
confidence: 99%
“…These fixed points, along with local linearizations of the RNN dynamics about those fixed points, provide a mechanistic description of how the network implements a computation. Identifying the fixed points of an RNN requires optimization tools that have previously been tailored for specific, "vanilla" or Hopfield-like RNN architectures (Sussillo & Barak, 2013, Katz & Reggia (2018). These approaches rely on hard-coded analytic derivatives (e.g., of the hyperbolic tangent function) and thus can be cumbersome for complicated models (e.g., gated architectures), for studies that involve multiple different models, or when frequent architectural changes are required during the development of a model.…”
mentioning
confidence: 99%
“…The optimisation algorithm we have used to find fixed points is based on Sussillo and Barak [55]; see [16] for an open-source Tensorflow toolbox for finding fixed points in arbitrary RNN architectures and [25] for an alternative method to identify fixed points. The key idea is to define a scalar function whose minima correspond to fixed points of the ESN dynamics.…”
Section: Finding Fixed Points Of the Dynamicsmentioning
confidence: 99%
“…This viewpoint has already been used to understand the difficulties for RNNs to capture longer time dependencies (Bengio, Frasconi, & Simard, 1993;Doya, 1993). In particular, recent work has highlighted the important role played by fixed points in RNN state spaces, that are defined as hidden states that updates to themself for a given input (Katz & Reggia, 2017;Sussillo & Barak, 2013). This line of work has argued that locating such fixed points efficiently could provide insights into RNN dynamics and input-output properties.…”
Section: Introductionmentioning
confidence: 99%