Search engine for discovering works of Art, research articles, and books related to Art and Culture
ShareThis
Javascript must be enabled to continue!

Architectural Bias in Recurrent Neural Networks: Fractal Analysis

View through CrossRef
We have recently shown that when initialized with “small” weights, recurrent neural networks (RNNs) with standard sigmoid-type activation functions are inherently biased toward Markov models; even prior to any training, RNN dynamics can be readily used to extract finite memory machines (Hammer & Tiňo, 2002; Tiňo, Čerňanský, &Beňušková, 2002a, 2002b). Following Christiansen and Chater (1999), we refer to this phenomenon as the architectural bias of RNNs. In this article, we extend our work on the architectural bias in RNNs by performing a rigorous fractal analysis of recurrent activation patterns. We assume the network is driven by sequences obtained by traversing an underlying finite-state transition diagram&a scenario that has been frequently considered in the past, for example, when studying RNN-based learning and implementation of regular grammars and finite-state transducers. We obtain lower and upper bounds on various types of fractal dimensions, such as box counting and Hausdorff dimensions. It turns out that not only can the recurrent activations inside RNNs with small initial weights be explored to build Markovian predictive models, but also the activations form fractal clusters, the dimension of which can be bounded by the scaled entropy of the underlying driving source. The scaling factors are fixed and are given by the RNN parameters.
Title: Architectural Bias in Recurrent Neural Networks: Fractal Analysis
Description:
We have recently shown that when initialized with “small” weights, recurrent neural networks (RNNs) with standard sigmoid-type activation functions are inherently biased toward Markov models; even prior to any training, RNN dynamics can be readily used to extract finite memory machines (Hammer & Tiňo, 2002; Tiňo, Čerňanský, &Beňušková, 2002a, 2002b).
Following Christiansen and Chater (1999), we refer to this phenomenon as the architectural bias of RNNs.
In this article, we extend our work on the architectural bias in RNNs by performing a rigorous fractal analysis of recurrent activation patterns.
We assume the network is driven by sequences obtained by traversing an underlying finite-state transition diagram&a scenario that has been frequently considered in the past, for example, when studying RNN-based learning and implementation of regular grammars and finite-state transducers.
We obtain lower and upper bounds on various types of fractal dimensions, such as box counting and Hausdorff dimensions.
It turns out that not only can the recurrent activations inside RNNs with small initial weights be explored to build Markovian predictive models, but also the activations form fractal clusters, the dimension of which can be bounded by the scaled entropy of the underlying driving source.
The scaling factors are fixed and are given by the RNN parameters.

Related Results

Bicomplex Projection Rule for Complex-Valued Hopfield Neural Networks
Bicomplex Projection Rule for Complex-Valued Hopfield Neural Networks
A complex-valued Hopfield neural network (CHNN) with a multistate activation function is a multistate model of neural associative memory. The weight parameters need a lot of memory...
An Introduction to Fractal Dynamics
An Introduction to Fractal Dynamics
Fractal dynamics is a unique, systems-based approach of looking at and thinking about organizations. In this view, the organization is viewed as part of a pervasive organizational ...
Noise Robust Projection Rule for Klein Hopfield Neural Networks
Noise Robust Projection Rule for Klein Hopfield Neural Networks
Multistate Hopfield models, such as complex-valued Hopfield neural networks (CHNNs), have been used as multistate neural associative memories. Quaternion-valued Hopfield neural net...
Nudge and bias in subjective ratings? The role of icon sets in determining ratings of icon characteristics
Nudge and bias in subjective ratings? The role of icon sets in determining ratings of icon characteristics
AbstractSubjective ratings have been central to the evaluation of icon characteristics. The current study examined biases in ratings in relation to the context in which icons are p...
Synaptic Dynamics in Analog VLSI
Synaptic Dynamics in Analog VLSI
Synapses are crucial elements for computation and information transfer in both real and artificial neural systems. Recent experimental findings and theoretical models of pulse-base...
Redundancy-Aware Pruning of Convolutional Neural Networks
Redundancy-Aware Pruning of Convolutional Neural Networks
Pruning is an effective way to slim and speed up convolutional neural networks. Generally previous work directly pruned neural networks in the original feature space without consid...
Adaptive Integration in the Visual Cortex by Depressing Recurrent Cortical Circuits
Adaptive Integration in the Visual Cortex by Depressing Recurrent Cortical Circuits
Neurons in the visual cortex receive a large amount of input from recurrent connections, yet the functional role of these connections remains unclear. Here we explore networks with...
Digital cosmopoiesis in architectural pedagogy: An analysis through Frascari
Digital cosmopoiesis in architectural pedagogy: An analysis through Frascari
Abstract This article derives from three observations of architectural drawing: the current ubiquitousness of digitization, the ongoing disputation of digitization in architectural...

Recent Results

Bowl
Bowl
Stoneware with natural ash glaze around inside walls (Tokoname ware), Kamakura period (1185–1333), Japan...
Anglo-Saxon art
Anglo-Saxon art
David M. Wilson, Anglo-Saxons, 1984, Overlook Press...
Mitoraj, urok Gorgony
Mitoraj, urok Gorgony
Agnieszka Dębska, Criticism and interpretation, 2003, Muzeum Narodowe w Warszawie...

Back to Top