You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: content/constraint-satisfaction.md
+5-9Lines changed: 5 additions & 9 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -5,19 +5,15 @@ bibfile = "ccnlab.json"
5
5
6
6
**Constraint satisfaction** is one of the most important concepts for understanding the power of [[bidirectional connectivity]] in the [[neocortex]], and is central to the promise of the [[Axon]] approach, which is one of the few neural network models to incorporate extensive bidirectional connectivity in a way that tames its wild side while retaining its many benefits.
7
7
8
-
The constraint satisfaction problem (CSP) is defined as finding a set of values for N variables that satisfy a set of constraints defined over those variables ([[@Tsang14]]). A classic example is the N-queens problem, where you need to place N chess queens on a board such that no two queens should threaten each other. Another such problem is the _travelling salesman problem_ (TSP), analyzed by [[@^HopfieldTank85]] using a bidirectionally-connected Hopfield network, which involves finding a minimum distance route between _N_ cities.
8
+
{id="figure_hopfield" style="height:30em"}
9
+

9
10
10
-
Thus, the CSP is essentially a problem of [[search]]ing over all possible states to find the one(s) that best fit the set of constraints imposed. As the number of states increases, the number of possible states explodes exponentially due to the [[curse of dimensionality]],
11
-
12
-
A bidirectionally-connected neural network can implement this search process in a highly efficient manner, by integrating all of the constraints in a single gradient-based computational step over _dedicated-parallel_ representations, effectively performing a _stochastic gradient descent_ process over possible solution states. This is essentially the same strategy used in [[error-backpropagation]] learning, to search over the high-dimensional space of possible representations, as explained in [[search]].
11
+
The constraint satisfaction problem (CSP) is defined as finding a set of values for N variables that satisfy a set of constraints defined over those variables ([[@Tsang14]]). A classic example is the N-queens problem, where you need to place N chess queens on a board such that no two queens should threaten each other. Another such problem is the _travelling salesman problem_ (TSP), analyzed by [[@^HopfieldTank85]] using a bidirectionally-connected Hopfield network, which involves finding a minimum distance route between _N_ cities ([[#figure_hopfield]]).
13
12
14
-
<!--- TODO: old: -->
15
-
16
-
In the context of the [[curse of dimensionality]] problem, bidirectional connectivity enables the network to rapidly converge through parallel processing on a set of representations that satisfy constraints communicated throughout the network, where each active neuron contributes its own constraint, and is in turn subject to the constraints from the other neurons. Given the _small world_ nature of neural connectivity, each neuron is effectively only a few synapses away from any other neuron, so effectively any constraint anywhere can be felt anywhere else in the network, in principle.
13
+
Thus, the CSP is essentially a problem of [[search]]ing over all possible states to find the one(s) that best fit the set of constraints imposed. As the number of states increases, the number of possible states explodes exponentially due to the [[curse of dimensionality]],
17
14
18
-
This parallel, distributed constraint satisfaction process is effectively performing a massive search over the entire space of possible ways of representing the current combination of external and internal constraints, in a way that would otherwise be impossible in any kind of more serial process, due to the combinatorial explosion of possible such representations.
15
+
A bidirectionally-connected neural network can implement this search process in a highly efficient manner, by integrating all of the constraints in a single gradient-based computational step over _dedicated-parallel_ representations, effectively performing a _stochastic gradient descent_ process over possible solution states. This is essentially the same strategy used in [[error-backpropagation]] learning, to search over the high-dimensional space of possible representations, as explained in [[search]]. Mathematically, it is effectively a process of [[error backpropagation#backpropagation to activations]].
19
16
20
17
Purely feedforward networks do not adapt their representations dynamically as they process the current set of inputs, and instead just generate a representation in one sweep, based on the learned weights. Thus, they are not optimizing these representations to find the most _satisfying_ way of interpreting the current situation. By contrast, the iterative back-and-forth interactions among bidirectionally-connected neurons ends up optimizing the active representation, which then provides the basis for subsequent learning.
21
18
22
-
Mathematically, it is effectively a process of [[error backpropagation#backpropagation to activations]], where the error gradient that is otherwise used to drive learning is actually used to drive the activation state of the network, and then this activation state, via the [[temporal derivative]] learning principle, drives learning according to this gradient.
Copy file name to clipboardExpand all lines: content/references.md
+2Lines changed: 2 additions & 0 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1044,6 +1044,8 @@
1044
1044
1045
1045
<pid="ParnaudeauBolkanKellendonk18">Parnaudeau, S., Bolkan, S.S., & Kellendonk, C. (2018). The mediodorsal thalamus: An essential partner of the prefrontal cortex for cognition. <i>Biological Psychiatry, 83</i>, 648-656. <ahref="http://www.sciencedirect.com/science/article/pii/S0006322317321935">http://www.sciencedirect.com/science/article/pii/S0006322317321935</a><ahref="http://doi.org/10.1016/j.biopsych.2017.11.008"> http://doi.org/10.1016/j.biopsych.2017.11.008</a></p>
1046
1046
1047
+
<pid="Pashler94">Pashler, H. (1994). Dual-task interference in simple tasks: data and theory. <i>Psychological bulletin, 116</i>, 220-244. <ahref="http://www.ncbi.nlm.nih.gov/pubmed/7972591">http://www.ncbi.nlm.nih.gov/pubmed/7972591</a></p>
1048
+
1047
1049
<pid="PateriaSubagdjaTanEtAl21">Pateria, S., Subagdja, B., Tan, A., & Quek, C. (2021). Hierarchical Reinforcement Learning: A Comprehensive Survey. <i>ACM Comput. Surv., 54</i>, 109:1–109:35. <ahref="https://dl.acm.org/doi/10.1145/3453160">https://dl.acm.org/doi/10.1145/3453160</a><ahref="http://doi.org/10.1145/3453160"> http://doi.org/10.1145/3453160</a></p>
1048
1050
1049
1051
<pid="PauliHazyOReilly12">Pauli, W.M., Hazy, T.E., & O'Reilly, R.C. (2012). Expectancy, ambiguity, and behavioral flexibility: separable and complementary roles of the orbital frontal cortex and amygdala in processing reward expectancies. <i>Journal of Cognitive Neuroscience, 24</i>, 351-366. <ahref="http://www.ncbi.nlm.nih.gov/pubmed/22004047">http://www.ncbi.nlm.nih.gov/pubmed/22004047</a></p>
0 commit comments