A statistical method is applied to explore the unique characteristics of a certain class of neural network autoassociative memory with N neurons and first-order synaptic interconnections. The memory matrix is constructed to store M = αN vectors based on the outer-product learning algorithm. We theoretically prove that, by setting all the diagonal terms of the memory matrix to be M and letting the input error ratio ρ = 0, the probability of successful recall Pr steadily decreases as α increases, but as α increases past 1.0, Pr begins to increase slowly. When 0 < ρ ≤ 0.5, the network exhibits strong error-correction capability if α ≤ 0.15 and this capability is shown to rapidly decrease as α increases. The network essentially loses all its error-correction capability at α = 2, regardless of the value of ρ. When 0 < ρ ≤ 0.5, and under the constraint of Pr > 0.99, the tradeoff between the number of stable states and their attraction force is analyzed and the memory capacity is shown to be 0.15N at best.

This content is only available as a PDF.
You do not currently have access to this content.