Jump to content

Talk:Hopfield network: Difference between revisions

Page contents not supported in other languages.
From Wikipedia, the free encyclopedia
Content deleted Content added
Amshali (talk | contribs)
energy
Line 44: Line 44:
I T.A a neural networks course...
I T.A a neural networks course...
you can easily see this be derivating E w.r.t S_j to get h_j
you can easily see this be derivating E w.r.t S_j to get h_j

== energy ==

I have an argument on energy function. Sometimes threshold is a more complicated function and we cannot easily incorporate it into Energy function. I mean as I have seen in [http://www.shef.ac.uk/psychology/gurney/notes/l5/l5.html "Associative memories - the Hopfield net"], it should not contain this term:
:<math>\sum_i{\theta_i\ s_i}</math>
Am I right?

Revision as of 07:04, 26 June 2007

Further Correction - symmetric weights

My error - I misread the context of his statement. The condition of symmetric weights guarantees that following the update rule makes energy a monotonically decreasing function, which guarantees convergence to local minima, however, non-symmetric weights do not seem to impare the use of the network as a content-addressable memory system.

Correction - connections need not be symmetric!

If you refer to the origina Hopfield paper ( citied at the bottom of the page ) he discusses the performance of networks with the "special case" of symmetric weights, but says that the network performs just as well with non-symmetric weights. Specifically he says: "The flow in phase space produced by this model algorithm has the properties necessary for a content-addressable memory whether or not Tij is symmetric" (Hopfield, 1982, p. 2556)

Connection between Hopfield Net and Ising model?

Hello!

I had some classes this week which involved the definitions of Hopfield networks and Ising model, and came here to look for further information/links.

There is a link in this article to Ising model, but nothing is written in the article body that explains the connections between the two concepts, maybe someone could fill that gap in?

(I'll try after I've studied enough to understand the connection myself).

Cheers


Definitions?

The relation between the a[i]'s and the s[i]'s is not clear. Are the a[i]'s just the updated values of the s[i]'s? In that case, why not call them both s[i]?

Another terminological matter: The article says

Hopfield nets can either have units that take on values of 1 or -1, or units that take on values of 1 or 0.

and goes on to give the updating rules in the two cases. This seems like to much attention to a trivial matter of scaling. I would suggest choosing one convention or the other for the article and then mentioning that the other convention is also used. --Macrakis 16:21, 15 August 2006 (UTC)[reply]

Energy formula

Currently energy is written as:

I feel this is incorrect. Either removing 1/2

or summing over all and

would fix the problem. But I'm not so confident to modify the main text. I'd appreciate if somebody could check it. -- i agree, and i've changed it, (before looking here) I T.A a neural networks course... you can easily see this be derivating E w.r.t S_j to get h_j

energy

I have an argument on energy function. Sometimes threshold is a more complicated function and we cannot easily incorporate it into Energy function. I mean as I have seen in "Associative memories - the Hopfield net", it should not contain this term:

Am I right?