Oja's learning rule, or simply Oja's rule, named after Finnish computer scientist Erkki Oja, is a model of how neurons in the brain or in artificial neural networks change connection strength, or learn, over time. It is a modification of the standard Hebb's Rule (see Hebbian learning) that, through multiplicative normalization, solves all stability problems and generates an algorithm for principal components analysis. This is a computational form of an effect which is believed to happen in biological neurons.
Attributes | Values |
---|---|
rdf:type | |
rdfs:label |
|
rdfs:comment |
|
dcterms:subject | |
Wikipage page ID |
|
Wikipage revision ID |
|
Link from a Wikipage to another Wikipage |
|
Link from a Wikipage to an external page | |
sameAs | |
dbp:wikiPageUsesTemplate | |
has abstract |
|
gold:hypernym | |
prov:wasDerivedFrom | |
page length (characters) of wiki page |
|
foaf:isPrimaryTopicOf | |
is Link from a Wikipage to another Wikipage of | |
is Wikipage disambiguates of | |
is known for of | |
is known for of | |
is foaf:primaryTopic of |