In statistics and signal processing, the orthogonality principle is a necessary and sufficient condition for the optimality of a Bayesian estimator. Loosely stated, the orthogonality principle says that the error vector of the optimal estimator (in a mean square error sense) is orthogonal to any possible estimator. The orthogonality principle is most commonly stated for linear estimators, but more general formulations are possible. Since the principle is a necessary and sufficient condition for optimality, it can be used to find the minimum mean square error estimator.
Attributes | Values |
---|---|
rdf:type | |
rdfs:label |
|
rdfs:comment |
|
foaf:depiction | |
dct:subject | |
Wikipage page ID |
|
Wikipage revision ID |
|
Link from a Wikipage to another Wikipage |
|
sameAs | |
dbp:wikiPageUsesTemplate | |
thumbnail | |
has abstract |
|
gold:hypernym | |
prov:wasDerivedFrom | |
page length (characters) of wiki page |
|
foaf:isPrimaryTopicOf | |
is Link from a Wikipage to another Wikipage of | |
is Wikipage redirect of | |
is foaf:primaryTopic of |