Fisher 정보는 어떤 종류의 정보입니까?


29

랜덤 변수 Xf(x|θ) 가 있다고 가정 해 봅시다 . 경우 θ0 실제 파라미터 있었다 상기 우도 함수를 최대화 제로 유도체 같아야한다. 이것이 최대 가능성 추정기의 기본 원리입니다.

내가 알기로 Fisher 정보는 다음과 같이 정의됩니다.

I(θ)=E[(θf(X|θ))2]

따라서 θ0 이 참 매개 변수이면 I(θ)=0 입니다. 그러나 θ0 이 참 매개 변수가 아닌 경우 더 많은 Fisher 정보를 갖게됩니다.

내 질문

  1. Fisher 정보가 주어진 MLE의 "오류"를 측정합니까? 다시 말해서, 긍정적 인 Fisher 정보가 존재한다고 MLE가 이상적이지 않다는 것을 의미하지 않습니까?
  2. "정보"에 대한이 정의는 Shannon이 사용하는 것과 어떻게 다릅니 까? 정보라고 부르는 이유는 무엇입니까?

Why do you write it Eθ? The expectation is over values of X distributed as if they came from your distribution with parameter θ.
Neil G

3
Also I(θ) is not zero at the true parameter.
Neil G

E (S)는 0 (즉, 점수 함수의 기대)이지만 Neil G가 쓴 것처럼 피셔 정보 (V (S))는 (보통) 0이 아닙니다.
Tal Galili

답변:


15

(θ)=logf(x;θ)
as a function of θ for θΘ, the parameter space. Assuming some regularity conditions we do not discuss here, we have Eθ(θ)=Eθ˙(θ)=0 (we will write derivatives with respect to the parameter as dots as here). The variance is the Fisher information
I(θ)=Eθ(˙(θ))2=Eθ¨(θ)
the last formula showing that it is the (negative) curvature of the loglikelihood function. One often finds the maximum likelihood estimator (mle) of θ by solving the likelihood equation ˙(θ)=0 when the Fisher information as the variance of the score ˙(θ) is large, then the solution to that equation will be very sensitive to the data, giving a hope for high precision of the mle. That is confirmed at least asymptotically, the asymptotic variance of the mle being the inverse of Fisher information.

How can we interpret this? (θ) is the likelihood information about the parameter θ from the sample. This can really only be interpreted in a relative sense, like when we use it to compare the plausibilities of two distinct possible parameter values via the likelihood ratio test (θ0)(θ1). The rate of change of the loglikelihood is the score function ˙(θ) tells us how fast the likelihood changes, and its variance I(θ) how much this varies from sample to sample, at a given paramiter value, say θ0. The equation (which is really surprising!)

I(θ)=Eθ¨(θ)
tells us there is a relationsship (equality) between the variability in the information (likelihood) for a given parameter value, θ0, and the curvature of the likelihood function for that parameter value. This is a surprising relationship between the variability (variance) of ths statistic ˙(θ)θ=θ0 and the expected change in likehood when we vary the parameter θ in some interval around θ0 (for the same data). This is really both strange, surprising and powerful!

So what is the likelihood function? We usually think of the statistical model {f(x;θ),θΘ} as a family of probability distributions for data x, indexed by the parameter θ some element in the parameter space Θ. We think of this model as being true if there exist some value θ0Θ such that the data x actually have the probability distribution f(x;θ0). So we get a statistical model by imbedding the true datagenerating probability distribution f(x;θ0) in a family of probability distributions. But, it is clear that such an imbedding can be done in many different ways, and each such imbedding will be a "true" model, and they will give different likelihood functions. And, without such an imbedding, there is no likelihood function. It seems that we really do need some help, some principles for how to choose an imbedding wisely!

So, what does this mean? It means that the choice of likelihood function tells us how we would expect the data to change, if the truth changed a little bit. But, this cannot really be verified by the data, as the data only gives information about the true model function f(x;θ0) which actually generated the data, and not nothing about all the other elements in the choosen model. This way we see that choice of likelihood function is similar to choice of a prior in Bayesian analysis, it injects non-data information into the analysis. Let us look at this in a simple (somewhat artificial) example, and look at the effect of imbedding f(x;θ0) in a model in different ways.

Let us assume that X1,,Xn are iid as N(μ=10,σ2=1). So, that is the true, data-generating distribution. Now, let us embed this in a model in two different ways, model A and model B.

A:X1,,Xn iid N(μ,σ2=1),μRB:X1,,Xn iid N(μ,μ/10),μ>0
you can check that this coincides for μ=10.

The loglikelihood functions become

A(μ)=n2log(2π)12i(xiμ)2B(μ)=n2log(2π)n2log(μ/10)102i(xiμ)2μ

The score functions: (loglikelihood derivatives):

˙A(μ)=n(x¯μ)˙B(μ)=n2μ102i(xiμ)215n
and the curvatures
¨A(μ)=n¨B(μ)=n2μ2+102i2xi2μ3
so, the Fisher information do really depend on the imbedding. Now, we calculate the Fisher information at the true value μ=10,
IA(μ=10)=n,IB(μ=10)=n(1200+20202000)>n
so the Fisher information about the parameter is somewhat larger in model B.

This illustrates that, in some sense, the Fisher information tells us how fast the information from the data about the parameter would have changed if the governing parameter changed in the way postulated by the imbedding in a model family. The explanation of higher information in model B is that our model family B postulates that if the expectation would have increased, then the variance too would have increased. So that, under model B, the sample variance will also carry information about μ, which it will not do under model A.

Also, this example illustrates that we really do need some theory for helping us in how to construct model families.


1
great explanation. Why do you say \Eθ˙(θ)=0? it's a function of θ - isn't it 0 only when evaluated at the true parameter θ0?
ihadanny

1
Yes, what you say is true, @idadanny It is zero when evaluated at the true parameter value.
kjetil b halvorsen

Thanks again @kjetil - so just one more question: is the surprising relationship between the variance of the score and the curvature of the likelihood true for every θ? or only in the neighborhood of the true parameter θ0?
ihadanny

Again, that trelationship is true for the true parameter value. But for that to be of much help, there must be continuity, so that it is approximately true in some neighborhood, since we will use it at the estimated value θ^, not only at the true (unknown) value.
kjetil b halvorsen

so, the relationship holds for the true parameter θ0, it almost holds for θmle since we assume that it's in the neighborhood of θ0, but for a general θ1 it does not hold, right?
ihadanny

31

Let's think in terms of the negative log-likelihood function . The negative score is its gradient with respect to the parameter value. At the true parameter, the score is zero. Otherwise, it gives the direction towards the minimum (or in the case of non-convex , a saddle point or local minimum or maximum).

The Fisher information measures the curvature of around θ if the data follows θ. In other words, it tells you how much wiggling the parameter would affect your log-likelihood.

Consider that you had a big model with millions of parameters. And you had a small thumb drive on which to store your model. How should you prioritize how many bits of each parameter to store? The right answer is to allocate bits according the Fisher information (Rissanen wrote about this). If the Fisher information of a parameter is zero, that parameter doesn't matter.

We call it "information" because the Fisher information measures how much this parameter tells us about the data.


A colloquial way to think about it is this: Suppose the parameters are driving a car, and the data is in the back seat correcting the driver. The annoyingness of the data is the Fisher information. If the data lets the driver drive, the Fisher information is zero; if the data is constantly making corrections, it's big. In this sense, the Fisher information is the amount of information going from the data to the parameters.

Consider what happens if you make the steering wheel more sensitive. This is equivalent to a reparametrization. In that case, the data doesn't want to be so loud for fear of the car oversteering. This kind of reparametrization decreases the Fisher information.


20

Complementary to @NeilG's nice answer (+1) and to address your specific questions:

  1. I would say it counts the "precision" rather than the "error" itself.

Remember that the Hessian of the log-likelihood evaluated at the ML estimates is the observed Fisher information. The estimated standard errors are the square roots of the diagonal elements of the inverse of the observed Fisher information matrix. Stemming from this the Fisher information is the trace of the Fisher information matrix. Given that the Fisher Information matrix I is a Hermitian positive-semidefinite matrix matrix then the diagonal entries Ij,j of it are real and non-negative; as a direct consequence it trace tr(I) must be positive. This means that you can have only "non-ideal" estimators according to your assertion. So no, a positive Fisher information is not related to how ideal is your MLE.

  1. The definition differs in the way we interpreter the notion of information in both cases. Having said that, the two measurements are closely related.

The inverse of Fisher information is the minimum variance of an unbiased estimator (Cramér–Rao bound). In that sense the information matrix indicates how much information about the estimated coefficients is contained in the data. On the contrary the Shannon entropy was taken from thermodynamics. It relates the information content of a particular value of a variable as p·log2(p) where p is the probability of the variable taking on the value. Both are measurements of how "informative" a variable is. In the first case though you judge this information in terms of precision while in the second case in terms of disorder; different sides, same coin! :D

To recap: The inverse of the Fisher information matrix I evaluated at the ML estimator values is the asymptotic or approximate covariance matrix. As this ML estimator values are found in a local minimum graphically the Fisher information shows how deep is that minimum and who much wiggle room you have around it. I found this paper by Lutwak et al. on Extensions of Fisher information and Stam’s inequality an informative read on this matter. The Wikipedia articles on the Fisher Information Metric and on Jensen–Shannon divergence are also good to get you started.

당사 사이트를 사용함과 동시에 당사의 쿠키 정책개인정보 보호정책을 읽고 이해하였음을 인정하는 것으로 간주합니다.
Licensed under cc by-sa 3.0 with attribution required.