Generalized Bayesian Cramér-Rao Inequality via Information Geometry of Relative α-Entropy

2020 
The relative α-entropy is the Renyi analog of relative entropy and arises prominently in information-theoretic problems. Recent information geometric investigations on this quantity have enabled the generalization of the Cramer-Rao inequality, which provides a lower bound for the variance of an estimator of an escort of the underlying parametric probability distribution. However, this framework remains unexamined in the Bayesian framework. In this paper, we propose a general Riemannian metric based on relative α-entropy to obtain a generalized Bayesian Cramer-Rao inequality. This establishes a lower bound for the variance of an unbiased estimator for the α-escort distribution starting from an unbiased estimator for the underlying distribution. We show that in the limiting case when the entropy order approaches unity, this framework reduces to the conventional Bayesian Cramer-Rao inequality. Further, in the absence of priors, the same framework yields the deterministic Cramer-Rao inequality.
    • Correction
    • Source
    • Cite
    • Save
    • Machine Reading By IdeaReader
    27
    References
    4
    Citations
    NaN
    KQI
    []