TY - JOUR
T1 - Geometry of EM and related iterative algorithms
AU - Hino, Hideitsu
AU - Akaho, Shotaro
AU - Murata, Noboru
N1 - Publisher Copyright:
© 2022, The Author(s), under exclusive licence to Springer Nature Singapore Pte Ltd.
PY - 2023/12
Y1 - 2023/12
N2 - The Expectation–Maximization (EM) algorithm is a simple meta-algorithm that has been used for many years as a methodology for statistical inference when there are missing measurements in the observed data or when the data is composed of observables and unobservables. Its general properties are well studied, and also, there are countless ways to apply it to individual problems. In this paper, we introduce the em algorithm, an information geometric formulation of the EM algorithm, and its extensions and applications to various problems. Specifically, we will see that it is possible to formulate an outlier–robust inference algorithm, an algorithm for calculating channel capacity, parameter estimation methods on probability simplex, particular multivariate analysis methods such as principal component analysis in a space of probability models and modal regression, matrix factorization, and learning generative models, which have recently attracted attention in deep learning, from the geometric perspective provided by Amari.
AB - The Expectation–Maximization (EM) algorithm is a simple meta-algorithm that has been used for many years as a methodology for statistical inference when there are missing measurements in the observed data or when the data is composed of observables and unobservables. Its general properties are well studied, and also, there are countless ways to apply it to individual problems. In this paper, we introduce the em algorithm, an information geometric formulation of the EM algorithm, and its extensions and applications to various problems. Specifically, we will see that it is possible to formulate an outlier–robust inference algorithm, an algorithm for calculating channel capacity, parameter estimation methods on probability simplex, particular multivariate analysis methods such as principal component analysis in a space of probability models and modal regression, matrix factorization, and learning generative models, which have recently attracted attention in deep learning, from the geometric perspective provided by Amari.
KW - Bregman divergence
KW - EM algorithm
KW - Generative models
KW - Information geometry
KW - Information theory
KW - Robust statistics
KW - em algorithm
UR - http://www.scopus.com/inward/record.url?scp=85142249708&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85142249708&partnerID=8YFLogxK
U2 - 10.1007/s41884-022-00080-y
DO - 10.1007/s41884-022-00080-y
M3 - Article
AN - SCOPUS:85142249708
SN - 2511-2481
VL - 7
SP - 39
EP - 77
JO - Information Geometry
JF - Information Geometry
ER -