Zum Hauptinhalt springen Zur Suche springen Zur Hauptnavigation springen
Haben Sie Fragen? Einfach anrufen, wir helfen gerne: Tel. 089/210233-0
oder besuchen Sie unser Ladengeschäft in der Pacellistraße 5 (Maxburg) 80333 München
+++ Versandkostenfreie Lieferung innerhalb Deutschlands
Haben Sie Fragen? Tel. 089/210233-0

Minimum Gamma-Divergence for Regression and Classification Problems

48,14 €*

Versandkostenfrei

Produktnummer: 188789018bf1b243f7b03884b4e3838cde
Autor: Eguchi, Shinto
Themengebiete: AdaBoost Boltzmann Machine Gamma-Divergence Geometric Mean Multi-Class Classification Poisson Distribution
Veröffentlichungsdatum: 12.03.2025
EAN: 9789819788798
Sprache: Englisch
Seitenzahl: 118
Produktart: Kartoniert / Broschiert
Verlag: Springer Singapore
Produktinformationen "Minimum Gamma-Divergence for Regression and Classification Problems"
This book introduces the gamma-divergence, a measure of distance between probability distributions that was proposed by Fujisawa and Eguchi in 2008. The gamma-divergence has been extensively explored to provide robust estimation when the power index ? is positive. The gamma-divergence can be defined even when the power index ? is negative, as long as the condition of integrability is satisfied. Thus, the authors consider the gamma-divergence defined on a set of discrete distributions. The arithmetic, geometric, and harmonic means for the distribution ratios are closely connected with the gamma-divergence with a negative ?. In particular, the authors call the geometric-mean (GM) divergence the gamma-divergence when ? is equal to -1.The book begins by providing an overview of the gamma-divergence and its properties. It then goes on to discuss the applications of the gamma-divergence in various areas, including machine learning, statistics, and ecology. Bernoulli, categorical, Poisson, negative binomial, and Boltzmann distributions are discussed as typical examples. Furthermore, regression analysis models that explicitly or implicitly assume these distributions as the dependent variable in generalized linear models are discussed to apply the minimum gamma-divergence method.In ensemble learning, AdaBoost is derived by the exponential loss function in the weighted majority vote manner. It is pointed out that the exponential loss function is deeply connected to the GM divergence. In the Boltzmann machine, the maximum likelihood has to use approximation methods such as mean field approximation because of the intractable computation of the partition function. However, by considering the GM divergence and the exponential loss, it is shown that the calculation of the partition function is not necessary, and it can be executed without variational inference.
Bücherregal gefüllt mit juristischen Werken

Sie möchten lieber vor Ort einkaufen?

Sie haben Fragen zu diesem oder anderen Produkten oder möchten einfach gerne analog im Laden stöbern? Wir sind gerne für Sie da und beraten Sie auch telefonisch.

Juristische Fachbuchhandlung
Georg Blendl

Parcellistraße 5 (Maxburg)
8033 München

Montag - Freitag: 8:15 -18 Uhr
Samstags geschlossen