Skip to main content

Questions tagged [kullback-leibler]

An asymmetric measure of distance (or dissimilarity) between probability distributions. It might be interpreted as the expected value of the log likelihood ratio under the alternative hypothesis.

Filter by
Sorted by
Tagged with
3 votes
2 answers
142 views

I'm having a fundamental disconnect between my intuition for KL divergence and the standard notation $D_{KL}(P \parallel Q)$. My intuition, which I believe is correct, is based on "excess ...
Charlie Parker's user avatar
0 votes
0 answers
61 views

For the following binary hypothesis testing problem $$ \begin{aligned} H_0: \boldsymbol{y} \sim f(\boldsymbol{y} | H_0)\\ H_1: \boldsymbol{y} \sim f(\boldsymbol{y} | H_1) \end{aligned} $$ where $\...
colter's user avatar
  • 1
13 votes
1 answer
576 views

Not a technical question, more of a curiosity from someone outside of Statistics/Probability. The paper from Berk (1966), "Limiting Behavior of Posterior Distributions when the Model is Incorrect&...
Joao Francisco Cabral Perez's user avatar
4 votes
1 answer
125 views

I learnt from the course 18.650 MIT OCW that we need i.i.d samples to derive MLE from KL divergence. But in the GLM framework the catch is when we model the mean of the selected distribution basically ...
Kavalali's user avatar
  • 373
1 vote
0 answers
93 views

My question is regarding the paradigm of deep learning, I do not get where does the cost functions come from? For example for a classification task are we treating the encoder as the expected value of ...
Kavalali's user avatar
  • 373
4 votes
0 answers
120 views

Exact hierarchical decomposition of KL divergence into marginals and higher‑order interactions In the standard set‑up, you compare a joint distribution $$ P(X_1,\dots,X_k) $$ to an independent ...
Will's user avatar
  • 416
0 votes
0 answers
22 views

I came across an article that stated the following: However, from this discussion, mutual information is not equivalent to Kullback–Leibler divergence. I assume only one interpretation can be correct ...
anna6931's user avatar
  • 151
3 votes
1 answer
182 views

I'm interested in estimating $D_\mathrm{KL}(q \parallel p) = \int q(x) \log \frac{q(x)}{p(x)}\,\mathrm dx$, where $p$ is a multivariate Gaussian and $q$ is an implicit distribution parameterized by a ...
Kaiwen's user avatar
  • 307
0 votes
0 answers
91 views

I am reading a paper "Complex-Valued Variational Autoencoder: A Novel Deep Generative Model for Direct Representation of Complex Spectra" In this paper, the author calculate the KL ...
Jiatong LI's user avatar
1 vote
1 answer
135 views

Assume $(E,\mathcal E,\lambda)$ is a $\sigma$-finite measure space and $\nu$ is a probability measure on $(E,\mathcal E)$ with $\nu\ll\lambda$. Furthermore, assume that $\mu=\sum_{i=0}^{n-1}\delta_{...
0xbadf00d's user avatar
  • 223
1 vote
1 answer
165 views

Suppose that we have two independent identically distributed samples. The first sample looks like $x_1 , \ldots, x_n$ with $x_i \in \mathbb{R}^d$ for every $i$. The second sample looks like $y_1, \...
温泽海's user avatar
  • 808
0 votes
0 answers
63 views

I would like to know the following which has been stated in some literature, but never explicitly proved Consider a setup consisting of a binary vector of random variables of length n say $\vec{v}=(...
chemo's user avatar
  • 1
1 vote
1 answer
394 views

I am comparing the similarity between multiple distributions based on the output of different machine-learning models. I am applying the generalised JS divergence (wiki): $$ JSD_{\pi_1,...,\pi_n}(p_1,....
Edi's user avatar
  • 175
1 vote
0 answers
133 views

Unlike a real distance measure, relative entropy is not symmetric in the sense that $D(p(x)∥q(x)) \neq D(q(x)∥p(x))$. It turns out that many information measures can be expressed by relative entropies....
허정윤's user avatar
1 vote
0 answers
29 views

I read this question Why do we use Kullback-Leibler divergence rather than cross entropy in the t-SNE objective function? and I cannot fully understand the answer. If we're using KL divergence for the ...
COTHE's user avatar
  • 11

15 30 50 per page
1
2 3 4 5
37