KL Divergence

Entropy of distribution P is H(P)=\sum_i{p_i\log{\frac{1}{p_i}}}, which reflects the amount of uncertainty in P. Uniform distributions always have the largest entropy.
If we do not have prior knowledge about P and guess it to be Q, then we actually add extra uncertainty and have cross entropy as H(P, Q)=\sum_i{p_i\log{\frac{1}{q_i}}}. In another aspect, cross entropy itself is a good alternative to MSE loss with sigmoid function as demonstrated.
The discrepancy between H(P) and H(P, Q) is relative entropy, also known as KL divergence, formulated as KL(P||Q)=H(P, Q)-H(P)=\sum_i{p_i\log{\frac{p_i}{q_i}}}.
KL divergence is non-negative proved by using Jensen’s Inequality. Besides, KL divergence is asymmetric (KL(P||Q)\ne KL(Q||P)). However, we can define a symmetric variant as KL'(P||Q)=(KL(P||Q)+KL(Q||P))/2. More properties can be referred here.

©著作权归作者所有,转载或内容合作请联系作者
平台声明:文章内容(如有图片或视频亦包括在内)由作者上传并发布,文章内容仅代表作者本人观点,简书系信息发布平台,仅提供信息存储服务。

推荐阅读更多精彩内容

  • rljs by sennchi Timeline of History Part One The Cognitiv...
    sennchi阅读 12,140评论 0 10
  • 【观点回顾】 【收盘综述】 周二指数小幅高开后冲高回落,全天呈现震荡格局,早盘题材股反弹相对强势,强于主板,因权重...
    看市者鹏凯阅读 2,750评论 0 0
  • 当我睁开眼的时候,寝室里已经是灰暗的一片了,头很晕,我坐着靠在墙的一边,大约两分钟后,我睁开眼,定神四面看。寝...
    熊小love阅读 1,612评论 0 0
  • 没有反思的人生不值得过——苏格拉底 #周检视#第10周20180512 本周关键词:拥有成长思维 本周总评:7 检...
    笑元素阅读 1,267评论 0 1
  • 清晨的六点十分,一夜无眠的我终于将消息发送给了W,以此来结束我们两年的感情。亲爱的W,感谢你长久以来对我的照顾,没...
    爱榴莲不留恋阅读 1,332评论 0 0