Gael Varoquaux 12/8/2019

Comparing distributions: Kernels estimate good representations, l1 distances give good tests

Read Original

This technical article revisits the classic statistical problem of two-sample testing, examining whether two observed datasets are drawn from the same underlying distribution. It explains the use of kernel mean embeddings and Maximum Mean Discrepancy (MMD) to compare distributions, and argues that metrics based on L1 geometry can provide superior testing power compared to traditional kernel methods. The content is based on a NeurIPS 2019 conference paper and delves into the mathematical framework of Integral Probability Metrics and Reproducible Kernel Hilbert Spaces.

Comparing distributions: Kernels estimate good representations, l1 distances give good tests

Comments

No comments yet

Be the first to share your thoughts!

Browser Extension

Get instant access to AllDevBlogs from your browser

Top of the Week

2
Designing Design Systems
TkDodo Dominik Dorfmeister 2 votes
4
Introducing RSC Explorer
Dan Abramov 1 votes
6
Fragments Dec 11
Martin Fowler 1 votes
7
Adding Type Hints to my Blog
Daniel Feldroy 1 votes
8
Refactoring English: Month 12
Michael Lynch 1 votes
10