## Signal processing/spectral uncertainties

TBD. see wikipedia for now.

## Entropic uncertainty

\[ g(y)\approx \int _{{-\infty }}^{\infty }\exp(-2\pi ixy)f(x)\,dx,\qquad f(x)\approx \int _{{-\infty }}^{\infty }\exp(2\pi ixy)g(y)\,dy~, \]

where the “≈” indicates convergence in \(L_2\), and normalized so that (by Plancherel’s theorem),

\[ \int _{{-\infty }}^{\infty }|f(x)|^{2}\,dx=\int _{{-\infty }}^{\infty }|g(y)|^{2}\,dy=1~. \]

He showed that for any such functions the sum of the Shannon entropies is non-negative,

\[ H(|f|^{2})+H(|g|^{2})\equiv -\int _{{-\infty }}^{\infty }|f(x)|^{2}\log |f(x)|^{2}\,dx-\int _{{-\infty }}^{\infty }|g(y)|^{2}\log |g(y)|^{2}\,dy\geq 0. \]

\[ H(|f|^{2})+H(|g|^{2})\geq \log {\frac e2}~, \]

was conjectured by Hirschman and Everett, proven in 1975 by Beckner and in the same year interpreted by as a generalized quantum mechanical uncertainty principle by Białynicki-Birula and Mycielski. The equality holds in the case of Gaussian distributions.