###### tags: `one-offs` `rejection sampling` `isoperimetry`
# Isoperimetric Surrogates for the Gaussian
**Overview**: In this note, I outline some calculations relating to some one-dimensional probability measures which are close in character to the standard Gaussian measure, but which retain certain tractabilities which may be of interest.
## Motivation
To begin with, I was interested in cooking up some probability measures in one dimension which qualitatively resemble the Gaussian measure (in terms of e.g. concentration, isoperimetry, etc.), but are a bit more { tractable / tangible / etc. } in terms of analytic PDF, CDF, etc.
Anyways, it is known that the isoperimetric profile (which I won't describe, for now) of the standard Gaussian satisfies
\begin{align}
I\left(p\right) &=\left( \varphi\circ\Phi^{-1} \right) \left(p\right) \\
&\sim p \cdot \left(2\cdot\log\left(\frac{1}{p}\right)\right)^{1/2} \quad \mathrm{as} \,p\to0^{+}.
\end{align}
Using some well-known relationships, one can then reverse-engineer probability measures with this form of isoperimetric profile. In particular, for symmetric, log-concave probability measures in dimension one with density $f$ and cumulative distribution function $F$, it is known that
\begin{align}
I=f\circ F^{-1}\leadsto F^{'}=I\circ F.
\end{align}
As such, taking $I\left(p\right) = p \cdot \left(2\cdot\log\left( \frac{1}{p}\right) \right)^{1/2}$ leads to
\begin{align}
\frac{\mathrm{d}F}{F\cdot\left(2\cdot\log\left(\frac{1}{F}\right)\right)^{1/2}} =\mathrm{d}x\quad\leadsto\quad F\left(x\right)=\exp\left(-\frac{1}{2}\left(x-\sqrt{2\cdot\log2}\right)^{2}\right)
\end{align}
for $x\leqslant0$, and hence by symmetry, $F\left(x\right)=1-\exp\left(-\frac{1}{2}\left(x+\sqrt{2\cdot\log2}\right)^{2}\right)$ for $x\geqslant0$.
## Generalising and Checking
This suggests considering symmetric probability measures satisfying
\begin{align}
\forall r \geqslant 0, \quad \mathbf{P}\left(\left|X\right|\geqslant r\right)=\exp\left(\frac{1}{2}a^{2}-\frac{1}{2}\left(a+r\right)^{2}\right)
\end{align}
for some $a > 0$. These measures have density
\begin{align}
\mu_{a}\left(\mathrm{d}x\right) &= f_{a}\left(x\right)\:\mathrm{d}x \\
f_{a}\left(x\right) &= \frac{1}{2} \cdot \left(a+\left|x\right|\right)\cdot\exp\left(\frac{1}{2}a^{2}-\frac{1}{2}\left(a+\left|x\right|\right)^{2}\right)
\end{align}
and cumulative distribution function
\begin{align}
F_{a}\left(x\right)=\begin{cases}
\frac{1}{2}\cdot\exp\left(\frac{1}{2}a^{2}-\frac{1}{2}\left(a+\left|x\right|\right)^{2}\right) & x\leqslant0\\
1-\frac{1}{2}\cdot\exp\left(\frac{1}{2}a^{2}-\frac{1}{2}\left(a+\left|x\right|\right)^{2}\right) & x\geqslant0
\end{cases}.
\end{align}
In particular, the isoperimetric profile can then be read off as (for $p\in\left[0,\frac{1}{2}\right]$)
\begin{align}
I_{a}\left(p\right) &=\left(f_{a}\circ F_{a}^{-1}\right)\left(p\right) \\
&=p\cdot\left(2\cdot\log\left(\frac{\alpha_{a}}{p}\right)\right)^{1/2}
\end{align}
with $\alpha_{a}=\frac{1}{2} \cdot \exp \left( \frac{1}{2}a^{2} \right)$.
So, having constructed these measures with a goal in mind, we ask the natural question: will this probability measure be qualitatively close to the Gaussian in all relevant ways?
## Log-Concavity
There is one important detail which turns out to be sensitive to the precise choice of $a$: the log-concavity of the measure. Many important geometric properties of probability measures are particularly easy to work with for log-concave measures, and so it is of interest to ensure this. It will turn out that $a=1$ is a critical value for the log-concavity of $\mu_{a}$. Note that just differentiating $\log f_{a}$ twice to check for concavity is a risky strategy, due to the non-smoothness of $f_a$. Still, write
\begin{align}
-\log f_{a}\left(x\right) &=-\log\left(\frac{1}{2}\cdot\left(a+\left|x\right|\right)\cdot\exp\left(\frac{1}{2}a^{2}-\frac{1}{2}\left(a+\left|x\right|\right)^{2}\right)\right) \\
&=-\log\left(1+\frac{\left|x\right|}{a}\right)+\frac{1}{2}\left(a+\left|x\right|\right)^{2}-\frac{1}{2}a^{2}-\log a+\log2.
\end{align}
Now, $x\mapsto-\log\left(1+\frac{\left|x\right|}{a}\right)$ is convex on either side of $x=0$, but not globally. However, one can show that $x\mapsto\frac{\left|x\right|}{a}-\log\left(1+\frac{\left|x\right|}{a}\right)$ **is** globally convex, and so we regroup terms to see that
\begin{align}
-\log f_{a}\left(x\right) =&\left(\frac{\left|x\right|}{a}-\log\left(1+\frac{\left|x\right|}{a}\right)\right)+\frac{1}{2}\left(\left(a-a^{-1}\right)+\left|x\right|\right)^{2} \\
&-\frac{1}{2}\left(a-a^{-1}\right)^{2}-\log a+\log2.
\end{align}
We have just claimed that $x\mapsto\frac{\left|x\right|}{a}-\log\left(1+\frac{\left|x\right|}{a}\right)$ is convex, and one can check directly that $x\mapsto\left(b+\left|x\right|\right)^{2}$ is convex precisely when $b\geqslant0$; the remaining terms are constant in $x$ and so the claim follows. For $a\in\left(0,1\right)$, one can check directly by inspection that $-\log f_{a}$ fails to be convex in a neighbourhood of the origin.
## Next Steps
We now know that for $a\geqslant1$, log-concavity (even $1$-strong log-concavity) holds for $\mu_{a}$, and so these are reasonable model probability measures with which to do some concrete calculations. Sampling is also straightforward by e.g. inversion, i.e. draw $h\sim\mathrm{Exp}\left(1\right)$, $s\sim\mathrm{Unif}\left(\left\{ \pm1\right\} \right)$, and set $x=s\cdot\left(\left(a^{2}+2\cdot h\right)^{1/2}-a\right)$.
Moving beyond this, I can think of a few interesting things to play around with; including:
1. Since simulation is straightforward, it is (at least conceptually) interesting to consider using $\mu_{a}$ for rejection sampling for Gaussians; how should one set $a$ optimally?
2. What are the qualitative differences between the $\mu_{a}$ in the limits $a\to0^{+}$, $a\to\infty$?
3. Transport maps between $\mu_{a}$ and the standard Gaussian measure ought to have reasonably good regularity (e.g. Lipschitz constants). For which values of $a$ are the regularity properties best?
4. Do there exist any modifications (e.g. exponential tiltings, mixtures, conjugate families, higher-dimensional analogues, ...) of the $\mu_{a}$ which are interesting or tractable?