# Monotone likelihood ratio

In statistics, the monotone likelihood ratio property is a property of the ratio of two probability density functions (PDFs). Formally, distributions ƒ(x) and g(x) bear the property if

${\displaystyle {\text{for every }}x_{1}>x_{0},\quad {\frac {f(x_{1})}{g(x_{1})}}\geq {\frac {f(x_{0})}{g(x_{0})}}}$
A monotonic likelihood ratio in distributions ${\displaystyle f(x)}$ and ${\displaystyle g(x)}$

The ratio of the density functions above is increasing in the parameter ${\displaystyle x}$, so ${\displaystyle f(x)/g(x)}$ satisfies the monotone likelihood ratio property.

that is, if the ratio is nondecreasing in the argument ${\displaystyle x}$.

If the functions are first-differentiable, the property may sometimes be stated

${\displaystyle {\frac {\partial }{\partial x}}\left({\frac {f(x)}{g(x)}}\right)\geq 0}$

For two distributions that satisfy the definition with respect to some argument x, we say they "have the MLRP in x." For a family of distributions that all satisfy the definition with respect to some statistic T(X), we say they "have the MLR in T(X)."

## Intuition

The MLRP is used to represent a data-generating process that enjoys a straightforward relationship between the magnitude of some observed variable and the distribution it draws from. If ${\displaystyle f(x)}$ satisfies the MLRP with respect to ${\displaystyle g(x)}$, the higher the observed value ${\displaystyle x}$, the more likely it was drawn from distribution ${\displaystyle f}$ rather than ${\displaystyle g}$. As usual for monotonic relationships, the likelihood ratio's monotonicity comes in handy in statistics, particularly when using maximum-likelihood estimation. Also, distribution families with MLR have a number of well-behaved stochastic properties, such as first-order stochastic dominance and increasing hazard ratios. Unfortunately, as is also usual, the strength of this assumption comes at the price of realism. Many processes in the world do not exhibit a monotonic correspondence between input and output.

### Example: Working hard or slacking off

Suppose you are working on a project, and you can either work hard or slack off. Call your choice of effort ${\displaystyle e}$ and the quality of the resulting project ${\displaystyle q}$. If the MLRP holds for the distribution of q conditional on your effort ${\displaystyle e}$, the higher the quality the more likely you worked hard. Conversely, the lower the quality the more likely you slacked off.

1. Choose effort ${\displaystyle e\in \{H,L\}}$ where H means high, L means low
2. Observe ${\displaystyle q}$ drawn from ${\displaystyle f(q\mid e)}$. By Bayes' law with a uniform prior,
${\displaystyle \Pr[e=H\mid q]={\frac {f(q\mid H)}{f(q\mid H)+f(q\mid L)}}}$
3. Suppose ${\displaystyle f(q\mid e)}$ satisfies the MLRP. Rearranging, the probability the worker worked hard is
${\displaystyle {\frac {1}{1+f(q\mid L)/f(q\mid H)}}}$
which, thanks to the MLRP, is monotonically increasing in ${\displaystyle q}$ (because ${\displaystyle f(q\mid L)/f(q\mid H)}$ is decreasing in ${\displaystyle q}$). Hence if some employer is doing a "performance review" he can infer his employee's behavior from the merits of his work.

## Families of distributions satisfying MLR

Statistical models often assume that data are generated by a distribution from some family of distributions and seek to determine that distribution. This task is simplified if the family has the monotone likelihood ratio property (MLRP).

A family of density functions ${\displaystyle \{f_{\theta }(x)\}_{\theta \in \Theta }}$ indexed by a parameter ${\displaystyle \theta }$ taking values in an ordered set ${\displaystyle \Theta }$ is said to have a monotone likelihood ratio (MLR) in the statistic ${\displaystyle T(X)}$ if for any ${\displaystyle \theta _{1}<\theta _{2}}$,

${\displaystyle {\frac {f_{\theta _{2}}(X=x_{1},x_{2},x_{3},\dots )}{f_{\theta _{1}}(X=x_{1},x_{2},x_{3},\dots )}}}$  is a non-decreasing function of ${\displaystyle T(X)}$.

Then we say the family of distributions "has MLR in ${\displaystyle T(X)}$".

### List of families

Family${\displaystyle T(X)}$  in which ${\displaystyle f_{\theta }(X)}$ has the MLR
Exponential${\displaystyle [\lambda ]}$${\displaystyle \sum x_{i}}$ observations
Binomial${\displaystyle [n,p]}$${\displaystyle \sum x_{i}}$ observations
Poisson${\displaystyle [\lambda ]}$${\displaystyle \sum x_{i}}$ observations
Normal${\displaystyle [\mu ,\sigma ]}$if ${\displaystyle \sigma }$ known, ${\displaystyle \sum x_{i}}$ observations

### Hypothesis testing

If the family of random variables has the MLRP in ${\displaystyle T(X)}$, a uniformly most powerful test can easily be determined for the hypothesis ${\displaystyle H_{0}:\theta \leq \theta _{0}}$ versus ${\displaystyle H_{1}:\theta >\theta _{0}}$.

### Example: Effort and output

Example: Let ${\displaystyle e}$ be an input into a stochastic technology – worker's effort, for instance – and ${\displaystyle y}$ its output, the likelihood of which is described by a probability density function ${\displaystyle f(y;e).}$ Then the monotone likelihood ratio property (MLRP) of the family ${\displaystyle f}$ is expressed as follows: for any ${\displaystyle e_{1},e_{2}}$, the fact that ${\displaystyle e_{2}>e_{1}}$ implies that the ratio ${\displaystyle f(y;e_{2})/f(y;e_{1})}$ is increasing in ${\displaystyle y}$.

## Relation to other statistical properties

Monotone likelihoods are used in several areas of statistical theory, including point estimation and hypothesis testing, as well as in probability models.

### Exponential families

One-parameter exponential families have monotone likelihood-functions. In particular, the one-dimensional exponential family of probability density functions or probability mass functions with

${\displaystyle f_{\theta }(x)=c(\theta )h(x)\exp(\pi (\theta )T(x))}$

has a monotone non-decreasing likelihood ratio in the sufficient statistic T(x), provided that ${\displaystyle \pi (\theta )}$ is non-decreasing.

### Most powerful tests: The Karlin–Rubin theorem

Monotone likelihood functions are used to construct uniformly most powerful tests, according to the Karlin–Rubin theorem.[1] Consider a scalar measurement having a probability density function parameterized by a scalar parameter θ, and define the likelihood ratio ${\displaystyle \ell (x)=f_{\theta _{1}}(x)/f_{\theta _{0}}(x)}$. If ${\displaystyle \ell (x)}$ is monotone non-decreasing, in ${\displaystyle x}$, for any pair ${\displaystyle \theta _{1}\geq \theta _{0}}$ (meaning that the greater ${\displaystyle x}$ is, the more likely ${\displaystyle H_{1}}$ is), then the threshold test:

${\displaystyle \varphi (x)={\begin{cases}1&{\text{if }}x>x_{0}\\0&{\text{if }}x
where ${\displaystyle x_{0}}$ is chosen so that ${\displaystyle \operatorname {E} _{\theta _{0}}\varphi (X)=\alpha }$

is the UMP test of size α for testing ${\displaystyle H_{0}:\theta \leq \theta _{0}{\text{ vs. }}H_{1}:\theta >\theta _{0}.}$

Note that exactly the same test is also UMP for testing ${\displaystyle H_{0}:\theta =\theta _{0}{\text{ vs. }}H_{1}:\theta >\theta _{0}.}$

### Median unbiased estimation

Monotone likelihood-functions are used to construct median-unbiased estimators, using methods specified by Johann Pfanzagl and others.[2][3] One such procedure is an analogue of the Rao–Blackwell procedure for mean-unbiased estimators: The procedure holds for a smaller class of probability distributions than does the Rao–Blackwell procedure for mean-unbiased estimation but for a larger class of loss functions.[4]

### Lifetime analysis: Survival analysis and reliability

If a family of distributions ${\displaystyle f_{\theta }(x)}$ has the monotone likelihood ratio property in ${\displaystyle T(X)}$,

1. the family has monotone decreasing hazard rates in ${\displaystyle \theta }$ (but not necessarily in ${\displaystyle T(X)}$)
2. the family exhibits the first-order (and hence second-order) stochastic dominance in ${\displaystyle x}$, and the best Bayesian update of ${\displaystyle \theta }$ is increasing in ${\displaystyle T(X)}$.

But not conversely: neither monotone hazard rates nor stochastic dominance imply the MLRP.

#### Proofs

Let distribution family ${\displaystyle f_{\theta }}$ satisfy MLR in x, so that for ${\displaystyle \theta _{1}>\theta _{0}}$ and ${\displaystyle x_{1}>x_{0}}$:

${\displaystyle {\frac {f_{\theta _{1}}(x_{1})}{f_{\theta _{0}}(x_{1})}}\geq {\frac {f_{\theta _{1}}(x_{0})}{f_{\theta _{0}}(x_{0})}},}$

or equivalently:

${\displaystyle f_{\theta _{1}}(x_{1})f_{\theta _{0}}(x_{0})\geq f_{\theta _{1}}(x_{0})f_{\theta _{0}}(x_{1}).\,}$

Integrating this expression twice, we obtain:

 1. To ${\displaystyle x_{1}}$ with respect to ${\displaystyle x_{0}}$ {\displaystyle {\begin{aligned}&\int _{\min _{x}\in X}^{x_{1}}f_{\theta _{1}}(x_{1})f_{\theta _{0}}(x_{0})\,dx_{0}\\[6pt]\geq {}&\int _{\min _{x}\in X}^{x_{1}}f_{\theta _{1}}(x_{0})f_{\theta _{0}}(x_{1})\,dx_{0}\end{aligned}}} integrate and rearrange to obtain ${\displaystyle {\frac {f_{\theta _{1}}}{f_{\theta _{0}}}}(x)\geq {\frac {F_{\theta _{1}}}{F_{\theta _{0}}}}(x)}$ 2. From ${\displaystyle x_{0}}$ with respect to ${\displaystyle x_{1}}$ {\displaystyle {\begin{aligned}&\int _{x_{0}}^{\max _{x}\in X}f_{\theta _{1}}(x_{1})f_{\theta _{0}}(x_{0})\,dx_{1}\\[6pt]\geq {}&\int _{x_{0}}^{\max _{x}\in X}f_{\theta _{1}}(x_{0})f_{\theta _{0}}(x_{1})\,dx_{1}\end{aligned}}} integrate and rearrange to obtain ${\displaystyle {\frac {1-F_{\theta _{1}}(x)}{1-F_{\theta _{0}}(x)}}\geq {\frac {f_{\theta _{1}}}{f_{\theta _{0}}}}(x)}$

#### First-order stochastic dominance

Combine the two inequalities above to get first-order dominance:

${\displaystyle F_{\theta _{1}}(x)\leq F_{\theta _{0}}(x)\ \forall x}$

#### Monotone hazard rate

Use only the second inequality above to get a monotone hazard rate:

${\displaystyle {\frac {f_{\theta _{1}}(x)}{1-F_{\theta _{1}}(x)}}\leq {\frac {f_{\theta _{0}}(x)}{1-F_{\theta _{0}}(x)}}\ \forall x}$

## Uses

### Economics

The MLR is an important condition on the type distribution of agents in mechanism design. Most solutions to mechanism design models assume a type distribution to satisfy the MLR to take advantage of a common solution method.

## References

1. Casella, G.; Berger, R.L. (2008), Statistical Inference, Brooks/Cole. ISBN 0-495-39187-5 (Theorem 8.3.17)
2. Pfanzagl, Johann. "On optimal median unbiased estimators in the presence of nuisance parameters." The Annals of Statistics (1979): 187–193.
3. Brown, L. D.; Cohen, Arthur; Strawderman, W. E. A Complete Class Theorem for Strict Monotone Likelihood Ratio With Applications. Ann. Statist. 4 (1976), no. 4, 712–722. doi:10.1214/aos/1176343543. http://projecteuclid.org/euclid.aos/1176343543.
4. Page 713: Brown, L. D.; Cohen, Arthur; Strawderman, W. E. A Complete Class Theorem for Strict Monotone Likelihood Ratio With Applications. Ann. Statist. 4 (1976), no. 4, 712–722. doi:10.1214/aos/1176343543. http://projecteuclid.org/euclid.aos/1176343543.