generalized mean#Quadratic

{{Short description|N-th root of the arithmetic mean of the given numbers raised to the power n}}

{{More citations needed|date=June 2020}}

File:Generalized means of 1, x.svg

In mathematics, generalised means (or power mean or Hölder mean from Otto Hölder) are a family of functions for aggregating sets of numbers. These include as special cases the Pythagorean means (arithmetic, geometric, and harmonic means).

Definition

If {{mvar|p}} is a non-zero real number, and x_1, \dots, x_n are positive real numbers, then the generalized mean or power mean with exponent {{mvar|p}} of these positive real numbers is{{cite journal|last=de Carvalho|first=Miguel|title=Mean, what do you Mean?|journal=The American Statistician|year=2016|volume=70|issue=3|pages=764‒776|doi=10.1080/00031305.2016.1148632|url=https://zenodo.org/record/895400|hdl=20.500.11820/fd7a8991-69a4-4fe5-876f-abcd2957a88c|hdl-access=free}}

M_p(x_1,\dots,x_n) = \left( \frac{1}{n} \sum_{i=1}^n x_i^p \right)^{{1}/{p}} .

(See Norm (mathematics)#p-norm). For {{math|1=p = 0}} we set it equal to the geometric mean (which is the limit of means with exponents approaching zero, as proved below):

M_0(x_1, \dots, x_n) = \left(\prod_{i=1}^n x_i\right)^{1/n} .

Furthermore, for a sequence of positive weights {{mvar|wi}} we define the weighted power mean as

M_p(x_1,\dots,x_n) = \left(\frac{\sum_{i=1}^n w_i x_i^p}{\sum_{i=1}^n w_i} \right)^{{1}/{p}}

and when {{math|1=p = 0}}, it is equal to the weighted geometric mean:

M_0(x_1,\dots,x_n) = \left(\prod_{i=1}^n x_i^{w_i}\right)^{1 / \sum_{i=1}^n w_i} .

The unweighted means correspond to setting all {{math|1=wi = 1}}.

Special cases

A few particular values of {{mvar|p}} yield special cases with their own names:{{MathWorld|title=Power Mean|urlname=PowerMean}} (retrieved 2019-08-17)

;minimum :M_{-\infty}(x_1,\dots,x_n) = \lim_{p\to-\infty} M_p(x_1,\dots,x_n) = \min \{x_1,\dots,x_n\}

;Image:MathematicalMeans.svgharmonic mean :M_{-1}(x_1,\dots,x_n) = \frac{n}{\frac{1}{x_1}+\dots+\frac{1}{x_n}}

;geometric mean M_0(x_1,\dots,x_n) = \lim_{p\to0} M_p(x_1,\dots,x_n) = \sqrt[n]{x_1\cdot\dots\cdot x_n}

;arithmetic mean :M_1(x_1,\dots,x_n) = \frac{x_1 + \dots + x_n}{n}

;root mean square{{anchor|Quadratic}}
or quadratic mean{{cite book |last1=Thompson |first1=Sylvanus P. |title=Calculus Made Easy |date=1965 |publisher=Macmillan International Higher Education |isbn=9781349004874 |page=185 |url=https://books.google.com/books?id=6VJdDwAAQBAJ&pg=PA185 |access-date=5 July 2020 }}{{Dead link|date=May 2024 |bot=InternetArchiveBot |fix-attempted=yes }}{{cite book |last1=Jones |first1=Alan R. |title=Probability, Statistics and Other Frightening Stuff |date=2018 |publisher=Routledge |isbn=9781351661386 |page=48 |url=https://books.google.com/books?id=OvtsDwAAQBAJ&pg=PA48 |access-date=5 July 2020}} :M_2(x_1,\dots,x_n) = \sqrt{\frac{x_1^2 + \dots + x_n^2}{n}}

;cubic mean :M_3(x_1,\dots,x_n) = \sqrt[3]{\frac{x_1^3 + \dots + x_n^3}{n}}

;maximum :M_{+\infty}(x_1,\dots,x_n) = \lim_{p\to\infty} M_p(x_1,\dots,x_n) = \max \{x_1,\dots,x_n\}

{{Math proof|title=Proof of \lim_{p \to 0} M_p = M_0 (geometric mean)|proof=For the purpose of the proof, we will assume without loss of generality that

w_i \in [0,1]

and

\sum_{i=1}^n w_i = 1.

We can rewrite the definition of M_p using the exponential function as

M_p(x_1,\dots,x_n) = \exp{\left( \ln{\left[\left(\sum_{i=1}^n w_ix_{i}^p \right)^{1/p}\right]} \right) } = \exp{\left( \frac{\ln{\left(\sum_{i=1}^n w_ix_{i}^p \right)}}{p} \right) }

In the limit {{math|p → 0}}, we can apply L'Hôpital's rule to the argument of the exponential function. We assume that p \isin \mathbb{R} but {{math|p ≠ 0}}, and that the sum of {{mvar|wi}} is equal to 1 (without loss in generality);{{Cite book |title=Handbook of Means and Their Inequalities (Mathematics and Its Applications)}} Differentiating the numerator and denominator with respect to {{mvar|p}}, we have

\begin{align}

\lim_{p \to 0} \frac{\ln{\left(\sum_{i=1}^n w_ix_{i}^p \right)}}{p} &= \lim_{p \to 0} \frac{\frac{\sum_{i=1}^n w_i x_i^p \ln{x_i}}{\sum_{j=1}^n w_j x_j^p}}{1} \\

&= \lim_{p \to 0} \frac{\sum_{i=1}^n w_i x_i^p \ln{x_i}}{\sum_{j=1}^n w_j x_j^p} \\

&= \frac{\sum_{i=1}^n w_i \ln{x_i}}{\sum_{j=1}^n w_j} \\

&= \sum_{i=1}^n w_i \ln{x_i} \\

&= \ln{\left(\prod_{i=1}^n x_i^{w_i} \right)}

\end{align}

By the continuity of the exponential function, we can substitute back into the above relation to obtain

\lim_{p \to 0} M_p(x_1,\dots,x_n) = \exp{\left( \ln{\left(\prod_{i=1}^n x_i^{w_i} \right)} \right)} = \prod_{i=1}^n x_i^{w_i} = M_0(x_1,\dots,x_n)

as desired.P. S. Bullen: Handbook of Means and Their Inequalities. Dordrecht, Netherlands: Kluwer, 2003, pp. 175-177}}

{{Proof|title= Proof of \lim_{p \to \infty} M_p = M_\infty and \lim_{p \to -\infty} M_p = M_{-\infty} |proof=

Assume (possibly after relabeling and combining terms together) that x_1 \geq \dots \geq x_n. Then

\begin{align}

\lim_{p \to \infty} M_p(x_1,\dots,x_n) &= \lim_{p \to \infty} \left( \sum_{i=1}^n w_i x_i^p \right)^{1/p} \\

&= x_1 \lim_{p \to \infty} \left( \sum_{i=1}^n w_i \left( \frac{x_i}{x_1} \right)^p \right)^{1/p} \\

&= x_1 = M_\infty (x_1,\dots,x_n).

\end{align}

The formula for M_{-\infty} follows from

M_{-\infty} (x_1,\dots,x_n) = \frac{1}{M_\infty (1/x_1,\dots,1/x_n)} = x_n.

}}

Properties

Let x_1, \dots, x_n be a sequence of positive real numbers, then the following properties hold:{{cite journal|last=Sýkora|first=Stanislav|year=2009|title=Mathematical means and averages: basic properties|journal=Stan's Library |location=Castano Primo, Italy|volume=III |doi=10.3247/SL3Math09.001 }}

  1. \min(x_1, \dots, x_n) \le M_p(x_1, \dots, x_n) \le \max(x_1, \dots, x_n).{{block indent|left=1|text= Each generalized mean always lies between the smallest and largest of the {{mvar|x}} values.}}

    1. M_p(x_1, \dots, x_n) = M_p(P(x_1, \dots, x_n)), where P is a permutation operator.{{block indent|left=1|text= Each generalized mean is a symmetric function of its arguments; permuting the arguments of a generalized mean does not change its value.}}

      1. M_p(b x_1, \dots, b x_n) = b \cdot M_p(x_1, \dots, x_n).{{block indent|left=1|text= Like most means, the generalized mean is a homogeneous function of its arguments {{math|x1, ..., xn}}. That is, if {{mvar|b}} is a positive real number, then the generalized mean with exponent {{mvar|p}} of the numbers b\cdot x_1,\dots, b\cdot x_n is equal to {{mvar|b}} times the generalized mean of the numbers {{math|x1, ..., xn}}.}}

        1. M_p(x_1, \dots, x_{n \cdot k}) = M_p\left[M_p(x_1, \dots, x_{k}), M_p(x_{k + 1}, \dots, x_{2 \cdot k}), \dots, M_p(x_{(n - 1) \cdot k + 1}, \dots, x_{n \cdot k})\right].{{block indent|left=1|text= Like the quasi-arithmetic means, the computation of the mean can be split into computations of equal sized sub-blocks. This enables use of a divide and conquer algorithm to calculate the means, when desirable.}}

= Generalized mean inequality =

{{QM_AM_GM_HM_inequality_visual_proof.svg}}

In general, if {{math|p < q}}, then

M_p(x_1, \dots, x_n) \le M_q(x_1, \dots, x_n)

and the two means are equal if and only if {{math|1= x1 = x2 = ... = xn}}.

The inequality is true for real values of {{mvar|p}} and {{mvar|q}}, as well as positive and negative infinity values.

It follows from the fact that, for all real {{mvar|p}},

\frac{\partial}{\partial p}M_p(x_1, \dots, x_n) \geq 0

which can be proved using Jensen's inequality.

In particular, for {{mvar|p}} in {{math|{−1, 0, 1}}}, the generalized mean inequality implies the Pythagorean means inequality as well as the inequality of arithmetic and geometric means.

Proof of the weighted inequality

We will prove the weighted power mean inequality. For the purpose of the proof we will assume the following without loss of generality:

\begin{align}

w_i \in [0, 1] \\

\sum_{i=1}^nw_i = 1

\end{align}

The proof for unweighted power means can be easily obtained by substituting {{math|1= wi = 1/n}}.

=Equivalence of inequalities between means of opposite signs=

Suppose an average between power means with exponents {{mvar|p}} and {{mvar|q}} holds:

\left(\sum_{i=1}^n w_i x_i^p\right)^{1/p} \geq \left(\sum_{i=1}^n w_i x_i^q\right)^{1/q}

applying this, then:

\left(\sum_{i=1}^n\frac{w_i}{x_i^p}\right)^{1/p} \geq \left(\sum_{i=1}^n\frac{w_i}{x_i^q}\right)^{1/q}

We raise both sides to the power of −1 (strictly decreasing function in positive reals):

\left(\sum_{i=1}^nw_ix_i^{-p}\right)^{-1/p}

= \left(\frac{1}{\sum_{i=1}^nw_i\frac{1}{x_i^p}}\right)^{1/p}

\leq \left(\frac{1}{\sum_{i=1}^nw_i\frac{1}{x_i^q}}\right)^{1/q}

= \left(\sum_{i=1}^nw_ix_i^{-q}\right)^{-1/q}

We get the inequality for means with exponents {{math|−p}} and {{math|−q}}, and we can use the same reasoning backwards, thus proving the inequalities to be equivalent, which will be used in some of the later proofs.

=Geometric mean=

For any {{math|q > 0}} and non-negative weights summing to 1, the following inequality holds:

\left(\sum_{i=1}^n w_i x_i^{-q}\right)^{-1/q} \leq \prod_{i=1}^n x_i^{w_i} \leq \left(\sum_{i=1}^n w_i x_i^q\right)^{1/q}.

The proof follows from Jensen's inequality, making use of the fact the logarithm is concave:

\log \prod_{i=1}^n x_i^{w_i} = \sum_{i=1}^n w_i\log x_i \leq \log \sum_{i=1}^n w_i x_i.

By applying the exponential function to both sides and observing that as a strictly increasing function it preserves the sign of the inequality, we get

\prod_{i=1}^n x_i^{w_i} \leq \sum_{i=1}^n w_i x_i.

Taking {{mvar|q}}-th powers of the {{mvar|xi}} yields

\begin{align}

&\prod_{i=1}^n x_i^{q{\cdot}w_i} \leq \sum_{i=1}^n w_i x_i^q \\

&\prod_{i=1}^n x_i^{w_i} \leq \left(\sum_{i=1}^n w_i x_i^q\right)^{1/q}.\end{align}

Thus, we are done for the inequality with positive {{mvar|q}}; the case for negatives is identical but for the swapped signs in the last step:

\prod_{i=1}^n x_i^{-q{\cdot}w_i} \leq \sum_{i=1}^n w_i x_i^{-q}.

Of course, taking each side to the power of a negative number {{math|-1/q}} swaps the direction of the inequality.

\prod_{i=1}^n x_i^{w_i} \geq \left(\sum_{i=1}^n w_i x_i^{-q}\right)^{-1/q}.

=Inequality between any two power means=

We are to prove that for any {{math|p < q}} the following inequality holds:

\left(\sum_{i=1}^n w_i x_i^p\right)^{1/p} \leq \left(\sum_{i=1}^nw_ix_i^q\right)^{1/q}

if {{mvar|p}} is negative, and {{mvar|q}} is positive, the inequality is equivalent to the one proved above:

\left(\sum_{i=1}^nw_i x_i^p\right)^{1/p} \leq \prod_{i=1}^n x_i^{w_i} \leq \left(\sum_{i=1}^n w_i x_i^q\right)^{1/q}

The proof for positive {{mvar|p}} and {{mvar|q}} is as follows: Define the following function: {{math|f : R+R+}} f(x)=x^{\frac{q}{p}}. {{mvar|f}} is a power function, so it does have a second derivative:

f''(x) = \left(\frac{q}{p} \right) \left( \frac{q}{p}-1 \right)x^{\frac{q}{p}-2}

which is strictly positive within the domain of {{mvar|f}}, since {{math|q > p}}, so we know {{mvar|f}} is convex.

Using this, and the Jensen's inequality we get:

\begin{align}

f \left( \sum_{i=1}^nw_ix_i^p \right) &\leq \sum_{i=1}^nw_if(x_i^p) \\[3pt]

\left(\sum_{i=1}^n w_i x_i^p\right)^{q/p} &\leq \sum_{i=1}^nw_ix_i^q

\end{align}

after raising both side to the power of {{math|1/q}} (an increasing function, since {{math|1/q}} is positive) we get the inequality which was to be proven:

\left(\sum_{i=1}^n w_i x_i^p\right)^{1/p} \leq \left(\sum_{i=1}^n w_i x_i^q\right)^{1/q}

Using the previously shown equivalence we can prove the inequality for negative {{mvar|p}} and {{mvar|q}} by replacing them with {{mvar|−q}} and {{mvar|−p}}, respectively.

Generalized ''f''-mean

{{Main|Generalized f-mean|l1=Generalized {{mvar|f}}-mean}}

The power mean could be generalized further to the generalized f-mean:

M_f(x_1,\dots,x_n) = f^{-1} \left({\frac{1}{n}\cdot\sum_{i=1}^n{f(x_i)}}\right)

This covers the geometric mean without using a limit with {{math|1= f(x) {{=}} log(x)}}. The power mean is obtained for {{mvar|1= f(x) {{=}} xp}}. Properties of these means are studied in de Carvalho (2016).

Applications

=Signal processing=

A power mean serves a non-linear moving average which is shifted towards small signal values for small {{mvar|p}} and emphasizes big signal values for big {{mvar|p}}. Given an efficient implementation of a moving arithmetic mean called smooth one can implement a moving power mean according to the following Haskell code.

powerSmooth :: Floating a => ([a] -> [a]) -> a -> [a] -> [a]

powerSmooth smooth p = map (** recip p) . smooth . map (**p)

See also

Notes

{{notelist}}

{{reflist|group=note}}

References

{{reflist}}

Further reading

  • {{cite book|first1=P. S. |last1=Bullen|title=Handbook of Means and Their Inequalities|location=Dordrecht, Netherlands|publisher=Kluwer|year=2003|chapter=Chapter III - The Power Means|pages=175–265}}