From Wikipedia, the free encyclopedia
Mathematical identity proof [ edit ]
The variance article presents a proof of the standard deviation equivalent formula in terms of E (X ). Here is another proof, given in terms of the sample points x i .
For the variance (i.e., the square of the standard deviation), assuming a finite population with equal probabilities at all points, we have:
σ
2
=
1
N
∑
i
=
1
N
(
x
i
−
x
¯
)
2
{\displaystyle \sigma ^{2}={\frac {1}{N}}\sum _{i=1}^{N}(x_{i}-{\overline {x}})^{2}}
Expanding this, we get:
σ
2
=
1
N
[
(
x
1
−
x
¯
)
2
+
(
x
2
−
x
¯
)
2
+
⋯
+
(
x
n
−
x
¯
)
2
]
{\displaystyle \sigma ^{2}={\frac {1}{N}}\left[(x_{1}-{\overline {x}})^{2}+(x_{2}-{\overline {x}})^{2}+\cdots +(x_{n}-{\overline {x}})^{2}\right]}
Simplifying, we get:
σ
2
=
1
N
[
(
x
1
2
−
2
x
1
x
¯
+
x
¯
2
)
+
(
x
2
2
−
2
x
2
x
¯
+
x
¯
2
)
+
⋯
+
(
x
n
2
−
2
x
n
x
¯
+
x
¯
2
)
]
=
1
N
[
x
1
2
+
x
2
2
+
⋯
+
x
n
2
−
2
x
1
x
¯
−
2
x
2
x
¯
−
⋯
−
2
x
n
x
¯
+
x
¯
2
+
⋯
+
x
¯
2
]
=
1
N
[
x
1
2
+
x
2
2
+
⋯
+
x
n
2
−
2
x
¯
(
x
1
+
x
2
+
⋯
+
x
n
)
+
N
x
¯
2
]
=
1
N
[
x
1
2
+
x
2
2
+
⋯
+
x
n
2
]
−
2
x
¯
1
N
(
x
1
+
x
2
+
⋯
+
x
n
)
+
x
¯
2
=
1
N
(
∑
i
=
1
N
x
i
2
)
−
2
x
¯
(
1
N
∑
i
=
1
N
x
i
)
+
x
¯
2
=
1
N
(
∑
i
=
1
N
x
i
2
)
−
2
x
¯
2
+
x
¯
2
=
1
N
(
∑
i
=
1
N
x
i
2
)
−
x
¯
2
{\displaystyle {\begin{aligned}\sigma ^{2}&={\frac {1}{N}}\left[(x_{1}^{2}-2x_{1}{\overline {x}}+{\overline {x}}^{2})+(x_{2}^{2}-2x_{2}{\overline {x}}+{\overline {x}}^{2})+\cdots +(x_{n}^{2}-2x_{n}{\overline {x}}+{\overline {x}}^{2})\right]\\&={\frac {1}{N}}\left[x_{1}^{2}+x_{2}^{2}+\cdots +x_{n}^{2}\,-\,2x_{1}{\overline {x}}-2x_{2}{\overline {x}}-\cdots -2x_{n}{\overline {x}}\,+\,{\overline {x}}^{2}+\cdots +{\overline {x}}^{2}\right]\\&={\frac {1}{N}}\left[x_{1}^{2}+x_{2}^{2}+\cdots +x_{n}^{2}\,-\,2{\overline {x}}(x_{1}+x_{2}+\cdots +x_{n})\,+\,N{\overline {x}}^{2}\right]\\&={\frac {1}{N}}\left[x_{1}^{2}+x_{2}^{2}+\cdots +x_{n}^{2}\right]-2{\overline {x}}{\frac {1}{N}}(x_{1}+x_{2}+\cdots +x_{n})+{\overline {x}}^{2}\\&={\frac {1}{N}}\left(\sum _{i=1}^{N}x_{i}^{2}\right)-2{\overline {x}}\left({\frac {1}{N}}\sum _{i=1}^{N}x_{i}\right)+{\overline {x}}^{2}\\&={\frac {1}{N}}\left(\sum _{i=1}^{N}x_{i}^{2}\right)-2{\overline {x}}^{2}+{\overline {x}}^{2}\\&={\frac {1}{N}}\left(\sum _{i=1}^{N}x_{i}^{2}\right)-{\overline {x}}^{2}\end{aligned}}}
Expanding the last term, we get:
σ
2
=
1
N
∑
i
=
1
N
x
i
2
−
(
1
N
∑
i
=
1
N
x
i
)
2
{\displaystyle \sigma ^{2}={\frac {1}{N}}\sum _{i=1}^{N}x_{i}^{2}\,-\,\left({\frac {1}{N}}\sum _{i=1}^{N}x_{i}\right)^{2}}
So then for the standard deviation, we get:
σ
=
1
N
∑
i
=
1
N
(
x
i
−
x
¯
)
2
=
1
N
(
∑
i
=
1
N
x
i
2
)
−
x
¯
2
=
1
N
∑
i
=
1
N
x
i
2
−
(
1
N
∑
i
=
1
N
x
i
)
2
.
{\displaystyle \sigma ={\sqrt {{\frac {1}{N}}\sum _{i=1}^{N}(x_{i}-{\overline {x}})^{2}}}={\sqrt {{\frac {1}{N}}\left(\sum _{i=1}^{N}x_{i}^{2}\right)-{\overline {x}}^{2}}}={\sqrt {{\frac {1}{N}}\sum _{i=1}^{N}x_{i}^{2}-\left({\frac {1}{N}}\sum _{i=1}^{N}x_{i}\right)^{2}}}\;.}
— Loadmaster (talk ) 18:12, 22 February 2013 (UTC)
This can also be written as:
σ
=
N
∑
i
=
1
N
x
i
2
−
(
∑
i
=
1
N
x
i
)
2
N
2
.
{\displaystyle \sigma ={\sqrt {\frac {N\sum _{i=1}^{N}x_{i}^{2}-\left(\sum _{i=1}^{N}x_{i}\right)^{2}}{N^{2}}}}\;.}