In probability theory, it is possible to approximate the moments of a function f of a random variable X using Taylor expansions, provided that f is sufficiently differentiable and that the moments of X are finite.
E
[
f
(
X
)
]
=
E
[
f
(
μ
X
+
(
X
−
μ
X
)
)
]
≈
E
[
f
(
μ
X
)
+
f
′
(
μ
X
)
(
X
−
μ
X
)
+
1
2
f
″
(
μ
X
)
(
X
−
μ
X
)
2
]
.
Notice that
E
[
X
−
μ
X
]
=
0
, the 2nd term disappears. Also
E
[
(
X
−
μ
X
)
2
]
is
σ
X
2
. Therefore,
E
[
f
(
X
)
]
≈
f
(
μ
X
)
+
f
″
(
μ
X
)
2
σ
X
2
where
μ
X
and
σ
X
2
are the mean and variance of X respectively.
It is possible to generalize this to functions of more than one variable using multivariate Taylor expansions. For example,
E
[
X
Y
]
≈
E
[
X
]
E
[
Y
]
−
cov
[
X
,
Y
]
E
[
Y
]
2
+
E
[
X
]
E
[
Y
]
3
var
[
Y
]
Analogously,
var
[
f
(
X
)
]
≈
(
f
′
(
E
[
X
]
)
)
2
var
[
X
]
=
(
f
′
(
μ
X
)
)
2
σ
X
2
.
The above is using a first order approximation unlike for the method used in estimating the first moment. It will be a poor approximation in cases where
f
(
X
)
is highly non-linear. This is a special case of the delta method. For example,
var
[
X
Y
]
≈
var
[
X
]
E
[
Y
]
2
−
2
E
[
X
]
E
[
Y
]
3
cov
[
X
,
Y
]
+
E
[
X
]
2
E
[
Y
]
4
var
[
Y
]
.