Intuitive explanation of variance and moment in ProbabilityIntuition behind moments of random variablesUsefulness of VarianceWhat is variance intutively? and is there a math proof for variance ? how is variance and covarience relatedFor System dependent on normally distributed parameter, are deviations added or variations?Intuitive meaning of Pearson Product-moment correlation coefficient FormulaIntuitive explanation of the tower property of conditional expectationMotivation behind the ingredients of First Cohomology group $H^1$Intuition behind Variance forumlaIntuitive Approach to Sheaf and Cech CohomologyWhat is an intuitive explanation for Birkhoff's ergodic theorem?Understanding expectation, variance and standard deviationhow to interpret the variance of a variance?Intuition behind moments of random variablesFind variance and general formula for for r$^th$ moment for random variable uniform over (0,1)
What does the same-ish mean?
What is the opposite of "eschatology"?
Is it a bad idea to plug the other end of ESD strap to wall ground?
Ambiguity in the definition of entropy
What historical events would have to change in order to make 19th century "steampunk" technology possible?
Using "tail" to follow a file without displaying the most recent lines
Avoiding the "not like other girls" trope?
How dangerous is XSS
How many wives did king shaul have
Forgetting the musical notes while performing in concert
Are British MPs missing the point, with these 'Indicative Votes'?
How to travel to Japan while expressing milk?
How to remove border from elements in the last row?
OP Amp not amplifying audio signal
Theorists sure want true answers to this!
How do I exit BASH while loop using modulus operator?
Implication of namely
What is the fastest integer factorization to break RSA?
Unlock My Phone! February 2018
Why is it a bad idea to hire a hitman to eliminate most corrupt politicians?
Can compressed videos be decoded back to their uncompresed original format?
Where would I need my direct neural interface to be implanted?
Is there a hemisphere-neutral way of specifying a season?
Processor speed limited at 0.4 Ghz
Intuitive explanation of variance and moment in Probability
Intuition behind moments of random variablesUsefulness of VarianceWhat is variance intutively? and is there a math proof for variance ? how is variance and covarience relatedFor System dependent on normally distributed parameter, are deviations added or variations?Intuitive meaning of Pearson Product-moment correlation coefficient FormulaIntuitive explanation of the tower property of conditional expectationMotivation behind the ingredients of First Cohomology group $H^1$Intuition behind Variance forumlaIntuitive Approach to Sheaf and Cech CohomologyWhat is an intuitive explanation for Birkhoff's ergodic theorem?Understanding expectation, variance and standard deviationhow to interpret the variance of a variance?Intuition behind moments of random variablesFind variance and general formula for for r$^th$ moment for random variable uniform over (0,1)
$begingroup$
While I understand the intuition behind expectation, I don't really understand the meaning of variance and moment.
What is a good way to think of those two terms?
probability intuition probability-theory
$endgroup$
add a comment |
$begingroup$
While I understand the intuition behind expectation, I don't really understand the meaning of variance and moment.
What is a good way to think of those two terms?
probability intuition probability-theory
$endgroup$
$begingroup$
Why the edit, more than 4 years later?
$endgroup$
– Did
Dec 20 '14 at 13:05
$begingroup$
@Did -- Sorry, was checking something and forgot edits bump up questions :(
$endgroup$
– Pandora
Dec 20 '14 at 13:13
add a comment |
$begingroup$
While I understand the intuition behind expectation, I don't really understand the meaning of variance and moment.
What is a good way to think of those two terms?
probability intuition probability-theory
$endgroup$
While I understand the intuition behind expectation, I don't really understand the meaning of variance and moment.
What is a good way to think of those two terms?
probability intuition probability-theory
probability intuition probability-theory
edited Dec 20 '14 at 12:32
Pandora
asked Sep 24 '10 at 17:14
PandoraPandora
2,72113157
2,72113157
$begingroup$
Why the edit, more than 4 years later?
$endgroup$
– Did
Dec 20 '14 at 13:05
$begingroup$
@Did -- Sorry, was checking something and forgot edits bump up questions :(
$endgroup$
– Pandora
Dec 20 '14 at 13:13
add a comment |
$begingroup$
Why the edit, more than 4 years later?
$endgroup$
– Did
Dec 20 '14 at 13:05
$begingroup$
@Did -- Sorry, was checking something and forgot edits bump up questions :(
$endgroup$
– Pandora
Dec 20 '14 at 13:13
$begingroup$
Why the edit, more than 4 years later?
$endgroup$
– Did
Dec 20 '14 at 13:05
$begingroup$
Why the edit, more than 4 years later?
$endgroup$
– Did
Dec 20 '14 at 13:05
$begingroup$
@Did -- Sorry, was checking something and forgot edits bump up questions :(
$endgroup$
– Pandora
Dec 20 '14 at 13:13
$begingroup$
@Did -- Sorry, was checking something and forgot edits bump up questions :(
$endgroup$
– Pandora
Dec 20 '14 at 13:13
add a comment |
4 Answers
4
active
oldest
votes
$begingroup$
There are many good answers possible, depending on how one's intuition works and one's experience with related things like averages, $L^p$ norms, etc. One answer that builds on your intuition of an expectation is to think of the higher central moments as weighted expectations. The central third moment, for example, is defined as the expectation of the third powers of deviations from the mean. Split a third power into two parts: one is the deviation from the mean (let's call it, after common statistical parlance, the "residual") and the other is the squared deviation from the mean (ergo, the squared residual). Treat the latter as a weight. As the residual gets larger in size, these weights get larger (in a quadratic fashion). The third central moment is the weighted "average" of the residuals. (I have to put "average" in quotes because, strictly speaking, to deserve being called an average one would require the weights themselves to average out to 1, but they don't usually do that.) Thus, the third central moment gives disproportionately greater emphasis to larger residuals compared to smaller ones. (It is possible for that emphasis to be so great that the third central moment doesn't even exist: if the probability doesn't decay rapidly enough with the size of the residual, the weights can swamp the probability, with the net effect of creating an infinitely great third central moment.)
You can understand, and even analyze, all other central moments in the same way, even fractional moments and absolute moments. It should be immediately clear, for instance, that the higher the moment, the more the weights are emphasized at extreme values: higher central moments measure (average) behavior of the probability distribution at greater distances from the mean.
$endgroup$
add a comment |
$begingroup$
Variance of a r.v. $X$ is $Var(X) = mathbbE[(X-mathbbE(X))^2]$. Variance is used as a measure for the dispersion of a random variable. If you are interested in how far a random variable is from what you expect; perhaps you would consider $mathbbE(|X-mathbbE(X)|)$ but see: https://stats.stackexchange.com/questions/118/standard-deviation-why-square-the-difference-instead-of-taking-the-absolute-val for why this we use the square instead.
For all except the first moment (i.e. the mean) the central moments (centred about the mean) are far more interesting.
As for the $r$th (central) moment of $X$, that is $mathbbE[(X-mathbbE(X))^r]$. The lower moments ($r<5$) are related to properties of the distribution:
$r=0$ is $1$
$r=1$ is $0$
$r=2$ is the variance which is a measure of the spread from the mean.
$r=3$ is related to skewness which is a measure of the asymmetry of the distribution. See: http://en.wikipedia.org/wiki/Skewness
$r=4$ is related to kurtois which is a measure of the 'peakedness' of a distribution. See: http://en.wikipedia.org/wiki/Kurtosis
For $rgeq5$ they become slightly harder to compute. The $5th$ central moment is a measure for the asymmetry of the distributions tails, but I would use skewness instead.
$endgroup$
2
$begingroup$
$r=3$ and $r=4$ are only related to skewness and kurtosis. For these moments to merit these names, they have to be normalized by an appropriate measure of spread (to wit, the standard deviation).
$endgroup$
– whuber
Sep 24 '10 at 18:32
$begingroup$
Yes that is true. Thanks edited.
$endgroup$
– alext87
Sep 24 '10 at 18:40
1
$begingroup$
Thanks; +1 for that. You left off the punch line, though, and it might help to make it explicit by stating how one usually thinks of skewness (asymmetry) and kurtosis (peakedness or fat-tailedness). That would make it obvious that your answer really does address the request for "ways to think of" the moments.
$endgroup$
– whuber
Sep 24 '10 at 18:46
$begingroup$
That was the wiki page's job. Though in hindsight it makes it a better answer. Edited +1. :) Whuber, out of curiousity, have you used the higher (central) moments for any practical work?
$endgroup$
– alext87
Sep 24 '10 at 19:04
$begingroup$
It depends on what you mean. If you mean for data analysis, the answer is a resounding no: even for quite large datasets, the higher moments are extremely sensitive to outliers and are not robust enough for drawing useful conclusions. Even the use of the third and fourth moments is awfully old-fashioned for that reason, but these moments can occasionally be useful measurements of the skewness and tailedness of an empirical distribution. Modern statistics tends to use extreme quantiles or order statistics to characterize the tails.
$endgroup$
– whuber
Sep 30 '10 at 4:38
add a comment |
$begingroup$
The localization and the dispersion probabilistic measures can be "seen" as the corresponding momentums of mechanical systems of "probabilistic masses".
The expectation has the following mechanical interpretation. Given that $F(x)$ is the "probabilistic mass" contained in the interval $0le Xlt x$ (in one dimention), the mathematical expectation of the random variable $X$ is the static momentum with respect to the origin of the "probabilistic masses" system.
The variance is the mechanical analog of the inertia momentum of the "probabilistic masses" system with respect to the center of masses.
The variance of the random variable $X$ is the 2nd order momentum of $X-m$ where $m$ is the expectation of $X$, i.e. $m=E(X)=displaystyleint_0^x x, dF(x)$ ($F(x)$ is the cumulative distribution function).
The $k$-order momentum is the expectation of $(X-m)^k$.
$endgroup$
add a comment |
$begingroup$
The variance is the expected squared deviation between the random variable and its expectation.
It measures how much the random variable scatters or spreads around its expectation.
$endgroup$
$begingroup$
This is the definition, but I was hoping for some intuitive insight.
$endgroup$
– Pandora
Sep 24 '10 at 17:20
$begingroup$
I was hoping that my second sentence would provide some intuition. It contains the common way of thinking of the variance.
$endgroup$
– Rasmus
Sep 24 '10 at 17:22
$begingroup$
+1 for making the effort in the second sentence. Note, though, that the variance only indirectly measures spread (because it's not even in the same units of measurement): its square root, the standard deviation, more directly reflects spread.
$endgroup$
– whuber
Sep 24 '10 at 18:28
add a comment |
Your Answer
StackExchange.ifUsing("editor", function ()
return StackExchange.using("mathjaxEditing", function ()
StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix)
StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
);
);
, "mathjax-editing");
StackExchange.ready(function()
var channelOptions =
tags: "".split(" "),
id: "69"
;
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function()
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled)
StackExchange.using("snippets", function()
createEditor();
);
else
createEditor();
);
function createEditor()
StackExchange.prepareEditor(
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader:
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
,
noCode: true, onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
);
);
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f5392%2fintuitive-explanation-of-variance-and-moment-in-probability%23new-answer', 'question_page');
);
Post as a guest
Required, but never shown
4 Answers
4
active
oldest
votes
4 Answers
4
active
oldest
votes
active
oldest
votes
active
oldest
votes
$begingroup$
There are many good answers possible, depending on how one's intuition works and one's experience with related things like averages, $L^p$ norms, etc. One answer that builds on your intuition of an expectation is to think of the higher central moments as weighted expectations. The central third moment, for example, is defined as the expectation of the third powers of deviations from the mean. Split a third power into two parts: one is the deviation from the mean (let's call it, after common statistical parlance, the "residual") and the other is the squared deviation from the mean (ergo, the squared residual). Treat the latter as a weight. As the residual gets larger in size, these weights get larger (in a quadratic fashion). The third central moment is the weighted "average" of the residuals. (I have to put "average" in quotes because, strictly speaking, to deserve being called an average one would require the weights themselves to average out to 1, but they don't usually do that.) Thus, the third central moment gives disproportionately greater emphasis to larger residuals compared to smaller ones. (It is possible for that emphasis to be so great that the third central moment doesn't even exist: if the probability doesn't decay rapidly enough with the size of the residual, the weights can swamp the probability, with the net effect of creating an infinitely great third central moment.)
You can understand, and even analyze, all other central moments in the same way, even fractional moments and absolute moments. It should be immediately clear, for instance, that the higher the moment, the more the weights are emphasized at extreme values: higher central moments measure (average) behavior of the probability distribution at greater distances from the mean.
$endgroup$
add a comment |
$begingroup$
There are many good answers possible, depending on how one's intuition works and one's experience with related things like averages, $L^p$ norms, etc. One answer that builds on your intuition of an expectation is to think of the higher central moments as weighted expectations. The central third moment, for example, is defined as the expectation of the third powers of deviations from the mean. Split a third power into two parts: one is the deviation from the mean (let's call it, after common statistical parlance, the "residual") and the other is the squared deviation from the mean (ergo, the squared residual). Treat the latter as a weight. As the residual gets larger in size, these weights get larger (in a quadratic fashion). The third central moment is the weighted "average" of the residuals. (I have to put "average" in quotes because, strictly speaking, to deserve being called an average one would require the weights themselves to average out to 1, but they don't usually do that.) Thus, the third central moment gives disproportionately greater emphasis to larger residuals compared to smaller ones. (It is possible for that emphasis to be so great that the third central moment doesn't even exist: if the probability doesn't decay rapidly enough with the size of the residual, the weights can swamp the probability, with the net effect of creating an infinitely great third central moment.)
You can understand, and even analyze, all other central moments in the same way, even fractional moments and absolute moments. It should be immediately clear, for instance, that the higher the moment, the more the weights are emphasized at extreme values: higher central moments measure (average) behavior of the probability distribution at greater distances from the mean.
$endgroup$
add a comment |
$begingroup$
There are many good answers possible, depending on how one's intuition works and one's experience with related things like averages, $L^p$ norms, etc. One answer that builds on your intuition of an expectation is to think of the higher central moments as weighted expectations. The central third moment, for example, is defined as the expectation of the third powers of deviations from the mean. Split a third power into two parts: one is the deviation from the mean (let's call it, after common statistical parlance, the "residual") and the other is the squared deviation from the mean (ergo, the squared residual). Treat the latter as a weight. As the residual gets larger in size, these weights get larger (in a quadratic fashion). The third central moment is the weighted "average" of the residuals. (I have to put "average" in quotes because, strictly speaking, to deserve being called an average one would require the weights themselves to average out to 1, but they don't usually do that.) Thus, the third central moment gives disproportionately greater emphasis to larger residuals compared to smaller ones. (It is possible for that emphasis to be so great that the third central moment doesn't even exist: if the probability doesn't decay rapidly enough with the size of the residual, the weights can swamp the probability, with the net effect of creating an infinitely great third central moment.)
You can understand, and even analyze, all other central moments in the same way, even fractional moments and absolute moments. It should be immediately clear, for instance, that the higher the moment, the more the weights are emphasized at extreme values: higher central moments measure (average) behavior of the probability distribution at greater distances from the mean.
$endgroup$
There are many good answers possible, depending on how one's intuition works and one's experience with related things like averages, $L^p$ norms, etc. One answer that builds on your intuition of an expectation is to think of the higher central moments as weighted expectations. The central third moment, for example, is defined as the expectation of the third powers of deviations from the mean. Split a third power into two parts: one is the deviation from the mean (let's call it, after common statistical parlance, the "residual") and the other is the squared deviation from the mean (ergo, the squared residual). Treat the latter as a weight. As the residual gets larger in size, these weights get larger (in a quadratic fashion). The third central moment is the weighted "average" of the residuals. (I have to put "average" in quotes because, strictly speaking, to deserve being called an average one would require the weights themselves to average out to 1, but they don't usually do that.) Thus, the third central moment gives disproportionately greater emphasis to larger residuals compared to smaller ones. (It is possible for that emphasis to be so great that the third central moment doesn't even exist: if the probability doesn't decay rapidly enough with the size of the residual, the weights can swamp the probability, with the net effect of creating an infinitely great third central moment.)
You can understand, and even analyze, all other central moments in the same way, even fractional moments and absolute moments. It should be immediately clear, for instance, that the higher the moment, the more the weights are emphasized at extreme values: higher central moments measure (average) behavior of the probability distribution at greater distances from the mean.
answered Sep 24 '10 at 18:43
whuberwhuber
7,2232134
7,2232134
add a comment |
add a comment |
$begingroup$
Variance of a r.v. $X$ is $Var(X) = mathbbE[(X-mathbbE(X))^2]$. Variance is used as a measure for the dispersion of a random variable. If you are interested in how far a random variable is from what you expect; perhaps you would consider $mathbbE(|X-mathbbE(X)|)$ but see: https://stats.stackexchange.com/questions/118/standard-deviation-why-square-the-difference-instead-of-taking-the-absolute-val for why this we use the square instead.
For all except the first moment (i.e. the mean) the central moments (centred about the mean) are far more interesting.
As for the $r$th (central) moment of $X$, that is $mathbbE[(X-mathbbE(X))^r]$. The lower moments ($r<5$) are related to properties of the distribution:
$r=0$ is $1$
$r=1$ is $0$
$r=2$ is the variance which is a measure of the spread from the mean.
$r=3$ is related to skewness which is a measure of the asymmetry of the distribution. See: http://en.wikipedia.org/wiki/Skewness
$r=4$ is related to kurtois which is a measure of the 'peakedness' of a distribution. See: http://en.wikipedia.org/wiki/Kurtosis
For $rgeq5$ they become slightly harder to compute. The $5th$ central moment is a measure for the asymmetry of the distributions tails, but I would use skewness instead.
$endgroup$
2
$begingroup$
$r=3$ and $r=4$ are only related to skewness and kurtosis. For these moments to merit these names, they have to be normalized by an appropriate measure of spread (to wit, the standard deviation).
$endgroup$
– whuber
Sep 24 '10 at 18:32
$begingroup$
Yes that is true. Thanks edited.
$endgroup$
– alext87
Sep 24 '10 at 18:40
1
$begingroup$
Thanks; +1 for that. You left off the punch line, though, and it might help to make it explicit by stating how one usually thinks of skewness (asymmetry) and kurtosis (peakedness or fat-tailedness). That would make it obvious that your answer really does address the request for "ways to think of" the moments.
$endgroup$
– whuber
Sep 24 '10 at 18:46
$begingroup$
That was the wiki page's job. Though in hindsight it makes it a better answer. Edited +1. :) Whuber, out of curiousity, have you used the higher (central) moments for any practical work?
$endgroup$
– alext87
Sep 24 '10 at 19:04
$begingroup$
It depends on what you mean. If you mean for data analysis, the answer is a resounding no: even for quite large datasets, the higher moments are extremely sensitive to outliers and are not robust enough for drawing useful conclusions. Even the use of the third and fourth moments is awfully old-fashioned for that reason, but these moments can occasionally be useful measurements of the skewness and tailedness of an empirical distribution. Modern statistics tends to use extreme quantiles or order statistics to characterize the tails.
$endgroup$
– whuber
Sep 30 '10 at 4:38
add a comment |
$begingroup$
Variance of a r.v. $X$ is $Var(X) = mathbbE[(X-mathbbE(X))^2]$. Variance is used as a measure for the dispersion of a random variable. If you are interested in how far a random variable is from what you expect; perhaps you would consider $mathbbE(|X-mathbbE(X)|)$ but see: https://stats.stackexchange.com/questions/118/standard-deviation-why-square-the-difference-instead-of-taking-the-absolute-val for why this we use the square instead.
For all except the first moment (i.e. the mean) the central moments (centred about the mean) are far more interesting.
As for the $r$th (central) moment of $X$, that is $mathbbE[(X-mathbbE(X))^r]$. The lower moments ($r<5$) are related to properties of the distribution:
$r=0$ is $1$
$r=1$ is $0$
$r=2$ is the variance which is a measure of the spread from the mean.
$r=3$ is related to skewness which is a measure of the asymmetry of the distribution. See: http://en.wikipedia.org/wiki/Skewness
$r=4$ is related to kurtois which is a measure of the 'peakedness' of a distribution. See: http://en.wikipedia.org/wiki/Kurtosis
For $rgeq5$ they become slightly harder to compute. The $5th$ central moment is a measure for the asymmetry of the distributions tails, but I would use skewness instead.
$endgroup$
2
$begingroup$
$r=3$ and $r=4$ are only related to skewness and kurtosis. For these moments to merit these names, they have to be normalized by an appropriate measure of spread (to wit, the standard deviation).
$endgroup$
– whuber
Sep 24 '10 at 18:32
$begingroup$
Yes that is true. Thanks edited.
$endgroup$
– alext87
Sep 24 '10 at 18:40
1
$begingroup$
Thanks; +1 for that. You left off the punch line, though, and it might help to make it explicit by stating how one usually thinks of skewness (asymmetry) and kurtosis (peakedness or fat-tailedness). That would make it obvious that your answer really does address the request for "ways to think of" the moments.
$endgroup$
– whuber
Sep 24 '10 at 18:46
$begingroup$
That was the wiki page's job. Though in hindsight it makes it a better answer. Edited +1. :) Whuber, out of curiousity, have you used the higher (central) moments for any practical work?
$endgroup$
– alext87
Sep 24 '10 at 19:04
$begingroup$
It depends on what you mean. If you mean for data analysis, the answer is a resounding no: even for quite large datasets, the higher moments are extremely sensitive to outliers and are not robust enough for drawing useful conclusions. Even the use of the third and fourth moments is awfully old-fashioned for that reason, but these moments can occasionally be useful measurements of the skewness and tailedness of an empirical distribution. Modern statistics tends to use extreme quantiles or order statistics to characterize the tails.
$endgroup$
– whuber
Sep 30 '10 at 4:38
add a comment |
$begingroup$
Variance of a r.v. $X$ is $Var(X) = mathbbE[(X-mathbbE(X))^2]$. Variance is used as a measure for the dispersion of a random variable. If you are interested in how far a random variable is from what you expect; perhaps you would consider $mathbbE(|X-mathbbE(X)|)$ but see: https://stats.stackexchange.com/questions/118/standard-deviation-why-square-the-difference-instead-of-taking-the-absolute-val for why this we use the square instead.
For all except the first moment (i.e. the mean) the central moments (centred about the mean) are far more interesting.
As for the $r$th (central) moment of $X$, that is $mathbbE[(X-mathbbE(X))^r]$. The lower moments ($r<5$) are related to properties of the distribution:
$r=0$ is $1$
$r=1$ is $0$
$r=2$ is the variance which is a measure of the spread from the mean.
$r=3$ is related to skewness which is a measure of the asymmetry of the distribution. See: http://en.wikipedia.org/wiki/Skewness
$r=4$ is related to kurtois which is a measure of the 'peakedness' of a distribution. See: http://en.wikipedia.org/wiki/Kurtosis
For $rgeq5$ they become slightly harder to compute. The $5th$ central moment is a measure for the asymmetry of the distributions tails, but I would use skewness instead.
$endgroup$
Variance of a r.v. $X$ is $Var(X) = mathbbE[(X-mathbbE(X))^2]$. Variance is used as a measure for the dispersion of a random variable. If you are interested in how far a random variable is from what you expect; perhaps you would consider $mathbbE(|X-mathbbE(X)|)$ but see: https://stats.stackexchange.com/questions/118/standard-deviation-why-square-the-difference-instead-of-taking-the-absolute-val for why this we use the square instead.
For all except the first moment (i.e. the mean) the central moments (centred about the mean) are far more interesting.
As for the $r$th (central) moment of $X$, that is $mathbbE[(X-mathbbE(X))^r]$. The lower moments ($r<5$) are related to properties of the distribution:
$r=0$ is $1$
$r=1$ is $0$
$r=2$ is the variance which is a measure of the spread from the mean.
$r=3$ is related to skewness which is a measure of the asymmetry of the distribution. See: http://en.wikipedia.org/wiki/Skewness
$r=4$ is related to kurtois which is a measure of the 'peakedness' of a distribution. See: http://en.wikipedia.org/wiki/Kurtosis
For $rgeq5$ they become slightly harder to compute. The $5th$ central moment is a measure for the asymmetry of the distributions tails, but I would use skewness instead.
edited Apr 13 '17 at 12:44
Community♦
1
1
answered Sep 24 '10 at 17:47
alext87alext87
2,173818
2,173818
2
$begingroup$
$r=3$ and $r=4$ are only related to skewness and kurtosis. For these moments to merit these names, they have to be normalized by an appropriate measure of spread (to wit, the standard deviation).
$endgroup$
– whuber
Sep 24 '10 at 18:32
$begingroup$
Yes that is true. Thanks edited.
$endgroup$
– alext87
Sep 24 '10 at 18:40
1
$begingroup$
Thanks; +1 for that. You left off the punch line, though, and it might help to make it explicit by stating how one usually thinks of skewness (asymmetry) and kurtosis (peakedness or fat-tailedness). That would make it obvious that your answer really does address the request for "ways to think of" the moments.
$endgroup$
– whuber
Sep 24 '10 at 18:46
$begingroup$
That was the wiki page's job. Though in hindsight it makes it a better answer. Edited +1. :) Whuber, out of curiousity, have you used the higher (central) moments for any practical work?
$endgroup$
– alext87
Sep 24 '10 at 19:04
$begingroup$
It depends on what you mean. If you mean for data analysis, the answer is a resounding no: even for quite large datasets, the higher moments are extremely sensitive to outliers and are not robust enough for drawing useful conclusions. Even the use of the third and fourth moments is awfully old-fashioned for that reason, but these moments can occasionally be useful measurements of the skewness and tailedness of an empirical distribution. Modern statistics tends to use extreme quantiles or order statistics to characterize the tails.
$endgroup$
– whuber
Sep 30 '10 at 4:38
add a comment |
2
$begingroup$
$r=3$ and $r=4$ are only related to skewness and kurtosis. For these moments to merit these names, they have to be normalized by an appropriate measure of spread (to wit, the standard deviation).
$endgroup$
– whuber
Sep 24 '10 at 18:32
$begingroup$
Yes that is true. Thanks edited.
$endgroup$
– alext87
Sep 24 '10 at 18:40
1
$begingroup$
Thanks; +1 for that. You left off the punch line, though, and it might help to make it explicit by stating how one usually thinks of skewness (asymmetry) and kurtosis (peakedness or fat-tailedness). That would make it obvious that your answer really does address the request for "ways to think of" the moments.
$endgroup$
– whuber
Sep 24 '10 at 18:46
$begingroup$
That was the wiki page's job. Though in hindsight it makes it a better answer. Edited +1. :) Whuber, out of curiousity, have you used the higher (central) moments for any practical work?
$endgroup$
– alext87
Sep 24 '10 at 19:04
$begingroup$
It depends on what you mean. If you mean for data analysis, the answer is a resounding no: even for quite large datasets, the higher moments are extremely sensitive to outliers and are not robust enough for drawing useful conclusions. Even the use of the third and fourth moments is awfully old-fashioned for that reason, but these moments can occasionally be useful measurements of the skewness and tailedness of an empirical distribution. Modern statistics tends to use extreme quantiles or order statistics to characterize the tails.
$endgroup$
– whuber
Sep 30 '10 at 4:38
2
2
$begingroup$
$r=3$ and $r=4$ are only related to skewness and kurtosis. For these moments to merit these names, they have to be normalized by an appropriate measure of spread (to wit, the standard deviation).
$endgroup$
– whuber
Sep 24 '10 at 18:32
$begingroup$
$r=3$ and $r=4$ are only related to skewness and kurtosis. For these moments to merit these names, they have to be normalized by an appropriate measure of spread (to wit, the standard deviation).
$endgroup$
– whuber
Sep 24 '10 at 18:32
$begingroup$
Yes that is true. Thanks edited.
$endgroup$
– alext87
Sep 24 '10 at 18:40
$begingroup$
Yes that is true. Thanks edited.
$endgroup$
– alext87
Sep 24 '10 at 18:40
1
1
$begingroup$
Thanks; +1 for that. You left off the punch line, though, and it might help to make it explicit by stating how one usually thinks of skewness (asymmetry) and kurtosis (peakedness or fat-tailedness). That would make it obvious that your answer really does address the request for "ways to think of" the moments.
$endgroup$
– whuber
Sep 24 '10 at 18:46
$begingroup$
Thanks; +1 for that. You left off the punch line, though, and it might help to make it explicit by stating how one usually thinks of skewness (asymmetry) and kurtosis (peakedness or fat-tailedness). That would make it obvious that your answer really does address the request for "ways to think of" the moments.
$endgroup$
– whuber
Sep 24 '10 at 18:46
$begingroup$
That was the wiki page's job. Though in hindsight it makes it a better answer. Edited +1. :) Whuber, out of curiousity, have you used the higher (central) moments for any practical work?
$endgroup$
– alext87
Sep 24 '10 at 19:04
$begingroup$
That was the wiki page's job. Though in hindsight it makes it a better answer. Edited +1. :) Whuber, out of curiousity, have you used the higher (central) moments for any practical work?
$endgroup$
– alext87
Sep 24 '10 at 19:04
$begingroup$
It depends on what you mean. If you mean for data analysis, the answer is a resounding no: even for quite large datasets, the higher moments are extremely sensitive to outliers and are not robust enough for drawing useful conclusions. Even the use of the third and fourth moments is awfully old-fashioned for that reason, but these moments can occasionally be useful measurements of the skewness and tailedness of an empirical distribution. Modern statistics tends to use extreme quantiles or order statistics to characterize the tails.
$endgroup$
– whuber
Sep 30 '10 at 4:38
$begingroup$
It depends on what you mean. If you mean for data analysis, the answer is a resounding no: even for quite large datasets, the higher moments are extremely sensitive to outliers and are not robust enough for drawing useful conclusions. Even the use of the third and fourth moments is awfully old-fashioned for that reason, but these moments can occasionally be useful measurements of the skewness and tailedness of an empirical distribution. Modern statistics tends to use extreme quantiles or order statistics to characterize the tails.
$endgroup$
– whuber
Sep 30 '10 at 4:38
add a comment |
$begingroup$
The localization and the dispersion probabilistic measures can be "seen" as the corresponding momentums of mechanical systems of "probabilistic masses".
The expectation has the following mechanical interpretation. Given that $F(x)$ is the "probabilistic mass" contained in the interval $0le Xlt x$ (in one dimention), the mathematical expectation of the random variable $X$ is the static momentum with respect to the origin of the "probabilistic masses" system.
The variance is the mechanical analog of the inertia momentum of the "probabilistic masses" system with respect to the center of masses.
The variance of the random variable $X$ is the 2nd order momentum of $X-m$ where $m$ is the expectation of $X$, i.e. $m=E(X)=displaystyleint_0^x x, dF(x)$ ($F(x)$ is the cumulative distribution function).
The $k$-order momentum is the expectation of $(X-m)^k$.
$endgroup$
add a comment |
$begingroup$
The localization and the dispersion probabilistic measures can be "seen" as the corresponding momentums of mechanical systems of "probabilistic masses".
The expectation has the following mechanical interpretation. Given that $F(x)$ is the "probabilistic mass" contained in the interval $0le Xlt x$ (in one dimention), the mathematical expectation of the random variable $X$ is the static momentum with respect to the origin of the "probabilistic masses" system.
The variance is the mechanical analog of the inertia momentum of the "probabilistic masses" system with respect to the center of masses.
The variance of the random variable $X$ is the 2nd order momentum of $X-m$ where $m$ is the expectation of $X$, i.e. $m=E(X)=displaystyleint_0^x x, dF(x)$ ($F(x)$ is the cumulative distribution function).
The $k$-order momentum is the expectation of $(X-m)^k$.
$endgroup$
add a comment |
$begingroup$
The localization and the dispersion probabilistic measures can be "seen" as the corresponding momentums of mechanical systems of "probabilistic masses".
The expectation has the following mechanical interpretation. Given that $F(x)$ is the "probabilistic mass" contained in the interval $0le Xlt x$ (in one dimention), the mathematical expectation of the random variable $X$ is the static momentum with respect to the origin of the "probabilistic masses" system.
The variance is the mechanical analog of the inertia momentum of the "probabilistic masses" system with respect to the center of masses.
The variance of the random variable $X$ is the 2nd order momentum of $X-m$ where $m$ is the expectation of $X$, i.e. $m=E(X)=displaystyleint_0^x x, dF(x)$ ($F(x)$ is the cumulative distribution function).
The $k$-order momentum is the expectation of $(X-m)^k$.
$endgroup$
The localization and the dispersion probabilistic measures can be "seen" as the corresponding momentums of mechanical systems of "probabilistic masses".
The expectation has the following mechanical interpretation. Given that $F(x)$ is the "probabilistic mass" contained in the interval $0le Xlt x$ (in one dimention), the mathematical expectation of the random variable $X$ is the static momentum with respect to the origin of the "probabilistic masses" system.
The variance is the mechanical analog of the inertia momentum of the "probabilistic masses" system with respect to the center of masses.
The variance of the random variable $X$ is the 2nd order momentum of $X-m$ where $m$ is the expectation of $X$, i.e. $m=E(X)=displaystyleint_0^x x, dF(x)$ ($F(x)$ is the cumulative distribution function).
The $k$-order momentum is the expectation of $(X-m)^k$.
edited Mar 28 at 14:15
answered Sep 25 '10 at 23:22
Américo TavaresAmérico Tavares
32.6k1181206
32.6k1181206
add a comment |
add a comment |
$begingroup$
The variance is the expected squared deviation between the random variable and its expectation.
It measures how much the random variable scatters or spreads around its expectation.
$endgroup$
$begingroup$
This is the definition, but I was hoping for some intuitive insight.
$endgroup$
– Pandora
Sep 24 '10 at 17:20
$begingroup$
I was hoping that my second sentence would provide some intuition. It contains the common way of thinking of the variance.
$endgroup$
– Rasmus
Sep 24 '10 at 17:22
$begingroup$
+1 for making the effort in the second sentence. Note, though, that the variance only indirectly measures spread (because it's not even in the same units of measurement): its square root, the standard deviation, more directly reflects spread.
$endgroup$
– whuber
Sep 24 '10 at 18:28
add a comment |
$begingroup$
The variance is the expected squared deviation between the random variable and its expectation.
It measures how much the random variable scatters or spreads around its expectation.
$endgroup$
$begingroup$
This is the definition, but I was hoping for some intuitive insight.
$endgroup$
– Pandora
Sep 24 '10 at 17:20
$begingroup$
I was hoping that my second sentence would provide some intuition. It contains the common way of thinking of the variance.
$endgroup$
– Rasmus
Sep 24 '10 at 17:22
$begingroup$
+1 for making the effort in the second sentence. Note, though, that the variance only indirectly measures spread (because it's not even in the same units of measurement): its square root, the standard deviation, more directly reflects spread.
$endgroup$
– whuber
Sep 24 '10 at 18:28
add a comment |
$begingroup$
The variance is the expected squared deviation between the random variable and its expectation.
It measures how much the random variable scatters or spreads around its expectation.
$endgroup$
The variance is the expected squared deviation between the random variable and its expectation.
It measures how much the random variable scatters or spreads around its expectation.
answered Sep 24 '10 at 17:17
RasmusRasmus
14.3k14479
14.3k14479
$begingroup$
This is the definition, but I was hoping for some intuitive insight.
$endgroup$
– Pandora
Sep 24 '10 at 17:20
$begingroup$
I was hoping that my second sentence would provide some intuition. It contains the common way of thinking of the variance.
$endgroup$
– Rasmus
Sep 24 '10 at 17:22
$begingroup$
+1 for making the effort in the second sentence. Note, though, that the variance only indirectly measures spread (because it's not even in the same units of measurement): its square root, the standard deviation, more directly reflects spread.
$endgroup$
– whuber
Sep 24 '10 at 18:28
add a comment |
$begingroup$
This is the definition, but I was hoping for some intuitive insight.
$endgroup$
– Pandora
Sep 24 '10 at 17:20
$begingroup$
I was hoping that my second sentence would provide some intuition. It contains the common way of thinking of the variance.
$endgroup$
– Rasmus
Sep 24 '10 at 17:22
$begingroup$
+1 for making the effort in the second sentence. Note, though, that the variance only indirectly measures spread (because it's not even in the same units of measurement): its square root, the standard deviation, more directly reflects spread.
$endgroup$
– whuber
Sep 24 '10 at 18:28
$begingroup$
This is the definition, but I was hoping for some intuitive insight.
$endgroup$
– Pandora
Sep 24 '10 at 17:20
$begingroup$
This is the definition, but I was hoping for some intuitive insight.
$endgroup$
– Pandora
Sep 24 '10 at 17:20
$begingroup$
I was hoping that my second sentence would provide some intuition. It contains the common way of thinking of the variance.
$endgroup$
– Rasmus
Sep 24 '10 at 17:22
$begingroup$
I was hoping that my second sentence would provide some intuition. It contains the common way of thinking of the variance.
$endgroup$
– Rasmus
Sep 24 '10 at 17:22
$begingroup$
+1 for making the effort in the second sentence. Note, though, that the variance only indirectly measures spread (because it's not even in the same units of measurement): its square root, the standard deviation, more directly reflects spread.
$endgroup$
– whuber
Sep 24 '10 at 18:28
$begingroup$
+1 for making the effort in the second sentence. Note, though, that the variance only indirectly measures spread (because it's not even in the same units of measurement): its square root, the standard deviation, more directly reflects spread.
$endgroup$
– whuber
Sep 24 '10 at 18:28
add a comment |
Thanks for contributing an answer to Mathematics Stack Exchange!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
Use MathJax to format equations. MathJax reference.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f5392%2fintuitive-explanation-of-variance-and-moment-in-probability%23new-answer', 'question_page');
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
$begingroup$
Why the edit, more than 4 years later?
$endgroup$
– Did
Dec 20 '14 at 13:05
$begingroup$
@Did -- Sorry, was checking something and forgot edits bump up questions :(
$endgroup$
– Pandora
Dec 20 '14 at 13:13