Intuitive explanation of variance and moment in ProbabilityIntuition behind moments of random variablesUsefulness of VarianceWhat is variance intutively? and is there a math proof for variance ? how is variance and covarience relatedFor System dependent on normally distributed parameter, are deviations added or variations?Intuitive meaning of Pearson Product-moment correlation coefficient FormulaIntuitive explanation of the tower property of conditional expectationMotivation behind the ingredients of First Cohomology group $H^1$Intuition behind Variance forumlaIntuitive Approach to Sheaf and Cech CohomologyWhat is an intuitive explanation for Birkhoff's ergodic theorem?Understanding expectation, variance and standard deviationhow to interpret the variance of a variance?Intuition behind moments of random variablesFind variance and general formula for for r$^th$ moment for random variable uniform over (0,1)

What does the same-ish mean?

What is the opposite of "eschatology"?

Is it a bad idea to plug the other end of ESD strap to wall ground?

Ambiguity in the definition of entropy

What historical events would have to change in order to make 19th century "steampunk" technology possible?

Using "tail" to follow a file without displaying the most recent lines

Avoiding the "not like other girls" trope?

How dangerous is XSS

How many wives did king shaul have

Forgetting the musical notes while performing in concert

Are British MPs missing the point, with these 'Indicative Votes'?

How to travel to Japan while expressing milk?

How to remove border from elements in the last row?

OP Amp not amplifying audio signal

Theorists sure want true answers to this!

How do I exit BASH while loop using modulus operator?

Implication of namely

What is the fastest integer factorization to break RSA?

Unlock My Phone! February 2018

Why is it a bad idea to hire a hitman to eliminate most corrupt politicians?

Can compressed videos be decoded back to their uncompresed original format?

Where would I need my direct neural interface to be implanted?

Is there a hemisphere-neutral way of specifying a season?

Processor speed limited at 0.4 Ghz



Intuitive explanation of variance and moment in Probability


Intuition behind moments of random variablesUsefulness of VarianceWhat is variance intutively? and is there a math proof for variance ? how is variance and covarience relatedFor System dependent on normally distributed parameter, are deviations added or variations?Intuitive meaning of Pearson Product-moment correlation coefficient FormulaIntuitive explanation of the tower property of conditional expectationMotivation behind the ingredients of First Cohomology group $H^1$Intuition behind Variance forumlaIntuitive Approach to Sheaf and Cech CohomologyWhat is an intuitive explanation for Birkhoff's ergodic theorem?Understanding expectation, variance and standard deviationhow to interpret the variance of a variance?Intuition behind moments of random variablesFind variance and general formula for for r$^th$ moment for random variable uniform over (0,1)













17












$begingroup$


While I understand the intuition behind expectation, I don't really understand the meaning of variance and moment.



What is a good way to think of those two terms?










share|cite|improve this question











$endgroup$











  • $begingroup$
    Why the edit, more than 4 years later?
    $endgroup$
    – Did
    Dec 20 '14 at 13:05










  • $begingroup$
    @Did -- Sorry, was checking something and forgot edits bump up questions :(
    $endgroup$
    – Pandora
    Dec 20 '14 at 13:13















17












$begingroup$


While I understand the intuition behind expectation, I don't really understand the meaning of variance and moment.



What is a good way to think of those two terms?










share|cite|improve this question











$endgroup$











  • $begingroup$
    Why the edit, more than 4 years later?
    $endgroup$
    – Did
    Dec 20 '14 at 13:05










  • $begingroup$
    @Did -- Sorry, was checking something and forgot edits bump up questions :(
    $endgroup$
    – Pandora
    Dec 20 '14 at 13:13













17












17








17


12



$begingroup$


While I understand the intuition behind expectation, I don't really understand the meaning of variance and moment.



What is a good way to think of those two terms?










share|cite|improve this question











$endgroup$




While I understand the intuition behind expectation, I don't really understand the meaning of variance and moment.



What is a good way to think of those two terms?







probability intuition probability-theory






share|cite|improve this question















share|cite|improve this question













share|cite|improve this question




share|cite|improve this question








edited Dec 20 '14 at 12:32







Pandora

















asked Sep 24 '10 at 17:14









PandoraPandora

2,72113157




2,72113157











  • $begingroup$
    Why the edit, more than 4 years later?
    $endgroup$
    – Did
    Dec 20 '14 at 13:05










  • $begingroup$
    @Did -- Sorry, was checking something and forgot edits bump up questions :(
    $endgroup$
    – Pandora
    Dec 20 '14 at 13:13
















  • $begingroup$
    Why the edit, more than 4 years later?
    $endgroup$
    – Did
    Dec 20 '14 at 13:05










  • $begingroup$
    @Did -- Sorry, was checking something and forgot edits bump up questions :(
    $endgroup$
    – Pandora
    Dec 20 '14 at 13:13















$begingroup$
Why the edit, more than 4 years later?
$endgroup$
– Did
Dec 20 '14 at 13:05




$begingroup$
Why the edit, more than 4 years later?
$endgroup$
– Did
Dec 20 '14 at 13:05












$begingroup$
@Did -- Sorry, was checking something and forgot edits bump up questions :(
$endgroup$
– Pandora
Dec 20 '14 at 13:13




$begingroup$
@Did -- Sorry, was checking something and forgot edits bump up questions :(
$endgroup$
– Pandora
Dec 20 '14 at 13:13










4 Answers
4






active

oldest

votes


















7












$begingroup$

There are many good answers possible, depending on how one's intuition works and one's experience with related things like averages, $L^p$ norms, etc. One answer that builds on your intuition of an expectation is to think of the higher central moments as weighted expectations. The central third moment, for example, is defined as the expectation of the third powers of deviations from the mean. Split a third power into two parts: one is the deviation from the mean (let's call it, after common statistical parlance, the "residual") and the other is the squared deviation from the mean (ergo, the squared residual). Treat the latter as a weight. As the residual gets larger in size, these weights get larger (in a quadratic fashion). The third central moment is the weighted "average" of the residuals. (I have to put "average" in quotes because, strictly speaking, to deserve being called an average one would require the weights themselves to average out to 1, but they don't usually do that.) Thus, the third central moment gives disproportionately greater emphasis to larger residuals compared to smaller ones. (It is possible for that emphasis to be so great that the third central moment doesn't even exist: if the probability doesn't decay rapidly enough with the size of the residual, the weights can swamp the probability, with the net effect of creating an infinitely great third central moment.)



You can understand, and even analyze, all other central moments in the same way, even fractional moments and absolute moments. It should be immediately clear, for instance, that the higher the moment, the more the weights are emphasized at extreme values: higher central moments measure (average) behavior of the probability distribution at greater distances from the mean.






share|cite|improve this answer









$endgroup$




















    10












    $begingroup$

    Variance of a r.v. $X$ is $Var(X) = mathbbE[(X-mathbbE(X))^2]$. Variance is used as a measure for the dispersion of a random variable. If you are interested in how far a random variable is from what you expect; perhaps you would consider $mathbbE(|X-mathbbE(X)|)$ but see: https://stats.stackexchange.com/questions/118/standard-deviation-why-square-the-difference-instead-of-taking-the-absolute-val for why this we use the square instead.



    For all except the first moment (i.e. the mean) the central moments (centred about the mean) are far more interesting.



    As for the $r$th (central) moment of $X$, that is $mathbbE[(X-mathbbE(X))^r]$. The lower moments ($r<5$) are related to properties of the distribution:



    $r=0$ is $1$



    $r=1$ is $0$



    $r=2$ is the variance which is a measure of the spread from the mean.



    $r=3$ is related to skewness which is a measure of the asymmetry of the distribution. See: http://en.wikipedia.org/wiki/Skewness



    $r=4$ is related to kurtois which is a measure of the 'peakedness' of a distribution. See: http://en.wikipedia.org/wiki/Kurtosis



    For $rgeq5$ they become slightly harder to compute. The $5th$ central moment is a measure for the asymmetry of the distributions tails, but I would use skewness instead.






    share|cite|improve this answer











    $endgroup$








    • 2




      $begingroup$
      $r=3$ and $r=4$ are only related to skewness and kurtosis. For these moments to merit these names, they have to be normalized by an appropriate measure of spread (to wit, the standard deviation).
      $endgroup$
      – whuber
      Sep 24 '10 at 18:32










    • $begingroup$
      Yes that is true. Thanks edited.
      $endgroup$
      – alext87
      Sep 24 '10 at 18:40






    • 1




      $begingroup$
      Thanks; +1 for that. You left off the punch line, though, and it might help to make it explicit by stating how one usually thinks of skewness (asymmetry) and kurtosis (peakedness or fat-tailedness). That would make it obvious that your answer really does address the request for "ways to think of" the moments.
      $endgroup$
      – whuber
      Sep 24 '10 at 18:46










    • $begingroup$
      That was the wiki page's job. Though in hindsight it makes it a better answer. Edited +1. :) Whuber, out of curiousity, have you used the higher (central) moments for any practical work?
      $endgroup$
      – alext87
      Sep 24 '10 at 19:04










    • $begingroup$
      It depends on what you mean. If you mean for data analysis, the answer is a resounding no: even for quite large datasets, the higher moments are extremely sensitive to outliers and are not robust enough for drawing useful conclusions. Even the use of the third and fourth moments is awfully old-fashioned for that reason, but these moments can occasionally be useful measurements of the skewness and tailedness of an empirical distribution. Modern statistics tends to use extreme quantiles or order statistics to characterize the tails.
      $endgroup$
      – whuber
      Sep 30 '10 at 4:38


















    4












    $begingroup$

    The localization and the dispersion probabilistic measures can be "seen" as the corresponding momentums of mechanical systems of "probabilistic masses".



    The expectation has the following mechanical interpretation. Given that $F(x)$ is the "probabilistic mass" contained in the interval $0le Xlt x$ (in one dimention), the mathematical expectation of the random variable $X$ is the static momentum with respect to the origin of the "probabilistic masses" system.



    The variance is the mechanical analog of the inertia momentum of the "probabilistic masses" system with respect to the center of masses.



    The variance of the random variable $X$ is the 2nd order momentum of $X-m$ where $m$ is the expectation of $X$, i.e. $m=E(X)=displaystyleint_0^x x, dF(x)$ ($F(x)$ is the cumulative distribution function).



    The $k$-order momentum is the expectation of $(X-m)^k$.






    share|cite|improve this answer











    $endgroup$




















      2












      $begingroup$

      The variance is the expected squared deviation between the random variable and its expectation.



      It measures how much the random variable scatters or spreads around its expectation.






      share|cite|improve this answer









      $endgroup$












      • $begingroup$
        This is the definition, but I was hoping for some intuitive insight.
        $endgroup$
        – Pandora
        Sep 24 '10 at 17:20










      • $begingroup$
        I was hoping that my second sentence would provide some intuition. It contains the common way of thinking of the variance.
        $endgroup$
        – Rasmus
        Sep 24 '10 at 17:22











      • $begingroup$
        +1 for making the effort in the second sentence. Note, though, that the variance only indirectly measures spread (because it's not even in the same units of measurement): its square root, the standard deviation, more directly reflects spread.
        $endgroup$
        – whuber
        Sep 24 '10 at 18:28











      Your Answer





      StackExchange.ifUsing("editor", function ()
      return StackExchange.using("mathjaxEditing", function ()
      StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix)
      StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
      );
      );
      , "mathjax-editing");

      StackExchange.ready(function()
      var channelOptions =
      tags: "".split(" "),
      id: "69"
      ;
      initTagRenderer("".split(" "), "".split(" "), channelOptions);

      StackExchange.using("externalEditor", function()
      // Have to fire editor after snippets, if snippets enabled
      if (StackExchange.settings.snippets.snippetsEnabled)
      StackExchange.using("snippets", function()
      createEditor();
      );

      else
      createEditor();

      );

      function createEditor()
      StackExchange.prepareEditor(
      heartbeatType: 'answer',
      autoActivateHeartbeat: false,
      convertImagesToLinks: true,
      noModals: true,
      showLowRepImageUploadWarning: true,
      reputationToPostImages: 10,
      bindNavPrevention: true,
      postfix: "",
      imageUploader:
      brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
      contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
      allowUrls: true
      ,
      noCode: true, onDemand: true,
      discardSelector: ".discard-answer"
      ,immediatelyShowMarkdownHelp:true
      );



      );













      draft saved

      draft discarded


















      StackExchange.ready(
      function ()
      StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f5392%2fintuitive-explanation-of-variance-and-moment-in-probability%23new-answer', 'question_page');

      );

      Post as a guest















      Required, but never shown

























      4 Answers
      4






      active

      oldest

      votes








      4 Answers
      4






      active

      oldest

      votes









      active

      oldest

      votes






      active

      oldest

      votes









      7












      $begingroup$

      There are many good answers possible, depending on how one's intuition works and one's experience with related things like averages, $L^p$ norms, etc. One answer that builds on your intuition of an expectation is to think of the higher central moments as weighted expectations. The central third moment, for example, is defined as the expectation of the third powers of deviations from the mean. Split a third power into two parts: one is the deviation from the mean (let's call it, after common statistical parlance, the "residual") and the other is the squared deviation from the mean (ergo, the squared residual). Treat the latter as a weight. As the residual gets larger in size, these weights get larger (in a quadratic fashion). The third central moment is the weighted "average" of the residuals. (I have to put "average" in quotes because, strictly speaking, to deserve being called an average one would require the weights themselves to average out to 1, but they don't usually do that.) Thus, the third central moment gives disproportionately greater emphasis to larger residuals compared to smaller ones. (It is possible for that emphasis to be so great that the third central moment doesn't even exist: if the probability doesn't decay rapidly enough with the size of the residual, the weights can swamp the probability, with the net effect of creating an infinitely great third central moment.)



      You can understand, and even analyze, all other central moments in the same way, even fractional moments and absolute moments. It should be immediately clear, for instance, that the higher the moment, the more the weights are emphasized at extreme values: higher central moments measure (average) behavior of the probability distribution at greater distances from the mean.






      share|cite|improve this answer









      $endgroup$

















        7












        $begingroup$

        There are many good answers possible, depending on how one's intuition works and one's experience with related things like averages, $L^p$ norms, etc. One answer that builds on your intuition of an expectation is to think of the higher central moments as weighted expectations. The central third moment, for example, is defined as the expectation of the third powers of deviations from the mean. Split a third power into two parts: one is the deviation from the mean (let's call it, after common statistical parlance, the "residual") and the other is the squared deviation from the mean (ergo, the squared residual). Treat the latter as a weight. As the residual gets larger in size, these weights get larger (in a quadratic fashion). The third central moment is the weighted "average" of the residuals. (I have to put "average" in quotes because, strictly speaking, to deserve being called an average one would require the weights themselves to average out to 1, but they don't usually do that.) Thus, the third central moment gives disproportionately greater emphasis to larger residuals compared to smaller ones. (It is possible for that emphasis to be so great that the third central moment doesn't even exist: if the probability doesn't decay rapidly enough with the size of the residual, the weights can swamp the probability, with the net effect of creating an infinitely great third central moment.)



        You can understand, and even analyze, all other central moments in the same way, even fractional moments and absolute moments. It should be immediately clear, for instance, that the higher the moment, the more the weights are emphasized at extreme values: higher central moments measure (average) behavior of the probability distribution at greater distances from the mean.






        share|cite|improve this answer









        $endgroup$















          7












          7








          7





          $begingroup$

          There are many good answers possible, depending on how one's intuition works and one's experience with related things like averages, $L^p$ norms, etc. One answer that builds on your intuition of an expectation is to think of the higher central moments as weighted expectations. The central third moment, for example, is defined as the expectation of the third powers of deviations from the mean. Split a third power into two parts: one is the deviation from the mean (let's call it, after common statistical parlance, the "residual") and the other is the squared deviation from the mean (ergo, the squared residual). Treat the latter as a weight. As the residual gets larger in size, these weights get larger (in a quadratic fashion). The third central moment is the weighted "average" of the residuals. (I have to put "average" in quotes because, strictly speaking, to deserve being called an average one would require the weights themselves to average out to 1, but they don't usually do that.) Thus, the third central moment gives disproportionately greater emphasis to larger residuals compared to smaller ones. (It is possible for that emphasis to be so great that the third central moment doesn't even exist: if the probability doesn't decay rapidly enough with the size of the residual, the weights can swamp the probability, with the net effect of creating an infinitely great third central moment.)



          You can understand, and even analyze, all other central moments in the same way, even fractional moments and absolute moments. It should be immediately clear, for instance, that the higher the moment, the more the weights are emphasized at extreme values: higher central moments measure (average) behavior of the probability distribution at greater distances from the mean.






          share|cite|improve this answer









          $endgroup$



          There are many good answers possible, depending on how one's intuition works and one's experience with related things like averages, $L^p$ norms, etc. One answer that builds on your intuition of an expectation is to think of the higher central moments as weighted expectations. The central third moment, for example, is defined as the expectation of the third powers of deviations from the mean. Split a third power into two parts: one is the deviation from the mean (let's call it, after common statistical parlance, the "residual") and the other is the squared deviation from the mean (ergo, the squared residual). Treat the latter as a weight. As the residual gets larger in size, these weights get larger (in a quadratic fashion). The third central moment is the weighted "average" of the residuals. (I have to put "average" in quotes because, strictly speaking, to deserve being called an average one would require the weights themselves to average out to 1, but they don't usually do that.) Thus, the third central moment gives disproportionately greater emphasis to larger residuals compared to smaller ones. (It is possible for that emphasis to be so great that the third central moment doesn't even exist: if the probability doesn't decay rapidly enough with the size of the residual, the weights can swamp the probability, with the net effect of creating an infinitely great third central moment.)



          You can understand, and even analyze, all other central moments in the same way, even fractional moments and absolute moments. It should be immediately clear, for instance, that the higher the moment, the more the weights are emphasized at extreme values: higher central moments measure (average) behavior of the probability distribution at greater distances from the mean.







          share|cite|improve this answer












          share|cite|improve this answer



          share|cite|improve this answer










          answered Sep 24 '10 at 18:43









          whuberwhuber

          7,2232134




          7,2232134





















              10












              $begingroup$

              Variance of a r.v. $X$ is $Var(X) = mathbbE[(X-mathbbE(X))^2]$. Variance is used as a measure for the dispersion of a random variable. If you are interested in how far a random variable is from what you expect; perhaps you would consider $mathbbE(|X-mathbbE(X)|)$ but see: https://stats.stackexchange.com/questions/118/standard-deviation-why-square-the-difference-instead-of-taking-the-absolute-val for why this we use the square instead.



              For all except the first moment (i.e. the mean) the central moments (centred about the mean) are far more interesting.



              As for the $r$th (central) moment of $X$, that is $mathbbE[(X-mathbbE(X))^r]$. The lower moments ($r<5$) are related to properties of the distribution:



              $r=0$ is $1$



              $r=1$ is $0$



              $r=2$ is the variance which is a measure of the spread from the mean.



              $r=3$ is related to skewness which is a measure of the asymmetry of the distribution. See: http://en.wikipedia.org/wiki/Skewness



              $r=4$ is related to kurtois which is a measure of the 'peakedness' of a distribution. See: http://en.wikipedia.org/wiki/Kurtosis



              For $rgeq5$ they become slightly harder to compute. The $5th$ central moment is a measure for the asymmetry of the distributions tails, but I would use skewness instead.






              share|cite|improve this answer











              $endgroup$








              • 2




                $begingroup$
                $r=3$ and $r=4$ are only related to skewness and kurtosis. For these moments to merit these names, they have to be normalized by an appropriate measure of spread (to wit, the standard deviation).
                $endgroup$
                – whuber
                Sep 24 '10 at 18:32










              • $begingroup$
                Yes that is true. Thanks edited.
                $endgroup$
                – alext87
                Sep 24 '10 at 18:40






              • 1




                $begingroup$
                Thanks; +1 for that. You left off the punch line, though, and it might help to make it explicit by stating how one usually thinks of skewness (asymmetry) and kurtosis (peakedness or fat-tailedness). That would make it obvious that your answer really does address the request for "ways to think of" the moments.
                $endgroup$
                – whuber
                Sep 24 '10 at 18:46










              • $begingroup$
                That was the wiki page's job. Though in hindsight it makes it a better answer. Edited +1. :) Whuber, out of curiousity, have you used the higher (central) moments for any practical work?
                $endgroup$
                – alext87
                Sep 24 '10 at 19:04










              • $begingroup$
                It depends on what you mean. If you mean for data analysis, the answer is a resounding no: even for quite large datasets, the higher moments are extremely sensitive to outliers and are not robust enough for drawing useful conclusions. Even the use of the third and fourth moments is awfully old-fashioned for that reason, but these moments can occasionally be useful measurements of the skewness and tailedness of an empirical distribution. Modern statistics tends to use extreme quantiles or order statistics to characterize the tails.
                $endgroup$
                – whuber
                Sep 30 '10 at 4:38















              10












              $begingroup$

              Variance of a r.v. $X$ is $Var(X) = mathbbE[(X-mathbbE(X))^2]$. Variance is used as a measure for the dispersion of a random variable. If you are interested in how far a random variable is from what you expect; perhaps you would consider $mathbbE(|X-mathbbE(X)|)$ but see: https://stats.stackexchange.com/questions/118/standard-deviation-why-square-the-difference-instead-of-taking-the-absolute-val for why this we use the square instead.



              For all except the first moment (i.e. the mean) the central moments (centred about the mean) are far more interesting.



              As for the $r$th (central) moment of $X$, that is $mathbbE[(X-mathbbE(X))^r]$. The lower moments ($r<5$) are related to properties of the distribution:



              $r=0$ is $1$



              $r=1$ is $0$



              $r=2$ is the variance which is a measure of the spread from the mean.



              $r=3$ is related to skewness which is a measure of the asymmetry of the distribution. See: http://en.wikipedia.org/wiki/Skewness



              $r=4$ is related to kurtois which is a measure of the 'peakedness' of a distribution. See: http://en.wikipedia.org/wiki/Kurtosis



              For $rgeq5$ they become slightly harder to compute. The $5th$ central moment is a measure for the asymmetry of the distributions tails, but I would use skewness instead.






              share|cite|improve this answer











              $endgroup$








              • 2




                $begingroup$
                $r=3$ and $r=4$ are only related to skewness and kurtosis. For these moments to merit these names, they have to be normalized by an appropriate measure of spread (to wit, the standard deviation).
                $endgroup$
                – whuber
                Sep 24 '10 at 18:32










              • $begingroup$
                Yes that is true. Thanks edited.
                $endgroup$
                – alext87
                Sep 24 '10 at 18:40






              • 1




                $begingroup$
                Thanks; +1 for that. You left off the punch line, though, and it might help to make it explicit by stating how one usually thinks of skewness (asymmetry) and kurtosis (peakedness or fat-tailedness). That would make it obvious that your answer really does address the request for "ways to think of" the moments.
                $endgroup$
                – whuber
                Sep 24 '10 at 18:46










              • $begingroup$
                That was the wiki page's job. Though in hindsight it makes it a better answer. Edited +1. :) Whuber, out of curiousity, have you used the higher (central) moments for any practical work?
                $endgroup$
                – alext87
                Sep 24 '10 at 19:04










              • $begingroup$
                It depends on what you mean. If you mean for data analysis, the answer is a resounding no: even for quite large datasets, the higher moments are extremely sensitive to outliers and are not robust enough for drawing useful conclusions. Even the use of the third and fourth moments is awfully old-fashioned for that reason, but these moments can occasionally be useful measurements of the skewness and tailedness of an empirical distribution. Modern statistics tends to use extreme quantiles or order statistics to characterize the tails.
                $endgroup$
                – whuber
                Sep 30 '10 at 4:38













              10












              10








              10





              $begingroup$

              Variance of a r.v. $X$ is $Var(X) = mathbbE[(X-mathbbE(X))^2]$. Variance is used as a measure for the dispersion of a random variable. If you are interested in how far a random variable is from what you expect; perhaps you would consider $mathbbE(|X-mathbbE(X)|)$ but see: https://stats.stackexchange.com/questions/118/standard-deviation-why-square-the-difference-instead-of-taking-the-absolute-val for why this we use the square instead.



              For all except the first moment (i.e. the mean) the central moments (centred about the mean) are far more interesting.



              As for the $r$th (central) moment of $X$, that is $mathbbE[(X-mathbbE(X))^r]$. The lower moments ($r<5$) are related to properties of the distribution:



              $r=0$ is $1$



              $r=1$ is $0$



              $r=2$ is the variance which is a measure of the spread from the mean.



              $r=3$ is related to skewness which is a measure of the asymmetry of the distribution. See: http://en.wikipedia.org/wiki/Skewness



              $r=4$ is related to kurtois which is a measure of the 'peakedness' of a distribution. See: http://en.wikipedia.org/wiki/Kurtosis



              For $rgeq5$ they become slightly harder to compute. The $5th$ central moment is a measure for the asymmetry of the distributions tails, but I would use skewness instead.






              share|cite|improve this answer











              $endgroup$



              Variance of a r.v. $X$ is $Var(X) = mathbbE[(X-mathbbE(X))^2]$. Variance is used as a measure for the dispersion of a random variable. If you are interested in how far a random variable is from what you expect; perhaps you would consider $mathbbE(|X-mathbbE(X)|)$ but see: https://stats.stackexchange.com/questions/118/standard-deviation-why-square-the-difference-instead-of-taking-the-absolute-val for why this we use the square instead.



              For all except the first moment (i.e. the mean) the central moments (centred about the mean) are far more interesting.



              As for the $r$th (central) moment of $X$, that is $mathbbE[(X-mathbbE(X))^r]$. The lower moments ($r<5$) are related to properties of the distribution:



              $r=0$ is $1$



              $r=1$ is $0$



              $r=2$ is the variance which is a measure of the spread from the mean.



              $r=3$ is related to skewness which is a measure of the asymmetry of the distribution. See: http://en.wikipedia.org/wiki/Skewness



              $r=4$ is related to kurtois which is a measure of the 'peakedness' of a distribution. See: http://en.wikipedia.org/wiki/Kurtosis



              For $rgeq5$ they become slightly harder to compute. The $5th$ central moment is a measure for the asymmetry of the distributions tails, but I would use skewness instead.







              share|cite|improve this answer














              share|cite|improve this answer



              share|cite|improve this answer








              edited Apr 13 '17 at 12:44









              Community

              1




              1










              answered Sep 24 '10 at 17:47









              alext87alext87

              2,173818




              2,173818







              • 2




                $begingroup$
                $r=3$ and $r=4$ are only related to skewness and kurtosis. For these moments to merit these names, they have to be normalized by an appropriate measure of spread (to wit, the standard deviation).
                $endgroup$
                – whuber
                Sep 24 '10 at 18:32










              • $begingroup$
                Yes that is true. Thanks edited.
                $endgroup$
                – alext87
                Sep 24 '10 at 18:40






              • 1




                $begingroup$
                Thanks; +1 for that. You left off the punch line, though, and it might help to make it explicit by stating how one usually thinks of skewness (asymmetry) and kurtosis (peakedness or fat-tailedness). That would make it obvious that your answer really does address the request for "ways to think of" the moments.
                $endgroup$
                – whuber
                Sep 24 '10 at 18:46










              • $begingroup$
                That was the wiki page's job. Though in hindsight it makes it a better answer. Edited +1. :) Whuber, out of curiousity, have you used the higher (central) moments for any practical work?
                $endgroup$
                – alext87
                Sep 24 '10 at 19:04










              • $begingroup$
                It depends on what you mean. If you mean for data analysis, the answer is a resounding no: even for quite large datasets, the higher moments are extremely sensitive to outliers and are not robust enough for drawing useful conclusions. Even the use of the third and fourth moments is awfully old-fashioned for that reason, but these moments can occasionally be useful measurements of the skewness and tailedness of an empirical distribution. Modern statistics tends to use extreme quantiles or order statistics to characterize the tails.
                $endgroup$
                – whuber
                Sep 30 '10 at 4:38












              • 2




                $begingroup$
                $r=3$ and $r=4$ are only related to skewness and kurtosis. For these moments to merit these names, they have to be normalized by an appropriate measure of spread (to wit, the standard deviation).
                $endgroup$
                – whuber
                Sep 24 '10 at 18:32










              • $begingroup$
                Yes that is true. Thanks edited.
                $endgroup$
                – alext87
                Sep 24 '10 at 18:40






              • 1




                $begingroup$
                Thanks; +1 for that. You left off the punch line, though, and it might help to make it explicit by stating how one usually thinks of skewness (asymmetry) and kurtosis (peakedness or fat-tailedness). That would make it obvious that your answer really does address the request for "ways to think of" the moments.
                $endgroup$
                – whuber
                Sep 24 '10 at 18:46










              • $begingroup$
                That was the wiki page's job. Though in hindsight it makes it a better answer. Edited +1. :) Whuber, out of curiousity, have you used the higher (central) moments for any practical work?
                $endgroup$
                – alext87
                Sep 24 '10 at 19:04










              • $begingroup$
                It depends on what you mean. If you mean for data analysis, the answer is a resounding no: even for quite large datasets, the higher moments are extremely sensitive to outliers and are not robust enough for drawing useful conclusions. Even the use of the third and fourth moments is awfully old-fashioned for that reason, but these moments can occasionally be useful measurements of the skewness and tailedness of an empirical distribution. Modern statistics tends to use extreme quantiles or order statistics to characterize the tails.
                $endgroup$
                – whuber
                Sep 30 '10 at 4:38







              2




              2




              $begingroup$
              $r=3$ and $r=4$ are only related to skewness and kurtosis. For these moments to merit these names, they have to be normalized by an appropriate measure of spread (to wit, the standard deviation).
              $endgroup$
              – whuber
              Sep 24 '10 at 18:32




              $begingroup$
              $r=3$ and $r=4$ are only related to skewness and kurtosis. For these moments to merit these names, they have to be normalized by an appropriate measure of spread (to wit, the standard deviation).
              $endgroup$
              – whuber
              Sep 24 '10 at 18:32












              $begingroup$
              Yes that is true. Thanks edited.
              $endgroup$
              – alext87
              Sep 24 '10 at 18:40




              $begingroup$
              Yes that is true. Thanks edited.
              $endgroup$
              – alext87
              Sep 24 '10 at 18:40




              1




              1




              $begingroup$
              Thanks; +1 for that. You left off the punch line, though, and it might help to make it explicit by stating how one usually thinks of skewness (asymmetry) and kurtosis (peakedness or fat-tailedness). That would make it obvious that your answer really does address the request for "ways to think of" the moments.
              $endgroup$
              – whuber
              Sep 24 '10 at 18:46




              $begingroup$
              Thanks; +1 for that. You left off the punch line, though, and it might help to make it explicit by stating how one usually thinks of skewness (asymmetry) and kurtosis (peakedness or fat-tailedness). That would make it obvious that your answer really does address the request for "ways to think of" the moments.
              $endgroup$
              – whuber
              Sep 24 '10 at 18:46












              $begingroup$
              That was the wiki page's job. Though in hindsight it makes it a better answer. Edited +1. :) Whuber, out of curiousity, have you used the higher (central) moments for any practical work?
              $endgroup$
              – alext87
              Sep 24 '10 at 19:04




              $begingroup$
              That was the wiki page's job. Though in hindsight it makes it a better answer. Edited +1. :) Whuber, out of curiousity, have you used the higher (central) moments for any practical work?
              $endgroup$
              – alext87
              Sep 24 '10 at 19:04












              $begingroup$
              It depends on what you mean. If you mean for data analysis, the answer is a resounding no: even for quite large datasets, the higher moments are extremely sensitive to outliers and are not robust enough for drawing useful conclusions. Even the use of the third and fourth moments is awfully old-fashioned for that reason, but these moments can occasionally be useful measurements of the skewness and tailedness of an empirical distribution. Modern statistics tends to use extreme quantiles or order statistics to characterize the tails.
              $endgroup$
              – whuber
              Sep 30 '10 at 4:38




              $begingroup$
              It depends on what you mean. If you mean for data analysis, the answer is a resounding no: even for quite large datasets, the higher moments are extremely sensitive to outliers and are not robust enough for drawing useful conclusions. Even the use of the third and fourth moments is awfully old-fashioned for that reason, but these moments can occasionally be useful measurements of the skewness and tailedness of an empirical distribution. Modern statistics tends to use extreme quantiles or order statistics to characterize the tails.
              $endgroup$
              – whuber
              Sep 30 '10 at 4:38











              4












              $begingroup$

              The localization and the dispersion probabilistic measures can be "seen" as the corresponding momentums of mechanical systems of "probabilistic masses".



              The expectation has the following mechanical interpretation. Given that $F(x)$ is the "probabilistic mass" contained in the interval $0le Xlt x$ (in one dimention), the mathematical expectation of the random variable $X$ is the static momentum with respect to the origin of the "probabilistic masses" system.



              The variance is the mechanical analog of the inertia momentum of the "probabilistic masses" system with respect to the center of masses.



              The variance of the random variable $X$ is the 2nd order momentum of $X-m$ where $m$ is the expectation of $X$, i.e. $m=E(X)=displaystyleint_0^x x, dF(x)$ ($F(x)$ is the cumulative distribution function).



              The $k$-order momentum is the expectation of $(X-m)^k$.






              share|cite|improve this answer











              $endgroup$

















                4












                $begingroup$

                The localization and the dispersion probabilistic measures can be "seen" as the corresponding momentums of mechanical systems of "probabilistic masses".



                The expectation has the following mechanical interpretation. Given that $F(x)$ is the "probabilistic mass" contained in the interval $0le Xlt x$ (in one dimention), the mathematical expectation of the random variable $X$ is the static momentum with respect to the origin of the "probabilistic masses" system.



                The variance is the mechanical analog of the inertia momentum of the "probabilistic masses" system with respect to the center of masses.



                The variance of the random variable $X$ is the 2nd order momentum of $X-m$ where $m$ is the expectation of $X$, i.e. $m=E(X)=displaystyleint_0^x x, dF(x)$ ($F(x)$ is the cumulative distribution function).



                The $k$-order momentum is the expectation of $(X-m)^k$.






                share|cite|improve this answer











                $endgroup$















                  4












                  4








                  4





                  $begingroup$

                  The localization and the dispersion probabilistic measures can be "seen" as the corresponding momentums of mechanical systems of "probabilistic masses".



                  The expectation has the following mechanical interpretation. Given that $F(x)$ is the "probabilistic mass" contained in the interval $0le Xlt x$ (in one dimention), the mathematical expectation of the random variable $X$ is the static momentum with respect to the origin of the "probabilistic masses" system.



                  The variance is the mechanical analog of the inertia momentum of the "probabilistic masses" system with respect to the center of masses.



                  The variance of the random variable $X$ is the 2nd order momentum of $X-m$ where $m$ is the expectation of $X$, i.e. $m=E(X)=displaystyleint_0^x x, dF(x)$ ($F(x)$ is the cumulative distribution function).



                  The $k$-order momentum is the expectation of $(X-m)^k$.






                  share|cite|improve this answer











                  $endgroup$



                  The localization and the dispersion probabilistic measures can be "seen" as the corresponding momentums of mechanical systems of "probabilistic masses".



                  The expectation has the following mechanical interpretation. Given that $F(x)$ is the "probabilistic mass" contained in the interval $0le Xlt x$ (in one dimention), the mathematical expectation of the random variable $X$ is the static momentum with respect to the origin of the "probabilistic masses" system.



                  The variance is the mechanical analog of the inertia momentum of the "probabilistic masses" system with respect to the center of masses.



                  The variance of the random variable $X$ is the 2nd order momentum of $X-m$ where $m$ is the expectation of $X$, i.e. $m=E(X)=displaystyleint_0^x x, dF(x)$ ($F(x)$ is the cumulative distribution function).



                  The $k$-order momentum is the expectation of $(X-m)^k$.







                  share|cite|improve this answer














                  share|cite|improve this answer



                  share|cite|improve this answer








                  edited Mar 28 at 14:15

























                  answered Sep 25 '10 at 23:22









                  Américo TavaresAmérico Tavares

                  32.6k1181206




                  32.6k1181206





















                      2












                      $begingroup$

                      The variance is the expected squared deviation between the random variable and its expectation.



                      It measures how much the random variable scatters or spreads around its expectation.






                      share|cite|improve this answer









                      $endgroup$












                      • $begingroup$
                        This is the definition, but I was hoping for some intuitive insight.
                        $endgroup$
                        – Pandora
                        Sep 24 '10 at 17:20










                      • $begingroup$
                        I was hoping that my second sentence would provide some intuition. It contains the common way of thinking of the variance.
                        $endgroup$
                        – Rasmus
                        Sep 24 '10 at 17:22











                      • $begingroup$
                        +1 for making the effort in the second sentence. Note, though, that the variance only indirectly measures spread (because it's not even in the same units of measurement): its square root, the standard deviation, more directly reflects spread.
                        $endgroup$
                        – whuber
                        Sep 24 '10 at 18:28















                      2












                      $begingroup$

                      The variance is the expected squared deviation between the random variable and its expectation.



                      It measures how much the random variable scatters or spreads around its expectation.






                      share|cite|improve this answer









                      $endgroup$












                      • $begingroup$
                        This is the definition, but I was hoping for some intuitive insight.
                        $endgroup$
                        – Pandora
                        Sep 24 '10 at 17:20










                      • $begingroup$
                        I was hoping that my second sentence would provide some intuition. It contains the common way of thinking of the variance.
                        $endgroup$
                        – Rasmus
                        Sep 24 '10 at 17:22











                      • $begingroup$
                        +1 for making the effort in the second sentence. Note, though, that the variance only indirectly measures spread (because it's not even in the same units of measurement): its square root, the standard deviation, more directly reflects spread.
                        $endgroup$
                        – whuber
                        Sep 24 '10 at 18:28













                      2












                      2








                      2





                      $begingroup$

                      The variance is the expected squared deviation between the random variable and its expectation.



                      It measures how much the random variable scatters or spreads around its expectation.






                      share|cite|improve this answer









                      $endgroup$



                      The variance is the expected squared deviation between the random variable and its expectation.



                      It measures how much the random variable scatters or spreads around its expectation.







                      share|cite|improve this answer












                      share|cite|improve this answer



                      share|cite|improve this answer










                      answered Sep 24 '10 at 17:17









                      RasmusRasmus

                      14.3k14479




                      14.3k14479











                      • $begingroup$
                        This is the definition, but I was hoping for some intuitive insight.
                        $endgroup$
                        – Pandora
                        Sep 24 '10 at 17:20










                      • $begingroup$
                        I was hoping that my second sentence would provide some intuition. It contains the common way of thinking of the variance.
                        $endgroup$
                        – Rasmus
                        Sep 24 '10 at 17:22











                      • $begingroup$
                        +1 for making the effort in the second sentence. Note, though, that the variance only indirectly measures spread (because it's not even in the same units of measurement): its square root, the standard deviation, more directly reflects spread.
                        $endgroup$
                        – whuber
                        Sep 24 '10 at 18:28
















                      • $begingroup$
                        This is the definition, but I was hoping for some intuitive insight.
                        $endgroup$
                        – Pandora
                        Sep 24 '10 at 17:20










                      • $begingroup$
                        I was hoping that my second sentence would provide some intuition. It contains the common way of thinking of the variance.
                        $endgroup$
                        – Rasmus
                        Sep 24 '10 at 17:22











                      • $begingroup$
                        +1 for making the effort in the second sentence. Note, though, that the variance only indirectly measures spread (because it's not even in the same units of measurement): its square root, the standard deviation, more directly reflects spread.
                        $endgroup$
                        – whuber
                        Sep 24 '10 at 18:28















                      $begingroup$
                      This is the definition, but I was hoping for some intuitive insight.
                      $endgroup$
                      – Pandora
                      Sep 24 '10 at 17:20




                      $begingroup$
                      This is the definition, but I was hoping for some intuitive insight.
                      $endgroup$
                      – Pandora
                      Sep 24 '10 at 17:20












                      $begingroup$
                      I was hoping that my second sentence would provide some intuition. It contains the common way of thinking of the variance.
                      $endgroup$
                      – Rasmus
                      Sep 24 '10 at 17:22





                      $begingroup$
                      I was hoping that my second sentence would provide some intuition. It contains the common way of thinking of the variance.
                      $endgroup$
                      – Rasmus
                      Sep 24 '10 at 17:22













                      $begingroup$
                      +1 for making the effort in the second sentence. Note, though, that the variance only indirectly measures spread (because it's not even in the same units of measurement): its square root, the standard deviation, more directly reflects spread.
                      $endgroup$
                      – whuber
                      Sep 24 '10 at 18:28




                      $begingroup$
                      +1 for making the effort in the second sentence. Note, though, that the variance only indirectly measures spread (because it's not even in the same units of measurement): its square root, the standard deviation, more directly reflects spread.
                      $endgroup$
                      – whuber
                      Sep 24 '10 at 18:28

















                      draft saved

                      draft discarded
















































                      Thanks for contributing an answer to Mathematics Stack Exchange!


                      • Please be sure to answer the question. Provide details and share your research!

                      But avoid


                      • Asking for help, clarification, or responding to other answers.

                      • Making statements based on opinion; back them up with references or personal experience.

                      Use MathJax to format equations. MathJax reference.


                      To learn more, see our tips on writing great answers.




                      draft saved


                      draft discarded














                      StackExchange.ready(
                      function ()
                      StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f5392%2fintuitive-explanation-of-variance-and-moment-in-probability%23new-answer', 'question_page');

                      );

                      Post as a guest















                      Required, but never shown





















































                      Required, but never shown














                      Required, but never shown












                      Required, but never shown







                      Required, but never shown

































                      Required, but never shown














                      Required, but never shown












                      Required, but never shown







                      Required, but never shown







                      Popular posts from this blog

                      Triangular numbers and gcdProving sum of a set is $0 pmod n$ if $n$ is odd, or $fracn2 pmod n$ if $n$ is even?Is greatest common divisor of two numbers really their smallest linear combination?GCD, LCM RelationshipProve a set of nonnegative integers with greatest common divisor 1 and closed under addition has all but finite many nonnegative integers.all pairs of a and b in an equation containing gcdTriangular Numbers Modulo $k$ - Hit All Values?Understanding the Existence and Uniqueness of the GCDGCD and LCM with logical symbolsThe greatest common divisor of two positive integers less than 100 is equal to 3. Their least common multiple is twelve times one of the integers.Suppose that for all integers $x$, $x|a$ and $x|b$ if and only if $x|c$. Then $c = gcd(a,b)$Which is the gcd of 2 numbers which are multiplied and the result is 600000?

                      Ingelân Ynhâld Etymology | Geografy | Skiednis | Polityk en bestjoer | Ekonomy | Demografy | Kultuer | Klimaat | Sjoch ek | Keppelings om utens | Boarnen, noaten en referinsjes Navigaasjemenuwww.gov.ukOffisjele webside fan it regear fan it Feriene KeninkrykOffisjele webside fan it Britske FerkearsburoNederlânsktalige ynformaasje fan it Britske FerkearsburoOffisjele webside fan English Heritage, de organisaasje dy't him ynset foar it behâld fan it Ingelske kultuergoedYnwennertallen fan alle Britske stêden út 'e folkstelling fan 2011Notes en References, op dizze sideEngland

                      Հադիս Բովանդակություն Անվանում և նշանակություն | Դասակարգում | Աղբյուրներ | Նավարկման ցանկ