Examples of a statistic that is not independent of sample's distribution?












13












$begingroup$


This is the definition for statistic on wikipedia




More formally, statistical theory defines a statistic as a function of a sample where the function itself is independent of the sample's distribution; that is, the function can be stated before realization of the data. The term statistic is used both for the function and for the value of the function on a given sample.




I think I understand most of this definition, however the part - where the function is independent of the sample's distribution I haven't been able to sort out.



My understanding of statistic so far



A sample is a set of realizations of some number of independent, identically distributed (iid) random variables with distribution F (10 realizations of a roll of a 20-sided fair dice, 100 realizations of 5 rolls of a 6-sided fair dice, randomly draw 100 people from a population).



A function, whose domain is that set, and whose range is the real numbers (or maybe it can produce other things, like a vector or other mathematical object...) would be considered a statistic.



When I think of examples, mean, median, variance all make sense in this context. They are a function on set of realizations (blood pressure measurements from a random sample). I can also see how a linear regression model could be considered a statistic $y_{i} = alpha + beta cdot x_{i}$ - is this not just a function on a set of realizations?



Where I'm confused



Assuming that my understanding from above is correct, I haven't been able to understand where a function might not be independent of the sample's distribution. I've been trying to think of an example to make sense of it, but no luck. Any insight would be much appreciated!










share|cite|improve this question









$endgroup$

















    13












    $begingroup$


    This is the definition for statistic on wikipedia




    More formally, statistical theory defines a statistic as a function of a sample where the function itself is independent of the sample's distribution; that is, the function can be stated before realization of the data. The term statistic is used both for the function and for the value of the function on a given sample.




    I think I understand most of this definition, however the part - where the function is independent of the sample's distribution I haven't been able to sort out.



    My understanding of statistic so far



    A sample is a set of realizations of some number of independent, identically distributed (iid) random variables with distribution F (10 realizations of a roll of a 20-sided fair dice, 100 realizations of 5 rolls of a 6-sided fair dice, randomly draw 100 people from a population).



    A function, whose domain is that set, and whose range is the real numbers (or maybe it can produce other things, like a vector or other mathematical object...) would be considered a statistic.



    When I think of examples, mean, median, variance all make sense in this context. They are a function on set of realizations (blood pressure measurements from a random sample). I can also see how a linear regression model could be considered a statistic $y_{i} = alpha + beta cdot x_{i}$ - is this not just a function on a set of realizations?



    Where I'm confused



    Assuming that my understanding from above is correct, I haven't been able to understand where a function might not be independent of the sample's distribution. I've been trying to think of an example to make sense of it, but no luck. Any insight would be much appreciated!










    share|cite|improve this question









    $endgroup$















      13












      13








      13


      2



      $begingroup$


      This is the definition for statistic on wikipedia




      More formally, statistical theory defines a statistic as a function of a sample where the function itself is independent of the sample's distribution; that is, the function can be stated before realization of the data. The term statistic is used both for the function and for the value of the function on a given sample.




      I think I understand most of this definition, however the part - where the function is independent of the sample's distribution I haven't been able to sort out.



      My understanding of statistic so far



      A sample is a set of realizations of some number of independent, identically distributed (iid) random variables with distribution F (10 realizations of a roll of a 20-sided fair dice, 100 realizations of 5 rolls of a 6-sided fair dice, randomly draw 100 people from a population).



      A function, whose domain is that set, and whose range is the real numbers (or maybe it can produce other things, like a vector or other mathematical object...) would be considered a statistic.



      When I think of examples, mean, median, variance all make sense in this context. They are a function on set of realizations (blood pressure measurements from a random sample). I can also see how a linear regression model could be considered a statistic $y_{i} = alpha + beta cdot x_{i}$ - is this not just a function on a set of realizations?



      Where I'm confused



      Assuming that my understanding from above is correct, I haven't been able to understand where a function might not be independent of the sample's distribution. I've been trying to think of an example to make sense of it, but no luck. Any insight would be much appreciated!










      share|cite|improve this question









      $endgroup$




      This is the definition for statistic on wikipedia




      More formally, statistical theory defines a statistic as a function of a sample where the function itself is independent of the sample's distribution; that is, the function can be stated before realization of the data. The term statistic is used both for the function and for the value of the function on a given sample.




      I think I understand most of this definition, however the part - where the function is independent of the sample's distribution I haven't been able to sort out.



      My understanding of statistic so far



      A sample is a set of realizations of some number of independent, identically distributed (iid) random variables with distribution F (10 realizations of a roll of a 20-sided fair dice, 100 realizations of 5 rolls of a 6-sided fair dice, randomly draw 100 people from a population).



      A function, whose domain is that set, and whose range is the real numbers (or maybe it can produce other things, like a vector or other mathematical object...) would be considered a statistic.



      When I think of examples, mean, median, variance all make sense in this context. They are a function on set of realizations (blood pressure measurements from a random sample). I can also see how a linear regression model could be considered a statistic $y_{i} = alpha + beta cdot x_{i}$ - is this not just a function on a set of realizations?



      Where I'm confused



      Assuming that my understanding from above is correct, I haven't been able to understand where a function might not be independent of the sample's distribution. I've been trying to think of an example to make sense of it, but no luck. Any insight would be much appreciated!







      mathematical-statistics definition






      share|cite|improve this question













      share|cite|improve this question











      share|cite|improve this question




      share|cite|improve this question










      asked Mar 11 at 9:55









      Jake KirschJake Kirsch

      805




      805






















          2 Answers
          2






          active

          oldest

          votes


















          43












          $begingroup$

          That definition is a somewhat awkward way to state it. A "statistic" is any function of the observable values. All that definition means is that a statistic is a function only of the observable values, not a function of the distribution or any of its parameters. For example, if $X_1, X_2, ..., X_n sim text{N}(mu, 1)$ then a statistic would be any function $T(X_1,...,X_n)$ whereas a function $H(X_1,....,X_n, mu)$ would not be a statistic, since it depends on $mu$. Here are some further examples:



          $$begin{equation} begin{aligned}
          text{Statistic} & & & & & bar{X}_n = frac{1}{n} sum_{i=1}^n X_i, \[12pt]
          text{Statistic} & & & & & S_n^2 = frac{1}{n} sum_{i=1}^n (X_i - bar{X}_n)^2, \[12pt]
          text{Not a statistic} & & & & & D_n = bar{X}_n - mu, \[12pt]
          text{Not a statistic} & & & & & p_i = text{N}(x_i | mu, 1), \[12pt]
          text{Not a statistic} & & & & & Q = 10 mu. \[12pt]
          end{aligned} end{equation}$$



          Every statistic is a function only of the observable values, and not of their distribution or its parameters. So there are no examples of a statistic that is a function of the distribution or its parameters (any such function would not be a statistic). However, it is important to note that the distribution of a statistic (as opposed to the statistic itself) will generally depend on the underlying distribution of the values. (This is true for all statistics other than ancillary statistics.)





          What about a function where the parameters are known? In the comments below, Alecos asks an excellent follow-up question. What about a function that uses a fixed hypothesised value of the parameter? For example, what about the statistic $sqrt{n} (bar{x} - mu)$ where $mu = mu_0$ is taken to be equal to a known hypothesised value $mu_0 in mathbb{R}$. Here the function is indeed a statistic, so long as it is defined on the appropriately restricted domain. So the function $H_0: mathbb{R}^n rightarrow mathbb{R}$ with $H_0(x_1,...,x_n) = sqrt{n} (bar{x} - mu_0)$ would be a statistic, but the function $H: mathbb{R}^{n+1} rightarrow mathbb{R}$ with $H(x_1,...,x_n, mu) = sqrt{n} (bar{x} - mu)$ would not be a statistic.






          share|cite|improve this answer











          $endgroup$









          • 1




            $begingroup$
            Very helpful answer, considering the underlying statistical parameter as part of the non-statistic was particularly helpful.
            $endgroup$
            – Jake Kirsch
            Mar 11 at 10:56






          • 4




            $begingroup$
            @CarlWitthoft I don't get your point. If it's a function of the observable values, then it's a statistic. It may be a function of a smaller subset of the values; that can still be a useful thing to consider. If you want to estimate the mean and you have $10^{10}$ observations, you might still look at $(X_1+X_2+dots+X_{1000})/1000$ if the cost of processing data is high and the cost of error is small. Or for some reason you might want to consider two independent estimates of the mean, and could consider $(X_1+dots+X_{n/2})/(n/2)$ and $(X_{n/2+1}+dots+X_n)/(n/2)$. These are still statistics.
            $endgroup$
            – James Martin
            Mar 11 at 14:06








          • 4




            $begingroup$
            Those examples seem entirely valid to me. Are you saying the idea of dividing data into a training set and a validation set is not valid?
            $endgroup$
            – James Martin
            Mar 11 at 14:53








          • 2




            $begingroup$
            I'm a little confused by that as well. Let me attempt to describe @CarlWitthoft point. It would still be a statistic in terms of mathematical definition, but I could see a case where a consultant takes a 'statistic' of observations, but arbitrarily decides to remove a few results (consultants do this all the time right?). This would be 'valid' in the sense it's still a function on observations, however the way that statistic may be presented and interpreted likely wouldn't be valid.
            $endgroup$
            – Jake Kirsch
            Mar 11 at 15:41








          • 2




            $begingroup$
            @Carl Withhoft: With respect to the point you are making, it is important to distinguish between a statistic (which need not include all the data, and may not encompass all the information in the sample) and a sufficient statistic (which will encompass all the information with respect to some parameter). Statistical theory already has well-developed concepts like sufficiency that capture the idea that a statistic includes all relevant information in the sample. It is not necessary, or desirable, to try to build that requirement into the definition of a "statistic".
            $endgroup$
            – Ben
            Mar 11 at 21:11



















          4












          $begingroup$

          I interpret that as saying that you should decide before you see the data what statistic you are going to calculate. So, for instance, if you're going to take out outliers, you should decide before you see the data what constitutes an "outlier". If you decide after you see the data, then your function is dependent on the data.






          share|cite|improve this answer









          $endgroup$













          • $begingroup$
            this is also helpful! So making a decision on which observations to include in the function after knowing what observations are available, which is more or less what I was describing in my comment on the previous answer.
            $endgroup$
            – Jake Kirsch
            Mar 11 at 19:37






          • 2




            $begingroup$
            (+1) It might be worth noting that this important because if you define a rule a prior about what constitutes a data point that will be dropped, it is (relatively) easy to derive a distribution for statistic (i.e., truncated mean, etc.). It's really hard to derive a distribution for a measure that involves dropping data points for reasons that are not cleanly defined before hand.
            $endgroup$
            – Cliff AB
            Mar 11 at 23:41











          Your Answer





          StackExchange.ifUsing("editor", function () {
          return StackExchange.using("mathjaxEditing", function () {
          StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
          StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
          });
          });
          }, "mathjax-editing");

          StackExchange.ready(function() {
          var channelOptions = {
          tags: "".split(" "),
          id: "65"
          };
          initTagRenderer("".split(" "), "".split(" "), channelOptions);

          StackExchange.using("externalEditor", function() {
          // Have to fire editor after snippets, if snippets enabled
          if (StackExchange.settings.snippets.snippetsEnabled) {
          StackExchange.using("snippets", function() {
          createEditor();
          });
          }
          else {
          createEditor();
          }
          });

          function createEditor() {
          StackExchange.prepareEditor({
          heartbeatType: 'answer',
          autoActivateHeartbeat: false,
          convertImagesToLinks: false,
          noModals: true,
          showLowRepImageUploadWarning: true,
          reputationToPostImages: null,
          bindNavPrevention: true,
          postfix: "",
          imageUploader: {
          brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
          contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
          allowUrls: true
          },
          onDemand: true,
          discardSelector: ".discard-answer"
          ,immediatelyShowMarkdownHelp:true
          });


          }
          });














          draft saved

          draft discarded


















          StackExchange.ready(
          function () {
          StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstats.stackexchange.com%2fquestions%2f396815%2fexamples-of-a-statistic-that-is-not-independent-of-samples-distribution%23new-answer', 'question_page');
          }
          );

          Post as a guest















          Required, but never shown

























          2 Answers
          2






          active

          oldest

          votes








          2 Answers
          2






          active

          oldest

          votes









          active

          oldest

          votes






          active

          oldest

          votes









          43












          $begingroup$

          That definition is a somewhat awkward way to state it. A "statistic" is any function of the observable values. All that definition means is that a statistic is a function only of the observable values, not a function of the distribution or any of its parameters. For example, if $X_1, X_2, ..., X_n sim text{N}(mu, 1)$ then a statistic would be any function $T(X_1,...,X_n)$ whereas a function $H(X_1,....,X_n, mu)$ would not be a statistic, since it depends on $mu$. Here are some further examples:



          $$begin{equation} begin{aligned}
          text{Statistic} & & & & & bar{X}_n = frac{1}{n} sum_{i=1}^n X_i, \[12pt]
          text{Statistic} & & & & & S_n^2 = frac{1}{n} sum_{i=1}^n (X_i - bar{X}_n)^2, \[12pt]
          text{Not a statistic} & & & & & D_n = bar{X}_n - mu, \[12pt]
          text{Not a statistic} & & & & & p_i = text{N}(x_i | mu, 1), \[12pt]
          text{Not a statistic} & & & & & Q = 10 mu. \[12pt]
          end{aligned} end{equation}$$



          Every statistic is a function only of the observable values, and not of their distribution or its parameters. So there are no examples of a statistic that is a function of the distribution or its parameters (any such function would not be a statistic). However, it is important to note that the distribution of a statistic (as opposed to the statistic itself) will generally depend on the underlying distribution of the values. (This is true for all statistics other than ancillary statistics.)





          What about a function where the parameters are known? In the comments below, Alecos asks an excellent follow-up question. What about a function that uses a fixed hypothesised value of the parameter? For example, what about the statistic $sqrt{n} (bar{x} - mu)$ where $mu = mu_0$ is taken to be equal to a known hypothesised value $mu_0 in mathbb{R}$. Here the function is indeed a statistic, so long as it is defined on the appropriately restricted domain. So the function $H_0: mathbb{R}^n rightarrow mathbb{R}$ with $H_0(x_1,...,x_n) = sqrt{n} (bar{x} - mu_0)$ would be a statistic, but the function $H: mathbb{R}^{n+1} rightarrow mathbb{R}$ with $H(x_1,...,x_n, mu) = sqrt{n} (bar{x} - mu)$ would not be a statistic.






          share|cite|improve this answer











          $endgroup$









          • 1




            $begingroup$
            Very helpful answer, considering the underlying statistical parameter as part of the non-statistic was particularly helpful.
            $endgroup$
            – Jake Kirsch
            Mar 11 at 10:56






          • 4




            $begingroup$
            @CarlWitthoft I don't get your point. If it's a function of the observable values, then it's a statistic. It may be a function of a smaller subset of the values; that can still be a useful thing to consider. If you want to estimate the mean and you have $10^{10}$ observations, you might still look at $(X_1+X_2+dots+X_{1000})/1000$ if the cost of processing data is high and the cost of error is small. Or for some reason you might want to consider two independent estimates of the mean, and could consider $(X_1+dots+X_{n/2})/(n/2)$ and $(X_{n/2+1}+dots+X_n)/(n/2)$. These are still statistics.
            $endgroup$
            – James Martin
            Mar 11 at 14:06








          • 4




            $begingroup$
            Those examples seem entirely valid to me. Are you saying the idea of dividing data into a training set and a validation set is not valid?
            $endgroup$
            – James Martin
            Mar 11 at 14:53








          • 2




            $begingroup$
            I'm a little confused by that as well. Let me attempt to describe @CarlWitthoft point. It would still be a statistic in terms of mathematical definition, but I could see a case where a consultant takes a 'statistic' of observations, but arbitrarily decides to remove a few results (consultants do this all the time right?). This would be 'valid' in the sense it's still a function on observations, however the way that statistic may be presented and interpreted likely wouldn't be valid.
            $endgroup$
            – Jake Kirsch
            Mar 11 at 15:41








          • 2




            $begingroup$
            @Carl Withhoft: With respect to the point you are making, it is important to distinguish between a statistic (which need not include all the data, and may not encompass all the information in the sample) and a sufficient statistic (which will encompass all the information with respect to some parameter). Statistical theory already has well-developed concepts like sufficiency that capture the idea that a statistic includes all relevant information in the sample. It is not necessary, or desirable, to try to build that requirement into the definition of a "statistic".
            $endgroup$
            – Ben
            Mar 11 at 21:11
















          43












          $begingroup$

          That definition is a somewhat awkward way to state it. A "statistic" is any function of the observable values. All that definition means is that a statistic is a function only of the observable values, not a function of the distribution or any of its parameters. For example, if $X_1, X_2, ..., X_n sim text{N}(mu, 1)$ then a statistic would be any function $T(X_1,...,X_n)$ whereas a function $H(X_1,....,X_n, mu)$ would not be a statistic, since it depends on $mu$. Here are some further examples:



          $$begin{equation} begin{aligned}
          text{Statistic} & & & & & bar{X}_n = frac{1}{n} sum_{i=1}^n X_i, \[12pt]
          text{Statistic} & & & & & S_n^2 = frac{1}{n} sum_{i=1}^n (X_i - bar{X}_n)^2, \[12pt]
          text{Not a statistic} & & & & & D_n = bar{X}_n - mu, \[12pt]
          text{Not a statistic} & & & & & p_i = text{N}(x_i | mu, 1), \[12pt]
          text{Not a statistic} & & & & & Q = 10 mu. \[12pt]
          end{aligned} end{equation}$$



          Every statistic is a function only of the observable values, and not of their distribution or its parameters. So there are no examples of a statistic that is a function of the distribution or its parameters (any such function would not be a statistic). However, it is important to note that the distribution of a statistic (as opposed to the statistic itself) will generally depend on the underlying distribution of the values. (This is true for all statistics other than ancillary statistics.)





          What about a function where the parameters are known? In the comments below, Alecos asks an excellent follow-up question. What about a function that uses a fixed hypothesised value of the parameter? For example, what about the statistic $sqrt{n} (bar{x} - mu)$ where $mu = mu_0$ is taken to be equal to a known hypothesised value $mu_0 in mathbb{R}$. Here the function is indeed a statistic, so long as it is defined on the appropriately restricted domain. So the function $H_0: mathbb{R}^n rightarrow mathbb{R}$ with $H_0(x_1,...,x_n) = sqrt{n} (bar{x} - mu_0)$ would be a statistic, but the function $H: mathbb{R}^{n+1} rightarrow mathbb{R}$ with $H(x_1,...,x_n, mu) = sqrt{n} (bar{x} - mu)$ would not be a statistic.






          share|cite|improve this answer











          $endgroup$









          • 1




            $begingroup$
            Very helpful answer, considering the underlying statistical parameter as part of the non-statistic was particularly helpful.
            $endgroup$
            – Jake Kirsch
            Mar 11 at 10:56






          • 4




            $begingroup$
            @CarlWitthoft I don't get your point. If it's a function of the observable values, then it's a statistic. It may be a function of a smaller subset of the values; that can still be a useful thing to consider. If you want to estimate the mean and you have $10^{10}$ observations, you might still look at $(X_1+X_2+dots+X_{1000})/1000$ if the cost of processing data is high and the cost of error is small. Or for some reason you might want to consider two independent estimates of the mean, and could consider $(X_1+dots+X_{n/2})/(n/2)$ and $(X_{n/2+1}+dots+X_n)/(n/2)$. These are still statistics.
            $endgroup$
            – James Martin
            Mar 11 at 14:06








          • 4




            $begingroup$
            Those examples seem entirely valid to me. Are you saying the idea of dividing data into a training set and a validation set is not valid?
            $endgroup$
            – James Martin
            Mar 11 at 14:53








          • 2




            $begingroup$
            I'm a little confused by that as well. Let me attempt to describe @CarlWitthoft point. It would still be a statistic in terms of mathematical definition, but I could see a case where a consultant takes a 'statistic' of observations, but arbitrarily decides to remove a few results (consultants do this all the time right?). This would be 'valid' in the sense it's still a function on observations, however the way that statistic may be presented and interpreted likely wouldn't be valid.
            $endgroup$
            – Jake Kirsch
            Mar 11 at 15:41








          • 2




            $begingroup$
            @Carl Withhoft: With respect to the point you are making, it is important to distinguish between a statistic (which need not include all the data, and may not encompass all the information in the sample) and a sufficient statistic (which will encompass all the information with respect to some parameter). Statistical theory already has well-developed concepts like sufficiency that capture the idea that a statistic includes all relevant information in the sample. It is not necessary, or desirable, to try to build that requirement into the definition of a "statistic".
            $endgroup$
            – Ben
            Mar 11 at 21:11














          43












          43








          43





          $begingroup$

          That definition is a somewhat awkward way to state it. A "statistic" is any function of the observable values. All that definition means is that a statistic is a function only of the observable values, not a function of the distribution or any of its parameters. For example, if $X_1, X_2, ..., X_n sim text{N}(mu, 1)$ then a statistic would be any function $T(X_1,...,X_n)$ whereas a function $H(X_1,....,X_n, mu)$ would not be a statistic, since it depends on $mu$. Here are some further examples:



          $$begin{equation} begin{aligned}
          text{Statistic} & & & & & bar{X}_n = frac{1}{n} sum_{i=1}^n X_i, \[12pt]
          text{Statistic} & & & & & S_n^2 = frac{1}{n} sum_{i=1}^n (X_i - bar{X}_n)^2, \[12pt]
          text{Not a statistic} & & & & & D_n = bar{X}_n - mu, \[12pt]
          text{Not a statistic} & & & & & p_i = text{N}(x_i | mu, 1), \[12pt]
          text{Not a statistic} & & & & & Q = 10 mu. \[12pt]
          end{aligned} end{equation}$$



          Every statistic is a function only of the observable values, and not of their distribution or its parameters. So there are no examples of a statistic that is a function of the distribution or its parameters (any such function would not be a statistic). However, it is important to note that the distribution of a statistic (as opposed to the statistic itself) will generally depend on the underlying distribution of the values. (This is true for all statistics other than ancillary statistics.)





          What about a function where the parameters are known? In the comments below, Alecos asks an excellent follow-up question. What about a function that uses a fixed hypothesised value of the parameter? For example, what about the statistic $sqrt{n} (bar{x} - mu)$ where $mu = mu_0$ is taken to be equal to a known hypothesised value $mu_0 in mathbb{R}$. Here the function is indeed a statistic, so long as it is defined on the appropriately restricted domain. So the function $H_0: mathbb{R}^n rightarrow mathbb{R}$ with $H_0(x_1,...,x_n) = sqrt{n} (bar{x} - mu_0)$ would be a statistic, but the function $H: mathbb{R}^{n+1} rightarrow mathbb{R}$ with $H(x_1,...,x_n, mu) = sqrt{n} (bar{x} - mu)$ would not be a statistic.






          share|cite|improve this answer











          $endgroup$



          That definition is a somewhat awkward way to state it. A "statistic" is any function of the observable values. All that definition means is that a statistic is a function only of the observable values, not a function of the distribution or any of its parameters. For example, if $X_1, X_2, ..., X_n sim text{N}(mu, 1)$ then a statistic would be any function $T(X_1,...,X_n)$ whereas a function $H(X_1,....,X_n, mu)$ would not be a statistic, since it depends on $mu$. Here are some further examples:



          $$begin{equation} begin{aligned}
          text{Statistic} & & & & & bar{X}_n = frac{1}{n} sum_{i=1}^n X_i, \[12pt]
          text{Statistic} & & & & & S_n^2 = frac{1}{n} sum_{i=1}^n (X_i - bar{X}_n)^2, \[12pt]
          text{Not a statistic} & & & & & D_n = bar{X}_n - mu, \[12pt]
          text{Not a statistic} & & & & & p_i = text{N}(x_i | mu, 1), \[12pt]
          text{Not a statistic} & & & & & Q = 10 mu. \[12pt]
          end{aligned} end{equation}$$



          Every statistic is a function only of the observable values, and not of their distribution or its parameters. So there are no examples of a statistic that is a function of the distribution or its parameters (any such function would not be a statistic). However, it is important to note that the distribution of a statistic (as opposed to the statistic itself) will generally depend on the underlying distribution of the values. (This is true for all statistics other than ancillary statistics.)





          What about a function where the parameters are known? In the comments below, Alecos asks an excellent follow-up question. What about a function that uses a fixed hypothesised value of the parameter? For example, what about the statistic $sqrt{n} (bar{x} - mu)$ where $mu = mu_0$ is taken to be equal to a known hypothesised value $mu_0 in mathbb{R}$. Here the function is indeed a statistic, so long as it is defined on the appropriately restricted domain. So the function $H_0: mathbb{R}^n rightarrow mathbb{R}$ with $H_0(x_1,...,x_n) = sqrt{n} (bar{x} - mu_0)$ would be a statistic, but the function $H: mathbb{R}^{n+1} rightarrow mathbb{R}$ with $H(x_1,...,x_n, mu) = sqrt{n} (bar{x} - mu)$ would not be a statistic.







          share|cite|improve this answer














          share|cite|improve this answer



          share|cite|improve this answer








          edited Mar 11 at 21:17

























          answered Mar 11 at 10:05









          BenBen

          26.8k230124




          26.8k230124








          • 1




            $begingroup$
            Very helpful answer, considering the underlying statistical parameter as part of the non-statistic was particularly helpful.
            $endgroup$
            – Jake Kirsch
            Mar 11 at 10:56






          • 4




            $begingroup$
            @CarlWitthoft I don't get your point. If it's a function of the observable values, then it's a statistic. It may be a function of a smaller subset of the values; that can still be a useful thing to consider. If you want to estimate the mean and you have $10^{10}$ observations, you might still look at $(X_1+X_2+dots+X_{1000})/1000$ if the cost of processing data is high and the cost of error is small. Or for some reason you might want to consider two independent estimates of the mean, and could consider $(X_1+dots+X_{n/2})/(n/2)$ and $(X_{n/2+1}+dots+X_n)/(n/2)$. These are still statistics.
            $endgroup$
            – James Martin
            Mar 11 at 14:06








          • 4




            $begingroup$
            Those examples seem entirely valid to me. Are you saying the idea of dividing data into a training set and a validation set is not valid?
            $endgroup$
            – James Martin
            Mar 11 at 14:53








          • 2




            $begingroup$
            I'm a little confused by that as well. Let me attempt to describe @CarlWitthoft point. It would still be a statistic in terms of mathematical definition, but I could see a case where a consultant takes a 'statistic' of observations, but arbitrarily decides to remove a few results (consultants do this all the time right?). This would be 'valid' in the sense it's still a function on observations, however the way that statistic may be presented and interpreted likely wouldn't be valid.
            $endgroup$
            – Jake Kirsch
            Mar 11 at 15:41








          • 2




            $begingroup$
            @Carl Withhoft: With respect to the point you are making, it is important to distinguish between a statistic (which need not include all the data, and may not encompass all the information in the sample) and a sufficient statistic (which will encompass all the information with respect to some parameter). Statistical theory already has well-developed concepts like sufficiency that capture the idea that a statistic includes all relevant information in the sample. It is not necessary, or desirable, to try to build that requirement into the definition of a "statistic".
            $endgroup$
            – Ben
            Mar 11 at 21:11














          • 1




            $begingroup$
            Very helpful answer, considering the underlying statistical parameter as part of the non-statistic was particularly helpful.
            $endgroup$
            – Jake Kirsch
            Mar 11 at 10:56






          • 4




            $begingroup$
            @CarlWitthoft I don't get your point. If it's a function of the observable values, then it's a statistic. It may be a function of a smaller subset of the values; that can still be a useful thing to consider. If you want to estimate the mean and you have $10^{10}$ observations, you might still look at $(X_1+X_2+dots+X_{1000})/1000$ if the cost of processing data is high and the cost of error is small. Or for some reason you might want to consider two independent estimates of the mean, and could consider $(X_1+dots+X_{n/2})/(n/2)$ and $(X_{n/2+1}+dots+X_n)/(n/2)$. These are still statistics.
            $endgroup$
            – James Martin
            Mar 11 at 14:06








          • 4




            $begingroup$
            Those examples seem entirely valid to me. Are you saying the idea of dividing data into a training set and a validation set is not valid?
            $endgroup$
            – James Martin
            Mar 11 at 14:53








          • 2




            $begingroup$
            I'm a little confused by that as well. Let me attempt to describe @CarlWitthoft point. It would still be a statistic in terms of mathematical definition, but I could see a case where a consultant takes a 'statistic' of observations, but arbitrarily decides to remove a few results (consultants do this all the time right?). This would be 'valid' in the sense it's still a function on observations, however the way that statistic may be presented and interpreted likely wouldn't be valid.
            $endgroup$
            – Jake Kirsch
            Mar 11 at 15:41








          • 2




            $begingroup$
            @Carl Withhoft: With respect to the point you are making, it is important to distinguish between a statistic (which need not include all the data, and may not encompass all the information in the sample) and a sufficient statistic (which will encompass all the information with respect to some parameter). Statistical theory already has well-developed concepts like sufficiency that capture the idea that a statistic includes all relevant information in the sample. It is not necessary, or desirable, to try to build that requirement into the definition of a "statistic".
            $endgroup$
            – Ben
            Mar 11 at 21:11








          1




          1




          $begingroup$
          Very helpful answer, considering the underlying statistical parameter as part of the non-statistic was particularly helpful.
          $endgroup$
          – Jake Kirsch
          Mar 11 at 10:56




          $begingroup$
          Very helpful answer, considering the underlying statistical parameter as part of the non-statistic was particularly helpful.
          $endgroup$
          – Jake Kirsch
          Mar 11 at 10:56




          4




          4




          $begingroup$
          @CarlWitthoft I don't get your point. If it's a function of the observable values, then it's a statistic. It may be a function of a smaller subset of the values; that can still be a useful thing to consider. If you want to estimate the mean and you have $10^{10}$ observations, you might still look at $(X_1+X_2+dots+X_{1000})/1000$ if the cost of processing data is high and the cost of error is small. Or for some reason you might want to consider two independent estimates of the mean, and could consider $(X_1+dots+X_{n/2})/(n/2)$ and $(X_{n/2+1}+dots+X_n)/(n/2)$. These are still statistics.
          $endgroup$
          – James Martin
          Mar 11 at 14:06






          $begingroup$
          @CarlWitthoft I don't get your point. If it's a function of the observable values, then it's a statistic. It may be a function of a smaller subset of the values; that can still be a useful thing to consider. If you want to estimate the mean and you have $10^{10}$ observations, you might still look at $(X_1+X_2+dots+X_{1000})/1000$ if the cost of processing data is high and the cost of error is small. Or for some reason you might want to consider two independent estimates of the mean, and could consider $(X_1+dots+X_{n/2})/(n/2)$ and $(X_{n/2+1}+dots+X_n)/(n/2)$. These are still statistics.
          $endgroup$
          – James Martin
          Mar 11 at 14:06






          4




          4




          $begingroup$
          Those examples seem entirely valid to me. Are you saying the idea of dividing data into a training set and a validation set is not valid?
          $endgroup$
          – James Martin
          Mar 11 at 14:53






          $begingroup$
          Those examples seem entirely valid to me. Are you saying the idea of dividing data into a training set and a validation set is not valid?
          $endgroup$
          – James Martin
          Mar 11 at 14:53






          2




          2




          $begingroup$
          I'm a little confused by that as well. Let me attempt to describe @CarlWitthoft point. It would still be a statistic in terms of mathematical definition, but I could see a case where a consultant takes a 'statistic' of observations, but arbitrarily decides to remove a few results (consultants do this all the time right?). This would be 'valid' in the sense it's still a function on observations, however the way that statistic may be presented and interpreted likely wouldn't be valid.
          $endgroup$
          – Jake Kirsch
          Mar 11 at 15:41






          $begingroup$
          I'm a little confused by that as well. Let me attempt to describe @CarlWitthoft point. It would still be a statistic in terms of mathematical definition, but I could see a case where a consultant takes a 'statistic' of observations, but arbitrarily decides to remove a few results (consultants do this all the time right?). This would be 'valid' in the sense it's still a function on observations, however the way that statistic may be presented and interpreted likely wouldn't be valid.
          $endgroup$
          – Jake Kirsch
          Mar 11 at 15:41






          2




          2




          $begingroup$
          @Carl Withhoft: With respect to the point you are making, it is important to distinguish between a statistic (which need not include all the data, and may not encompass all the information in the sample) and a sufficient statistic (which will encompass all the information with respect to some parameter). Statistical theory already has well-developed concepts like sufficiency that capture the idea that a statistic includes all relevant information in the sample. It is not necessary, or desirable, to try to build that requirement into the definition of a "statistic".
          $endgroup$
          – Ben
          Mar 11 at 21:11




          $begingroup$
          @Carl Withhoft: With respect to the point you are making, it is important to distinguish between a statistic (which need not include all the data, and may not encompass all the information in the sample) and a sufficient statistic (which will encompass all the information with respect to some parameter). Statistical theory already has well-developed concepts like sufficiency that capture the idea that a statistic includes all relevant information in the sample. It is not necessary, or desirable, to try to build that requirement into the definition of a "statistic".
          $endgroup$
          – Ben
          Mar 11 at 21:11













          4












          $begingroup$

          I interpret that as saying that you should decide before you see the data what statistic you are going to calculate. So, for instance, if you're going to take out outliers, you should decide before you see the data what constitutes an "outlier". If you decide after you see the data, then your function is dependent on the data.






          share|cite|improve this answer









          $endgroup$













          • $begingroup$
            this is also helpful! So making a decision on which observations to include in the function after knowing what observations are available, which is more or less what I was describing in my comment on the previous answer.
            $endgroup$
            – Jake Kirsch
            Mar 11 at 19:37






          • 2




            $begingroup$
            (+1) It might be worth noting that this important because if you define a rule a prior about what constitutes a data point that will be dropped, it is (relatively) easy to derive a distribution for statistic (i.e., truncated mean, etc.). It's really hard to derive a distribution for a measure that involves dropping data points for reasons that are not cleanly defined before hand.
            $endgroup$
            – Cliff AB
            Mar 11 at 23:41
















          4












          $begingroup$

          I interpret that as saying that you should decide before you see the data what statistic you are going to calculate. So, for instance, if you're going to take out outliers, you should decide before you see the data what constitutes an "outlier". If you decide after you see the data, then your function is dependent on the data.






          share|cite|improve this answer









          $endgroup$













          • $begingroup$
            this is also helpful! So making a decision on which observations to include in the function after knowing what observations are available, which is more or less what I was describing in my comment on the previous answer.
            $endgroup$
            – Jake Kirsch
            Mar 11 at 19:37






          • 2




            $begingroup$
            (+1) It might be worth noting that this important because if you define a rule a prior about what constitutes a data point that will be dropped, it is (relatively) easy to derive a distribution for statistic (i.e., truncated mean, etc.). It's really hard to derive a distribution for a measure that involves dropping data points for reasons that are not cleanly defined before hand.
            $endgroup$
            – Cliff AB
            Mar 11 at 23:41














          4












          4








          4





          $begingroup$

          I interpret that as saying that you should decide before you see the data what statistic you are going to calculate. So, for instance, if you're going to take out outliers, you should decide before you see the data what constitutes an "outlier". If you decide after you see the data, then your function is dependent on the data.






          share|cite|improve this answer









          $endgroup$



          I interpret that as saying that you should decide before you see the data what statistic you are going to calculate. So, for instance, if you're going to take out outliers, you should decide before you see the data what constitutes an "outlier". If you decide after you see the data, then your function is dependent on the data.







          share|cite|improve this answer












          share|cite|improve this answer



          share|cite|improve this answer










          answered Mar 11 at 17:29









          AcccumulationAcccumulation

          1,68626




          1,68626












          • $begingroup$
            this is also helpful! So making a decision on which observations to include in the function after knowing what observations are available, which is more or less what I was describing in my comment on the previous answer.
            $endgroup$
            – Jake Kirsch
            Mar 11 at 19:37






          • 2




            $begingroup$
            (+1) It might be worth noting that this important because if you define a rule a prior about what constitutes a data point that will be dropped, it is (relatively) easy to derive a distribution for statistic (i.e., truncated mean, etc.). It's really hard to derive a distribution for a measure that involves dropping data points for reasons that are not cleanly defined before hand.
            $endgroup$
            – Cliff AB
            Mar 11 at 23:41


















          • $begingroup$
            this is also helpful! So making a decision on which observations to include in the function after knowing what observations are available, which is more or less what I was describing in my comment on the previous answer.
            $endgroup$
            – Jake Kirsch
            Mar 11 at 19:37






          • 2




            $begingroup$
            (+1) It might be worth noting that this important because if you define a rule a prior about what constitutes a data point that will be dropped, it is (relatively) easy to derive a distribution for statistic (i.e., truncated mean, etc.). It's really hard to derive a distribution for a measure that involves dropping data points for reasons that are not cleanly defined before hand.
            $endgroup$
            – Cliff AB
            Mar 11 at 23:41
















          $begingroup$
          this is also helpful! So making a decision on which observations to include in the function after knowing what observations are available, which is more or less what I was describing in my comment on the previous answer.
          $endgroup$
          – Jake Kirsch
          Mar 11 at 19:37




          $begingroup$
          this is also helpful! So making a decision on which observations to include in the function after knowing what observations are available, which is more or less what I was describing in my comment on the previous answer.
          $endgroup$
          – Jake Kirsch
          Mar 11 at 19:37




          2




          2




          $begingroup$
          (+1) It might be worth noting that this important because if you define a rule a prior about what constitutes a data point that will be dropped, it is (relatively) easy to derive a distribution for statistic (i.e., truncated mean, etc.). It's really hard to derive a distribution for a measure that involves dropping data points for reasons that are not cleanly defined before hand.
          $endgroup$
          – Cliff AB
          Mar 11 at 23:41




          $begingroup$
          (+1) It might be worth noting that this important because if you define a rule a prior about what constitutes a data point that will be dropped, it is (relatively) easy to derive a distribution for statistic (i.e., truncated mean, etc.). It's really hard to derive a distribution for a measure that involves dropping data points for reasons that are not cleanly defined before hand.
          $endgroup$
          – Cliff AB
          Mar 11 at 23:41


















          draft saved

          draft discarded




















































          Thanks for contributing an answer to Cross Validated!


          • Please be sure to answer the question. Provide details and share your research!

          But avoid



          • Asking for help, clarification, or responding to other answers.

          • Making statements based on opinion; back them up with references or personal experience.


          Use MathJax to format equations. MathJax reference.


          To learn more, see our tips on writing great answers.




          draft saved


          draft discarded














          StackExchange.ready(
          function () {
          StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstats.stackexchange.com%2fquestions%2f396815%2fexamples-of-a-statistic-that-is-not-independent-of-samples-distribution%23new-answer', 'question_page');
          }
          );

          Post as a guest















          Required, but never shown





















































          Required, but never shown














          Required, but never shown












          Required, but never shown







          Required, but never shown

































          Required, but never shown














          Required, but never shown












          Required, but never shown







          Required, but never shown







          Popular posts from this blog

          How to change which sound is reproduced for terminal bell?

          Can I use Tabulator js library in my java Spring + Thymeleaf project?

          Title Spacing in Bjornstrup Chapter, Removing Chapter Number From Contents