Which estimation technique minimizes the MAPE?












3












$begingroup$


Suppose we have two estimation techniques:




  • Linear Least Squares, which aims to minimize squared residuals

  • Least Absolute Deviation, which aims to minimize absolute residuals


We have a model, which purpose is prediction. More specifically, we want the model to perform good based on the prediction evaluation measure: Mean Absolute Prediction Error (MAPE). Assuming we have no information of the data, can we say something about which estimation technique will likely perform better?










share|cite|improve this question











$endgroup$

















    3












    $begingroup$


    Suppose we have two estimation techniques:




    • Linear Least Squares, which aims to minimize squared residuals

    • Least Absolute Deviation, which aims to minimize absolute residuals


    We have a model, which purpose is prediction. More specifically, we want the model to perform good based on the prediction evaluation measure: Mean Absolute Prediction Error (MAPE). Assuming we have no information of the data, can we say something about which estimation technique will likely perform better?










    share|cite|improve this question











    $endgroup$















      3












      3








      3





      $begingroup$


      Suppose we have two estimation techniques:




      • Linear Least Squares, which aims to minimize squared residuals

      • Least Absolute Deviation, which aims to minimize absolute residuals


      We have a model, which purpose is prediction. More specifically, we want the model to perform good based on the prediction evaluation measure: Mean Absolute Prediction Error (MAPE). Assuming we have no information of the data, can we say something about which estimation technique will likely perform better?










      share|cite|improve this question











      $endgroup$




      Suppose we have two estimation techniques:




      • Linear Least Squares, which aims to minimize squared residuals

      • Least Absolute Deviation, which aims to minimize absolute residuals


      We have a model, which purpose is prediction. More specifically, we want the model to perform good based on the prediction evaluation measure: Mean Absolute Prediction Error (MAPE). Assuming we have no information of the data, can we say something about which estimation technique will likely perform better?







      forecasting estimation prediction






      share|cite|improve this question















      share|cite|improve this question













      share|cite|improve this question




      share|cite|improve this question








      edited Jan 8 at 14:50









      Stephan Kolassa

      44k692161




      44k692161










      asked Jan 8 at 10:04









      K. RoelofsK. Roelofs

      187




      187






















          1 Answer
          1






          active

          oldest

          votes


















          4












          $begingroup$

          If the probability density of your future distribution is positively skewed, then typically (though not always; von Hippel, 2005) the median will be lower than its mean. So a technique that aims at the median as a point forecast will be biased low. Since the MAPE usually prefers a low biased prediction, such a technique will usually perform better in terms of the MAPE in such a situation.



          Note that there are a couple of caveats in this description. The chain of reasoning is not perfect, because you can find pathological counter-examples for at least two steps. Nevertheless, it should work in most practical cases.



          You may be better off using a custom optimization routine that directly attempts to minimize the MAPE. The problem being, of course, that the MAPE is not differentiable at perfect forecasts. Alternatively, you could try to estimate full predictive densities and then output the (-1)-median of this density as a point forecast, which is the functional that minimizes the MAPE in expectation (Gneiting, 2011, p. 752 with $beta=-1$).



          You may be interested in What are the shortcomings of the Mean Absolute Percentage Error (MAPE)?






          share|cite|improve this answer









          $endgroup$













          • $begingroup$
            The Mean Absolute Prediction Error is something else than the Mean Absolute Percentage Error. Mean Absolute Prediction Error is just the mean of the absolute prediction errors.
            $endgroup$
            – K. Roelofs
            Jan 8 at 14:44






          • 1




            $begingroup$
            Ah. Sorry. I misread. (So did three upvoters.) Question: since you already consider techniques that minimize the mean absolute error, what else are you looking for? Of course a technique can overfit in minimizing the MAE just as it can overfit in minimizing the MSE, so is control of overfitting your question? If so, the very same methods apply to MAE optimizers: cross-validation and regularization. What am I missing?
            $endgroup$
            – Stephan Kolassa
            Jan 8 at 14:50













          Your Answer





          StackExchange.ifUsing("editor", function () {
          return StackExchange.using("mathjaxEditing", function () {
          StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
          StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
          });
          });
          }, "mathjax-editing");

          StackExchange.ready(function() {
          var channelOptions = {
          tags: "".split(" "),
          id: "65"
          };
          initTagRenderer("".split(" "), "".split(" "), channelOptions);

          StackExchange.using("externalEditor", function() {
          // Have to fire editor after snippets, if snippets enabled
          if (StackExchange.settings.snippets.snippetsEnabled) {
          StackExchange.using("snippets", function() {
          createEditor();
          });
          }
          else {
          createEditor();
          }
          });

          function createEditor() {
          StackExchange.prepareEditor({
          heartbeatType: 'answer',
          autoActivateHeartbeat: false,
          convertImagesToLinks: false,
          noModals: true,
          showLowRepImageUploadWarning: true,
          reputationToPostImages: null,
          bindNavPrevention: true,
          postfix: "",
          imageUploader: {
          brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
          contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
          allowUrls: true
          },
          onDemand: true,
          discardSelector: ".discard-answer"
          ,immediatelyShowMarkdownHelp:true
          });


          }
          });














          draft saved

          draft discarded


















          StackExchange.ready(
          function () {
          StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstats.stackexchange.com%2fquestions%2f386111%2fwhich-estimation-technique-minimizes-the-mape%23new-answer', 'question_page');
          }
          );

          Post as a guest















          Required, but never shown

























          1 Answer
          1






          active

          oldest

          votes








          1 Answer
          1






          active

          oldest

          votes









          active

          oldest

          votes






          active

          oldest

          votes









          4












          $begingroup$

          If the probability density of your future distribution is positively skewed, then typically (though not always; von Hippel, 2005) the median will be lower than its mean. So a technique that aims at the median as a point forecast will be biased low. Since the MAPE usually prefers a low biased prediction, such a technique will usually perform better in terms of the MAPE in such a situation.



          Note that there are a couple of caveats in this description. The chain of reasoning is not perfect, because you can find pathological counter-examples for at least two steps. Nevertheless, it should work in most practical cases.



          You may be better off using a custom optimization routine that directly attempts to minimize the MAPE. The problem being, of course, that the MAPE is not differentiable at perfect forecasts. Alternatively, you could try to estimate full predictive densities and then output the (-1)-median of this density as a point forecast, which is the functional that minimizes the MAPE in expectation (Gneiting, 2011, p. 752 with $beta=-1$).



          You may be interested in What are the shortcomings of the Mean Absolute Percentage Error (MAPE)?






          share|cite|improve this answer









          $endgroup$













          • $begingroup$
            The Mean Absolute Prediction Error is something else than the Mean Absolute Percentage Error. Mean Absolute Prediction Error is just the mean of the absolute prediction errors.
            $endgroup$
            – K. Roelofs
            Jan 8 at 14:44






          • 1




            $begingroup$
            Ah. Sorry. I misread. (So did three upvoters.) Question: since you already consider techniques that minimize the mean absolute error, what else are you looking for? Of course a technique can overfit in minimizing the MAE just as it can overfit in minimizing the MSE, so is control of overfitting your question? If so, the very same methods apply to MAE optimizers: cross-validation and regularization. What am I missing?
            $endgroup$
            – Stephan Kolassa
            Jan 8 at 14:50


















          4












          $begingroup$

          If the probability density of your future distribution is positively skewed, then typically (though not always; von Hippel, 2005) the median will be lower than its mean. So a technique that aims at the median as a point forecast will be biased low. Since the MAPE usually prefers a low biased prediction, such a technique will usually perform better in terms of the MAPE in such a situation.



          Note that there are a couple of caveats in this description. The chain of reasoning is not perfect, because you can find pathological counter-examples for at least two steps. Nevertheless, it should work in most practical cases.



          You may be better off using a custom optimization routine that directly attempts to minimize the MAPE. The problem being, of course, that the MAPE is not differentiable at perfect forecasts. Alternatively, you could try to estimate full predictive densities and then output the (-1)-median of this density as a point forecast, which is the functional that minimizes the MAPE in expectation (Gneiting, 2011, p. 752 with $beta=-1$).



          You may be interested in What are the shortcomings of the Mean Absolute Percentage Error (MAPE)?






          share|cite|improve this answer









          $endgroup$













          • $begingroup$
            The Mean Absolute Prediction Error is something else than the Mean Absolute Percentage Error. Mean Absolute Prediction Error is just the mean of the absolute prediction errors.
            $endgroup$
            – K. Roelofs
            Jan 8 at 14:44






          • 1




            $begingroup$
            Ah. Sorry. I misread. (So did three upvoters.) Question: since you already consider techniques that minimize the mean absolute error, what else are you looking for? Of course a technique can overfit in minimizing the MAE just as it can overfit in minimizing the MSE, so is control of overfitting your question? If so, the very same methods apply to MAE optimizers: cross-validation and regularization. What am I missing?
            $endgroup$
            – Stephan Kolassa
            Jan 8 at 14:50
















          4












          4








          4





          $begingroup$

          If the probability density of your future distribution is positively skewed, then typically (though not always; von Hippel, 2005) the median will be lower than its mean. So a technique that aims at the median as a point forecast will be biased low. Since the MAPE usually prefers a low biased prediction, such a technique will usually perform better in terms of the MAPE in such a situation.



          Note that there are a couple of caveats in this description. The chain of reasoning is not perfect, because you can find pathological counter-examples for at least two steps. Nevertheless, it should work in most practical cases.



          You may be better off using a custom optimization routine that directly attempts to minimize the MAPE. The problem being, of course, that the MAPE is not differentiable at perfect forecasts. Alternatively, you could try to estimate full predictive densities and then output the (-1)-median of this density as a point forecast, which is the functional that minimizes the MAPE in expectation (Gneiting, 2011, p. 752 with $beta=-1$).



          You may be interested in What are the shortcomings of the Mean Absolute Percentage Error (MAPE)?






          share|cite|improve this answer









          $endgroup$



          If the probability density of your future distribution is positively skewed, then typically (though not always; von Hippel, 2005) the median will be lower than its mean. So a technique that aims at the median as a point forecast will be biased low. Since the MAPE usually prefers a low biased prediction, such a technique will usually perform better in terms of the MAPE in such a situation.



          Note that there are a couple of caveats in this description. The chain of reasoning is not perfect, because you can find pathological counter-examples for at least two steps. Nevertheless, it should work in most practical cases.



          You may be better off using a custom optimization routine that directly attempts to minimize the MAPE. The problem being, of course, that the MAPE is not differentiable at perfect forecasts. Alternatively, you could try to estimate full predictive densities and then output the (-1)-median of this density as a point forecast, which is the functional that minimizes the MAPE in expectation (Gneiting, 2011, p. 752 with $beta=-1$).



          You may be interested in What are the shortcomings of the Mean Absolute Percentage Error (MAPE)?







          share|cite|improve this answer












          share|cite|improve this answer



          share|cite|improve this answer










          answered Jan 8 at 10:52









          Stephan KolassaStephan Kolassa

          44k692161




          44k692161












          • $begingroup$
            The Mean Absolute Prediction Error is something else than the Mean Absolute Percentage Error. Mean Absolute Prediction Error is just the mean of the absolute prediction errors.
            $endgroup$
            – K. Roelofs
            Jan 8 at 14:44






          • 1




            $begingroup$
            Ah. Sorry. I misread. (So did three upvoters.) Question: since you already consider techniques that minimize the mean absolute error, what else are you looking for? Of course a technique can overfit in minimizing the MAE just as it can overfit in minimizing the MSE, so is control of overfitting your question? If so, the very same methods apply to MAE optimizers: cross-validation and regularization. What am I missing?
            $endgroup$
            – Stephan Kolassa
            Jan 8 at 14:50




















          • $begingroup$
            The Mean Absolute Prediction Error is something else than the Mean Absolute Percentage Error. Mean Absolute Prediction Error is just the mean of the absolute prediction errors.
            $endgroup$
            – K. Roelofs
            Jan 8 at 14:44






          • 1




            $begingroup$
            Ah. Sorry. I misread. (So did three upvoters.) Question: since you already consider techniques that minimize the mean absolute error, what else are you looking for? Of course a technique can overfit in minimizing the MAE just as it can overfit in minimizing the MSE, so is control of overfitting your question? If so, the very same methods apply to MAE optimizers: cross-validation and regularization. What am I missing?
            $endgroup$
            – Stephan Kolassa
            Jan 8 at 14:50


















          $begingroup$
          The Mean Absolute Prediction Error is something else than the Mean Absolute Percentage Error. Mean Absolute Prediction Error is just the mean of the absolute prediction errors.
          $endgroup$
          – K. Roelofs
          Jan 8 at 14:44




          $begingroup$
          The Mean Absolute Prediction Error is something else than the Mean Absolute Percentage Error. Mean Absolute Prediction Error is just the mean of the absolute prediction errors.
          $endgroup$
          – K. Roelofs
          Jan 8 at 14:44




          1




          1




          $begingroup$
          Ah. Sorry. I misread. (So did three upvoters.) Question: since you already consider techniques that minimize the mean absolute error, what else are you looking for? Of course a technique can overfit in minimizing the MAE just as it can overfit in minimizing the MSE, so is control of overfitting your question? If so, the very same methods apply to MAE optimizers: cross-validation and regularization. What am I missing?
          $endgroup$
          – Stephan Kolassa
          Jan 8 at 14:50






          $begingroup$
          Ah. Sorry. I misread. (So did three upvoters.) Question: since you already consider techniques that minimize the mean absolute error, what else are you looking for? Of course a technique can overfit in minimizing the MAE just as it can overfit in minimizing the MSE, so is control of overfitting your question? If so, the very same methods apply to MAE optimizers: cross-validation and regularization. What am I missing?
          $endgroup$
          – Stephan Kolassa
          Jan 8 at 14:50




















          draft saved

          draft discarded




















































          Thanks for contributing an answer to Cross Validated!


          • Please be sure to answer the question. Provide details and share your research!

          But avoid



          • Asking for help, clarification, or responding to other answers.

          • Making statements based on opinion; back them up with references or personal experience.


          Use MathJax to format equations. MathJax reference.


          To learn more, see our tips on writing great answers.




          draft saved


          draft discarded














          StackExchange.ready(
          function () {
          StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstats.stackexchange.com%2fquestions%2f386111%2fwhich-estimation-technique-minimizes-the-mape%23new-answer', 'question_page');
          }
          );

          Post as a guest















          Required, but never shown





















































          Required, but never shown














          Required, but never shown












          Required, but never shown







          Required, but never shown

































          Required, but never shown














          Required, but never shown












          Required, but never shown







          Required, but never shown







          Popular posts from this blog

          Biblatex bibliography style without URLs when DOI exists (in Overleaf with Zotero bibliography)

          How to change which sound is reproduced for terminal bell?

          Can I use Tabulator js library in my java Spring + Thymeleaf project?