There are countless ways to develop a model to predict the future behavior of almost any data that can be defined as a time series. The precision of these methods is persistently getting better and better. However, there is one question in my mind to which I've not been able to find an answer, and that is why we use exogenous inputs in time series modeling?

The reasoning behind an autoregressive time series model is pretty solid. It can and certainly have in most cases resulted in acceptable models, not particularly excellent but acceptable and we can use the outputs of the same model as inputs for future time steps.

There is a fact that introducing an exogenous input to the model, for example, using a set of precipitation data to an autoregressive river discharge estimation model, can significantly increase the precision of the model, for both the training and testing phase. But using another variable in a model that is also autocorrelated seems really wrong. Now, to forecast the future we need to develop another model to predict that exogenous variable as well, since we don't have future data for that excess input! Even if a time series model can be trained to tremendously good precision, we have essentially taken away its forecasting capability and I think this makes the model completely useless. What is the reasoning behind such an approach? Am I missing something?

More Sina Masoumzadeh's questions See All
Similar questions and discussions