The fpp3 package is a convenience meta-package associated with the textbook Forecasting: Principles and Practice. It loads the core tidy time-series packages used throughout the course and provides direct access to datasets used in examples. More importantly, its construction reflects the underlying principles of forecasting in a way that is both coherent and extensible. The workflow it enforces is not arbitrary — it mirrors how forecasting problems are actually approached in practice. Its versatility is well aligned with the types of tasks I expect to encounter both in this course and in future professional settings, where reproducibility, clarity, and interpretability matter as much as model performance.
A tsibble is a tidy time-series data structure that explicitly represents time through an index variable and supports multiple time series through key variables, enabling time-aware operations and modeling. This is important because the temporal nature of forecasting is its entire underlying principle. If a series cannot be parsed correctly over time — or related back to its own history in a coherent way — then any confidence interval or forecast output risks losing its intended meaning. The tsibble framework enforces this discipline by design, making it much harder to accidentally violate the assumptions that time-series analysis depends on.
The feasts package focuses on exploratory analysis and diagnostics for time series, including decomposition methods, autocorrelation analysis, and seasonal visualization. This is the stage where the pieces are tied together: before fitting models, you need to understand what structure is actually present in the data. These tools allow patterns to be surfaced and communicated clearly, and they support visualizations that are capable of conveying complex behavior without unnecessary abstraction.
The fable package provides a framework for fitting, comparing, and forecasting statistical time-series models using tsibbles, including ARIMA, ETS, and time-series linear models. The key point here is not just that models can be fit, but that they can be compared and evaluated within a consistent structure. This makes feature extraction and model behavior more explicit and precise, and it allows forecasts to be treated as first-class objects that integrate cleanly with the rest of the workflow.
For Part II, I pulled a monthly time series from FRED using the fredr package. The point here wasn’t just to download data, but to practice the full loop the course emphasizes: retrieve, structure correctly, visualize, and diagnose before moving toward modeling. If the data structure is wrong at the start, everything downstream becomes performative — plots and models may run, but they lose their interpretability.
After setting my API key, I requested the series and immediately converted it into a monthly tsibble. Even when FRED provides monthly observations, the raw date field can encourage treating the index like generic calendar dates. Converting the index to yearmonth() makes the temporal resolution explicit and prevents mismatches in seasonal tools, lag interpretation, and model assumptions. I used transmute() to keep only the time index and value so that every downstream operation remains interpretable.
Hyndman and Khandakar discuss automatic forecasting workflows and model selection in R (hyndman2008?).
@article{hyndman2008,title = {Automatic Time Series Forecasting: The forecast Package for R},author = {Hyndman, Rob J. and Khandakar, Yeasmin},journal = {Journal of Statistical Software},year = {2008},volume = {27},number = {3},pages = {1--22},doi = {10.18637/jss.v027.i03}}