Let’s revisit the spot where we paused: tackling the responsibility of multi-step time-series forecasting.
We attempted an initial solution that functioned as a makeshift fix. The mannequin was programmed to deliver a solitary forecast, in line with the predetermined timeline. To generate a longer-term forecast, we could simply refeed the model with its own prediction, incrementing the input sequence by one step.
In contrast, the latest model could be trained to predict a customizable number of observations instantaneously. Although the underlying framework remains crucial, its simplicity means it can serve as a starting point for future endeavors.
We utilise the same data as previously. vic_elec
from tsibbledata
.
Compared to the final result, however, dataset
class has to vary. Prior to production, each product line had a distinct target.y
It’s now a vector, equivalent to the input. x
. And identical to n_timesteps
Is now optional and allows specifying a custom enter sequence size when needed. n_forecast
, to configure goal dimension.
In our instance, n_timesteps
and n_forecast
Are assigned the same value, yet there’s no requirement for them to share equal footing. You can also effectively apply patterns from week-long sequences to forecast developments over a single day or a month.
Aside from the truth that .getitem()
now returns a vector for y
in addition to x
There isn’t much to say regarding dataset development.
The following code enables the arrangement of information within a pipeline.
The mannequin supplants the original single linear layer responsible for producing the final prediction with a small network comprising two linear layers, optionally paired with dropout.
In ahead()
We initially employ the RNN, consistent with our previous exposition, utilizing outputs
solely; or extra particularly, the output
The concept of a singular, definitive moment in time, akin to the ultimate time step? What is the purpose behind this brief snippet? torch
RNN returns.)
With the introduction of mannequin instantiation, an additional configuration parameter has been implemented, governing the frequency of dropout between the two linear layers.
The coaching process remains unaltered in its entirety.
Epoch 1, Coaching: Loss - 0.65737; Validation: Loss - 0.54586 Epoch 2, Coaching: Loss - 0.43991; Validation: Loss - 0.50588 ... Epoch 30, Coaching: Loss - 0.20022; Validation: Loss - 0.33331
As the decline in losses on the coaching dataset suggests, it appears likely that the model is indeed learning something. It’s likely to continue progressing over a considerable period of time still. While we observe significant gains on the training set, this improvement does not translate as well to the validation set, suggesting that our model may be overfitting.
Now that we’ve trained and evaluated our model, we’re eager to see how well it performs on unseen data. For testing purposes, we’re focusing on the remarkably arduous month of January 2014 – an unusually demanding period due to a severe heatwave that generated extremely high and sustained demand.
Without the need for complex looping constructs, the analysis simplifies significantly.

Determining One-Week-Ahead Predictions for January 2014
The reforecasted outcome from iteratively refining predictions has been secured. The revised demand profiles over the day now appear strikingly authentic. What are the most critical stages that require immediate attention? Evidently, the mirrors should not be reflected within the forecast, nor to a greater extent than in the “loop approach”. In practice, the model’s behavior is revealed through attention-grabbing insights into its character: When primed with inputs oscillating around a higher mean, it tends to deviate from its average, only to revert back to its comfort zone once more?
Since we provided the option to incorporate dropout within the MLP, you may wonder whether this could improve forecast accuracy on the test set. It appears that the phenomenon did not materialize as expected in my investigations. How do communities naturally recognize when top demand is emerging without external indicators like temperature?
We will draw a more nuanced line of demarcation in our assessment. As initial predictions unfold, it’s evident that one crucial aspect was misjudged – namely, the unanticipated surge in demand that persisted for roughly two to three days. Within those brief seconds, all that was required of the community was to maintain its current, elevated trajectory. Will it be attention-grabbing to observe how this challenge is addressed by the architectures we discuss subsequently?
What if we leveraged temperature as an additional input variable, allowing us to fine-tune our model’s performance in response to varying environmental conditions? While coaching efficiency did indeed experience an enhancement, a noticeable impact on the validation and test procedures remained elusive. Regardless, the likelihood of finding the code useful remains high, as its applicability extends seamlessly to datasets featuring additional predictor variables. As a result, we include this information in the appendix for reference purposes.
Thanks for studying!
Picture by on