Show simple item record

dc.contributor.authorHaugsdal, Espen
dc.contributor.authorAune, Erlend
dc.contributor.authorRuocco, Massimiliano
dc.date.accessioned2024-01-08T09:05:45Z
dc.date.available2024-01-08T09:05:45Z
dc.date.created2023-08-29T16:10:58Z
dc.date.issued2023
dc.identifier.citationApplied intelligence (Boston). 2023, 26781-26796.en_US
dc.identifier.issn0924-669X
dc.identifier.urihttps://hdl.handle.net/11250/3110312
dc.description.abstractTime series forecasting is an important problem, with many real world applications. Transformer models have been successfully applied to natural language processing tasks, but have received relatively little attention for time series forecasting. Motivated by the differences between classification tasks and forecasting, we propose PI-Transformer, an adaptation of the Transformer architecture designed for time series forecasting, consisting of three parts: First, we propose a novel initialization method called Persistence Initialization, with the goal of increasing training stability of forecasting models by ensuring that the initial outputs of an untrained model are identical to the outputs of a simple baseline model. Second, we use ReZero normalization instead of Layer Normalization, in order to further tackle issues related to training stability. Third, we use Rotary positional encodings to provide a better inductive bias for forecasting. Multiple ablation studies show that the PI-Transformer is more accurate, learns faster, and scales better than regular Transformer models. Finally, PI-Transformer achieves competitive performance on the challenging M4 dataset, both when compared to the current state of the art, and to recently proposed Transformer models for time series forecasting.en_US
dc.language.isoengen_US
dc.publisherSpringeren_US
dc.rightsNavngivelse 4.0 Internasjonal*
dc.rights.urihttp://creativecommons.org/licenses/by/4.0/deed.no*
dc.titlePersistence Initialization: a novel adaptation of the Transformer architecture for time series forecastingen_US
dc.title.alternativePersistence Initialization: a novel adaptation of the Transformer architecture for time series forecastingen_US
dc.typePeer revieweden_US
dc.typeJournal articleen_US
dc.description.versionpublishedVersionen_US
dc.source.pagenumber26781-26796en_US
dc.source.journalApplied intelligence (Boston)en_US
dc.identifier.doi10.1007/s10489-023-04927-4
dc.identifier.cristin2170636
dc.relation.projectNorges forskningsråd: 312062en_US
cristin.ispublishedtrue
cristin.fulltextoriginal
cristin.qualitycode2


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record

Navngivelse 4.0 Internasjonal
Except where otherwise noted, this item's license is described as Navngivelse 4.0 Internasjonal