IDEAS home Printed from https://ideas.repec.org/a/gam/jeners/v17y2024i24p6452-d1549422.html
   My bibliography  Save this article

Partial Transfer Learning from Patch Transformer to Variate-Based Linear Forecasting Model

Author

Listed:
  • Le Hoang Anh

    (Department of Intelligent Electronics and Computer Engineering, Chonnam National University, Gwangju 61186, Republic of Korea
    These authors contributed equally to this work.)

  • Dang Thanh Vu

    (Research Center, AISeed Inc., Gwangju 61186, Republic of Korea
    These authors contributed equally to this work.)

  • Seungmin Oh

    (Department of Intelligent Electronics and Computer Engineering, Chonnam National University, Gwangju 61186, Republic of Korea)

  • Gwang-Hyun Yu

    (Department of Intelligent Electronics and Computer Engineering, Chonnam National University, Gwangju 61186, Republic of Korea)

  • Nguyen Bui Ngoc Han

    (Department of Electronic Convergence Engineering, Kwangwoon University, Seoul 01897, Republic of Korea)

  • Hyoung-Gook Kim

    (Department of Electronic Convergence Engineering, Kwangwoon University, Seoul 01897, Republic of Korea)

  • Jin-Sul Kim

    (Department of Intelligent Electronics and Computer Engineering, Chonnam National University, Gwangju 61186, Republic of Korea)

  • Jin-Young Kim

    (Department of Intelligent Electronics and Computer Engineering, Chonnam National University, Gwangju 61186, Republic of Korea)

Abstract

Transformer-based time series forecasting models use patch tokens for temporal patterns and variate tokens to learn covariates’ dependencies. While patch tokens inherently facilitate self-supervised learning, variate tokens are more suitable for linear forecasters as they help to mitigate distribution drift. However, the use of variate tokens prohibits masked model pretraining, as masking an entire series is absurd. To close this gap, we propose LSPatch-T (Long–Short Patch Transfer), a framework that transfers knowledge from short-length patch tokens into full-length variate tokens. A key implementation is that we selectively transfer a portion of the Transformer encoder to ensure the linear design of the downstream model. Additionally, we introduce a robust frequency loss to maintain consistency across different temporal ranges. The experimental results show that our approach outperforms Transformer-based baselines (Transformer, Informer, Crossformer, Autoformer, PatchTST, iTransformer) on three public datasets (ETT, Exchange, Weather), which is a promising step forward in generalizing time series forecasting models.

Suggested Citation

  • Le Hoang Anh & Dang Thanh Vu & Seungmin Oh & Gwang-Hyun Yu & Nguyen Bui Ngoc Han & Hyoung-Gook Kim & Jin-Sul Kim & Jin-Young Kim, 2024. "Partial Transfer Learning from Patch Transformer to Variate-Based Linear Forecasting Model," Energies, MDPI, vol. 17(24), pages 1-18, December.
  • Handle: RePEc:gam:jeners:v:17:y:2024:i:24:p:6452-:d:1549422
    as

    Download full text from publisher

    File URL: https://www.mdpi.com/1996-1073/17/24/6452/pdf
    Download Restriction: no

    File URL: https://www.mdpi.com/1996-1073/17/24/6452/
    Download Restriction: no
    ---><---

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:gam:jeners:v:17:y:2024:i:24:p:6452-:d:1549422. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    We have no bibliographic references for this item. You can help adding them by using this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: MDPI Indexing Manager (email available below). General contact details of provider: https://www.mdpi.com .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.