Please use this identifier to cite or link to this item: https://hdl.handle.net/10356/166684
Title: Architecture of electrical equipment health and monitoring (EHM) system based on influxDB platform
Authors: Nong, Chunkai
Keywords: Engineering::Electrical and electronic engineering::Computer hardware, software and systems
Issue Date: 2023
Publisher: Nanyang Technological University
Source: Nong, C. (2023). Architecture of electrical equipment health and monitoring (EHM) system based on influxDB platform. Master's thesis, Nanyang Technological University, Singapore. https://hdl.handle.net/10356/166684
Abstract: The Transformer architecture was originally developed for natural language processing tasks, such as machine translation and language modeling. The architecture consists of a self-attention mechanism that allows the model to attend to different parts of the input sequence, enabling it to capture long-range dependencies. Recently, the Transformer architecture has gained popularity in time series forecasting tasks, where the goal is to predict future values of a sequence based on its past values. One advantage of using the Transformer for time series forecasting is that it can handle variable-length input sequences, which is a common characteristic of time series data. Additionally, the self-attention mechanism allows the model to capture complex temporal relationships between different parts of the input sequence. There are several deep learning models based on the Transformer architecture that have been proposed for time series forecasting, including Informer, FEDformer, and others. The present study compares the performance of two deep learning models, Informer and FEDformer, for time series forecasting on two datasets, ETDataset and a real-world dataset. The results indicate that FEDformer outperforms Informer in both datasets. Furthermore, it was observed that the Transformer model, which is the basis for both Informer and FEDformer, performs well on certain features of the data but not on others. These findings suggest that while Transformer-based models may be effective for time series forecasting, their performance may vary depending on the characteristics of the data being analyzed. This study contributes to the growing body of research on deep learning models for time series forecasting and highlights the importance of selecting the appropriate model for a given dataset.
URI: https://hdl.handle.net/10356/166684
Schools: School of Electrical and Electronic Engineering 
Fulltext Permission: restricted
Fulltext Availability: With Fulltext
Appears in Collections:EEE Theses

Files in This Item:
File Description SizeFormat 
Final_Dissertation.pdf
  Restricted Access
2.96 MBAdobe PDFView/Open

Page view(s)

129
Updated on Apr 16, 2024

Download(s)

9
Updated on Apr 16, 2024

Google ScholarTM

Check

Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.