Trend-Aware Data Imputation Based on Generative Adversarial Network for Time Series

Trend-Aware Data Imputation Based on Generative Adversarial Network for Time Series

Han Li, Zhenxiong Liu, Jixiang Niu, Zhongguo Yang, Sikandar Ali
DOI: 10.4018/IJITSA.325212
Article PDF Download
Open access articles are freely available for download

Abstract

To solve the problems of generative adversarial network (GAN)-based imputation method for time series, which are ignoring the implied trends in data and using multi-stage training that may lead to high training complexity, this article proposes a trend-aware data imputation method based on GAN (TrendGAN). It implements an end-to-end training using de-noising auto-encoder (DAE). It also uses bidirectional gated recurrent unit (Bi-GRU) in the generator model to consider the bi-directional characteristics and supplement the features lost by de-noising auto-encoder and improves the discriminator's ability using Bi-GRU and hint vector. The authors conducted experiments on four real datasets. The results showed that all components introduced into the method contribute to enhancing the imputation accuracy, and the MSE values of TrendGAN are much lower than those of baseline methods when dealing with time series with random and continuous missing patterns. That is, TrendGAN is suitable for data imputation in complex scenarios with two missing patterns coexist, such as electric power and transportation.
Article Preview
Top

1. Trend-Aware Data Imputation Based On Generative Adversarial Network For Time Series

The world is full of multi-variate time series data, and time series analysis has already played an important role in various fields, such as stock price prediction (Li & Yang, 2020), urban applications (Tabassum et al., 2021), geolocation (Chatzigeorgakidis et al., 2020), financial data modelling (Dogariu et al., 2022), satellite monitoring (Yuan et al., 2023), fault anomaly detection (Patel et al., 2022), and IoT device maintenance (Alghamdi et al., 2022). Time series, however, are often incomplete for equipment fault, transmission error, human factor, and for other reasons, which affects the effectiveness of data analysis.

Traditional methods of data imputation mainly fall into two categories: deletion-based method and filling-based method. The deletion-based method creates the illusion of no missing values by deleting missing samples, which can ensure the integrity of the remaining data, but it causes a decrease in the scale of samples and transforms the deleted samples from partial missing state to complete missing state. Therefore, it is not applicable to time series with continuous changing trend (Xu et al., 2020). The filling-based method fills the missing data by generating new values, and it can be further subdivided into statistical-based method and machine learning-based method. There are many commonly used statistics-based methods, such as mean imputation (Wolbers et al., 2022), last observation carried forward (Sampoornam et al., 2022), median imputation (Hadeed et al., 2020), plural imputation (Memon et al., 2022), random imputation (Guillaume & Wilfried, 2018), next observation carried backward (Wu et al., 2022), Lagrange imputation (Essanhaji & Errachid, 2022), and so on. Meanwhile, the main technologies used in the machine learning-based methods include clustering (Lashmar et al., 2021), linear regression (Vance et al., 2022), matrix decomposition (Feng et al., 2023), correlation analysis (Zhang et al., 2021), and multiple imputation (Aleryani et al., 2022). These methods mainly focus on the processing of missing values of non-time series.

Time series is a chronologically arranged sequence of numerical data points (Ren et al., 2021) and has seen extensive applications in various domains of our daily lives, especially in industrial scenarios. Since time series is generally generated by end-users, edge devices, and different wearable devices, it is more inevitable for time series to suffer missing values.

Complete Article List

Search this Journal:
Reset
Volume 17: 1 Issue (2024)
Volume 16: 3 Issues (2023)
Volume 15: 3 Issues (2022)
Volume 14: 2 Issues (2021)
Volume 13: 2 Issues (2020)
Volume 12: 2 Issues (2019)
Volume 11: 2 Issues (2018)
Volume 10: 2 Issues (2017)
Volume 9: 2 Issues (2016)
Volume 8: 2 Issues (2015)
Volume 7: 2 Issues (2014)
Volume 6: 2 Issues (2013)
Volume 5: 2 Issues (2012)
Volume 4: 2 Issues (2011)
Volume 3: 2 Issues (2010)
Volume 2: 2 Issues (2009)
Volume 1: 2 Issues (2008)
View Complete Journal Contents Listing