THESIS
2024
1 online resource (xiii, 113 pages) : color illustrations
Abstract
Multivariate time series is a common data format within complex systems. Its lack of patterns and difficulty in reconstruction make it challenging for representation learning. To address these challenges in real-world datasets, this thesis proposes a graph causal information flow discovery framework for anomaly detection in multivariate time series, utilizing graph neural networks (GNNs) with a self-training algorithm to uncover nonlinear interdependencies within complex systems.
First, we quantify the effect of pseudo labeling on the error bound and convergence properties of GNNs. We propose a cautious plug-in strategy that iteratively pseudo-labels the top-k confident samples. We demonstrate the effectiveness and robustness of this strategy, successfully overcoming degradation proble...[
Read more ]
Multivariate time series is a common data format within complex systems. Its lack of patterns and difficulty in reconstruction make it challenging for representation learning. To address these challenges in real-world datasets, this thesis proposes a graph causal information flow discovery framework for anomaly detection in multivariate time series, utilizing graph neural networks (GNNs) with a self-training algorithm to uncover nonlinear interdependencies within complex systems.
First, we quantify the effect of pseudo labeling on the error bound and convergence properties of GNNs. We propose a cautious plug-in strategy that iteratively pseudo-labels the top-k confident samples. We demonstrate the effectiveness and robustness of this strategy, successfully overcoming degradation problems.
Second, to avoid correlation-based relationship discovery, we introduce causal inference into GNNs to enhance their message-passing capabilities. We estimate cause-effect relationships through conditional entropy and iteratively modify the causal structure accordingly. Our experiments show the superiority of this method in link prediction, and the explicit causal structure also improves node classification across various baselines.
Third, we construct a robust anomaly detection framework for time series. We compare an attention-based encoder-decoder network with principal component analysis, constructing principal and residual latent spaces based on attention scores without reconstruction.
Our experiments demonstrate the effectiveness of monitoring both spaces for anomaly detection.
Finally, all the modules are integrated into a comprehensive framework featuring a graph attention network with iterative causal structure learning. We verify the framework on open-source datasets, demonstrating its superiority in anomaly detection for multivariate time series.
Post a Comment