There is no commentSelect some text and then click Comment, or simply add a comment to this page from below to start a discussion.
Practical Adversarial Attacks on Spatiotemporal Traffic Forecasting Models. NIPS 2022.
Abstract
Existing traffic forecasting models assume a reliable and unbiased forecasting environment, which is not always available.
Investigate the vulnerability of spatiotemporal traffic forecasting models and propose a practical adversarial spatiotemporal attack framework
Extensive experiments show that the proposed framework achieves up to 67.8% performance degradation on baselines.
Introduction
Injecting slight adversarial perturbations on a few randomly selected nodes can significantly degrade the traffic forecasting accuracy of the whole system.
Figure 1: An illustration of adversarial attack against spatiotemporal forecasting models on the Bay Area traffic network in California, the data ranges from January 2017 to May 2017. (a) Adversarial attack of geo-distributed data. The malicious attacker may inject adversarial examples into a few randomly selected geo-distributed data sources. (e.g., roadway sensors) to mislead the prediction of the whole traffic forecasting system. (b) Accuracy drop of victim nodes. By adding less than 50% traffic speed perturbations to 10% victim nodes, we observe 60.4% accuracy drop of victim nodes in morning peak hour. (c) Accuracy drop of neighbouring nodes. Due to the information diffusion of spatiotemporal forecasting models, the adversarial attack also leads to up to about 47.23% accuracy drop for neighboring nodes
Adversarial attacks have been extensively studied in various application domains. Two major challenges prevent applying existing adversarial attack strategies to spatiotemporal traffic forecasting.
Limitation
Expensive and impractical to manipulate all data sources (hundreds of sensors and thousands of GPS devices). Identify the subset of salient victim nodes with a limited attack budget to maximize the attack.
Most existing adversarial attack strategies that focus on time-invariant label classification adversarial attack against traffic forecasting aims to disrupt the target model to make biased predictions of continuous "values".
Solution
Proposing a practical adversarial spatiotemporal attack framework that can disrupt the forecasting models.
Devising an iterative gradient-guided method to estimate node saliency, which helps to identify a small time-dependent set of victim nodes.
Spatiotemporal gradient descent scheme is proposed to guide the attack direction and generate real-valued adversarial traffic states.
Various attack settings, i.e., white-box attack, grey-box attack, and black-box attack.
Experimental studies on two real-world traffic datasets show that attacking 10% nodes in the traffic system can break down the MAE from 1.975 to 6.132.
Incorporating adversarial examples we generated with adversarial training can significantly improve the robustness of spatiotemporal traffic forecasting models.
Background
Traffic forecasting
denotes a traffic network at time step , where is a set of nodes and is a set of edges. as the spatiotemporal features associated to , where represents the -dimensional time-varying traffic conditions of node at .
where denotes the input and the traffic network in previous time steps. is the spatiotemporal traffic forecasting model parameterized by . is the estimation and is the ground-truth of .
Adversarial attack
Adversarial attack aims to mislead the model to derive biased predictions by generating the optimal adversarial example
where $x$ is the adversarial example with maximum bound under norm to guarantee the perturbation is imperceptible to human, and is the ground truth of clean example .
For instance, the adversarial example in FGSM
where is the Signum function and is the cross-entropy loss
The adversarial attack can be categorized into three classes
White-box attack. The attacker can fully access the target model, including the model architecture, the model parameters, gradients, model outputs, the input traffic states, and the corresponding labels.
Grey-box attack. The attacker can partially access the system, including the target model and the input traffic states, but without the labels.
Black-box attack. The attacker can only access the input traffic states, query the outputs of the target model or leverage a surrogate model to craft the adversarial examples.
Adversarial attack against spatiotemporal traffic forecasting
Adversarial traffic state is defined as
where is a diagonal matrix with ith diagonal element indicating whether node is a victim node, and is the perturbed spatiotemporal feature named adversarial spatiotemporal feature.
Since round truth (i.e., future traffic states) is unavailable at run-time. Practical adversarial spatiotemporal attack primarily falls into the grey-box attack setting.
Methodology
Identify time-dependent victim nodes
One unique characteristic that distinguishes attacking spatiotemporal forecasting from conventional classification tasks is the inaccessibility of ground truth at the test phase.
Surrogate label to guide the attack direction
where is a generalized function (e.g., , , ) , are random variables sampled from a probability distribution to increase the diversity of the attack direction.
Function parameter based on the pre-trained forecasting model parameter , and
where is the estimation function parameterized by . For simplicity, is derived from the pre-trained traffic forecasting model .
With the surrogate traffic state label , the derivation the time-dependent node saliency (TDNS) for each node.
where is the loss function and is the activation function. reveals the node-wise loss impact with the same degree of perturbations. Note depending on the time step , may vary.
A similar idea also has been adopted to identify static pixel saliency for image classification
From Eq. (8), is updated by gradient-based adversarial method [Aleksander Madry, Aleksandar Makelov, Ludwig Schmidt, Dimitris Tsipras, and Adrian Vladu. Towards deep learning models resistant to adversarial attacks. In ICLR 2018].
where is adversarial traffic states at -th iteration, is the step size, and is the project operation which clips the spatiotemporal feature with maximum perturbation bound . Note .
For each batch of data , the time-dependent node saliency gradient is derived by
is the batch size. is the activation function to compute the non-negative saliency score for each time step.
The set of victim node based on ,
Attack with adversarial traffic state
Based on the time-dependent victim set, adversarial attacks to spatiotemporal traffic forecasting models is conducted which is Spatiotemporal Projected Gradient Descent (STPGD) .
where is the adversarial traffic state at (−1)-th iteration in the iterative gradient descent, is the step size, and is the operation to bound adversarial features in a ball. Note .
In the testing phase, adversarial traffic states is injected
where
and
White-box attack. Since the adversaries can fully access the data and labels under the whitebox setting, the real ground truth traffic states to guide the generation of adversarial traffic states are directly used.
Black-box attack. The most restrictive black-box setting assumes limited accessibility to the target model and labels. Therefore, a surrogate model is first built, which can be learned from the training data. Then, adversarial traffic states based on the surrogate model are generated to attack the targeted traffic forecasting model.
Experiments
Experimental setup
Datasets
PEMS-BAY
METR-LA
The first 70% for training, the following 10% for validation, and the rest 20% for testing.
Target model
GraphWaveNet
Evaluation metrics
Global and local effect
where is a user-defined loss function
Mean Average Error (MAE)
Root Mean Square Error (RMSE)
Overall attack performance
(67.79%, 62.31%) and (19.88%, 14.55%) global performance degradation compared with the original forecasting results on PeMS-BAY and METR-LA dataset, respectively.