Interpreters for GNN-Based Vulnerability Detection: Are We There Yet?
Traditional vulnerability detection methods have limitations due to their need for extensive manual labor.
Using automated means for vulnerability detection has attracted research interest, especially deep learning, which has achieved remarkable results.
Since graphs can better convey the structural feature of code than text, \emph{graph neural network} (GNN) based vulnerability detection is significantly better than text-based approaches.
Therefore, GNN-based vulnerability detection approaches are becoming popular.
However, GNN models are close to black boxes for security analysts, so the models cannot provide clear evidence to explain why a code sample is detected as vulnerable or secure.
At this stage, many GNN interpreters have been proposed.
However, the explanations provided by these interpretations for vulnerability detection models are highly inconsistent and unconvincing to security experts.
To address the above issues, we propose principled guidelines to assess the quality of the interpretation approaches for GNN-based vulnerability detectors based on concerns in vulnerability detection, namely, stability, robustness, and effectiveness.
We conduct extensive experiments to evaluate the interpretation performance of six famous interpreters (\ie \emph{GNN-LRP}, \emph{DeepLIFT}, \emph{GradCAM}, \emph{GNNExplainer}, \emph{PGExplainer}, and \emph{SubGraphX}) on four vulnerability detectors (\ie \emph{DeepWukong}, \emph{Devign}, \emph{IVDetect}, and \emph{Reveal}).
The experimental results show that the target interpreters achieve poor performance in terms of effectiveness, stability, and robustness.
For effectiveness, we find that the instance-independent methods outperform others due to their deep insight into the detection model.
In terms of stability, the perturbation-based interpretation methods are more resilient to slight changes in model parameters as they are model-agnostic.
For robustness, the instance-independent approaches provide more consistent interpretation results for similar vulnerabilities.
Tue 18 JulDisplayed time zone: Pacific Time (US & Canada) change
15:30 - 17:00 | ISSTA Online 1: SE and Deep LearningTechnical Papers at Smith Classroom (Gates G10) Chair(s): Myra Cohen Iowa State University | ||
15:30 10mTalk | COME: Commit Message Generation with Modification Embedding Technical Papers Yichen He Beihang University, Liran Wang Beihang University, Kaiyi Wang Beihang University, Yupeng Zhang Beihang University, Hang Zhang Beihang University, Zhoujun Li Beihang University DOI | ||
15:40 10mTalk | CODEP: Grammatical Seq2Seq Model for General-Purpose Code Generation Technical Papers DOI Pre-print | ||
15:50 10mTalk | Towards More Realistic Evaluation for Neural Test Oracle Generation Technical Papers DOI Pre-print | ||
16:00 10mTalk | Detecting Condition-Related Bugs with Control Flow Graph Neural Network Technical Papers Jian Zhang Beihang University, Xu Wang Beihang University, Hongyu Zhang Chongqing University, Hailong Sun Beihang University, Xudong Liu Beihang University, Chunming Hu Beihang University, Yang Liu Nanyang Technological University DOI | ||
16:10 10mTalk | RefBERT: A Two-Stage Pre-trained Framework for Automatic Rename Refactoring Technical Papers Hao Liu Xiamen University, Yanlin Wang Sun Yat-sen University, Zhao Wei Tencent, Yong Xu Tencent, Juhong Wang Tencent, Hui Li Xiamen University, Rongrong Ji Xiamen University DOI Pre-print | ||
16:20 10mTalk | Interpreters for GNN-Based Vulnerability Detection: Are We There Yet? Technical Papers Yutao Hu Huazhong University of Science and Technology, Suyuan Wang Huazhong University of Science and Technology, Wenke Li Huazhong University of Science and Technology, Junru Peng Wuhan University, Yueming Wu Nanyang Technological University, Deqing Zou Huazhong University of Science and Technology, Hai Jin Huazhong University of Science and Technology DOI | ||
16:30 10mTalk | Towards Efficient Fine-Tuning of Pre-trained Code Models: An Experimental Study and Beyond Technical Papers Ensheng Shi Xi’an Jiaotong University, Yanlin Wang Sun Yat-sen University, Hongyu Zhang Chongqing University, Lun Du Microsoft Research, Shi Han Microsoft Research, Dongmei Zhang Microsoft Research, Hongbin Sun Xi’an Jiaotong University DOI |