CITED: A Decision Boundary-Aware Signature for GNNs Towards Model Extraction Defense
arXiv:2602.20418v1 Announce Type: new Abstract: Graph neural networks (GNNs) have demonstrated superior performance in various applications, such as recommendation systems and financial risk management. However, deploying large-scale GNN models locally is particularly challenging for users, as it requires significant computational resources and extensive property data. Consequently, Machine Learning as a Service (MLaaS) has become increasingly popular, offering a convenient way to deploy and access various models, including GNNs. However, an emerging threat known as Model Extraction Attacks (MEAs) presents significant risks, as adversaries can readily obtain surrogate GNN models exhibiting similar functionality. Specifically, attackers repeatedly query the target model using subgraph inputs to collect corresponding responses. These input-output pairs are subsequently utilized to train their own surrogate models at minimal cost. Many techniques have been proposed to defend against MEAs
arXiv:2602.20418v1 Announce Type: new Abstract: Graph neural networks (GNNs) have demonstrated superior performance in various applications, such as recommendation systems and financial risk management. However, deploying large-scale GNN models locally is particularly challenging for users, as it requires significant computational resources and extensive property data. Consequently, Machine Learning as a Service (MLaaS) has become increasingly popular, offering a convenient way to deploy and access various models, including GNNs. However, an emerging threat known as Model Extraction Attacks (MEAs) presents significant risks, as adversaries can readily obtain surrogate GNN models exhibiting similar functionality. Specifically, attackers repeatedly query the target model using subgraph inputs to collect corresponding responses. These input-output pairs are subsequently utilized to train their own surrogate models at minimal cost. Many techniques have been proposed to defend against MEAs, but most are limited to specific output levels (e.g., embedding or label) and suffer from inherent technical drawbacks. To address these limitations, we propose a novel ownership verification framework CITED which is a first-of-its-kind method to achieve ownership verification on both embedding and label levels. Moreover, CITED is a novel signature-based method that neither harms downstream performance nor introduces auxiliary models that reduce efficiency, while still outperforming all watermarking and fingerprinting approaches. Extensive experiments demonstrate the effectiveness and robustness of our CITED framework. Code is available at: https://github.com/LabRAI/CITED.
Executive Summary
This article proposes a novel ownership verification framework called CITED to defend against Model Extraction Attacks (MEAs) in Graph Neural Networks (GNNs). CITED is a signature-based method that achieves ownership verification on both embedding and label levels without harming downstream performance or introducing auxiliary models. The framework outperforms existing watermarking and fingerprinting approaches. Extensive experiments demonstrate its effectiveness and robustness. The authors provide a GitHub repository for the code. This framework addresses a critical concern in the deployment of GNN models in Machine Learning as a Service (MLaaS) and has significant implications for the security and integrity of these models.
Key Points
- ▸ Proposes a novel ownership verification framework CITED for GNNs to defend against MEAs
- ▸ Achieves ownership verification on both embedding and label levels without performance degradation
- ▸ Outperforms existing watermarking and fingerprinting approaches
Merits
Strength in Defense
CITED provides a robust defense against MEAs by verifying ownership on multiple levels, making it a strong countermeasure to these attacks.
Efficiency
CITED does not introduce auxiliary models or harm downstream performance, preserving the efficiency of GNN models.
Robustness
Extensive experiments demonstrate the effectiveness and robustness of CITED against various MEA settings.
Demerits
Limited Evaluation Scope
The evaluation of CITED focuses on specific MEA scenarios, and its performance in other scenarios is unclear.
Potential Overhead
The signature-based approach may incur additional computational overhead, which could impact the performance of GNN models in resource-constrained environments.
Expert Commentary
The CITED framework is a significant contribution to the field of GNN security, addressing a critical concern in the deployment of these models in MLaaS settings. The authors' approach to ownership verification on multiple levels provides a robust defense against MEAs, while preserving the efficiency of GNN models. However, the potential overhead of the signature-based approach and the limited evaluation scope of the paper require further investigation. The implications of CITED are far-reaching, with potential implications for the security and integrity of GNN models in MLaaS settings and potential regulatory requirements.
Recommendations
- ✓ Further investigation into the potential overhead of CITED in resource-constrained environments
- ✓ Expansion of the evaluation scope to include other MEA scenarios and GNN architectures