Academic

Analyzing Physical Adversarial Example Threats to Machine Learning in Election Systems

arXiv:2603.00481v1 Announce Type: new Abstract: Developments in the machine learning voting domain have shown both promising results and risks. Trained models perform well on ballot classification tasks (> 99% accuracy) but are at risk from adversarial example attacks that cause misclassifications. In this paper, we analyze an attacker who seeks to deploy adversarial examples against machine learning ballot classifiers to compromise a U.S. election. We first derive a probabilistic framework for determining the number of adversarial example ballots that must be printed to flip an election, in terms of the probability of each candidate winning and the total number of ballots cast. Second, it is an open question as to which type of adversarial example is most effective when physically printed in the voting domain. We analyze six different types of adversarial example attacks: l_infinity-APGD, l2-APGD, l1-APGD, l0 PGD, l0 + l_infinity PGD, and l0 + sigma-map PGD. Our experiments include p

arXiv:2603.00481v1 Announce Type: new Abstract: Developments in the machine learning voting domain have shown both promising results and risks. Trained models perform well on ballot classification tasks (> 99% accuracy) but are at risk from adversarial example attacks that cause misclassifications. In this paper, we analyze an attacker who seeks to deploy adversarial examples against machine learning ballot classifiers to compromise a U.S. election. We first derive a probabilistic framework for determining the number of adversarial example ballots that must be printed to flip an election, in terms of the probability of each candidate winning and the total number of ballots cast. Second, it is an open question as to which type of adversarial example is most effective when physically printed in the voting domain. We analyze six different types of adversarial example attacks: l_infinity-APGD, l2-APGD, l1-APGD, l0 PGD, l0 + l_infinity PGD, and l0 + sigma-map PGD. Our experiments include physical realizations of 144,000 adversarial examples through printing and scanning with four different machine learning models. We empirically demonstrate an analysis gap between the physical and digital domains, wherein attacks most effective in the digital domain (l2 and l_infinity) differ from those most effective in the physical domain (l1 and l2, depending on the model). By unifying a probabilistic election framework with digital and physical adversarial example evaluations, we move beyond prior close race analyses to explicitly quantify when and how adversarial ballot manipulation could alter outcomes.

Executive Summary

This article analyzes the vulnerability of machine learning-based voting systems to physical adversarial example attacks. The authors develop a probabilistic framework to determine the number of adversarial example ballots required to flip an election and evaluate six types of adversarial example attacks. The results show a gap between the physical and digital domains, with different attacks being effective in each domain. The study highlights the potential risks of adversarial ballot manipulation in elections and provides a foundation for further research in this area.

Key Points

  • Machine learning-based voting systems are vulnerable to physical adversarial example attacks
  • A probabilistic framework is developed to determine the number of adversarial example ballots required to flip an election
  • The study identifies a gap between the physical and digital domains in terms of effective adversarial example attacks

Merits

Comprehensive evaluation of adversarial example attacks

The study evaluates six different types of adversarial example attacks, providing a thorough understanding of the vulnerabilities of machine learning-based voting systems.

Demerits

Limited scope of the study

The study focuses on a specific type of attack and does not consider other potential vulnerabilities of machine learning-based voting systems.

Expert Commentary

This study provides a crucial analysis of the vulnerabilities of machine learning-based voting systems to physical adversarial example attacks. The findings highlight the importance of considering the physical domain when evaluating the security of these systems, as the most effective attacks in the digital domain may not be the same as those in the physical domain. The study's probabilistic framework provides a valuable tool for determining the potential impact of adversarial example attacks on election outcomes. However, further research is needed to fully understand the risks and to develop effective countermeasures.

Recommendations

  • Develop and implement robust security measures to detect and prevent adversarial example attacks in machine learning-based voting systems
  • Conduct further research to fully understand the vulnerabilities of these systems and to develop effective countermeasures

Sources