Towards Adversarial Policy Discovery via Evolutionary Program Synthesis

Research output: Contribution to conferencePaperpeer-review

Abstract

Recent work has shown that even superhuman reinforcement learning (RL) policies can be vulnerable to adversarial agents. Most existing approaches for generating such adversaries rely on RL-based methods similar to those used to train the original policy under attack, potentially limiting the diversity of discovered exploits. We present a proof of concept showing that genetic programming (GP) can evolve symbolic adversarial agents that expose flaws in trained RL policies. By framing adversarial discovery as a program synthesis task, our approach enables broader and more interpretable search than conventional methods. We evaluate this approach in two competitive game environments against agents trained by OpenAI, showing that GP-evolved agents can outperform RL-based adversaries. These early results suggest that GP is not only effective for discovering unconventional exploits, but may serve as a useful stress-testing tool for RL systems more generally.
Original languageEnglish
Publication statusPublished - 5 Sept 2025
Event24th UK Workshop in Computational Intelligence 2025 - Edinburgh, United Kingdom
Duration: 3 Sept 20255 Sept 2025

Conference

Conference24th UK Workshop in Computational Intelligence 2025
Abbreviated titleUKCI 2025
Country/TerritoryUnited Kingdom
CityEdinburgh
Period3/09/255/09/25

Fingerprint

Dive into the research topics of 'Towards Adversarial Policy Discovery via Evolutionary Program Synthesis'. Together they form a unique fingerprint.

Cite this