Simulating cyber attacks and countermeasures using Cyber Operations Research Gym (CybORG)

From ISLAB/CAISR
Jump to navigationJump to search
Title Simulating cyber attacks and countermeasures using Cyber Operations Research Gym (CybORG)
Summary Develop a simulation-based study of cyber attacks and countermeasures in CybORG, comparing baseline scripted defenses with reinforcement learning–based adaptive defenders to identify effective strategies for protecting networked systems.
Keywords
TimeFrame
References Quantitative Resilience Modeling for Autonomous Cyber Defense

https://rlj.cs.umass.edu/2025/papers/RLJ_RLC_2025_99.pdf

Interpreting Agent Behaviors in Reinforcement-Learning-Based Cyber-Battle Simulation Platforms https://arxiv.org/html/2506.08192v1

CybORG: https://github.com/cage-challenge/cyborg?utm_source=chatgpt.com

Prerequisites
Author
Supervisor Edison Pignaton de Freitas
Level Master
Status Open


Work in partnership with the Fraunhofer FKIE Institute in Bonn, Germany

Project Goal: Develop a simulation-based study of cyber attacks and countermeasures in CybORG, comparing baseline scripted defenses with reinforcement learning–based adaptive defenders to identify effective strategies for protecting networked systems.

Proposed Solution & Specific Tasks:

Set up CybORG environment: Install CybORG and familiarize with its baseline scenarios (e.g., red team exploiting vulnerable hosts, blue team defending services). Select representative scenarios (e.g., privilege escalation, lateral movement, persistence).

Model Attacks: Implement or tune red-team agents using scripted attack strategies (e.g., scanning, brute force, exploit chaining). Explore adaptive attackers using RL (e.g., Deep Q-Networks, PPO).

Design Defenses: Implement baseline defenses (patching, service monitoring, port blocking). Develop blue-team agents with reinforcement learning to dynamically choose defense actions (e.g., isolate host, restart service, deception).

Experimentation: Run controlled experiments where red-team agents attack and blue-team agents defend. Compare performance across different strategies (static rules vs. adaptive RL).

Analysis & Visualization: Collect metrics on attacker success rate, defender cost, time-to-compromise, and system availability. Visualize attack-defense dynamics over simulation episodes.

Evaluation Criteria Defensive Effectiveness: Reduction in attacker success rate (% of compromised hosts). Mean time to compromise (MTTC) improvements.

Efficiency: Resource overhead of defenses (CPU, memory, action cost in CybORG).

Adaptability: Ability of RL-based defenders to learn effective countermeasures against novel attack patterns.

Comparative Performance: Benchmark RL defenders vs. rule-based defenders. Benchmark scripted attackers vs. adaptive RL attackers.

Tools & Frameworks

Simulation Environment: CybORG* (Python-based cyber operations simulator).

Reinforcement Learning: Stable Baselines3 (PPO, DQN, A2C) or Ray RLlib for scalability.

Visualization & Analysis: Matplotlib, NetworkX (for network attack graphs), TensorBoard.

Expected Contributions A reproducible simulation framework for testing cyber attack/defense strategies in CybORG. Empirical insights on when adaptive defense outperforms static defense. Recommendations for deploying RL-based defenses in real-world cyber ranges.