Researchers have developed a hierarchical decision-making framework for unmanned aerial vehicle (UAV) search-and-rescue missions, leveraging a combination of rule-based high-level coaching and online goal-conditioned reinforcement learning (RL) control. This approach enables UAVs to adapt to complex scenarios with limited simulation training, stressing the importance of early adaptation in real-world deployments. The framework's design allows for a fixed, rule-based advisor to guide an RL controller, facilitating more effective decision-making in dynamic environments. By exploring the intersection of reinforcement learning and UAV operations, this work highlights the potential for state-aligned activity to shift the threat model from criminal to geopolitical, necessitating a revised approach to security and response1. This matters to practitioners because it underscores the need for a distinct playbook in addressing geopolitical threats that leverage reinforcement learning, particularly in high-stakes applications like search-and-rescue missions.