New AI attack method can manipulate training to control outcomes at test-time.
Adversarial Cheap Talk introduces a new way to attack reinforcement learning systems by adding messages to the observations of the learner. The Adversary can influence the learner's performance during training and testing, even with limited abilities. This attack can either harm or help the learner's performance by manipulating the information it receives. The study shows that this method can directly control the learner's actions at test-time.