25
Sep
This is a Plain English Papers summary of a research paper called New defenses still fall short against adversarial attacks on Go AIs. If you like these kinds of analysis, you should join AImodels.fyi or follow me on Twitter. Overview Previous research has shown that superhuman Go AI systems like KataGo can be defeated by simple adversarial strategies. This paper examines whether simple defenses can improve KataGo's performance against the worst-case scenarios. The paper tests three natural defenses: adversarial training on hand-constructed positions, iterated adversarial training, and changing the network architecture. Plain English Explanation The researchers wanted to see if…