Adversarial learning, where the noise is not purely random, but chosen to be the worst possible noise for you (subject to some rules of the game). This is in contrast to classic machine learning and statistics where the noise is purely random; Tyche is not “out to get you”.
As renewed in fame recently by the related (?) method of generative adversarial networks (although much older.)
The associated concept in normal human experience is Goodhardt’s law, which tells us that “people game the targets you set for them.”
🏗 discuss politics implied by treating the learning as a battle with a conniving adversary as opposed to an uncaringly random universe. I’m sure someone has done this well in a terribly eloquent blog post, but I haven’t found one I’d want to link to yet.
The toolset of adversarial techniques is broad. Game theory is an important one, but also computational complexity theory (how hard is to find adversarial inputs, or to learn despite them?) and lots of functional analysis and optimisation theory. Surely much other stuff I do not know because this is not really my field.
Applications are broad too — improving ML but also infosec, risk management etc.
References
Abernethy, Agarwal, Bartlett, et al. 2009.
“A Stochastic View of Optimal Regret Through Minimax Duality.” arXiv:0903.5328 [Cs, Stat].
Abernethy, Bartlett, and Hazan. 2011. “Blackwell Approachability and No-Regret Learning Are Equivalent.” In.
Arjovsky, Chintala, and Bottou. 2017.
“Wasserstein Generative Adversarial Networks.” In
International Conference on Machine Learning.
Bora, Jalal, Price, et al. 2017.
“Compressed Sensing Using Generative Models.” In
International Conference on Machine Learning.
Gebhart, Schrater, and Hylton. 2019.
“Characterizing the Shape of Activation Space in Deep Neural Networks.” arXiv:1901.09496 [Cs, Stat].
Ghosh, Kulharia, Namboodiri, et al. 2017.
“Multi-Agent Diverse Generative Adversarial Networks.” arXiv:1704.02906 [Cs, Stat].
Goodfellow, Ian, Pouget-Abadie, Mirza, et al. 2014.
“Generative Adversarial Nets.” In
Advances in Neural Information Processing Systems 27. NIPS’14.
Goodfellow, Ian J., Shlens, and Szegedy. 2014.
“Explaining and Harnessing Adversarial Examples.” arXiv:1412.6572 [Cs, Stat].
Grünwald, Peter D. 2023.
“The e-Posterior.” Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences.
Guo, Hong, Lin, et al. 2017.
“Relaxed Wasserstein with Applications to GANs.” arXiv:1705.07164 [Cs, Stat].
Ilyas, Engstrom, Santurkar, et al. 2019. “Adversarial Examples Are Not Bugs, They Are Features.” In Advances In Neural Information Processing Systems.
Jetchev, Bergmann, and Vollgraf. 2016.
“Texture Synthesis with Spatial Generative Adversarial Networks.” In
Advances in Neural Information Processing Systems 29.
Larsen, Sønderby, Larochelle, et al. 2015.
“Autoencoding Beyond Pixels Using a Learned Similarity Metric.” arXiv:1512.09300 [Cs, Stat].
Linial. 1994.
“Chapter 38 Game-Theoretic Aspects of Computing.” In
Handbook of Game Theory with Economic Applications.
Ohsawa. 2021.
“Unbiased Self-Play.” arXiv:2106.03007 [Cs, Econ, Stat].
Poole, Alemi, Sohl-Dickstein, et al. 2016.
“Improved Generator Objectives for GANs.” In
Advances in Neural Information Processing Systems 29.
Raghunathan, Cherian, and Jha. 2019.
“Game Theoretic Optimization via Gradient-Based Nikaido-Isoda Function.” arXiv:1905.05927 [Cs, Math, Stat].
Ramdas, Grünwald, Vovk, et al. 2023.
“Game-Theoretic Statistics and Safe Anytime-Valid Inference.” Statistical Science.
Sato, Akiyama, and Farmer. 2002.
“Chaos in Learning a Simple Two-Person Game.” Proceedings of the National Academy of Sciences.
Vervoort. 1996.
“Blackwell Games.” In
Statistics, Probability and Game Theory: Papers in Honor of David Blackwell.