Deep reinforcement learning in large discrete action spaces G Dulac-Arnold, R Evans, H van Hasselt, P Sunehag, T Lillicrap, J Hunt, ... arXiv preprint arXiv:1512.07679, 2015 | 521 | 2015 |
A Dual Approach to Scalable Verification of Deep Networks. K Dvijotham, R Stanforth, S Gowal, TA Mann, P Kohli UAI 1 (2), 3, 2018 | 377 | 2018 |
On the effectiveness of interval bound propagation for training verifiably robust models S Gowal, K Dvijotham, R Stanforth, R Bunel, C Qin, J Uesato, ... arXiv preprint arXiv:1810.12715, 2018 | 346 | 2018 |
Uncovering the limits of adversarial training against norm-bounded adversarial examples S Gowal, C Qin, J Uesato, T Mann, P Kohli arXiv preprint arXiv:2010.03593, 2020 | 188 | 2020 |
Fixing data augmentation to improve adversarial robustness SA Rebuffi, S Gowal, DA Calian, F Stimberg, O Wiles, T Mann arXiv preprint arXiv:2103.01946, 2021 | 124 | 2021 |
Scalable verified training for provably robust image classification S Gowal, KD Dvijotham, R Stanforth, R Bunel, C Qin, J Uesato, ... Proceedings of the IEEE/CVF International Conference on Computer Vision …, 2019 | 103 | 2019 |
Improving robustness using generated data S Gowal, SA Rebuffi, O Wiles, F Stimberg, DA Calian, TA Mann Advances in Neural Information Processing Systems 34, 4218-4233, 2021 | 94 | 2021 |
Robust reinforcement learning for continuous control with model misspecification DJ Mankowitz, N Levine, R Jeong, Y Shi, J Kay, A Abdolmaleki, ... arXiv preprint arXiv:1906.07516, 2019 | 78 | 2019 |
Data augmentation can improve robustness SA Rebuffi, S Gowal, DA Calian, F Stimberg, O Wiles, TA Mann Advances in Neural Information Processing Systems 34, 29935-29948, 2021 | 77 | 2021 |
Adaptive skills adaptive partitions (ASAP) DJ Mankowitz, TA Mann, S Mannor Advances in neural information processing systems 29, 2016 | 59 | 2016 |
An alternative surrogate loss for pgd-based adversarial testing S Gowal, J Uesato, C Qin, PS Huang, T Mann, P Kohli arXiv preprint arXiv:1910.09338, 2019 | 55 | 2019 |
Scaling up approximate value iteration with options: Better policies with fewer iterations T Mann, S Mannor International conference on machine learning, 127-135, 2014 | 55 | 2014 |
Approximate Value Iteration with Temporally Extended Actions SMDP Timothy A. Mann Journal of Artificial Intelligence Research 53, 375-438, 2015 | 54 | 2015 |
How hard is my MDP?" The distribution-norm to the rescue" OA Maillard, TA Mann, S Mannor Advances in Neural Information Processing Systems 27, 2014 | 54 | 2014 |
Achieving robustness in the wild via adversarial mixing with disentangled representations S Gowal, C Qin, PS Huang, T Cemgil, K Dvijotham, T Mann, P Kohli Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern …, 2020 | 44 | 2020 |
Soft-robust actor-critic policy-gradient E Derman, DJ Mankowitz, TA Mann, S Mannor arXiv preprint arXiv:1803.04848, 2018 | 42 | 2018 |
Learning robust options D Mankowitz, T Mann, PL Bacon, D Precup, S Mannor Proceedings of the AAAI Conference on Artificial Intelligence 32 (1), 2018 | 40 | 2018 |
Time regularized interrupting options DJ Mankowitz, TA Mann, S Mannor Internation Conference on Machine Learning, 2014 | 39* | 2014 |
A bayesian approach to robust reinforcement learning E Derman, D Mankowitz, T Mann, S Mannor Uncertainty in Artificial Intelligence, 648-658, 2020 | 33 | 2020 |
Off-policy model-based learning under unknown factored dynamics A Hallak, F Schnitzler, T Mann, S Mannor International Conference on Machine Learning, 711-719, 2015 | 31 | 2015 |