[Taken] Robustness of neural networks
The sudden rise of adversarial examples (i.e., input points intentionally crafted so to trick a model into misprediction) has shown that even state-of-the-art deep learning models can be extremely vulnerable to intelligent attacks. Unfortunately, the fragility of such models makes their deployment in safety-critical real-world applications (e.g., self-driving cars or eHealth) difficult to justify, hence … Read more