Abstract
Neural networks (NNs) play a crucial role in safety-critical fields, requiring robustness assurance. Bayesian Neural Networks
         (BNNs) address data uncertainty, providing probabilistic outputs. However, the literature on BNN robustness assessment is
         still limited, mainly focusing on adversarial examples, which are often impractical in real-world applications. This paper
         introduces a fresh perspective on BNN classifier robustness, considering natural input variations while accounting for prediction
         uncertainties. Our approach excludes predictions labeled as “unknown�, enabling practitioners to define alteration probabilities,
         penalize errors beyond a specified threshold, and tolerate varying error levels below it. We present a systematic approach
         for evaluating the robustness of BNNs, introducing new evaluation metrics that account for prediction uncertainty. We conduct
         a comparative study using two NNs – standard MLP and Bayesian MLP – on the MNIST dataset. Our results show that by leveraging
         estimated uncertainty, it is possible to enhance the system’s robustness
      
[download the pdf file] [DOI]