Peek into the Black-Box: Interpretable Neural Network using SAT Equations in Side-Channel Analysis

Authors

  • Trevor Yap School of Physical and Mathematical Sciences, Nanyang Technological University, Singapore
  • Adrien Benamira School of Physical and Mathematical Sciences, Nanyang Technological University, Singapore
  • Shivam Bhasin School of Physical and Mathematical Sciences, Nanyang Technological University, Singapore
  • Thomas Peyrin School of Physical and Mathematical Sciences, Nanyang Technological University, Singapore

DOI:

https://doi.org/10.46586/tches.v2023.i2.24-53

Keywords:

Side-channel, Neural Network, Deep Learning, Profiling attack, Interpretability, SAT

Abstract

Deep neural networks (DNN) have become a significant threat to the security of cryptographic implementations with regards to side-channel analysis (SCA), as they automatically combine the leakages without any preprocessing needed, leading to a more efficient attack. However, these DNNs for SCA remain mostly black-box algorithms that are very difficult to interpret. Benamira et al. recently proposed an interpretable neural network called Truth Table Deep Convolutional Neural Network (TT-DCNN), which is both expressive and easier to interpret. In particular, a TT-DCNN has a transparent inner structure that can entirely be transformed into SAT equations after training. In this work, we analyze the SAT equations extracted from a TT-DCNN when applied in SCA context, eventually obtaining the rules and decisions that the neural networks learned when retrieving the secret key from the cryptographic primitive (i.e., exact formula). As a result, we can pinpoint the critical rules that the neural network uses to locate the exact Points of Interest (PoIs). We validate our approach first on simulated traces for higher-order masking. However, applying TT-DCNN on real traces is not straightforward. We propose a method to adapt TT-DCNN for application on real SCA traces containing thousands of sample points. Experimental validation is performed on software-based ASCADv1 and hardware-based AES_HD_ext datasets. In addition, TT-DCNN is shown to be able to learn the exact countermeasure in a best-case setting.

Downloads

Published

2023-03-06

How to Cite

Yap, T., Benamira, A., Bhasin, S., & Peyrin, T. (2023). Peek into the Black-Box: Interpretable Neural Network using SAT Equations in Side-Channel Analysis. IACR Transactions on Cryptographic Hardware and Embedded Systems, 2023(2), 24–53. https://doi.org/10.46586/tches.v2023.i2.24-53

Issue

Section

Articles