
Explainable AI Achieves 83.5% Accuracy with Quantized Active Ingredients and Boltzmann Machines
Why It Matters
Higher accuracy combined with clearer feature attribution positions QBMs as a viable path toward transparent AI in high‑stakes sectors such as healthcare and finance.
Explainable AI Achieves 83.5% Accuracy with Quantized Active Ingredients and Boltzmann Machines
A Novel Approach to Explainable AI with Quantized Active Ingredients in Decision Making
Authors: A. M. A. S. D. Alagiyawanna, Asoka Karunananda, Thushari Silva, and A. Mahasinghe (University of Moratuwa, Sri Lanka)
The challenge of explaining how Artificial Intelligence systems reach decisions remains a significant hurdle, particularly within critical sectors like healthcare and finance where transparency is essential. The authors present a novel framework for explainable AI, comparing Quantized Boltzmann Machines (QBMs) with Classical Boltzmann Machines (CBMs). Their research leverages computing principles to enhance transparency in decision‑making processes, utilizing a binarised MNIST dataset and techniques such as gradient‑based saliency maps and SHAP values to evaluate feature importance. Results demonstrate that QBMs not only achieve superior classification accuracy—reaching 83.5 % compared to CBMs’ 54 %—but also provide more focused feature attributions, suggesting a clearer identification of the key factors driving predictions. This work highlights the potential of hybrid quantum‑classical models to deliver both improved performance and greater trustworthiness in AI systems.
The study involved training both QBMs and CBMs on a binarised and dimensionally reduced version of the MNIST dataset, preprocessed using Principal Component Analysis (PCA). To assess interpretability, the team employed gradient‑based saliency maps for the QBMs and SHAP (SHapley Additive exPlanations) for the CBMs. QBMs were constructed using hybrid quantum‑classical circuits incorporating strongly entangling layers, designed to create richer latent representations, whereas CBMs functioned as a classical baseline utilizing contrastive divergence. This innovative approach allowed a direct comparison of the explainability and performance characteristics of both models.
Experiments revealed that QBMs significantly outperformed CBMs in classification accuracy, achieving a score of 83.5 % versus 54 % for the classical models. Furthermore, the study demonstrated that QBMs exhibited more concentrated distributions in feature attributions, quantified by an entropy value of 1.27 versus 1.39 for CBMs. This indicates that QBMs not only provide superior predictive performance but also offer a clearer identification of the most influential features driving model predictions, effectively pinpointing the key factors behind their decisions. The work establishes that quantum‑classical hybrid models can simultaneously improve both accuracy and interpretability, paving the way for more trustworthy and explainable AI systems. By harnessing quantum parallelism, superposition, and entanglement, QBMs represent richer probability distributions with potentially fewer parameters than their classical counterparts.
To prepare the data, the team focused on binary classification using the MNIST dataset, initially restricting analysis to the digits zero and one due to their distinct visual characteristics. Each 28 × 28‑pixel image underwent normalisation, and PCA reduced the 784‑dimensional input to just four principal components. This dimensionality reduction facilitated efficient quantum computation and streamlined the subsequent analysis. The reduced‑dimension data was calculated using the formula
[
Z = U_k , X,
]
where (Z) represents the reduced data, (X) is the original data, and (U_k) is a matrix containing the first (k) principal components.
A key innovation lay in the quantum hidden‑layer design, implemented using the PennyLane software framework. Classical input features were encoded onto qubits via angle embedding, utilizing rotation around the Y‑axis (RY gate) to map the input vector into a valid quantum state. This quantum encoding allowed the circuit to explore complex feature interactions through superposition and entanglement, potentially capturing subtle correlations missed by classical models. The quantum state evolution was mathematically defined as
[
|\psi(x, \theta)\rangle = V(\theta), U_{\text{embed}}(x),|0\rangle^{\otimes 4},
]
where (V(\theta)) represents the parameterised quantum circuit and (U_{\text{embed}}(x)) is the embedding unitary. Feature extraction from the quantum layer involved measuring the expectation values of the Pauli‑Z operator on each qubit, resulting in a vector (z) representing the quantum‑encoded input data.
The team quantified feature‑attribution distributions using entropy, finding a value of 1.27 for QBMs versus 1.39 for CBMs, indicating more concentrated and therefore more easily interpretable distributions. This concentration suggests that QBMs effectively identify the most influential features, analogous to pinpointing the “active ingredient” in a complex system. Gradient‑based saliency maps were used to evaluate feature attributions in QBMs, while SHAP was employed for CBMs, allowing scientists to assess how each input feature contributes to the model’s output.
Results demonstrate that the hybrid quantum‑classical architecture of QBMs, incorporating strongly entangling layers, enables richer latent representations and facilitates more transparent decision‑making processes. The work illustrates that QBMs are capable of more effectively identifying key features influencing predictions compared to CBMs, offering a pathway toward more trustworthy and explainable AI systems.
ArXiv: https://arxiv.org/abs/2601.08733
Comments
Want to join the conversation?
Loading comments...