Access the full text.
Sign up today, get DeepDyve free for 14 days.
[ (2024)
Uncovering the hidden cost of model compressionProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition
[ (2021)
Brecq: Pushing the limit of post-training quantization by block reconstructionarXiv:2102.05426. Retrieved from https://arxiv.org/abs/2102.05426
[ (2022)
Investigating and mitigating effects of quantization on algorithmic biasLU-CS-EX
[ (2018)
Gender bias in coreference resolution: Evaluation and debiasing methodsarXiv:1804.06876. Retrieved from https://arxiv.org/abs/1804.06876
[ (2023)
Bias mimicking: A simple sampling approach for bias mitigationProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition
[ (2019)
Bert: Pre-training of deep bidirectional transformers for language understandingProceedings of naacL-HLT. Minneapolis
[ (2024)
DeepTensor: Low-rank tensor decomposition with deep network priorsIEEE Transactions on Pattern Analysis and Machine Intelligence, 46
[ (2020)
Up or down? Adaptive rounding for post-training quantizationProceedings of the International Conference on Machine Learning. PMLR
[ (2022)
The effect of model compression on fairness in facial expression recognitionProceedings of the International Conference on Pattern Recognition. Springer
[ (2023)
Deep learning model compression with rank reduction in tensor decompositionIEEE Transactions on Neural Networks and Learning Systems, 36
[ (2020)
FERMI: Fair empirical risk minimization via exponential rényi mutual informationarXiv:2304.03935. Retrieved from https://arxiv.org/abs/2304.03935
[ (2022)
FairNeuron: Improving deep neural network fairness with adversary games on selective neuronsProceedings of the 44th International Conference on Software Engineering
[ (2020)
Zeroq: A novel zero shot quantization frameworkProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition
[ (2020)
Post-training piecewise linear quantization for deep neural networksComputer Vision–ECCV 2020: 16th European Conference, 2020
[ (2021)
Fair attribute classification through latent space de-biasingProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition
[ (2024)
Rsmamba: Remote sensing image classification with state space modelIEEE Geoscience and Remote Sensing Letters, 21
[ (2019)
AI fairness 360: An extensible toolkit for detecting and mitigating algorithmic biasIBM Journal of Research and Development, 63
[ (2021)
A survey on bias and fairness in machine learningACM Computing Surveys, 54
[ (2017)
Data decisions and theoretical implications when adversarially learning fair representationsarXiv:1707.00075. Retrieved from https://arxiv.org/abs/1707.00075
[ (2019)
Data-free quantization through weight equalization and bias correctionProceedings of the IEEE/CVF International Conference on Computer Vision
[ (2023)
Last-layer fairness fine-tuning is simple and effective for neural networksarXiv:2304.03935. Retrieved from https://arxiv.org/abs/2304.03935
[ (2012)
Fairness through awarenessProceedings of the 3rd Innovations in Theoretical Computer Science Conference
[ (2015)
Simultaneous deep transfer across domains and tasksProceedings of the IEEE International Conference on Computer Vision
[ (2017)
To prune, or not to prune: Exploring the efficacy of pruning for model compressionarXiv:1710.01878. Retrieved from https://arxiv.org/abs/1710.01878
[ (2016)
Deep residual learning for image recognitionProceedings of the IEEE Conference on Computer Vision and Pattern Recognition
[ (2021)
Sparsity in deep learning: Pruning and growth for efficient inference and training in neural networksJournal of Machine Learning Research, 22
[ (2019)
Patient knowledge distillation for bert model compressionarXiv:1908.09355. Retrieved from https://arxiv.org/abs/1908.09355
[ (2020)
Fair generative modeling via weak supervisionProceedings of the International Conference on Machine Learning. PMLR
[ (2019)
Roberta: A robustly optimized bert pretraining approacharXiv:1907.11692. Retrieved from https://arxiv.org/abs/1907.11692
[ (2020)
Characterising bias in compressed modelsarXiv:2010.03058. Retrieved from https://arxiv.org/abs/2010.03058
[ (2024)
Multi-task learning in natural language processing: An overviewComputing Surveys, 56
[ (2024)
UniRepLKNet: A universal perception large-kernel convnet for audio video point cloud time-series and image recognitionProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition
[ (2024)
RUNNER: Responsible UNfair NEuron repair for enhancing deep neural network fairnessProceedings of the 46th IEEE/ACM International Conference on Software Engineering
[ (2024)
A review of modern recommender systems using generative models (gen-recsys)Proceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining
[ (2024)
VkD: Improving knowledge distillation using orthogonal projectionsProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition
[ (2024)
Recommender systems in the era of large language models (llms)IEEE Transactions on Knowledge and Data Engineering, 36
[ (2023)
To be robust and to be fair: Aligning fairness with robustnessarXiv:2304.00061. Retrieved from https://arxiv.org/abs/2304.00061
[ (2024)
Balancing act: Distribution-guided debiasing in diffusion modelsProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition
[ (2022)
Neuronfair: Interpretable white-box fairness testing through biased neuron identificationProceedings of the 44th International Conference on Software Engineering
[ (2024)
MixQuantBio: Towards extreme face and periocular recognition model compression with mixed-precision quantizationEngineering Applications of Artificial Intelligence, 137
[ (2023)
Compressed models decompress race biases: What quantized models forget for fair face recognitionProceedings of the 2023 International Conference of the Biometrics Special Interest Group. IEEE, 2023
[ (2018)
Mobilenetv2: Inverted residuals and linear bottlenecksProceedings of the IEEE Conference on Computer Vision and Pattern Recognition
[ (2024)
Medication recommendation system based on natural language processing for patient emotion analysisAcademic Journal of Science and Technology, 10
[ (2021)
Ethical adversaries: Towards mitigating unfairness with adversarial machine learningACM SIGKDD Explorations Newsletter, 23
[ (2022)
Quantface: Towards lightweight face recognition by synthetic data low-bit quantizationProceedings of the 2022 26th International Conference on Pattern Recognition. IEEE, 2022
[ (2021)
Fair feature distillation for visual recognitionProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition
[ (2023)
Variation-aware vision transformer quantizationarXiv:2307.00331. Retrieved from https://arxiv.org/abs/2307.00331
[ (2021)
Knowledge distillation: A surveyInternational Journal of Computer Vision, 129
[ (2015)
Deep learning face attributes in the wildProceedings of the IEEE International Conference on Computer Vision
[ (2024)
A survey on deep neural network pruning: Taxonomy, comparison, analysis, and recommendationsIEEE Transactions on Pattern Analysis and Machine Intelligence, 46
The increasing complexity of deep neural networks (DNNs) poses significant resource challenges for edge devices, prompting the development of compression technologies like model quantization. However, while improving model efficiency, quantization can introduce or perpetuate the original model’s bias. Existing debiasing methods for quantized models often incur additional costs. To address this issue, we propose FairQuanti, a novel quantization approach that leverages neuron role contribution to achieve fairness. By distinguishing between biased and normal neurons, FairQuanti employs mixed precision quantization to mitigate model bias during the quantization process. FairQuanti has four key differences from previous studies: (1) Neuron Roles - It formally defines biased and normal neuron roles, establishing a framework for feasible model quantization and bias mitigation; (2) Effectiveness - It introduces a fair quantization strategy that discriminatively quantizes neuron roles, balancing model accuracy and fairness through Bayesian optimization; (3) Generality - It applies to both structured and unstructured data across various quantization bit levels; (4) Robustness - It demonstrates resilience against adaptive attacks. Extensive experiments on five datasets (three structured and two unstructured) using five different models validate FairQuanti’s superior performance against eight baseline methods. Specifically, fairness metrics such as demographic parity (DP) improve by approximately 1.03 times, and the demographic parity ratio (DPR) improves by approximately 1.51 times compared to the baselines, with an average accuracy loss of less than 7.5% at 8-bit quantization. FairQuanti presents a promising solution for deploying fair and efficient deep models on resource-constrained devices and holds potential for application in large language models to reduce size and computational demands while minimizing bias. Our source code is available at https://github.com/Caozq2/FairQuanti.
ACM Transactions on Privacy and Security (TOPS) – Association for Computing Machinery
Published: Aug 23, 2025
Keywords: Deep neural network
Read and print from thousands of top scholarly journals.
Already have an account? Log in
Bookmark this article. You can see your Bookmarks on your DeepDyve Library.
To save an article, log in first, or sign up for a DeepDyve account if you don’t already have one.
Copy and paste the desired citation format or use the link below to download a file formatted for EndNote
Access the full text.
Sign up today, get DeepDyve free for 14 days.
All DeepDyve websites use cookies to improve your online experience. They were placed on your computer when you launched this website. You can change your cookie settings through your browser.