Please use this identifier to cite or link to this item:
http://hdl.handle.net/1942/48260Full metadata record
| DC Field | Value | Language |
|---|---|---|
| dc.contributor.author | LI, Bowen | - |
| dc.contributor.author | Huang, Kai | - |
| dc.contributor.author | CHEN, Siang | - |
| dc.contributor.author | Xiong, Dongliang | - |
| dc.contributor.author | Jiang, Haitian | - |
| dc.contributor.author | CLAESEN, Luc | - |
| dc.date.accessioned | 2026-01-27T13:48:21Z | - |
| dc.date.available | 2026-01-27T13:48:21Z | - |
| dc.date.issued | 2020 | - |
| dc.date.submitted | 2026-01-15T12:52:56Z | - |
| dc.identifier.citation | Jialin Pan, Sinno; Sugiyama, Masashi (Ed.). Proceedings of Machine Learning Research, p. 289 -304 | - |
| dc.identifier.issn | 2640-3498 | - |
| dc.identifier.uri | http://hdl.handle.net/1942/48260 | - |
| dc.description.abstract | Data free deep neural network quantization is a practical challenge, since the original training data is often unavailable due to some privacy, proprietary or transmission issues. The existing methods implicitly equate data-free with training-free and quantize model manually through analyzing the weights’ distribution. It leads to a significant accuracy drop in lower than 6-bit quantization. In this work, we propose the data free quantization-aware fine-tuning (DFQF), wherein no real training data is required, and the quantized network is fine-tuned with generated images. Specifically, we start with training a generator from the pre-trained full-precision network with inception score loss, batch-normalization statistics loss and adversarial loss to synthesize a fake image set. Then we fine-tune the quantized student network with the full-precision teacher network and the generated images by utilizing knowledge distillation (KD). The proposed DFQF outperforms state-of-the-art post-train quantization methods, and achieve W4A4 quantization of ResNet20 on the CIFAR10 dataset within 1% accuracy drop. | - |
| dc.language.iso | en | - |
| dc.relation.ispartofseries | Proceedings of Machine Learning Research | - |
| dc.rights | authors | - |
| dc.subject.other | Data-free | - |
| dc.subject.other | Quantization | - |
| dc.subject.other | Adversarial | - |
| dc.title | DFQF: Data Free Quantization-aware Fine-tuning | - |
| dc.type | Proceedings Paper | - |
| local.bibliographicCitation.authors | Jialin Pan, Sinno | - |
| local.bibliographicCitation.authors | Sugiyama, Masashi | - |
| local.bibliographicCitation.conferencedate | 2020, November 18-20 | - |
| local.bibliographicCitation.conferencename | ACML 2020 | - |
| local.bibliographicCitation.conferenceplace | Bangkok, Thailand | - |
| dc.identifier.epage | 304 | - |
| dc.identifier.spage | 289 | - |
| local.format.pages | 16 | - |
| local.bibliographicCitation.jcat | C1 | - |
| dc.relation.references | Ron Banner, Yury Nahshan, and Daniel Soudry. Post training 4-bit quantization of convolutional networks for rapid-deployment. In NeurIPS 2019, 8-14 December 2019, Vancouver, BC, Canada, pages 7948–7956, 2019. Shane T. Barratt and Rishi Sharma. A note on the inception score. CoRR, abs/1801.01973, 2018. Yoshua Bengio, Nicholas L´eonard, and Aaron C. Courville. Estimating or propagating gradients through stochastic neurons for conditional computation. CoRR, abs/1308.3432, 2013. Cristian Bucila, Rich Caruana, and Alexandru Niculescu-Mizil. Model compression. In Proceedings of the Twelfth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Philadelphia, PA, USA, August 20-23, 2006, pages 535–541, 2006. Yaohui Cai, Zhewei Yao, Zhen Dong, Amir Gholami, Michael W. Mahoney, and Kurt Keutzer. Zeroq: A novel zero shot quantization framework. CoRR, abs/2001.00281, 2020. Hanting Chen, Yunhe Wang, Chang Xu, Zhaohui Yang, Chuanjian Liu, Boxin Shi, Chunjing Xu, Chao Xu, and Qi Tian. Data-free learning of student networks. CoRR, abs/1904.01186, 2019. Jungwook Choi, Zhuo Wang, Swagath Venkataramani, Pierce I-Jen Chuang, Vijayalakshmi Srinivasan, and Kailash Gopalakrishnan. PACT: parameterized clipping activation for quantized neural networks. CoRR, abs/1805.06085, 2018. Emily L. Denton, Wojciech Zaremba, Joan Bruna, Yann LeCun, and Rob Fergus. Exploiting linear structure within convolutional networks for efficient evaluation. In NeurIPS 2014, December 8-13 2014, Montreal, Quebec, Canada, pages 1269–1277, 2014. DC Dowson and BV Landau. The fr´echet distance between multivariate normal distributions. Journal of multivariate analysis, 12(3):450–455, 1982. Gongfan Fang, Jie Song, Chengchao Shen, Xinchao Wang, Da Chen, and Mingli Song. Data-free adversarial distillation. CoRR, abs/1912.11006, 2019. Ian J. Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron C. Courville, and Yoshua Bengio. Generative adversarial nets. In NeurIPS 2014, December 8-13 2014, Montreal, Quebec, Canada, pages 2672–2680, 2014. Song Han, Huizi Mao, and William J. Dally. Deep compression: Compressing deep neural network with pruning, trained quantization and huffman coding. In ICLR 2016, San Juan, Puerto Rico, May 2-4, 2016, Conference Track Proceedings, 2016. Matan Haroush, Itay Hubara, Elad Hoffer, and Daniel Soudry. The knowledge within: Methods for data-free model compression. CoRR, abs/1912.01274, 2019. Geoffrey E. Hinton, Oriol Vinyals, and Jeffrey Dean. Distilling the knowledge in a neural network. CoRR, abs/1503.02531, 2015. Itay Hubara, Matthieu Courbariaux, Daniel Soudry, Ran El-Yaniv, and Yoshua Bengio. Binarized neural networks. In NeurIPS 2016, December 5-10, 2016, Barcelona, Spain, pages 4107–4115, 2016. Benoit Jacob, Skirmantas Kligys, Bo Chen, Menglong Zhu, Matthew Tang, Andrew G. Howard, Hartwig Adam, and Dmitry Kalenichenko. Quantization and training of neural networks for efficient integer-arithmetic-only inference. In CVPR 2018, Salt Lake City, UT, USA, June 18-22, 2018, pages 2704–2713, 2018. doi: 10.1109/CVPR.2018.00286. Yoon Kim. Convolutional neural networks for sentence classification. In EMNLP 2014, October 25-29, 2014, Doha, Qatar, A meeting of SIGDAT, a Special Interest Group of the ACL, pages 1746–1751, 2014. Raghuraman Krishnamoorthi. Quantizing deep convolutional networks for efficient inference: A whitepaper. CoRR, abs/1806.08342, 2018. Alex Krizhevsky. Learning multiple layers of features from tiny images. Technical report, 2009. Alex Krizhevsky, Ilya Sutskever, and Geoffrey E. Hinton. Imagenet classification with deep convolutional neural networks. In NeruPIS 2012. Proceedings of a meeting held December 3-6, 2012, Lake Tahoe, Nevada, United States, pages 1106–1114, 2012. Raphael Gontijo Lopes, Stefano Fenu, and Thad Starner. Data-free knowledge distillation for deep neural networks. CoRR, abs/1710.07535, 2017. Szymon Migacz. 8-bit inference with tensorrt. In NVIDIA GPU Technology Conference, 2017. Asit K. Mishra and Debbie Marr. Apprentice: Using knowledge distillation techniques to improve low-precision network accuracy. In ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, Conference Track Proceedings, 2018. Markus Nagel, Mart van Baalen, Tijmen Blankevoort, and Max Welling. Data-free quantization through weight equalization and bias correction. CoRR, abs/1906.04721, 2019. Antonio Polino, Razvan Pascanu, and Dan Alistarh. Model compression via distillation and quantization. In ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, Conference Track Proceedings, 2018. Alec Radford, Luke Metz, and Soumith Chintala. Unsupervised representation learning with deep convolutional generative adversarial networks. In ICLR 2016, San Juan, Puerto Rico, May 2-4, 2016, Conference Track Proceedings, 2016. Mohammad Rastegari, Vicente Ordonez, Joseph Redmon, and Ali Farhadi. Xnor-net: Imagenet classification using binary convolutional neural networks. In Computer Vision - ECCV 2016 - 14th European Conference, Amsterdam, The Netherlands, October 11-14, 2016, Proceedings, Part IV, pages 525–542, 2016. doi: 10.1007/978-3-319-46493-0\ 32. Tim Salimans, Ian J. Goodfellow, Wojciech Zaremba, Vicki Cheung, Alec Radford, and Xi Chen. Improved techniques for training gans. In NeurIPS 2016, December 5-10, 2016, Barcelona, Spain, pages 2226–2234, 2016. Wonyong Sung, Sungho Shin, and Kyuyeon Hwang. Resiliency of deep neural networks under quantization. CoRR, abs/1511.06488, 2015. Hongxu Yin, Pavlo Molchanov, Zhizhong Li, Jose M. Alvarez, Arun Mallya, Derek Hoiem, Niraj K. Jha, and Jan Kautz. Dreaming to distill: Data-free knowledge transfer via deepinversion. CoRR, abs/1912.08795, 2019. Ritchie Zhao, Yuwei Hu, Jordan Dotzel, Christopher De Sa, and Zhiru Zhang. Improving neural network quantization without retraining using outlier channel splitting. In ICML 2019, 9-15 June 2019, Long Beach, California, USA, pages 7543–7552, 2019. Aojun Zhou, Anbang Yao, Yiwen Guo, Lin Xu, and Yurong Chen. Incremental network quantization: Towards lossless cnns with low-precision weights. In ICLR 2017, Toulon, France, April 24-26, 2017, Conference Track Proceedings, 2017. Shuchang Zhou, Zekun Ni, Xinyu Zhou, He Wen, Yuxin Wu, and Yuheng Zou. Dorefa-net: Training low bitwidth convolutional neural networks with low bitwidth gradients. CoRR, abs/1606.06160, 2016. | - |
| local.type.refereed | Refereed | - |
| local.type.specified | Proceedings Paper | - |
| local.relation.ispartofseriesnr | 129 | - |
| dc.identifier.url | https://proceedings.mlr.press/v129/li20a/li20a.pdf | - |
| local.provider.type | - | |
| local.bibliographicCitation.btitle | Proceedings of Machine Learning Research | - |
| local.uhasselt.international | yes | - |
| item.fulltext | With Fulltext | - |
| item.accessRights | Open Access | - |
| item.fullcitation | LI, Bowen; Huang, Kai; CHEN, Siang; Xiong, Dongliang; Jiang, Haitian & CLAESEN, Luc (2020) DFQF: Data Free Quantization-aware Fine-tuning. In: Jialin Pan, Sinno; Sugiyama, Masashi (Ed.). Proceedings of Machine Learning Research, p. 289 -304. | - |
| item.contributor | LI, Bowen | - |
| item.contributor | Huang, Kai | - |
| item.contributor | CHEN, Siang | - |
| item.contributor | Xiong, Dongliang | - |
| item.contributor | Jiang, Haitian | - |
| item.contributor | CLAESEN, Luc | - |
| Appears in Collections: | Research publications | |
Files in This Item:
| File | Description | Size | Format | |
|---|---|---|---|---|
| DFQF - Data Free Quantization-aware Fine-tuning.pdf | Published version | 557.32 kB | Adobe PDF | View/Open |
Google ScholarTM
Check
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.