Stable Quantization-Aware Training with Adaptive Gradient Clipping

Jihoon Park, Seunghyun Lee, Byung Cheol Song

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

Quantization-Aware Training (QAT) uses batch normalization (BN) folding during fine-tuning, so it may not use the normalization effects of the BN layer. HAWQ, which achieved SOTA with QAT, has significant accuracy degradation as the training becomes longer. In this paper, we apply Adaptive Gradient Clipping (AGC) to stable quantization-aware training and improve accuracy by adding Dropout. Moreover, we have ablation studies about AGC.

Original languageEnglish
Title of host publication2023 International Conference on Electronics, Information, and Communication, ICEIC 2023
PublisherInstitute of Electrical and Electronics Engineers Inc.
ISBN (Electronic)9798350320213
DOIs
StatePublished - 2023
Event2023 International Conference on Electronics, Information, and Communication, ICEIC 2023 - Singapore, Singapore
Duration: 5 Feb 20238 Feb 2023

Publication series

Name2023 International Conference on Electronics, Information, and Communication, ICEIC 2023

Conference

Conference2023 International Conference on Electronics, Information, and Communication, ICEIC 2023
Country/TerritorySingapore
CitySingapore
Period5/02/238/02/23

Bibliographical note

Publisher Copyright:
© 2023 IEEE.

Fingerprint

Dive into the research topics of 'Stable Quantization-Aware Training with Adaptive Gradient Clipping'. Together they form a unique fingerprint.

Cite this