Abstract
Knowledge distillation (KD) is one of the most useful techniques for light-weight neural networks. Although neural networks have a clear purpose of embedding datasets into the low-dimensional space, the existing knowledge was quite far from this purpose and provided only limited information. We argue that good knowledge should be able to interpret the embedding procedure. This paper proposes a method of generating interpretable embedding procedure (IEP) knowledge based on principal component analysis, and distilling it based on a message passing neural network. Experimental results show that the student network trained by the proposed KD method improves 2.28% in the CIFAR100 dataset, which is higher performance than the state-of-the-art (SOTA) method. We also demonstrate that the embedding procedure knowledge is interpretable via visualization of the proposed KD process. The implemented code is available at https://github.com/sseung0703/IEPKT.
Original language | English |
---|---|
Title of host publication | 35th AAAI Conference on Artificial Intelligence, AAAI 2021 |
Publisher | Association for the Advancement of Artificial Intelligence |
Pages | 8297-8305 |
Number of pages | 9 |
ISBN (Electronic) | 9781713835974 |
State | Published - 2021 |
Event | 35th AAAI Conference on Artificial Intelligence, AAAI 2021 - Virtual, Online Duration: 2 Feb 2021 → 9 Feb 2021 |
Publication series
Name | 35th AAAI Conference on Artificial Intelligence, AAAI 2021 |
---|---|
Volume | 9B |
Conference
Conference | 35th AAAI Conference on Artificial Intelligence, AAAI 2021 |
---|---|
City | Virtual, Online |
Period | 2/02/21 → 9/02/21 |
Bibliographical note
Publisher Copyright:Copyright © 2021, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved.