Improved Compression of Artificial Neural Networks through Curvature-Aware Training

Publikation: Beitrag in Buch/Bericht/Sammelwerk/KonferenzbandAufsatz in KonferenzbandForschungPeer-Review

Autoren

Forschungs-netzwerk anzeigen

Details

OriginalspracheEnglisch
Titel des Sammelwerks2022 International Joint Conference on Neural Networks, IJCNN 2022 - Proceedings
Herausgeber (Verlag)Institute of Electrical and Electronics Engineers Inc.
ISBN (elektronisch)9781728186719
ISBN (Print)978-1-6654-9526-4
PublikationsstatusVeröffentlicht - 2022
Veranstaltung2022 International Joint Conference on Neural Networks, IJCNN 2022 - Padua, Italien
Dauer: 18 Juli 202223 Juli 2022

Publikationsreihe

NameProceedings of the International Joint Conference on Neural Networks
Band2022-July
ISSN (Print)2161-4393
ISSN (elektronisch)2161-4407

Abstract

Artificial neural networks achieve state-of-the-art performance in many branches of engineering. As such they are used for all kinds of tasks and nowadays are desired to be used on mobile devices like smartphones. Due to limited hardware resources or limited channel capacity on mobile devices, compression of neural network models to reduce storage or transmission costs is desired. Furthermore, reduced complexity is of interest. This work investigates introducing the curvature of the loss surface in the training of artificial neural networks and analyzes its benefit for the compression of neural networks through quantization and pruning of its weights. As proof-of-concept, three small LeNet-based neural networks were trained using a novel loss function consisting of a weighted average of the cross-entropy loss and the Frobenius norm of the hessian matrix. That way both, the loss as well as the local curvature, are minimized concurrently. Using the proposed method, mean test accuracies on the MNIST and FashionMNIST datasets after quantization were considerably improved by up to about 47.6 % for 1 bit quantization on MNIST and about 27.8 % on FashionMNIST compared to quantization after training without curvature information. Additionally, pruning was found to benefit from introducing curvature in the training as well with an increase of up to about 14.6 % mean test accuracy compared to pruning after training without curvature except for isolated cases. Training the artificial neural networks first without curvature information and subsequent training by only one epoch using curvature information allowed to increase the mean test accuracy after quantization at 1 bit by about 16 %. The proposed method can potentially improve the accuracy after compression irrespective of the compression method applied.

ASJC Scopus Sachgebiete

Zitieren

Improved Compression of Artificial Neural Networks through Curvature-Aware Training. / Hinrichs, Reemt; Liang, Kai; Lu, Ze et al.
2022 International Joint Conference on Neural Networks, IJCNN 2022 - Proceedings. Institute of Electrical and Electronics Engineers Inc., 2022. (Proceedings of the International Joint Conference on Neural Networks; Band 2022-July).

Publikation: Beitrag in Buch/Bericht/Sammelwerk/KonferenzbandAufsatz in KonferenzbandForschungPeer-Review

Hinrichs, R, Liang, K, Lu, Z & Ostermann, J 2022, Improved Compression of Artificial Neural Networks through Curvature-Aware Training. in 2022 International Joint Conference on Neural Networks, IJCNN 2022 - Proceedings. Proceedings of the International Joint Conference on Neural Networks, Bd. 2022-July, Institute of Electrical and Electronics Engineers Inc., 2022 International Joint Conference on Neural Networks, IJCNN 2022, Padua, Italien, 18 Juli 2022. https://doi.org/10.1109/IJCNN55064.2022.9892511
Hinrichs, R., Liang, K., Lu, Z., & Ostermann, J. (2022). Improved Compression of Artificial Neural Networks through Curvature-Aware Training. In 2022 International Joint Conference on Neural Networks, IJCNN 2022 - Proceedings (Proceedings of the International Joint Conference on Neural Networks; Band 2022-July). Institute of Electrical and Electronics Engineers Inc.. https://doi.org/10.1109/IJCNN55064.2022.9892511
Hinrichs R, Liang K, Lu Z, Ostermann J. Improved Compression of Artificial Neural Networks through Curvature-Aware Training. in 2022 International Joint Conference on Neural Networks, IJCNN 2022 - Proceedings. Institute of Electrical and Electronics Engineers Inc. 2022. (Proceedings of the International Joint Conference on Neural Networks). doi: 10.1109/IJCNN55064.2022.9892511
Hinrichs, Reemt ; Liang, Kai ; Lu, Ze et al. / Improved Compression of Artificial Neural Networks through Curvature-Aware Training. 2022 International Joint Conference on Neural Networks, IJCNN 2022 - Proceedings. Institute of Electrical and Electronics Engineers Inc., 2022. (Proceedings of the International Joint Conference on Neural Networks).
Download
@inproceedings{afea1594f9684c8780af871f4f09890b,
title = "Improved Compression of Artificial Neural Networks through Curvature-Aware Training",
abstract = "Artificial neural networks achieve state-of-the-art performance in many branches of engineering. As such they are used for all kinds of tasks and nowadays are desired to be used on mobile devices like smartphones. Due to limited hardware resources or limited channel capacity on mobile devices, compression of neural network models to reduce storage or transmission costs is desired. Furthermore, reduced complexity is of interest. This work investigates introducing the curvature of the loss surface in the training of artificial neural networks and analyzes its benefit for the compression of neural networks through quantization and pruning of its weights. As proof-of-concept, three small LeNet-based neural networks were trained using a novel loss function consisting of a weighted average of the cross-entropy loss and the Frobenius norm of the hessian matrix. That way both, the loss as well as the local curvature, are minimized concurrently. Using the proposed method, mean test accuracies on the MNIST and FashionMNIST datasets after quantization were considerably improved by up to about 47.6 % for 1 bit quantization on MNIST and about 27.8 % on FashionMNIST compared to quantization after training without curvature information. Additionally, pruning was found to benefit from introducing curvature in the training as well with an increase of up to about 14.6 % mean test accuracy compared to pruning after training without curvature except for isolated cases. Training the artificial neural networks first without curvature information and subsequent training by only one epoch using curvature information allowed to increase the mean test accuracy after quantization at 1 bit by about 16 %. The proposed method can potentially improve the accuracy after compression irrespective of the compression method applied.",
keywords = "compression, curvature, loss surface, neural networks",
author = "Reemt Hinrichs and Kai Liang and Ze Lu and Jorn Ostermann",
year = "2022",
doi = "10.1109/IJCNN55064.2022.9892511",
language = "English",
isbn = "978-1-6654-9526-4",
series = "Proceedings of the International Joint Conference on Neural Networks",
publisher = "Institute of Electrical and Electronics Engineers Inc.",
booktitle = "2022 International Joint Conference on Neural Networks, IJCNN 2022 - Proceedings",
address = "United States",
note = "2022 International Joint Conference on Neural Networks, IJCNN 2022 ; Conference date: 18-07-2022 Through 23-07-2022",

}

Download

TY - GEN

T1 - Improved Compression of Artificial Neural Networks through Curvature-Aware Training

AU - Hinrichs, Reemt

AU - Liang, Kai

AU - Lu, Ze

AU - Ostermann, Jorn

PY - 2022

Y1 - 2022

N2 - Artificial neural networks achieve state-of-the-art performance in many branches of engineering. As such they are used for all kinds of tasks and nowadays are desired to be used on mobile devices like smartphones. Due to limited hardware resources or limited channel capacity on mobile devices, compression of neural network models to reduce storage or transmission costs is desired. Furthermore, reduced complexity is of interest. This work investigates introducing the curvature of the loss surface in the training of artificial neural networks and analyzes its benefit for the compression of neural networks through quantization and pruning of its weights. As proof-of-concept, three small LeNet-based neural networks were trained using a novel loss function consisting of a weighted average of the cross-entropy loss and the Frobenius norm of the hessian matrix. That way both, the loss as well as the local curvature, are minimized concurrently. Using the proposed method, mean test accuracies on the MNIST and FashionMNIST datasets after quantization were considerably improved by up to about 47.6 % for 1 bit quantization on MNIST and about 27.8 % on FashionMNIST compared to quantization after training without curvature information. Additionally, pruning was found to benefit from introducing curvature in the training as well with an increase of up to about 14.6 % mean test accuracy compared to pruning after training without curvature except for isolated cases. Training the artificial neural networks first without curvature information and subsequent training by only one epoch using curvature information allowed to increase the mean test accuracy after quantization at 1 bit by about 16 %. The proposed method can potentially improve the accuracy after compression irrespective of the compression method applied.

AB - Artificial neural networks achieve state-of-the-art performance in many branches of engineering. As such they are used for all kinds of tasks and nowadays are desired to be used on mobile devices like smartphones. Due to limited hardware resources or limited channel capacity on mobile devices, compression of neural network models to reduce storage or transmission costs is desired. Furthermore, reduced complexity is of interest. This work investigates introducing the curvature of the loss surface in the training of artificial neural networks and analyzes its benefit for the compression of neural networks through quantization and pruning of its weights. As proof-of-concept, three small LeNet-based neural networks were trained using a novel loss function consisting of a weighted average of the cross-entropy loss and the Frobenius norm of the hessian matrix. That way both, the loss as well as the local curvature, are minimized concurrently. Using the proposed method, mean test accuracies on the MNIST and FashionMNIST datasets after quantization were considerably improved by up to about 47.6 % for 1 bit quantization on MNIST and about 27.8 % on FashionMNIST compared to quantization after training without curvature information. Additionally, pruning was found to benefit from introducing curvature in the training as well with an increase of up to about 14.6 % mean test accuracy compared to pruning after training without curvature except for isolated cases. Training the artificial neural networks first without curvature information and subsequent training by only one epoch using curvature information allowed to increase the mean test accuracy after quantization at 1 bit by about 16 %. The proposed method can potentially improve the accuracy after compression irrespective of the compression method applied.

KW - compression

KW - curvature

KW - loss surface

KW - neural networks

UR - http://www.scopus.com/inward/record.url?scp=85140788906&partnerID=8YFLogxK

U2 - 10.1109/IJCNN55064.2022.9892511

DO - 10.1109/IJCNN55064.2022.9892511

M3 - Conference contribution

AN - SCOPUS:85140788906

SN - 978-1-6654-9526-4

T3 - Proceedings of the International Joint Conference on Neural Networks

BT - 2022 International Joint Conference on Neural Networks, IJCNN 2022 - Proceedings

PB - Institute of Electrical and Electronics Engineers Inc.

T2 - 2022 International Joint Conference on Neural Networks, IJCNN 2022

Y2 - 18 July 2022 through 23 July 2022

ER -

Von denselben Autoren