Loading [MathJax]/extensions/tex2jax.js

Improved Compression of Artificial Neural Networks through Curvature-Aware Training

Research output: Chapter in book/report/conference proceedingConference contributionResearchpeer review

Authors

Research Organisations

Details

Original languageEnglish
Title of host publication2022 International Joint Conference on Neural Networks, IJCNN 2022 - Proceedings
PublisherInstitute of Electrical and Electronics Engineers Inc.
ISBN (electronic)9781728186719
ISBN (print)978-1-6654-9526-4
Publication statusPublished - 2022
Event2022 International Joint Conference on Neural Networks, IJCNN 2022 - Padua, Italy
Duration: 18 Jul 202223 Jul 2022

Publication series

NameProceedings of the International Joint Conference on Neural Networks
Volume2022-July
ISSN (Print)2161-4393
ISSN (electronic)2161-4407

Abstract

Artificial neural networks achieve state-of-the-art performance in many branches of engineering. As such they are used for all kinds of tasks and nowadays are desired to be used on mobile devices like smartphones. Due to limited hardware resources or limited channel capacity on mobile devices, compression of neural network models to reduce storage or transmission costs is desired. Furthermore, reduced complexity is of interest. This work investigates introducing the curvature of the loss surface in the training of artificial neural networks and analyzes its benefit for the compression of neural networks through quantization and pruning of its weights. As proof-of-concept, three small LeNet-based neural networks were trained using a novel loss function consisting of a weighted average of the cross-entropy loss and the Frobenius norm of the hessian matrix. That way both, the loss as well as the local curvature, are minimized concurrently. Using the proposed method, mean test accuracies on the MNIST and FashionMNIST datasets after quantization were considerably improved by up to about 47.6 % for 1 bit quantization on MNIST and about 27.8 % on FashionMNIST compared to quantization after training without curvature information. Additionally, pruning was found to benefit from introducing curvature in the training as well with an increase of up to about 14.6 % mean test accuracy compared to pruning after training without curvature except for isolated cases. Training the artificial neural networks first without curvature information and subsequent training by only one epoch using curvature information allowed to increase the mean test accuracy after quantization at 1 bit by about 16 %. The proposed method can potentially improve the accuracy after compression irrespective of the compression method applied.

Keywords

    compression, curvature, loss surface, neural networks

ASJC Scopus subject areas

Cite this

Improved Compression of Artificial Neural Networks through Curvature-Aware Training. / Hinrichs, Reemt; Liang, Kai; Lu, Ze et al.
2022 International Joint Conference on Neural Networks, IJCNN 2022 - Proceedings. Institute of Electrical and Electronics Engineers Inc., 2022. (Proceedings of the International Joint Conference on Neural Networks; Vol. 2022-July).

Research output: Chapter in book/report/conference proceedingConference contributionResearchpeer review

Hinrichs, R, Liang, K, Lu, Z & Ostermann, J 2022, Improved Compression of Artificial Neural Networks through Curvature-Aware Training. in 2022 International Joint Conference on Neural Networks, IJCNN 2022 - Proceedings. Proceedings of the International Joint Conference on Neural Networks, vol. 2022-July, Institute of Electrical and Electronics Engineers Inc., 2022 International Joint Conference on Neural Networks, IJCNN 2022, Padua, Italy, 18 Jul 2022. https://doi.org/10.1109/IJCNN55064.2022.9892511
Hinrichs, R., Liang, K., Lu, Z., & Ostermann, J. (2022). Improved Compression of Artificial Neural Networks through Curvature-Aware Training. In 2022 International Joint Conference on Neural Networks, IJCNN 2022 - Proceedings (Proceedings of the International Joint Conference on Neural Networks; Vol. 2022-July). Institute of Electrical and Electronics Engineers Inc.. https://doi.org/10.1109/IJCNN55064.2022.9892511
Hinrichs R, Liang K, Lu Z, Ostermann J. Improved Compression of Artificial Neural Networks through Curvature-Aware Training. In 2022 International Joint Conference on Neural Networks, IJCNN 2022 - Proceedings. Institute of Electrical and Electronics Engineers Inc. 2022. (Proceedings of the International Joint Conference on Neural Networks). doi: 10.1109/IJCNN55064.2022.9892511
Hinrichs, Reemt ; Liang, Kai ; Lu, Ze et al. / Improved Compression of Artificial Neural Networks through Curvature-Aware Training. 2022 International Joint Conference on Neural Networks, IJCNN 2022 - Proceedings. Institute of Electrical and Electronics Engineers Inc., 2022. (Proceedings of the International Joint Conference on Neural Networks).
Download
@inproceedings{afea1594f9684c8780af871f4f09890b,
title = "Improved Compression of Artificial Neural Networks through Curvature-Aware Training",
abstract = "Artificial neural networks achieve state-of-the-art performance in many branches of engineering. As such they are used for all kinds of tasks and nowadays are desired to be used on mobile devices like smartphones. Due to limited hardware resources or limited channel capacity on mobile devices, compression of neural network models to reduce storage or transmission costs is desired. Furthermore, reduced complexity is of interest. This work investigates introducing the curvature of the loss surface in the training of artificial neural networks and analyzes its benefit for the compression of neural networks through quantization and pruning of its weights. As proof-of-concept, three small LeNet-based neural networks were trained using a novel loss function consisting of a weighted average of the cross-entropy loss and the Frobenius norm of the hessian matrix. That way both, the loss as well as the local curvature, are minimized concurrently. Using the proposed method, mean test accuracies on the MNIST and FashionMNIST datasets after quantization were considerably improved by up to about 47.6 % for 1 bit quantization on MNIST and about 27.8 % on FashionMNIST compared to quantization after training without curvature information. Additionally, pruning was found to benefit from introducing curvature in the training as well with an increase of up to about 14.6 % mean test accuracy compared to pruning after training without curvature except for isolated cases. Training the artificial neural networks first without curvature information and subsequent training by only one epoch using curvature information allowed to increase the mean test accuracy after quantization at 1 bit by about 16 %. The proposed method can potentially improve the accuracy after compression irrespective of the compression method applied.",
keywords = "compression, curvature, loss surface, neural networks",
author = "Reemt Hinrichs and Kai Liang and Ze Lu and Jorn Ostermann",
year = "2022",
doi = "10.1109/IJCNN55064.2022.9892511",
language = "English",
isbn = "978-1-6654-9526-4",
series = "Proceedings of the International Joint Conference on Neural Networks",
publisher = "Institute of Electrical and Electronics Engineers Inc.",
booktitle = "2022 International Joint Conference on Neural Networks, IJCNN 2022 - Proceedings",
address = "United States",
note = "2022 International Joint Conference on Neural Networks, IJCNN 2022 ; Conference date: 18-07-2022 Through 23-07-2022",

}

Download

TY - GEN

T1 - Improved Compression of Artificial Neural Networks through Curvature-Aware Training

AU - Hinrichs, Reemt

AU - Liang, Kai

AU - Lu, Ze

AU - Ostermann, Jorn

PY - 2022

Y1 - 2022

N2 - Artificial neural networks achieve state-of-the-art performance in many branches of engineering. As such they are used for all kinds of tasks and nowadays are desired to be used on mobile devices like smartphones. Due to limited hardware resources or limited channel capacity on mobile devices, compression of neural network models to reduce storage or transmission costs is desired. Furthermore, reduced complexity is of interest. This work investigates introducing the curvature of the loss surface in the training of artificial neural networks and analyzes its benefit for the compression of neural networks through quantization and pruning of its weights. As proof-of-concept, three small LeNet-based neural networks were trained using a novel loss function consisting of a weighted average of the cross-entropy loss and the Frobenius norm of the hessian matrix. That way both, the loss as well as the local curvature, are minimized concurrently. Using the proposed method, mean test accuracies on the MNIST and FashionMNIST datasets after quantization were considerably improved by up to about 47.6 % for 1 bit quantization on MNIST and about 27.8 % on FashionMNIST compared to quantization after training without curvature information. Additionally, pruning was found to benefit from introducing curvature in the training as well with an increase of up to about 14.6 % mean test accuracy compared to pruning after training without curvature except for isolated cases. Training the artificial neural networks first without curvature information and subsequent training by only one epoch using curvature information allowed to increase the mean test accuracy after quantization at 1 bit by about 16 %. The proposed method can potentially improve the accuracy after compression irrespective of the compression method applied.

AB - Artificial neural networks achieve state-of-the-art performance in many branches of engineering. As such they are used for all kinds of tasks and nowadays are desired to be used on mobile devices like smartphones. Due to limited hardware resources or limited channel capacity on mobile devices, compression of neural network models to reduce storage or transmission costs is desired. Furthermore, reduced complexity is of interest. This work investigates introducing the curvature of the loss surface in the training of artificial neural networks and analyzes its benefit for the compression of neural networks through quantization and pruning of its weights. As proof-of-concept, three small LeNet-based neural networks were trained using a novel loss function consisting of a weighted average of the cross-entropy loss and the Frobenius norm of the hessian matrix. That way both, the loss as well as the local curvature, are minimized concurrently. Using the proposed method, mean test accuracies on the MNIST and FashionMNIST datasets after quantization were considerably improved by up to about 47.6 % for 1 bit quantization on MNIST and about 27.8 % on FashionMNIST compared to quantization after training without curvature information. Additionally, pruning was found to benefit from introducing curvature in the training as well with an increase of up to about 14.6 % mean test accuracy compared to pruning after training without curvature except for isolated cases. Training the artificial neural networks first without curvature information and subsequent training by only one epoch using curvature information allowed to increase the mean test accuracy after quantization at 1 bit by about 16 %. The proposed method can potentially improve the accuracy after compression irrespective of the compression method applied.

KW - compression

KW - curvature

KW - loss surface

KW - neural networks

UR - http://www.scopus.com/inward/record.url?scp=85140788906&partnerID=8YFLogxK

U2 - 10.1109/IJCNN55064.2022.9892511

DO - 10.1109/IJCNN55064.2022.9892511

M3 - Conference contribution

AN - SCOPUS:85140788906

SN - 978-1-6654-9526-4

T3 - Proceedings of the International Joint Conference on Neural Networks

BT - 2022 International Joint Conference on Neural Networks, IJCNN 2022 - Proceedings

PB - Institute of Electrical and Electronics Engineers Inc.

T2 - 2022 International Joint Conference on Neural Networks, IJCNN 2022

Y2 - 18 July 2022 through 23 July 2022

ER -

By the same author(s)