PhD forum: Why TanH is a hardware friendly activation function for CNNs - Université de Rennes Accéder directement au contenu
Communication Dans Un Congrès Année : 2017

PhD forum: Why TanH is a hardware friendly activation function for CNNs

Résumé

Convolutional Neural Networks (CNNs) [1] are the state of the art of image classification that improved accuracy and robustness of machine vision systems at the price of a very high computational cost. This motivated multiple research efforts to investigate the applicability of approximate computing and more particularly, fixed point-arithmetic for CNNs. In all this approaches, a recurrent problem is that the learned parameters in deep fraCNN layers have a significantly lower numerical dynamic range when compared to the feature maps, which prevents from using of a low bit-width representation in deep layers. In this paper, we demonstrate that using the TanH activation function is way to prevent this issue. To support this demonstration, three benchmark CNN models are trained with the TanH function. These models are then quantized using the same bit-width across all the layers. In the case of FPGA based accelerators, this approach infers the minimal amount of logic elements to deploy CNNs. © 2017 Association for Computing Machinery.
Fichier non déposé

Dates et versions

hal-01687764 , version 1 (18-01-2018)

Identifiants

Citer

K. Abdelouahab, Maxime Pelcat, F. Berry. PhD forum: Why TanH is a hardware friendly activation function for CNNs. 11th International Conference on Distributed Smart Cameras, ICDSC 2017, Sep 2017, Stanford, United States. ⟨10.1145/3131885.3131937⟩. ⟨hal-01687764⟩
136 Consultations
0 Téléchargements

Altmetric

Partager

Gmail Facebook X LinkedIn More