Resilient training of neural network classifiers with approximate computing techniques for hardware-optimised implementations

dc.creatorVitor Angelo Maria Ferreira Torres
dc.creatorFrank Sill Torres
dc.date.accessioned2025-05-21T13:22:57Z
dc.date.accessioned2025-09-09T00:54:49Z
dc.date.available2025-05-21T13:22:57Z
dc.date.issued2019
dc.identifier.doihttps://doi.org/10.1049/iet-cdt.2019.0036
dc.identifier.issn1751-8601
dc.identifier.urihttps://hdl.handle.net/1843/82405
dc.languageeng
dc.publisherUniversidade Federal de Minas Gerais
dc.relation.ispartofIET Computers and digital techniques articles
dc.rightsAcesso Restrito
dc.subjectAprendizado do computador
dc.subject.otherlarge scale unsupervised learning
dc.titleResilient training of neural network classifiers with approximate computing techniques for hardware-optimised implementations
dc.typeArtigo de periódico
local.citation.epage542
local.citation.issue6
local.citation.spage532
local.citation.volume13
local.description.resumoAs Machine Learning applications increase the demand for optimised implementations in both embedded and high-end processing platforms, the industry and research community have been responding with different approaches to implement these solutions. This work presents approximations to arithmetic operations and mathematical functions that, associated with a customised adaptive artificial neural networks training method, based on RMSProp, provide reliable and efficient implementations of classifiers. The proposed solution does not rely on mixed operations with higher precision or complex rounding methods that are commonly applied. The intention of this work is not to find the optimal simplifications for specific deep learning problems but to present an optimised framework that can be used as reliably as one implemented with precise operations, standard training algorithms and the same network structures and hyper-parameters. By simplifying the ‘half-precision’ floating point format and approximating exponentiation and square root operations, the authors’ work drastically reduces the field programmable gate array implementation complexity (e.g. −43 and −57% in two of the component resources). The reciprocal square root approximation is so simple it could be implemented only with combination logic. In a full software implementation for a mixed-precision platform, only two of the approximations compensate the processing overhead of precision conversions.
local.publisher.countryBrasil
local.publisher.departmentENG - DEPARTAMENTO DE ENGENHARIA ELETRÔNICA
local.publisher.initialsUFMG
local.url.externahttps://ietresearch.onlinelibrary.wiley.com/doi/abs/10.1049/iet-cdt.2019.0036

Arquivos

Licença do pacote

Agora exibindo 1 - 1 de 1
Carregando...
Imagem de Miniatura
Nome:
License.txt
Tamanho:
1.99 KB
Formato:
Plain Text
Descrição: