250 results on '"Binary neural network"'
Search Results
2. How to Train Accurate BNNs for Embedded Systems?
- Author
-
Putter, F. A. M. de, Corporaal, Henk, Pasricha, Sudeep, editor, and Shafique, Muhammad, editor
- Published
- 2024
- Full Text
- View/download PDF
3. Deep Learning and Neural Architecture Search for Optimizing Binary Neural Network Image Super Resolution.
- Author
-
Su, Yuanxin, Ang, Li-minn, Seng, Kah Phooi, and Smith, Jeremy
- Subjects
- *
DEEP learning , *HIGH resolution imaging , *COMPUTATIONAL complexity , *ARCHITECTURAL design - Abstract
The evolution of super-resolution (SR) technology has seen significant advancements through the adoption of deep learning methods. However, the deployment of such models by resource-constrained devices necessitates models that not only perform efficiently, but also conserve computational resources. Binary neural networks (BNNs) offer a promising solution by minimizing the data precision to binary levels, thus reducing the computational complexity and memory requirements. However, for BNNs, an effective architecture is essential due to their inherent limitations in representing information. Designing such architectures traditionally requires extensive computational resources and time. With the advancement in neural architecture search (NAS), differentiable NAS has emerged as an attractive solution for efficiently crafting network structures. In this paper, we introduce a novel and efficient binary network search method tailored for image super-resolution tasks. We adapt the search space specifically for super resolution to ensure it is optimally suited for the requirements of such tasks. Furthermore, we incorporate Libra Parameter Binarization (Libra-PB) to maximize information retention during forward propagation. Our experimental results demonstrate that the network structures generated by our method require only a third of the parameters, compared to conventional methods, and yet deliver comparable performance. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
4. CBin-NN: An Inference Engine for Binarized Neural Networks.
- Author
-
Sakr, Fouad, Berta, Riccardo, Doyle, Joseph, Capello, Alessio, Dabbous, Ali, Lazzaroni, Luca, and Bellotti, Francesco
- Subjects
ARTIFICIAL neural networks ,CONVOLUTIONAL neural networks ,FOOTPRINTS ,INTERNET of things - Abstract
Binarization is an extreme quantization technique that is attracting research in the Internet of Things (IoT) field, as it radically reduces the memory footprint of deep neural networks without a correspondingly significant accuracy drop. To support the effective deployment of Binarized Neural Networks (BNNs), we propose CBin-NN, a library of layer operators that allows the building of simple yet flexible convolutional neural networks (CNNs) with binary weights and activations. CBin-NN is platform-independent and is thus portable to virtually any software-programmable device. Experimental analysis on the CIFAR-10 dataset shows that our library, compared to a set of state-of-the-art inference engines, speeds up inference by 3.6 times and reduces the memory required to store model weights and activations by 7.5 times and 28 times, respectively, at the cost of slightly lower accuracy (2.5%). An ablation study stresses the importance of a Quantized Input Quantized Kernel Convolution layer to improve accuracy and reduce latency at the cost of a slight increase in model size. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
5. Multibit, Lead‐Free Cs2SnI6 Resistive Random Access Memory with Self‐Compliance for Improved Accuracy in Binary Neural Network Application.
- Author
-
Kumar, Ajit, Krishnaiah, Mokurala, Park, Jinwoo, Mishra, Dhananjay, Dash, Bidyashakti, Jo, Hyeon‐Bin, Lee, Geun, Youn, Sangwook, Kim, Hyungjin, and Jin, Sung Hun
- Subjects
- *
NONVOLATILE random-access memory , *RANDOM access memory , *CONVOLUTIONAL neural networks - Abstract
In the realm of neuromorphic computing, integrating Binary Neural Networks (BNN) with non‐volatile memory based on emerging materials can be a promising avenue for introducing novel functionalities. This study underscores the viability of lead‐free, air‐stable Cs2SnI6 (CSI) based resistive random access memory (RRAM) devices as synaptic weights in neuromorphic architectures, specifically for BNNs applications. Herein, hydrothermally synthesized CSI perovskites are explored as a resistive layer in RRAM devices either on the rigid or flexible substrate, highlighting reproducible multibit switching with self‐compliance, low‐ resistance‐state (LRS) variations, a decent On/Off ratio(or retention) of ≈103(or 104 s), and endurance exceeding 300 cycles. Moreover, a comprehensive evaluation with the 32 × 32 × 3 RGB CIFAR‐10 dataset reveals that binary convolutional neural networks (BCNN) trained solely on binary weight values can achieve competitive rates of accuracy comparable to those of their analog weight counterparts. These findings highlight the dominance of the LRS for CSI RRAM with self‐compliance in a weighted configuration and minimal influence of the high resistance state despite substantial fluctuations for flexible CSI RRAM under varying bending radii. With its unique electrical switching capabilities, the CSI RRAM is highly anticipated to emerge as a promising candidate for embedded AI systems, especially in IoT devices and wearables. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
6. High Speed Binary Neural Network Hardware Accelerator Relied on Optical NEMS.
- Author
-
Gholami, Yashar, Marvi, Fahimeh, Ghorbanloo, Romina, Eslami, Mohammad Reza, and Jafari, Kian
- Abstract
In this article, an electrostatically-actuated NEMS XOR gate is proposed based on photonic crystals for hardware implementation of binary neural networks. The device includes a 2D photonic crystal which is set on a movable electrode to implement the XOR logic using the transmission of specific wavelengths to the output. This design represents the importance of the proposed structure in which the logic gate operation is not dependent on the contact of its conductive layers. Consequently, one of the major issues in MEMS-based logic gates, which is due to the contact of the operating electrodes and may cause stiction problem, reducing the reliability of the system, can be tackled by the present approach. Furthermore, according to the simulation results, the functional characteristics of the present NEMS XOR gate are obtained as follows: pull-in voltage of Vp = 8V, operating voltage of Vo = 10V and switching time of ts = 4 μs. The results also show that the proposed design provides a classification error rate of between 1% to 12%, while used in neural network implementation. This error can be negligible compared to the state-of-the-art designs in neural network implementation. These appropriate parameters of the present NEMS gate make it a promising choice for the implementation of neural networks with a high network accuracy even in the presence of significant process variations. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
7. DCP–NAS: Discrepant Child–Parent Neural Architecture Search for 1-bit CNNs.
- Author
-
Li, Yanjing, Xu, Sheng, Cao, Xianbin, Zhuo, Li'an, Zhang, Baochang, Wang, Tian, and Guo, Guodong
- Subjects
- *
CONVOLUTIONAL neural networks , *NEWTON-Raphson method - Abstract
Neural architecture search (NAS) proves to be among the effective approaches for many tasks by generating an application-adaptive neural architecture, which is still challenged by high computational cost and memory consumption. At the same time, 1-bit convolutional neural networks (CNNs) with binary weights and activations show their potential for resource-limited embedded devices. One natural approach is to use 1-bit CNNs to reduce the computation and memory cost of NAS by taking advantage of the strengths of each in a unified framework, while searching the 1-bit CNNs is more challenging due to the more complicated processes involved. In this paper, we introduce Discrepant Child–Parent Neural Architecture Search (DCP–NAS) to efficiently search 1-bit CNNs, based on a new framework of searching the 1-bit model (Child) under the supervision of a real-valued model (Parent). Particularly, we first utilize a Parent model to calculate a tangent direction, based on which the tangent propagation method is introduced to search the optimized 1-bit Child. We further observe a coupling relationship between the weights and architecture parameters existing in such differentiable frameworks. To address the issue, we propose a decoupled optimization method to search an optimized architecture. Extensive experiments demonstrate that our DCP–NAS achieves much better results than prior arts on both CIFAR-10 and ImageNet datasets. In particular, the backbones achieved by our DCP–NAS achieve strong generalization performance on person re-identification and object detection. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
8. Enabling Binary Neural Network Training on the Edge.
- Author
-
WANG, ERWEI, DAVIS, JAMES J., MORO, DANIELE, ZIELINSKI, PIOTR, JIA JIE LIM, COELHO, CLAUDIONOR, CHATTERJEE, SATRAJIT, CHEUNG, PETER Y. K., and CONSTANTINIDES, GEORGE A.
- Subjects
MACHINE learning ,SIMPLE machines ,FOOTPRINTS - Abstract
The ever-growing computational demands of increasingly complex machine learning models frequently necessitate the use of powerful cloud-based infrastructure for their training. Binary neural networks are known to be promising candidates for on-device inference due to their extreme compute and memory savings over higher-precision alternatives. However, their existing training methods require the concurrent storage of high-precision activations for all layers, generally making learning on memory-constrained devices infeasible. In this article, we demonstrate that the backward propagation operations needed for binary neural network training are strongly robust to quantization, thereby making on-the-edge learning with modern models a practical proposition. We introduce a low-cost binary neural network training strategy exhibiting sizable memory footprint reductions while inducing little to no accuracy loss vs Courbariaux & Bengio's standard approach. These decreases are primarily enabled through the retention of activations exclusively in binary format. Against the latter algorithm, our drop-in replacement sees memory requirement reductions of 3-5x, while reaching similar test accuracy (±2 pp) in comparable time, across a range of small-scale models trained to classify popular datasets. We also demonstrate from-scratch ImageNet training of binarized ResNet-18, achieving a 3.78x memory reduction. Our work is open-source, and includes the Raspberry Pi-targeted prototype we used to verify our modeled memory decreases and capture the associated energy drops. Such savings will allow for unnecessary cloud offloading to be avoided, reducing latency, increasing energy efficiency, and safeguarding end-user privacy. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
9. A comprehensive review of Binary Neural Network.
- Author
-
Yuan, Chunyu and Agaian, Sos S.
- Abstract
Deep learning (DL) has recently changed the development of intelligent systems and is widely adopted in many real-life applications. Despite their various benefits and potentials, there is a high demand for DL processing in different computationally limited and energy-constrained devices. It is natural to study game-changing technologies such as Binary Neural Networks (BNN) to increase DL capabilities. Recently remarkable progress has been made in BNN since they can be implemented and embedded on tiny restricted devices and save a significant amount of storage, computation cost, and energy consumption. However, nearly all BNN acts trade with extra memory, computation cost, and higher performance. This article provides a complete overview of recent developments in BNN. This article focuses exclusively on 1-bit activations and weights 1-bit convolution networks, contrary to previous surveys in which low-bit works are mixed in. It conducted a complete investigation of BNN's development—from their predecessors to the latest BNN algorithms/techniques, presenting a broad design pipeline and discussing each module's variants. Along the way, it examines BNN (a) purpose: their early successes and challenges; (b) BNN optimization: selected representative works that contain essential optimization techniques; (c) deployment: open-source frameworks for BNN modeling and development; (d) terminal: efficient computing architectures and devices for BNN and (e) applications: diverse applications with BNN. Moreover, this paper discusses potential directions and future research opportunities in each section. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
10. A Binary Neural Network with Dual Attention for Plant Disease Classification.
- Author
-
Ma, Ping, Zhu, Junan, and Zhang, Gan
- Subjects
PLANT classification ,NOSOLOGY ,PLANT identification ,AGRICULTURE ,COST control ,PLANT diseases ,COMPUTATIONAL neuroscience - Abstract
Plant disease control has long been a critical issue in agricultural production and relies heavily on the identification of plant diseases, but traditional disease identification requires extensive experience. Most of the existing deep learning-based plant disease classification methods run on high-performance devices to meet the requirements for classification accuracy. However, agricultural applications have strict cost control and cannot be widely promoted. This paper presents a novel method for plant disease classification using a binary neural network with dual attention (DABNN), which can save computational resources and accelerate by using binary neural networks, and introduces a dual-attention mechanism to improve the accuracy of classification. To evaluate the effectiveness of our proposed approach, we conduct experiments on the PlantVillage dataset, which includes a range of diseases. The F 1 s c o r e and A c c u r a c y of our method reach 99.39% and 99.4%, respectively. Meanwhile, compared to AlexNet and VGG16, the C o m p u t a t i o n a l c o m p l e x i t y of our method is reduced by 72.3% and 98.7%, respectively. The P a r a m s s i z e of our algorithm is 5.4% of AlexNet and 2.3% of VGG16. The experimental results show that DABNN can identify various diseases effectively and accurately. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
11. Study of Rescaling Mechanism Utilization in Binary Neural Networks
- Author
-
Zharikov, Ilia, Ovcharenko, Kirill, Kacprzyk, Janusz, Series Editor, Kryzhanovsky, Boris, editor, Dunin-Barkowski, Witali, editor, Redko, Vladimir, editor, Tiumentsev, Yury, editor, and Klimov, Valentin, editor
- Published
- 2023
- Full Text
- View/download PDF
12. An Interpretable Loan Credit Evaluation Method Based on Rule Representation Learner
- Author
-
Chen, Zihao, Wang, Xiaomeng, Huang, Yuanjiang, Jia, Tao, Filipe, Joaquim, Editorial Board Member, Ghosh, Ashish, Editorial Board Member, Prates, Raquel Oliveira, Editorial Board Member, Zhou, Lizhu, Editorial Board Member, Sun, Yuqing, editor, Lu, Tun, editor, Guo, Yinzhang, editor, Song, Xiaoxia, editor, Fan, Hongfei, editor, Liu, Dongning, editor, Gao, Liping, editor, and Du, Bowen, editor
- Published
- 2023
- Full Text
- View/download PDF
13. Binary Neural Network for Video Action Recognition
- Author
-
Han, Hongfeng, Lu, Zhiwu, Wen, Ji-Rong, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Dang-Nguyen, Duc-Tien, editor, Gurrin, Cathal, editor, Larson, Martha, editor, Smeaton, Alan F., editor, Rudinac, Stevan, editor, Dao, Minh-Son, editor, Trattner, Christoph, editor, and Chen, Phoebe, editor
- Published
- 2023
- Full Text
- View/download PDF
14. Energy-Efficient Image Processing Using Binary Neural Networks with Hadamard Transform
- Author
-
Park, Jaeyoon, Lee, Sunggu, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Wang, Lei, editor, Gall, Juergen, editor, Chin, Tat-Jun, editor, Sato, Imari, editor, and Chellappa, Rama, editor
- Published
- 2023
- Full Text
- View/download PDF
15. PBCStereo: A Compressed Stereo Network with Pure Binary Convolutional Operations
- Author
-
Cai, Jiaxuan, Qi, Zhi, Fu, Keqi, Shi, Xulong, Li, Zan, Liu, Xuanyu, Liu, Hao, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Wang, Lei, editor, Gall, Juergen, editor, Chin, Tat-Jun, editor, Sato, Imari, editor, and Chellappa, Rama, editor
- Published
- 2023
- Full Text
- View/download PDF
16. Deep Learning and Neural Architecture Search for Optimizing Binary Neural Network Image Super Resolution
- Author
-
Yuanxin Su, Li-minn Ang, Kah Phooi Seng, and Jeremy Smith
- Subjects
deep learning ,neural architecture search ,binary neural network ,image super resolution ,Technology - Abstract
The evolution of super-resolution (SR) technology has seen significant advancements through the adoption of deep learning methods. However, the deployment of such models by resource-constrained devices necessitates models that not only perform efficiently, but also conserve computational resources. Binary neural networks (BNNs) offer a promising solution by minimizing the data precision to binary levels, thus reducing the computational complexity and memory requirements. However, for BNNs, an effective architecture is essential due to their inherent limitations in representing information. Designing such architectures traditionally requires extensive computational resources and time. With the advancement in neural architecture search (NAS), differentiable NAS has emerged as an attractive solution for efficiently crafting network structures. In this paper, we introduce a novel and efficient binary network search method tailored for image super-resolution tasks. We adapt the search space specifically for super resolution to ensure it is optimally suited for the requirements of such tasks. Furthermore, we incorporate Libra Parameter Binarization (Libra-PB) to maximize information retention during forward propagation. Our experimental results demonstrate that the network structures generated by our method require only a third of the parameters, compared to conventional methods, and yet deliver comparable performance.
- Published
- 2024
- Full Text
- View/download PDF
17. R-inmac: 10T SRAM based reconfigurable and efficient in-memory advance computation for edge devices.
- Author
-
Dhakad, Narendra Singh, Chittora, Eshika, Sharma, Vishal, and Vishvakarma, Santosh Kumar
- Subjects
STATIC random access memory ,BINARY operations ,ADDITION (Mathematics) ,COMPUTER systems - Abstract
This paper proposes a Reconfigurable In-Memory Advance Computing architecture using a novel 10 SRAM cell. In addition to basic logic operations, the proposed R-InMAC can also implement complex Boolean computing operations such as binary addition/subtraction, binary-to-gray, gray-to-binary conversion, 2's complement, less/greater than, and increment/decrement. Furthermore, content addressable memory (CAM) operation to search a binary string in a memory array is also proposed efficiently. It can search true and complementary data strings in a single cycle. The proposed R-InMAC architecture's reconfigurability allows it to be configured according to the needed operation and bit precision, making it ideal and energy-efficient. In addition, compared to the standard SRAM cells, the proposed 10T cell is suited for implementing the XNOR-based binary convolution operation required in Binary Neural Networks (BNNs) with improved latency of 58.89%. The optimized full adder of the proposed R-InMAC shows decrement in the area by 40%, static power by 28%, dynamic power by 55.2%, and latency by 25.3% as compared to conventional designs, making this work a promising candidate for modern edge AI compute in-memory systems. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
18. Regularizing Binary Neural Networks via Ensembling for Efficient Person Re-Identification
- Author
-
Ayse Serbetci and Yusuf Sinan Akgul
- Subjects
Binary neural network ,network regularization ,hash retrieval ,ensemble learning ,person re-identification ,Electrical engineering. Electronics. Nuclear engineering ,TK1-9971 - Abstract
This study aims to leverage Binary Neural Networks (BNN) to learn binary hash codes for efficient person re-identification (ReID). BNNs, which use binary weights and activations, show promise in speeding up the inference time in deep models. However, BNNs typically suffer from performance degradation mainly due to the discontinuity of the binarization operation. Proxy functions have been proposed to calculate the gradients in the backward propagation, but they lead to the gradient mismatch problem. In this study, we propose to address the gradient mismatch problem by designing a multi-branch ensemble model consisting of many weak hash code learners. Specifically, our design aggregates the gradients from multiple branches, which allows a better approximation of the gradients and regularizes the network. Our model adds little computational cost to the baseline BNN since a vast amount of network parameters are shared between the weak learners. Combining the efficiency of the BNNs and hash code learning, we obtain an effective ensemble model which is efficient both in feature extraction and ranking phases. Our experiments demonstrate that the proposed model outperforms a single BNN by more than %20 using nearly the same amount of floating point operations. Moreover, the proposed model outperforms a conventional ensemble of BNN by more than %7 while being nearly 10x and 2x more efficient in terms of CPU consumption and memory footprint, respectively. We explore the performance of BNNs for efficient person ReID as one of the first systems available in the literature. Moreover, we adopt the proposed ensemble model for further validation of the image classification task and observe that our method effectively regularizes BNNs, providing robustness to hyperparameter selection and producing more consistent results under different settings.
- Published
- 2023
- Full Text
- View/download PDF
19. A Systematic Literature Review on Binary Neural Networks
- Author
-
Ratshih Sayed, Haytham Azmi, Heba Shawkey, A. H. Khalil, and Mohamed Refky
- Subjects
Binary neural network ,convolutional neural network ,deep learning ,optimization approaches ,quantization ,systematic literature review ,Electrical engineering. Electronics. Nuclear engineering ,TK1-9971 - Abstract
This paper presents an extensive literature review on Binary Neural Network (BNN). BNN utilizes binary weights and activation function parameters to substitute the full-precision values. In digital implementations, BNN replaces the complex calculations of Convolutional Neural Networks (CNNs) with simple bitwise operations. BNN optimizes large computation and memory storage requirements, which leads to less area and power consumption compared to full-precision models. Although there are many advantages of BNN, the binarization process has a significant impact on the performance and accuracy of the generated models. To reflect the state-of-the-art in BNN and explore how to develop and improve BNN-based models, we conduct a systematic literature review on BNN with data extracted from 239 research studies. Our review discusses various BNN architectures and the optimization approaches developed to improve their performance. There are three main research directions in BNN: accuracy optimization, compression optimization, and acceleration optimization. The accuracy optimization approaches include quantization error reduction, special regularization, gradient error minimization, and network structure. The compression optimization approaches combine fractional BNN and pruning. The acceleration optimization approaches comprise computing in-memory, FPGA-based implementations, and ASIC-based implementations. At the end of our review, we present a comprehensive analysis of BNN applications and their evaluation metrics. Also, we shed some light on the most common BNN challenges and the future research trends of BNN.
- Published
- 2023
- Full Text
- View/download PDF
20. Double Quantification of Template and Network for Palmprint Recognition.
- Author
-
Lin, Qizhou, Leng, Lu, and Kim, Cheonshik
- Subjects
PALMPRINT recognition ,HAMMING distance ,BINARY codes ,BIOMETRY - Abstract
The outputs of deep hash network (DHN) are binary codes, so DHN has high retrieval efficiency in matching phase and can be used for high-speed palmprint recognition, which is a promising biometric modality. In this paper, the templates and network parameters are both quantized for fast and light-weight palmprint recognition. The parameters of DHN are binarized to compress the network weight and accelerate the speed. To avoid accuracy degradation caused by quantization, mutual information is leveraged to optimize the ambiguity in Hamming space to obtain a tri-valued hash code as a palmprint template. Kleene Logic's tri-valued Hamming distance measures the dissimilarity between palmprint templates. The ablation experiments are tested on the binarization of the network parameter, and the normalization and trivialization of the deep hash output value. The sufficient experiments conducted on several contact and contactless palmprint datasets confirm the multiple advantages of our method. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
21. Recurrent Bilinear Optimization for Binary Neural Networks
- Author
-
Xu, Sheng, Li, Yanjing, Wang, Tiancheng, Ma, Teli, Zhang, Baochang, Gao, Peng, Qiao, Yu, Lü, Jinhu, Guo, Guodong, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Avidan, Shai, editor, Brostow, Gabriel, editor, Cissé, Moustapha, editor, Farinella, Giovanni Maria, editor, and Hassner, Tal, editor
- Published
- 2022
- Full Text
- View/download PDF
22. Towards Accurate Binary Neural Networks via Modeling Contextual Dependencies
- Author
-
Xing, Xingrun, Li, Yangguang, Li, Wei, Ding, Wenrui, Jiang, Yalong, Wang, Yufeng, Shao, Jing, Liu, Chunlei, Liu, Xianglong, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Avidan, Shai, editor, Brostow, Gabriel, editor, Cissé, Moustapha, editor, Farinella, Giovanni Maria, editor, and Hassner, Tal, editor
- Published
- 2022
- Full Text
- View/download PDF
23. Approximations in Deep Learning
- Author
-
Dupuis, Etienne, Filip, Silviu, Sentieys, Olivier, Novo, David, O’Connor, Ian, Bosio, Alberto, Bosio, Alberto, editor, Ménard, Daniel, editor, and Sentieys, Olivier, editor
- Published
- 2022
- Full Text
- View/download PDF
24. S[formula omitted]NN: Time step reduction of spiking surrogate gradients for training energy efficient single-step spiking neural networks.
- Author
-
Suetake, Kazuma, Ikegawa, Shin-ichi, Saiin, Ryuji, and Sawada, Yoshihide
- Subjects
- *
ARTIFICIAL neural networks , *ENERGY consumption - Abstract
As the scales of neural networks increase, techniques that enable them to run with low computational cost and energy efficiency are required. From such demands, various efficient neural network paradigms, such as spiking neural networks (SNNs) or binary neural networks (BNNs), have been proposed. However, they have sticky drawbacks, such as degraded inference accuracy and latency. To solve these problems, we propose a single-step spiking neural network (S 3 NN), an energy-efficient neural network with low computational cost and high precision. The proposed S 3 NN processes the information between hidden layers by spikes as SNNs. Nevertheless, it has no temporal dimension so that there is no latency within training and inference phases as BNNs. Thus, the proposed S 3 NN has a lower computational cost than SNNs that require time-series processing. However, S 3 NN cannot adopt naïve backpropagation algorithms due to the non-differentiability nature of spikes. We deduce a suitable neuron model by reducing the surrogate gradient for multi-time step SNNs to a single-time step. We experimentally demonstrated that the obtained surrogate gradient allows S 3 NN to be trained appropriately. We also showed that the proposed S 3 NN could achieve comparable accuracy to full-precision networks while being highly energy-efficient. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
25. Distribution-Sensitive Information Retention for Accurate Binary Neural Network.
- Author
-
Qin, Haotong, Zhang, Xiangguo, Gong, Ruihao, Ding, Yifu, Xu, Yi, and Liu, Xianglong
- Subjects
- *
DEEP learning , *DISTILLATION , *STANDARDIZATION , *EMPIRICAL research - Abstract
Model binarization is an effective method of compressing neural networks and accelerating their inference process, which enables state-of-the-art models to run on resource-limited devices. Recently, advanced binarization methods have been greatly improved by minimizing the quantization error directly in the forward process. However, a significant performance gap still exists between the 1-bit model and the 32-bit one. The empirical study shows that binarization causes a great loss of information in the forward and backward propagation which harms the performance of binary neural networks (BNNs). We present a novel distribution-sensitive information retention network (DIR-Net) that retains the information in the forward and backward propagation by improving internal propagation and introducing external representations. The DIR-Net mainly relies on three technical contributions: (1) Information Maximized Binarization (IMB): minimizing the information loss and the binarization error of weights/activations simultaneously by weight balance and standardization; (2) Distribution-sensitive Two-stage Estimator (DTE): retaining the information of gradients by distribution-sensitive soft approximation by jointly considering the updating capability and accurate gradient; (3) Representation-align Binarization-aware Distillation (RBD): retaining the representation information by distilling the representations between full-precision and binarized networks. The DIR-Net investigates both forward and backward processes of BNNs from the unified information perspective, thereby providing new insight into the mechanism of network binarization. The three techniques in our DIR-Net are versatile and effective and can be applied in various structures to improve BNNs. Comprehensive experiments on the image classification and objective detection tasks show that our DIR-Net consistently outperforms the state-of-the-art binarization approaches under mainstream and compact architectures, such as ResNet, VGG, EfficientNet, DARTS, and MobileNet. Additionally, we conduct our DIR-Net on real-world resource-limited devices which achieves 11.1 × storage saving and 5.4 × speedup. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
26. Decomposition Method for Calculating the Weights of a Binary Neural Network.
- Author
-
Litvinenko, A., Kucherov, D., and Glybovets, M.
- Subjects
- *
DECOMPOSITION method , *ALGORITHMS - Abstract
A method for determining the weights of a binary neural network based on its decomposition into elementary modules is presented. The approach allows tuning the weight coefficients of all the network connections at the stage of its designing, which eliminates the implementation of time-consuming iterative algorithms for training the network during its operation. An algorithm and an example of calculating the weights are given. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
27. 'Ghost' and Attention in Binary Neural Network
- Author
-
Ruimin Sun, Wanbing Zou, and Yi Zhan
- Subjects
Binary neural network ,ghost feature map ,attention ,category information ,Electrical engineering. Electronics. Nuclear engineering ,TK1-9971 - Abstract
As the memory footprint requirement and computational scale concerned, the light-weighted Binary Neural Networks (BNNs) have great advantages in limited-resources platforms, such as AIoT (Artificial Intelligence in Internet of Things) edge terminals, wearable and portable devices, etc. However, the binarization process naturally brings considerable information losses and further deteriorates the accuracy. In this article, three aspects are introduced to better the binarized ReActNet accuracy performance with a more low-complex computation. Firstly, an improved Binarized Ghost Module (BGM) for the ReActNet is proposed to increase the feature maps information. At the same time, the computational scale of this structure is still kept at a very low level. Secondly, we propose a new Label-aware Loss Function (LLF) in the penultimate layer as a supervisor which takes the label information into consideration. This auxiliary loss function makes each category’s feature vectors more separate, and improve the final fully-connected layer’s classification accuracy accordingly. Thirdly, the Normalization-based Attention Module (NAM) method is adopted to regulate the activation flow. The module helps to avoid the gradient saturation problem. With these three approaches, our improved binarized network outperforms the other state-of-the-art methods. It can achieve 71.4% Top-1 accuracy on the ImageNet and 86.45% accuracy on the CIFAR-10 respectively. Meanwhile, its computational scale OPs is the least $0.86\times {10}^{8}$ compared with the other mainstream BNN models. The experimental results prove the effectiveness of our proposals, and the study is very helpful and promising for the future low-power hardware implementations.
- Published
- 2022
- Full Text
- View/download PDF
28. RAD-BNN: Regulating activation distribution for accurate binary neural network.
- Author
-
Yuan, Mingyu and Pei, Songwei
- Subjects
- *
CONVOLUTIONAL neural networks , *ENTROPY (Information theory) , *NETWORK performance , *LEARNING ability - Abstract
The outstanding performance of deep convolutional neural networks comes from their effective extraction and learning ability. Although binary neural networks (BNNs) have the obvious advantages of low storage and high efficiency over their full-precision counterparts on resource-constrained hardware devices, the accuracy degradation brought by binary quantization is still an unavoidable problem. The activation distribution in BNNs is a key factor affecting network performance. To elevate the accuracy of BNNs, in this paper, we propose to regulate the activation distribution to strengthen the representation ability of BNNs. We first propose an Information Entropy enhancement Basic block (IEBlock) to build a competitive baseline model with higher information entropy of output activation distribution. Specifically, we build the IEBlock by deliberately reorganizing the position of the elements in the normal basic block based on a deep analysis of the information flow. After that, we propose a Depth-aware Activation Distribution Amendment (DADA) module, which learns the interdependencies of feature channels to amend the activation distribution with information loss after binary convolution. Extensive experiments demonstrate that our method effectively improves the information entropy of binary activations and elevates the accuracy of BNNs. Our method has outperformed the state-of-the-art methods on CIFAR-10 and ImageNet datasets. Code is available at: https://github.com/tomorrow-rain/RAD-BNN • A BNN is implemented considering the information entropy of activations. • Depth-aware channel dependencies are built to regulate the activations. • SOTA results are achieved on CIFAR-10 and ImageNet datasets with the BNNs. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
29. Analysing the Adversarial Landscape of Binary Stochastic Networks
- Author
-
Tan, Yi Xiang Marcus, Elovici, Yuval, Binder, Alexander, Angrisani, Leopoldo, Series Editor, Arteaga, Marco, Series Editor, Panigrahi, Bijaya Ketan, Series Editor, Chakraborty, Samarjit, Series Editor, Chen, Jiming, Series Editor, Chen, Shanben, Series Editor, Chen, Tan Kay, Series Editor, Dillmann, Rüdiger, Series Editor, Duan, Haibin, Series Editor, Ferrari, Gianluigi, Series Editor, Ferre, Manuel, Series Editor, Hirche, Sandra, Series Editor, Jabbari, Faryar, Series Editor, Jia, Limin, Series Editor, Kacprzyk, Janusz, Series Editor, Khamis, Alaa, Series Editor, Kroeger, Torsten, Series Editor, Liang, Qilian, Series Editor, Martín, Ferran, Series Editor, Ming, Tan Cher, Series Editor, Minker, Wolfgang, Series Editor, Misra, Pradeep, Series Editor, Möller, Sebastian, Series Editor, Mukhopadhyay, Subhas, Series Editor, Ning, Cun-Zheng, Series Editor, Nishida, Toyoaki, Series Editor, Pascucci, Federica, Series Editor, Qin, Yong, Series Editor, Seng, Gan Woon, Series Editor, Speidel, Joachim, Series Editor, Veiga, Germano, Series Editor, Wu, Haitao, Series Editor, Zhang, Junjie James, Series Editor, Kim, Hyuncheol, editor, Kim, Kuinam J., editor, and Park, Suhyun, editor
- Published
- 2021
- Full Text
- View/download PDF
30. RB-Net: Training Highly Accurate and Efficient Binary Neural Networks With Reshaped Point-Wise Convolution and Balanced Activation.
- Author
-
Liu, Chunlei, Ding, Wenrui, Chen, Peng, Zhuang, Bohan, Wang, Yufeng, Zhao, Yang, Zhang, Baochang, and Han, Yuqi
- Subjects
- *
COMPUTATIONAL complexity - Abstract
In this paper, we find that the conventional convolution operation becomes the bottleneck for extremely efficient binary neural networks (BNNs). To address this issue, we open up a new direction by introducing a reshaped point-wise convolution (RPC) to replace the conventional one to build BNNs. Specifically, we conduct a point-wise convolution after rearranging the spatial information into depth, with which at least $2.25\times $ computation reduction can be achieved. Such an efficient RPC allows us to explore more powerful representational capacity of BNNs under a given computation complexity budget. Moreover, we propose to use a balanced activation (BA) to adjust the distribution of the scaled activations after binarization, which enables significant performance improvement of BNNs. After integrating RPC and BA, the proposed network, dubbed as RB-Net, strikes a good trade-off between accuracy and efficiency, achieving superior performance with lower computational cost against the state-of-the-art BNN methods. Specifically, our RB-Net achieves 66.8% Top-1 accuracy with ResNet-18 backbone on ImageNet, exceeding the state-of-the-art Real-to-Binary Net (65.4%) by 1.4% while achieving more than $3\times $ reduction (52M vs. 165M) in computational complexity. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
31. PB-GCN: Progressive binary graph convolutional networks for skeleton-based action recognition.
- Author
-
Zhao, Mengyi, Dai, Shuling, Zhu, Yanjun, Tang, Hao, Xie, Pan, Li, Yue, Liu, Chunlei, and Zhang, Baochang
- Subjects
- *
MEMORY - Abstract
Skeleton-based action recognition is an essential yet challenging visual task, whose accuracy has been remarkably improved due to the successful application of graph convolutional networks (GCNs). However, high computation cost and memory usage hinder their deployment on resource-constrained environment. To deal with the issue, in this paper, we introduce two novel progressive binary graph convolutional network for skeleton-based action recognition PB-GCN and PB-GCN * , which can obtain significant speed-up and memory saving. In PB-GCN, the filters are binarized, and in PB-GCN * , both filters and activations are binary. Specifically, we propose a progressive optimization, i.e., employing ternary models as the initialization of binary GCNs (BGCN) to improve the representational capability of binary models. Moreover, the center loss is exploited to improve the training procedure for better performance. Experimental results on two public benchmarks (i.e., Skeleton-Kinetics and NTU RGB + D) demonstrate that the accuracy of the proposed PB-GCN and PB-GCN * are comparable to their full-precision counterparts and outperforms the state-of-the-art methods, such as BWN, XNOR-Net, and Bi-Real Net. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
32. Synaptic Tunnel Field-Effect Transistors for Extremely-Low-Power Operation.
- Author
-
Lee, Jang Woo, Woo, Jae Seung, and Choi, Woo Young
- Subjects
TUNNEL field-effect transistors ,BINARY operations - Abstract
A synaptic cell composed of two tunnel field-effect transistors (TFETs) which is capable of XNOR operation for binary neural networks has been experimentally demonstrated. Our proposed synaptic TFETs feature lower current during inference and higher programming efficiency during weight transfer than conventional synaptic transistors. Moreover, the fabricated synaptic TFET arrays satisfy the neurobiological energy requirement $(\!\sim 10$ fJ per synaptic event) and low bit-error rate of $6.7\times10$ −7%. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
33. A Lightweight Collaborative Deep Neural Network for the Mobile Web in Edge Cloud.
- Author
-
Huang, Yakun, Qiao, Xiuquan, Ren, Pei, Liu, Ling, Pu, Calton, Dustdar, Schahram, and Chen, Junliang
- Subjects
ARTIFICIAL intelligence ,MOBILE learning ,EDGE computing ,MOBILE computing ,COMPUTING platforms ,DEEP learning - Abstract
Enabling deep learning technology on the mobile web can improve the user’s experience for achieving web artificial intelligence in various fields. However, heavy DNN models and limited computing resources of the mobile web are now unable to support executing computationally intensive DNNs when deploying in a cloud computing platform. With the help of promising edge computing, we propose a lightweight collaborative deep neural network for the mobile web, named LcDNN, which contributes to three aspects: (1) We design a composite collaborative DNN that reduces the model size, accelerates inference, and reduces mobile energy cost by executing a lightweight binary neural network (BNN) branch on the mobile web. (2) We provide a jointly training method for LcDNN and implement an energy-efficient inference library for executing the BNN branch on the mobile web. (3) To further promote the resource utilization of the edge cloud, we develop a DRL-based online scheduling scheme to obtain an optimal allocation for LcDNN. The experimental results show that LcDNN outperforms existing approaches for reducing the model size by about 16x to 29x. It also reduces the end-to-end latency and mobile energy cost with acceptable accuracy and improves the throughput and resource utilization of the edge cloud. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
34. Efficient Approximation of Filters for High-Accuracy Binary Convolutional Neural Networks
- Author
-
Park, Junyong, Moon, Yong-Hyuk, Lee, Yong-Ju, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Woeginger, Gerhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Bartoli, Adrien, editor, and Fusiello, Andrea, editor
- Published
- 2020
- Full Text
- View/download PDF
35. Binarized Neural Network for Single Image Super Resolution
- Author
-
Xin, Jingwei, Wang, Nannan, Jiang, Xinrui, Li, Jie, Huang, Heng, Gao, Xinbo, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Woeginger, Gerhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Vedaldi, Andrea, editor, Bischof, Horst, editor, Brox, Thomas, editor, and Frahm, Jan-Michael, editor
- Published
- 2020
- Full Text
- View/download PDF
36. Performance analysis and modeling of bio-hydrogen recovery from agro-industrial wastewater
- Author
-
SK Safdar Hossain, Syed Sadiq Ali, Chin Kui Cheng, and Bamidele Victor Ayodele
- Subjects
agro-industrial wastewater ,support vector machine ,Gaussian process regression ,binary neural network ,bio-hydrogen ,General Works - Abstract
Significant volumes of wastewater are routinely generated during agro-industry processing, amounting to millions of tonnes annually. In line with the circular economy concept, there could be a possibility of simultaneously treating the wastewater and recovering bio-energy resources such as bio-hydrogen. This study aimed to model the effect of different process parameters that could influence wastewater treatment and bio-energy recovery from agro-industrial wastewaters. Three agro-industrial wastewaters from dairy, chicken processing, and palm oil mills were investigated. Eight data-driven machine learning algorithms namely linear support vector machine (LSVM), quadratic support vector machine (QSVM), cubic support vector machine (CSVM), fine Gaussian support vector machine (FGSVM), binary neural network (BNN), rotation quadratic Gaussian process regression (RQGPR), exponential quadratic Gaussian process regression (EQGPR) and exponential Gaussian process regression (EGPR) were employed for the modeling process. The datasets obtained from the three agro-industrial processes were employed to train and test the models. The LSVM, QSVM, and CSVM did not show an impressive performance as indicated by the coefficient of determination (R2) < 0.7 for the prediction of hydrogen produced from wastewaters using the three agro-industrial processes. The LSVM, QSVM, and CSVM models were also characterized by high prediction errors. Superior performance was displayed by FGSVM, BNN, RQGPR, EQGPR, and EQGPR models as indicated by the high R2 > 0.9, an indication of better predictability with minimized prediction errors as indicated by the low root mean square error (RMSE), mean square error (MSE), and mean absolute error (MAE).
- Published
- 2022
- Full Text
- View/download PDF
37. Distillation-Guided Residual Learning for Binary Convolutional Neural Networks.
- Author
-
Ye, Jianming, Wang, Jingdong, and Zhang, Shiliang
- Subjects
- *
CONVOLUTIONAL neural networks , *DISTILLATION - Abstract
It is challenging to bridge the performance gap between binary convolutional neural network (BCNN) and floating-point CNN (FCNN). This performance gap is mainly caused by the inferior modeling capability and training strategy of BCNN, which leads to substantial residuals in intermediate feature maps between BCNN and FCNN. To minimize the performance gap, we enforce BCNN to produce similar intermediate feature maps with the ones of FCNN. This intuition leads to a more effective training strategy for BCNN, i.e., optimizing each binary convolutional block with blockwise distillation loss derived from FCNN. The goal of minimizing the residuals in intermediate feature maps also motivates us to update the binary convolutional block architecture to facilitate the optimization of blockwise distillation loss. Specifically, a lightweight shortcut branch is inserted into each binary convolutional block to complement residuals at each block. Benefited from its squeeze-and-interaction (SI) structure, this shortcut branch introduces a fraction of parameters, e.g., less than 10% overheads, but effectively boosts the modeling capability of binary convolution blocks in BCNN. Extensive experiments on ImageNet demonstrate the superior performance of our method in both classification efficiency and accuracy, e.g., BCNN trained with our methods achieves the accuracy of 60.45% on ImageNet, better than many state-of-the-art ones. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
38. BR-CIM: An Efficient Binary Representation Computation-In-Memory Design.
- Author
-
Yue, Zhiheng, Wang, Yabing, Qin, Yubin, Liu, Leibo, Wei, Shaojun, and Yin, Shouyi
- Subjects
- *
RANDOM access memory , *BINARY operations , *COMPOSITE columns , *ENERGY consumption , *COMPUTER architecture , *ARTIFICIAL intelligence - Abstract
Deep neural network (DNN) has recently attracted tremendous attention in various fields. But the computing operation requirement and the memory bottleneck limit the energy efficiency of hardware implementations. Binary quantization is proposed to relieve the pressure of hardware design. And the Computing-In-Memory (CIM) is regarded as a promising method to resolve the memory wall challenge. However, the binary computing paradigm is mismatched with the CIM scheme, which incurs complex circuits and peripheral to realize binary operation in previous works. To overcome previous issues, this work presents Binary Representation Computation-In-Memory (BR-CIM) with several key features. (1) A lightweight computation unit is realized within the 6T SRAM array to accelerate binary computing and enlarge signal margin; (2) The reconfigurable computing scheme and mapping method support extendable bit precision to satisfy the accuracy requirement of various applications; (3) Simultaneous computing and weight loading is supported by column circuitry, which shortens the data loading latency; Several experiments are conducted to estimate algorithm accuracy, the computing latency, and power consumption. The energy efficiency reaches up to 1280 TOPs/W for binary representation. And the algorithm accuracy achieves 97.82%/76.4% on MNIST/CIFAR-100 dataset. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
39. Trainable Communication Systems Based on the Binary Neural Network
- Author
-
Bo Che, Xinyi Li, Zhi Chen, and Qi He
- Subjects
autoencoder ,end-to-end learning ,communication system ,binary neural network ,low-complexity ,Communication. Mass media ,P87-96 - Abstract
End-to-end learning of the communication system regards the transmitter, channel, and receiver as a neural network-based autoencoder. This approach enables joint optimization of both the transmitter and receiver and can learn to communicate more efficiently than model-based ones. Despite the achieved success, high complexity is the major disadvantage that hinders its further development, while low-precision compression such as one-bit quantization is an effective solution. This study proposed an autoencoder communication system composed of binary neural networks (BNNs), which is based on bit operations and has a great potential to be applied to hardware platforms with very limited computing resources such as FPGAs. Several modifications are explored to further improve the performance. Experiments showed that the proposed BNN-based system can achieve a performance similar to that of the existing neural network-based autoencoder systems while largely reducing the storage and computation complexities.
- Published
- 2022
- Full Text
- View/download PDF
40. Maximizing Parallel Activation of Word-Lines in MRAM-Based Binary Neural Network Accelerators
- Author
-
Daehyun Ahn, Hyunmyung Oh, Hyungjun Kim, Yulhwa Kim, and Jae-Joon Kim
- Subjects
Magnetic RAM ,binary neural network ,device variation ,in-memory computing ,Electrical engineering. Electronics. Nuclear engineering ,TK1-9971 - Abstract
Magnetic RAM (MRAM)-based crossbar array has a great potential as a platform for in-memory binary neural network (BNN) computing. However, the number of word-lines that can be activated simultaneously is limited because of the low $I_{H}/I_{L}$ ratio of MRAM, which makes BNNs more vulnerable to the device variation. To address this issue, we propose an algorithm/hardware co-design methodology. First, we choose a promising memristor crossbar array (MCA) structure based on the sensitivity analysis to process variations. Since the selected MCA structure becomes more tolerant to the device variation when the number of 1 in input activation values decreases, we apply an input distribution regularization scheme to reduce the number of 1 in input of BNNs during training. We further improve the robustness against device variation by adopting the retraining scheme based on knowledge distillation. Experimental results show that the proposed method makes BNNs more tolerant to MRAM variation and increases the number of parallel word-line activation significantly; thereby achieving improved throughput and energy efficiency.
- Published
- 2021
- Full Text
- View/download PDF
41. SiO 2 Fin-Based Flash Synaptic Cells in AND Array Architecture for Binary Neural Networks.
- Author
-
Lee, Soochang, Kim, Hyeongsu, Lee, Sung-Tae, Park, Byung-Gook, and Lee, Jong-Ho
- Subjects
TISSUE arrays ,COMPUTER storage devices ,ELECTRIC fields ,LOW voltage systems ,FLASH memory ,COMPUTER architecture - Abstract
An oxide fin-based AND flash memory synaptic device is proposed and fabricated using a spacer patterning technology for a hardware-based binary neural network (BNN). A fin-like curved channel structure provides local electric field enhancement, which improves programming efficiency compared to planar-type flash synaptic devices. The fin-based AND flash cell exhibits a high on/off current ratio (>105) with sub-pA off current, and a low programming voltage (< 9 V) is used to achieve a sufficient dynamic range of synaptic weights (>103) for BNNs. Furthermore, a hardware-based BNN using novel two cell-based synaptic devices arranged in AND array architecture is proposed to implement parallel XNOR operation and bit-counting. Proposed BNN using the synapse model with measured dynamic range and retention property shows only < 0.5 % degradation of classification accuracy compared to the baseline accuracy, which is suitable to perform off-chip event-driven computation using parallel read-out operations. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
42. TP-ADMM: An Efficient Two-Stage Framework for Training Binary Neural Networks
- Author
-
Yuan, Yong, Chen, Chen, Hu, Xiyuan, Peng, Silong, Barbosa, Simone Diniz Junqueira, Editorial Board Member, Filipe, Joaquim, Editorial Board Member, Ghosh, Ashish, Editorial Board Member, Kotenko, Igor, Editorial Board Member, Zhou, Lizhu, Editorial Board Member, Gedeon, Tom, editor, Wong, Kok Wai, editor, and Lee, Minho, editor
- Published
- 2019
- Full Text
- View/download PDF
43. Human Body Posture Recognition Using Wearable Devices
- Author
-
Liu, Junxiu, Li, Mingxing, Luo, Yuling, Yang, Su, Qiu, Senhui, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Woeginger, Gerhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Tetko, Igor V., editor, Kůrková, Věra, editor, Karpov, Pavel, editor, and Theis, Fabian, editor
- Published
- 2019
- Full Text
- View/download PDF
44. An Energy-Efficient and High Throughput in-Memory Computing Bit-Cell With Excellent Robustness Under Process Variations for Binary Neural Network
- Author
-
Gobinda Saha, Zhewei Jiang, Sanjay Parihar, Cao Xi, Jack Higman, and Muhammed Ahosan Ul Karim
- Subjects
In-memory computing ,SRAM ,binary neural network ,nonideality ,process variation ,Electrical engineering. Electronics. Nuclear engineering ,TK1-9971 - Abstract
In-memory computing (IMC) is a promising approach for energy cost reduction due to data movement between memory and processor for running data-intensive deep learning applications on the computing systems. Together with Binary Neural Network (BNN), IMC provides a viable solution for running deep neural networks at the edge devices with stringent memory and energy constraints. In this paper, we propose a novel 10T bit-cell with a back-end-of-line (BEOL) metal-oxide-metal (MOM) capacitor laid on pitch for in-memory computing. Our IMC bit-cell, when arranged in a memory array, performs binary convolution (XNOR followed by Bit-count operations) and binary activation generation operations. We show, when binary layers of BNN are mapped into our IMC arrays for MNIST digit classification, 98.75% accuracy with energy efficiency of 2193 TOPS/W and throughput of 22857 GOPS can be obtained. We determine the memory array size considering the word-line and bit-line nonidealities and show how these impact classification accuracy. We analyze the impact of process variations on classification accuracy and show how word-line pulse tunability provided by our design can be used to improve the robustness of classification under process variations.
- Published
- 2020
- Full Text
- View/download PDF
45. Binary Neural Network
- Author
-
Jaeger, Dieter, editor and Jung, Ranu, editor
- Published
- 2022
- Full Text
- View/download PDF
46. A Time-Domain Binary CNN Engine With Error-Detection-Based Resilience in 28nm CMOS.
- Author
-
Cai, Zhikuang, Cheng, Boyang, Du, Yuxuan, Shang, Xinchao, and Shan, Weiwei
- Abstract
Due to the increasing demand of high energy-efficient processor for deep neural networks, traditional neural network engines with high-precision weights and activations that usually occupies huge on/off-chip resources with large power consumption are no longer suitable for Internet-of-Things applications. Binary neural networks (BNNs) reduce memory size and computation complexity, achieving drastically increased energy efficiency. In this brief, an energy-efficient time-domain binary neural network engine is optimized for image recognition, with time-domain accumulation (TD-MAC), timing error detection based adaptive voltage scaling design and the related approximate computing. The proposed key features are: 1) an error-tolerant adaptive voltage scaling system with TD-MAC chain truncation for aggressive power reduction, working from near-threshold to normal voltage; 2) architectural parallelism and data reuse with 100% TD-MAC utilization; 3) low power TD-MAC based on analog delay lines. Fabricated in a 28nm CMOS process, the whole system achieves a maximum 51.5TOPS/W energy efficiency at 0.42V and 25MHz, with 99.6% accuracy on MNIST dataset. When the length of the TD-MAC chain is truncated by configuration, with a 90% accuracy on MNIST and a 150MHz, the proposed BNN achieves a power-saving of 13.2% and a further energy efficiency increasing of 67.6%. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
47. E2BNet: MAC-free yet accurate 2-level binarized neural network accelerator for embedded systems.
- Author
-
Mirsalari, Seyed Ahmad, Nazari, Najmeh, Ansarmohammadi, Seyed Ali, Salehi, Mostafa E., and Ghiasi, Soheil
- Abstract
Deep neural networks are widely used in computer vision, pattern recognition, and speech recognition and achieve high accuracy at the cost of remarkable computation. High computational complexity and memory accesses of such networks create a big challenge for using them in resource-limited and low-power embedded systems. Several binary neural networks have been proposed that exploit only 1-bit values for both weights and activations. Binary neural networks substitute complex multiply-accumulation operations with bitwise logic operations to reduce computations and memory usage. However, these quantized neural networks suffer from accuracy loss, especially in big datasets. In this paper, we introduce a quantized neural network with 2-bit weights and activations that is more accurate compared to the state-of-the-art quantized neural networks, and also the accuracy is close to the full precision neural networks. Moreover, we propose E2BNet, an efficient MAC-free hardware architecture that increases power efficiency and throughput/W about 3.6 × and 1.5 × , respectively, compared to the state-of-the-art quantized neural networks. E2BNet processes more than 500 images/s on the ImageNet dataset that not only meet real-time requirements of images/video processing but also can be deployed on high frame rate video applications. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
48. MLUTNet: A Neural Network for Memory Based Reconfigurable Logic Device Architecture.
- Author
-
Zang, Xuechen and Nakatake, Shigetoshi
- Subjects
LOGIC devices ,PROGRAMMABLE logic devices ,MEMORY ,HARDWARE ,FEATURE selection - Abstract
Neural networks have been widely used and implemented on various hardware platforms, but high computational costs and low similarity of network structures relative to hardware structures are often obstacles to research. In this paper, we propose a novel neural network in combination with the structural features of a recently proposed memory-based programmable logic device, compare it with the standard structure, and test it on common datasets with full and binary precision, respectively. The experimental results reveal that the new structured network can provide almost consistent full-precision performance and binary-precision performance ranging from 61.0% to 78.8% after using sparser connections and about 50% reduction in the size of the weight matrix. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
49. A 55nm, 0.4V 5526-TOPS/W Compute-in-Memory Binarized CNN Accelerator for AIoT Applications.
- Author
-
Zhang, Hongtu, Shu, Yuhao, Jiang, Weixiong, Yin, Zihan, Zhao, Wenfeng, and Ha, Yajun
- Abstract
Binarized convolutional neural network (BCNN) is a promising and efficient technique toward the landscape of Artificial Intelligence of Things (AIoT) applications. In-Memory Computing (IMC) has widely been studied to accelerate the inference task of BCNN to maximize both throughput and energy efficiency. However, existing IMC circuits and architectures are only optimized for a fixed kernel size and nominal voltage operation, which poses practical limitations on optimal network architecture exploration and additional energy efficiency benefits. In this brief, we present a reconfigurable, near-threshold IMC-based BCNN accelerator design. The IMC-based accelerator architecture is scalable for different kernels sizes (3 × 3 × d and 5 × 5 × d) and achieves high resource utilization for both cases. Moreover, the IMC bitcell is optimized for reliable near-threshold operation. Implemented in a 55-nm CMOS process, our proposed reconfigurable IMC-based BCNN accelerator achieves 5526 TOPS/W energy efficiency at 0.4V, which is 6.38 × higher compared to the state-of-the-art designs. The inference accuracies of our proposed design are 97.73%, 82.56%, and 92.61% across three datasets (MNIST, CIFAR-10, and SVHN), respectively. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
50. Ultrafast and Energy-Efficient Ferrimagnetic XNOR Logic Gates for Binary Neural Networks.
- Author
-
Wang, Guanda, Zhang, Yue, Zhang, Zhizhong, Zheng, Zhenyi, Zhang, Kun, Wang, Jinkai, Klein, Jacques-Olivier, Ravelosona, Dafine, and Zhao, Weisheng
- Subjects
LOGIC circuits ,THRESHOLD logic ,ENERGY consumption ,MAGNETIC tunnelling ,NANOWIRES ,SILICON nanowires - Abstract
Ultrafast current-driven domain wall (DW) motions have been realized in ferrimagnetic (FiM) nanowires. However, the FiM dynamics can be significantly affected by the Joule-heating. In this work, we propose a highly efficient XNOR logic gate by properly leveraging the thermal effect on the FiM DW motions. Its functionality and advantageous performance have been confirmed by the micromagnetic simulations. Moreover, majority logic and full adder functions can also be reconfigured based on the proposed scheme. Lastly, a fully FiM DW based binary neural network (BNN) is built, which provides low energy consumption, short delay and excellent accuracy. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.