12 results on '"Shamma Nasrin"'
Search Results
2. List of contributors
- Author
-
Zahra Ahangari, Muhammed Ahosan Ul Karim, Navjeet Bagga, Writam Banerjee, Biswajit Baral, Ratul K. Baruah, Arighna Basak, Sriyanka Behera, Sekhar Bhattacharya, D.K. Bhattacharyya, Sudhansu Mohan Biswal, Avik Chakraborty, Manash Chanda, Joy Chowdhury, Jitendra Kumar Das, Sanghamitra Das, Shyamal K. Das, Gananath Dash, Taraprasanna Dash, Papiya Debnath, Arpan Deyasi, Shashank Kumar Dubey, Md Salim Equbal, Abhigyan Ganguly, Rupam Goswami, Richa Gupta, Anisul Haque, Hadi Heidari, N. Hoque, Aminul Islam, Raisul Islam, Devika Jena, Manasa Ranjan Jena, Chinmay Kumar Maiti, Eleena Mohapatra, Kamalakanta Mohapatra, Shamma Nasrin, Satya Ranjan Pattanaik, Vivek Raghuwanshi, Hafizur Rahman, Shubham Sahay, Angsuman Sarkar, Saheli Sarkhel, Nabin Sarmah, Ashok Kumar Sharma, Preeti Sharma, Rajnish Sharma, Santosh Sharma, Avtar Singh, Shree Prakash Tiwari, Rakesh Vaid, and Zheng Wang
- Published
- 2023
- Full Text
- View/download PDF
3. Emerging memories and their applications in neuromorphic computing
- Author
-
Zheng Wang, Shamma Nasrin, Raisul Islam, Anisul Haque, and Muhammed Ahosan Ul Karim
- Published
- 2023
- Full Text
- View/download PDF
4. Reconfigurable MoS2 Memtransistors for Continuous Learning in Spiking Neural Networks
- Author
-
Silu Guo, William A. Gaviria Rojas, Mark C. Hersam, Hadallia Bergeron, Stephanie E. Liu, Shamma Nasrin, Ahish Shylendra, Hong Sub Lee, Amit Ranjan Trivedi, Shaowei Li, Jiangtan Yuan, and Vinod K. Sangwan
- Subjects
Spiking neural network ,Computer science ,Mechanical Engineering ,Reconfigurability ,Bioengineering ,General Chemistry ,Condensed Matter Physics ,Computer architecture ,Neuromorphic engineering ,Learning curve ,Hardware acceleration ,Unsupervised learning ,General Materials Science ,Energy (signal processing) ,Efficient energy use - Abstract
Artificial intelligence and machine learning are growing computing paradigms, but current algorithms incur undesirable energy costs on conventional hardware platforms, thus motivating the exploration of more efficient neuromorphic architectures. Toward this end, we introduce here a memtransistor with gate-tunable dynamic learning behavior. By fabricating memtransistors from monolayer MoS2 grown on sapphire, the relative importance of the vertical field effect from the gate is enhanced, thereby heightening reconfigurability of the device response. Inspired by biological systems, gate pulses are used to modulate potentiation and depression, resulting in diverse learning curves and simplified spike-timing-dependent plasticity that facilitate unsupervised learning in simulated spiking neural networks. This capability also enables continuous learning, which is a previously underexplored cognitive concept in neuromorphic computing. Overall, this work demonstrates that the reconfigurability of memtransistors provides unique hardware accelerator opportunities for energy efficient artificial intelligence and machine learning.
- Published
- 2021
- Full Text
- View/download PDF
5. Higher order neural processing with input-adaptive dynamic weights on MoS2 memtransistor crossbars
- Author
-
Leila Rahimifard, Ahish Shylendra, Shamma Nasrin, Stephanie E. Liu, Vinod K. Sangwan, Mark C. Hersam, and Amit Ranjan Trivedi
- Abstract
The increasing complexity of deep learning systems has pushed conventional computing technologies to their limits. While the memristor is one of the prevailing technologies for deep learning acceleration, it is only suited for classical learning layers where only two operands, namely weights and inputs, are processed simultaneously. Meanwhile, to improve the computational efficiency of deep learning for emerging applications, a variety of non-traditional layers requiring concurrent processing of many operands are becoming popular. For example, hypernetworks improve their predictive robustness by simultaneously processing weights and inputs against the application context. Two-electrode memristor grids cannot directly map emerging layers’ higher-order multiplicative neural interactions. Addressing this unmet need, we present crossbar processing using dual-gated memtransistors based on two-dimensional semiconductor MoS2. Unlike the memristor, the resistance states of memtransistors can be persistently programmed and can be actively controlled by multiple gate electrodes. Thus, the discussed memtransistor crossbar enables several advanced inference architectures beyond a conventional passive crossbar. For example, we show that sneak paths can be effectively suppressed in memtransistor crossbars, whereas they limit size scalability in a passive memristor crossbar. Similarly, exploiting gate terminals to suppress crossbar weights dynamically reduces biasing power by ∼20% in memtransistor crossbars for a fully connected layer of AlexNet. On emerging layers such as hypernetworks, collocating multiple operations within the same crossbar cells reduces operating power by ∼15× on the considered network cases.
- Published
- 2022
- Full Text
- View/download PDF
6. Reconfigurable MoS
- Author
-
Jiangtan, Yuan, Stephanie E, Liu, Ahish, Shylendra, William A, Gaviria Rojas, Silu, Guo, Hadallia, Bergeron, Shaowei, Li, Hong-Sub, Lee, Shamma, Nasrin, Vinod K, Sangwan, Amit Ranjan, Trivedi, and Mark C, Hersam
- Subjects
Molybdenum ,Artificial Intelligence ,Computers ,Neural Networks, Computer ,Algorithms - Abstract
Artificial intelligence and machine learning are growing computing paradigms, but current algorithms incur undesirable energy costs on conventional hardware platforms, thus motivating the exploration of more efficient neuromorphic architectures. Toward this end, we introduce here a memtransistor with gate-tunable dynamic learning behavior. By fabricating memtransistors from monolayer MoS
- Published
- 2021
7. Low Power Restricted Boltzmann Machine Using Mixed-Mode Magneto-Tunneling Junctions
- Author
-
Amit Ranjan Trivedi, Shamma Nasrin, Supriyo Bandyopadhyay, and Justine L. Drobitch
- Subjects
Computer Science::Machine Learning ,010302 applied physics ,Physics ,Restricted Boltzmann machine ,Feature extraction ,Probabilistic logic ,Memristor ,Topology ,01 natural sciences ,Electronic, Optical and Magnetic Materials ,law.invention ,Power (physics) ,law ,0103 physical sciences ,Unsupervised learning ,Electrical and Electronic Engineering ,Magneto ,MNIST database - Abstract
This letter discusses mixed-mode magneto tunneling junction (m-MTJ)-based restricted Boltzmann machine (RBM). RBMs are unsupervised learning models, suitable for extracting features from high-dimensional data. The m-MTJ is actuated by the simultaneous actions of voltage-controlled magnetic anisotropy and voltage-controlled spin-transfer torque, where the switching of the free-layer is probabilistic and can be controlled by the two. Using m-MTJ-based activation functions, we present a novel low area/power RBM. We discuss online learning of the presented implementation to negate process variability. For MNIST hand-written dataset, the design achieves ~96% accuracy under expected variability in various components.
- Published
- 2019
- Full Text
- View/download PDF
8. Compute-in-Memory Upside Down: A Learning Operator Co-Design Perspective for Scalability
- Author
-
Shruthi Jaisimha, Amit Ranjan Trivedi, Priyesh Shukla, and Shamma Nasrin
- Subjects
Operator (computer programming) ,Memory management ,Computer engineering ,Computer science ,business.industry ,Deep learning ,Scalability ,Bayesian probability ,Artificial intelligence ,Bayesian inference ,business ,Implementation ,Efficient energy use - Abstract
This paper discusses the potential of model-hardware co-design to simplify the implementation complexity of compute-in-SRAM deep learning considerably. Although compute-in-SRAM has emerged as a promising approach to improve the energy efficiency of DNN processing, current implementations suffer due to complex and excessive mixed-signal peripherals, such as the need for parallel digital-to-analog converters (DACs) at each input port. Comparatively, our approach inherently obviates complex peripherals by co-designing learning operators to SRAM's operational constraints. For example, our co-designed implementation is DAC-free even for multibit precision DNN processing. Additionally, we also discuss the interaction of our compute-in-SRAM operator with Bayesian inference of DNNs. We show a synergistic interaction of Bayesian inference with our framework, where Bayesian methods allow achieving similar accuracy with much smaller network size. Although each iteration of sample-based Bayesian inference is computationally expensive, the cost is minimized by our compute-in-SRAM approach. Meanwhile, by reducing the network size, Bayesian methods reduce the footprint cost of compute-in-SRAM implementation, which is a crucial concern for the method. We characterize this interaction for deep learning-based pose (position and orientation) estimation for a drone.
- Published
- 2021
- Full Text
- View/download PDF
9. MF-Net: Compute-In-Memory SRAM for Multibit Precision Inference using Memory-immersed Data Conversion and Multiplication-free Operators
- Author
-
Gomes Wilfred, Ahmet Enis Cetin, Shamma Nasrin, Amit Ranjan Trivedi, and Diaa Badawi
- Subjects
FOS: Computer and information sciences ,Computer science ,020208 electrical & electronic engineering ,Successive approximation ADC ,02 engineering and technology ,computer.file_format ,Hardware_PERFORMANCEANDRELIABILITY ,Systems and Control (eess.SY) ,Electrical Engineering and Systems Science - Systems and Control ,Data conversion ,Computational science ,Parasitic capacitance ,CMOS ,Hardware Architecture (cs.AR) ,0202 electrical engineering, electronic engineering, information engineering ,Hardware_INTEGRATEDCIRCUITS ,FOS: Electrical engineering, electronic engineering, information engineering ,Overhead (computing) ,Static random-access memory ,Electrical and Electronic Engineering ,Computer Science - Hardware Architecture ,Tera ,computer ,MNIST database - Abstract
We propose a co-design approach for compute-in-memory inference for deep neural networks (DNN). We use multiplication-free function approximators based on $\ell _{1}$ norm along with a co-adapted processing array and compute flow. Using the approach, we overcame many deficiencies in the current art of in-SRAM DNN processing such as the need for digital-to-analog converters (DACs) at each operating SRAM row/column, the need for high precision analog-to-digital converters (ADCs), limited support for multi-bit precision weights, and limited vector-scale parallelism. Our co-adapted implementation seamlessly extends to multi-bit precision weights, it doesn’t require DACs, and it easily extends to higher vector-scale parallelism. We also propose an SRAM-immersed successive approximation ADC (SA-ADC), where we exploit the parasitic capacitance of bit lines of SRAM array as a capacitive DAC. Since the dominant area overhead in SA-ADC comes due to its capacitive DAC, by exploiting the intrinsic parasitic of SRAM array, our approach allows low area implementation of within-SRAM SA-ADC. Our $8\times 62$ SRAM macro, which requires a 5-bit ADC, achieves ~105 tera operations per second per Watt (TOPS/W) with 8-bit input/weight processing at 45 nm CMOS. Our $8\times 30$ SRAM macro, which requires a 4-bit ADC, achieves ~84 TOPS/W. SRAM macros that require lower ADC precision are more tolerant of process variability, however, have lower TOPS/W as well. We evaluated the accuracy and performance of our proposed network for MNIST, CIFAR10, and CIFAR100 datasets. We chose a network configuration which adaptively mixes multiplication-free and regular operators. The network configurations utilize the multiplication-free operator for more than 85% operations from the total. The selected configurations are 98.6% accurate for MNIST, 90.2% for CIFAR10, and 66.9% for CIFAR100. Since most of the operations in the considered configurations are based on proposed SRAM macros, our compute-in-memory’s efficiency benefits broadly translate to the system-level.
- Published
- 2021
- Full Text
- View/download PDF
10. Supported-BinaryNet: Bitcell Array-Based Weight Supports for Dynamic Accuracy-Energy Trade-Offs in SRAM-Based Binarized Neural Network
- Author
-
Amit Ranjan Trivedi, Srikanth Ramakrishna, Shamma Nasrin, and Theja Tulabandhula
- Subjects
Flexibility (engineering) ,Hardware_MEMORYSTRUCTURES ,Artificial neural network ,Computer science ,020208 electrical & electronic engineering ,Transistor ,02 engineering and technology ,Power (physics) ,law.invention ,Computer engineering ,law ,0202 electrical engineering, electronic engineering, information engineering ,Static random-access memory ,Dropout (neural networks) ,Energy (signal processing) ,MNIST database - Abstract
In this work, we introduce bitcell array-based support parameters to improve the prediction accuracy of SRAM-based binarized neural network (SRAM-BNN). Our approach enhances the training weight space of SRAM-BNN while requiring minimal overheads to a typical design. More flexibility of the weight space leads to higher prediction accuracy in our design. We adapt row digital-to-analog (DAC) converter, and computing flow in SRAM-BNN for bitcell array-based weight supports. Using the discussed interventions, our scheme also allows a dynamic trade-off of accuracy against energy to address dynamic energy constraints in typical real-time applications. Our approach reduces classification error in MNIST from 1.4% to 0.91%. To reduce the power overheads, we propose a dynamic drop out of support parameters, which also reduces the processing energy of the in-SRAM weight-input product Our architecture can dropout 52% of the bitcell array-based support parameters with only minimal accuracy degradation. We also characterize our design under varying degrees of process variability in the transistors.
- Published
- 2020
- Full Text
- View/download PDF
11. Bayesian reasoning machine on a magneto-tunneling junction network
- Author
-
Theja Tulabandhula, Supriyo Bandyopadhyay, Shamma Nasrin, Priyesh Shukla, Amit Ranjan Trivedi, and Justine L. Drobitch
- Subjects
Stochastic computing ,Materials science ,business.industry ,Mechanical Engineering ,Deep learning ,Bayesian probability ,Bayesian network ,Bioengineering ,02 engineering and technology ,General Chemistry ,010402 general chemistry ,021001 nanoscience & nanotechnology ,Bayesian inference ,01 natural sciences ,Stochastic programming ,0104 chemical sciences ,Neuromorphic engineering ,Computer engineering ,Mechanics of Materials ,Graph (abstract data type) ,General Materials Science ,Artificial intelligence ,Electrical and Electronic Engineering ,0210 nano-technology ,business - Abstract
The recent trend in adapting ultra-energy-efficient (but error-prone) nanomagnetic devices to non-Boolean computing and information processing (e.g. stochastic/probabilistic computing, neuromorphic, belief networks, etc) has resulted in rapid strides in new computing modalities. Of particular interest are Bayesian networks (BN) which may see revolutionary advances when adapted to a specific type of nanomagnetic devices. Here, we develop a novel nanomagnet-based computing substrate for BN that allows high-speed sampling from an arbitrary Bayesian graph. We show that magneto-tunneling junctions (MTJs) can be used for electrically programmable 'sub-nanosecond' probability sample generation by co-optimizing voltage-controlled magnetic anisotropy and spin transfer torque. We also discuss that just by engineering local magnetostriction in the soft layers of MTJs, one can stochastically couple them for programmable conditional sample generation as well. This obviates the need for extensive energy-inefficient hardware like OP-AMPS, gates, shift-registers, etc to generate the correlations. Based on the above findings, we present an architectural design and computation flow of the MTJ network to map an arbitrary Bayesian graph where we develop circuits to program and induce switching and interactions among MTJs. Our discussed framework can lead to a new generation of stochastic computing hardware for various other computing models, such as stochastic programming and Bayesian deep learning. This can spawn a novel genre of ultra-energy-efficient, extremely powerful computing paradigms, which is a transformational advance.
- Published
- 2020
12. Mixed-mode Magnetic Tunnel Junction-based Deep Belief Network
- Author
-
Shamma Nasrin, Supriyo Bandyopadhyay, Amit Ranjan Trivedi, and Justine L. Drobitch
- Subjects
Computer Science::Machine Learning ,010302 applied physics ,Restricted Boltzmann machine ,Computer science ,Activation function ,Probabilistic logic ,02 engineering and technology ,021001 nanoscience & nanotechnology ,01 natural sciences ,Power (physics) ,Deep belief network ,0103 physical sciences ,Unsupervised learning ,Hardware_ARITHMETICANDLOGICSTRUCTURES ,0210 nano-technology ,Cluster analysis ,Algorithm ,MNIST database - Abstract
We present a mixed-mode magneto tunneling junction (m-MTJ)-based Deep Belief Network (DBN). DBNs are unsupervised learning models, suitable for recognition and clustering. m-MTJ is a three-terminal magnetic device with probabilistic free layer switching controlled by the simultaneous actions of voltage-controlled magnetic anisotropy and spin-transfer torque. While DBNs achieve high prediction accuracy even with highly imprecise single-bit weights, the key complexity lies in their activation functions which are stochastic. Using an m-MTJ, we present a novel low area/power DBN neuron with stochastic activation function. We discuss an in-memory computing architecture that allows forward and backward flow of learning dynamics and online learning. Our design achieves ~88.80% accuracy for digit recognition in MNIST even under the worst case variability in nanoscaled m-MTJs.
- Published
- 2019
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.