69 results on '"Memory faults"'
Search Results
2. Analysis and Synthesis March Memory Tests
- Author
-
V. N. Yarmolik, V. A. Levantsevich, and D. V. Demenkovets
- Subjects
computer systems testing ,memory faults ,pattern sensitive faults ,march tests ,pseudo-exhaustive tests ,Information technology ,T58.5-58.64 - Abstract
The paper shows the relevance of testing storage devices in modern computing systems. Mathematical models of memory device faults and the efficiency of their detection, in particular, complex pattern sensitive faults of the PNPSFk type, based on classical march memory tests are presented. Limit estimates are given for the completeness of coverage of such faults depending on the number of memory cells involved in the fault. The necessity of synthesis of memory march tests characterized by high efficiency of PNPSFk failure detection is substantiated. The concept of a primitive providing conditions for activation and detection of various types of PNPSFk is defined. Examples of analysis and synthesis of memory march tests with different coverage of PNPSFk faults are given. The March OP memory test is synthesized, which is characterized by the maximum completeness of PNPSFk fault coverage and has the lowest time complexity compared to the known memory march tests, which provide the same comprehensiveness of coverage of complex memory device faults.
- Published
- 2021
3. Tolerating Defects in Low-Power Neural Network Accelerators Via Retraining-Free Weight Approximation.
- Author
-
HOSSEINI, FATEME S., FANRUO MENG, CHENGMO YANG, WUJIE WEN, and CAMMAROTA, ROSARIO
- Subjects
MEMORY ,OCCUPATIONAL retraining ,ALGORITHMS - Abstract
Hardware accelerators are essential to the accommodation of ever-increasing Deep Neural Network (DNN) workloads on the resource-constrained embedded devices. While accelerators facilitate fast and energyefficient DNN operations, their accuracy is threatened by faults in their on-chip and off-chip memories, where millions of DNN weights are held. The use of emerging Non-Volatile Memories (NVM) further exposes DNN accelerators to a non-negligible rate of permanent defects due to immature fabrication, limited endurance, and aging. To tolerate defects in NVM-based DNN accelerators, previous work either requires extra redundancy in hardware or performs defect-aware retraining, imposing significant overhead. In comparison, this paper proposes a set of algorithms that exploit the flexibility in setting the fault-free bits in weight memory to effectively approximate weight values, so as to mitigate defect-induced accuracy drop. These algorithms can be applied as a one-step solution when loading the weights to embedded devices. They only require trivial hardware support and impose negligible run-time overhead. Experiments on popular DNN models show that the proposed techniques successfully boost inference accuracy even in the face of elevated defect rates in the weight memory. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
4. Run Time Memory Error Recovery Process in Networking System.
- Author
-
Vitucci, Carlo, Danielsson, Jakob, Jägemar, Marcus, Larsson, Alf, Nolte, Thomas, Vitucci, Carlo, Danielsson, Jakob, Jägemar, Marcus, Larsson, Alf, and Nolte, Thomas
- Abstract
System memory errors have always been problematic; today, they cause more than forty percent of confirmed hardware errors in repair centers for both data centers and telecommunications network nodes. Therefore, it is somewhat expected that, in recent years, device manufacturers improved the hardware features to support hardware-assisted fault management implementation. For example, the new standard, DDR5, includes both data redundancy, the so-called Error Correcting Code (ECC), and physical redundancy, the post-package repair (PPR), as mandatory features. Production and repair centers mainly use physical redundancy to replace faulty memory rows. In contrast, field use still needs to be improved, mainly due to a need for integrated system solutions for network nodes. This paper aims to compensate for this shortcoming and presents a system solution for handling memory errors. It is a multi-technology proposition (mixed use of ECC and PPR) based on multi-layer (hardware, firmware, and software) error information exchange.
- Published
- 2023
- Full Text
- View/download PDF
5. Resilient Dynamic Programming.
- Author
-
Caminiti, Saverio, Finocchi, Irene, Fusco, Emanuele, and Silvestri, Francesco
- Subjects
- *
DYNAMIC programming , *MATHEMATICAL optimization , *MATHEMATICAL programming , *LINEAR programming , *INTEGER programming - Abstract
We investigate the design of dynamic programming algorithms in unreliable memories, i.e., in the presence of errors that lead the logical state of some bits to be read differently from how they were last written. Assuming that a limited number of memory faults can be inserted at run-time by an adversary with unbounded computational power, we obtain the first resilient algorithms for a broad range of dynamic programming problems, devising a general framework that can be applied to both iterative and recursive implementations. Besides all local dependency problems, where updates to table entries are determined by the contents of neighboring cells, we also settle challenging non-local problems, such as all-pairs shortest paths and matrix multiplication. All our algorithms are correct with high probability and match the running time of their standard non-resilient counterparts while tolerating a polynomial number of faults. The recursive algorithms are also cache-efficient and can tolerate faults at any level of the memory hierarchy. Our results exploit a careful combination of data replication, majority techniques, fingerprint computations, and lazy fault detection. To cope with the complex data access patterns induced by some of our algorithms, we also devise amplified fingerprints, which might be of independent interest in the design of resilient algorithms for different problems. [ABSTRACT FROM AUTHOR]
- Published
- 2017
- Full Text
- View/download PDF
6. Pseudoexhaustive memory testing based on March A type march tests
- Author
-
Ireneusz Mrozek, Vyacheslav N. Yarmolik, and S. V. Yarmolik
- Subjects
Test procedures ,Computer science ,Memory faults ,Binary number ,0102 computer and information sciences ,02 engineering and technology ,pseudo-exhaustive tests ,QA75.5-76.95 ,01 natural sciences ,Computing systems ,020202 computer hardware & architecture ,Set (abstract data type) ,embedded testing ,010201 computation theory & mathematics ,Electronic computers. Computer science ,0202 electrical engineering, electronic engineering, information engineering ,testing of computing systems ,march memory tests ,Coupon collector's problem ,Algorithm ,Memory testing ,multi-run testing - Abstract
The relevance of testing of memory devices of modern computing systems is shown. The methods and algorithms for implementing test procedures based on classical March tests are analyzed. Multiple March tests are highlighted to detect complex pattern-sensitive memory faults. To detect them, the necessary condition that test procedures must satisfy to deal complex faults, is substantiated. This condition is in the formation of a pseudo-exhaustive test for a given number of arbitrary memory cells. We study the effectiveness of single and double application of tests like MATS ++, March C– and March A, and also give its analytical estimates for a different number of k ≤ 10 memory cells participating in a malfunction. The applicability of the mathematical model of the combinatorial problem of the coupon collector for describing multiple memory testing is substantiated. The values of the average, minimum, and maximum multiplicity of multiple tests are presented to provide an exhaustive set of binary combinations for a given number of arbitrary memory cells. The validity of analytical estimates is experimentally shown and the high efficiency of the formation of a pseudo-exhaustive coverage by tests of the March A type is confirmed.
- Published
- 2020
7. W-ERA: One-Time Memory Repair with Wafer-Level Early Repair Analysis for Cost Reduction
- Author
-
Donghyun Han, Hayoung Lee, Sungho Kang, and Hogyeong Kim
- Subjects
Repair time ,Hardware_MEMORYSTRUCTURES ,Fault occurrence ,Total cost ,Computer science ,Memory faults ,020206 networking & telecommunications ,02 engineering and technology ,Repair rate ,020202 computer hardware & architecture ,Reliability engineering ,Cost reduction ,0202 electrical engineering, electronic engineering, information engineering ,Redundancy (engineering) - Abstract
Since the probability of fault occurrence on memory has increased with the advance of memory density and capacity, memory repairs in wafer-level and package-level are widely used with redundancy analysis (RA) to improve memory yield. However, as the costs for memory repair also have increased in proportion to the memory density and capacity, the repair costs have occupied a significant portion of the total costs. To address the problem, one-time memory repair with wafer-level early repair analysis (W-ERA) for cost reduction is proposed in this paper. The proposed W-ERA facilitates that all unrepairable memories are classified rapidly without searching memory repair solutions in wafer-level and repairable memory faults occurred in wafer-level are repaired in package-level with additional faults occurred in package-level simultaneously. It means, as the costs of memory repair can be highly reduced since memory repair is skipped in wafer-level, the total costs also can be highly reduced. In addition, memory redundancies can be efficiently used for memory repair in package-level and it results a high repair rate achievement.
- Published
- 2020
8. On the Memory Fault Resilience of TLS 1.3
- Author
-
Robin Leander Schröder, Michael Yonli, Lukas Brandstetter, and Marc Fischlin
- Subjects
Computer science ,Memory faults ,Computer security model ,Resilience (network) ,Computer security ,computer.software_genre ,Fault (power engineering) ,Protocol (object-oriented programming) ,computer ,Signature (logic) ,Key exchange - Abstract
Recently, Aranha et al. (Eurocrypt 2020) as well as Fischlin and Gunther (CT-RSA 2020) investigated the possibility to model memory fault attacks like Rowhammer in security games, and to deduce statements about the (in)security of schemes against such attacks. They looked into the fault-resistance of signature and AEAD schemes. Here, we extend the approach to the TLS 1.3 key exchange protocol.
- Published
- 2020
9. Modeling Memory Faults in Signature and Authenticated Encryption Schemes
- Author
-
Felix Günther, Marc Fischlin, and Jarecki, Stanislaw
- Subjects
Authenticated encryption ,business.industry ,Computer science ,Memory faults ,Cryptography ,02 engineering and technology ,Computer security model ,computer.software_genre ,Computer security ,020202 computer hardware & architecture ,Software ,0202 electrical engineering, electronic engineering, information engineering ,Malware ,020201 artificial intelligence & image processing ,business ,computer ,Host machine ,Randomness - Abstract
Memory fault attacks, inducing errors in computations, have been an ever-evolving threat to cryptographic schemes since their discovery for cryptography by Boneh et al. (Eurocrypt 1997). Initially requiring physical tampering with hardware, the software-based rowhammer attack put forward by Kim et al. (ISCA 2014) enabled fault attacks also through malicious software running on the same host machine. This led to concerning novel attack vectors, for example on deterministic signature schemes, whose approach to avoid dependency on (good) randomness renders them vulnerable to fault attacks. This has been demonstrated in realistic adversarial settings in a series of recent works. However, a unified formalism of different memory fault attacks, enabling also to argue the security of countermeasures, is missing yet. In this work, we suggest a generic extension for existing security models that enables a game-based treatment of cryptographic fault resilience. Our modeling specifies exemplary memory fault attack types of different strength, ranging from random bit-flip faults to differential (rowhammer-style) faults to full adversarial control on indicated memory variables. We apply our model first to deterministic signatures to revisit known fault attacks as well as to establish provable guarantees of fault resilience for proposed fault-attack countermeasures. In a second application to nonce-misuse resistant authenticated encryption, we provide the first fault-attack treatment of the SIV mode of operation and give a provably secure fault-resilient variant., Lecture Notes in Computer Science, 12006, ISSN:0302-9743, ISSN:1611-3349, Topics in Cryptology – CT-RSA 2020, ISBN:978-3-030-40185-6, ISBN:978-3-030-40186-3
- Published
- 2020
10. Data structures resilient to memory faults.
- Author
-
Ferraro-Petrillo, Umberto, Grandoni, Fabrizio, and Italiano, Giuseppe F.
- Subjects
DATA structures ,ENCYCLOPEDIAS & dictionaries ,COMPUTER storage devices ,ALGORITHM research ,COMPUTER algorithms ,RESEARCH - Abstract
We address the problem of implementing data structures resilient to memory faults, which may arbitrarily corrupt memory locations. In this framework, we focus on the implementation of dictionaries and perform a thorough experimental study using a testbed that we designed for this purpose. Our main discovery is that the best-known (asymptotically optimal) resilient data structures have very large space overheads. More precisely, most of the space used by these data structures is not due to key storage. This might not be acceptable in practice, since resilient data structures are meant for applications where a huge amount of data (often of the order of terabytes) has to be stored. Exploiting techniques developed in the context of resilient (static) sorting and searching, in combination with some new ideas, we designed and engineered an alternative implementation, which, while still guaranteeing optimal asymptotic time and space bounds, performs much better in terms of memory without compromising the time efficiency. [ABSTRACT FROM AUTHOR]
- Published
- 2013
- Full Text
- View/download PDF
11. Evaluating softcore GPU in SRAM-based FPGA under radiation-induced effects
- Author
-
Marcio M. Goncalves, Jose Rodrigo Azambuja, Fernanda Lima Kastensmidt, Michael Hübner, Marcelo Brandalero, Giani Braga, Hector Gerardo Munoz Hernandez, and Fabio Benevenuti
- Subjects
business.industry ,Computer science ,Memory faults ,Radiation induced ,Hardware_PERFORMANCEANDRELIABILITY ,Fault injection ,Condensed Matter Physics ,Atomic and Molecular Physics, and Optics ,Surfaces, Coatings and Films ,Electronic, Optical and Magnetic Materials ,Reliability (semiconductor) ,Embedded system ,Static random-access memory ,Electrical and Electronic Engineering ,Safety, Risk, Reliability and Quality ,business ,Field-programmable gate array - Abstract
This work investigates selective mitigation techniques to improve the reliability of a configurable open-source softcore GPU implemented in an SRAM-based FPGA against configuration memory faults. It explores the unhardened and hardened reliability curves of isolated groups of the target GPU's modules to guide the decision of the best candidates for selective hardening. Fault injection campaigns are performed on the target GPU running three case-study applications by accumulating radiation-induced effects in the FPGA's configuration memory. Individual modules' reliability curves are combined with their resource usages to selectively harden a larger area of the target GPU and increase reliability/area effectiveness. Results can help designers in hardening configurable GPUs for a given application more effectively.
- Published
- 2021
12. Resilient Dictionaries.
- Author
-
Finocchi, Irene, Grandoni, Fabrizio, and Italiano, Giuseppe F.
- Subjects
ENCYCLOPEDIAS & dictionaries ,INFORMATION resources management ,INFORMATION services management ,INFORMATION science ,PUBLICATIONS ,DATABASES - Abstract
We address the problem of designing data structures in the presence of faults that may arbitrarily corrupt memory locations. More precisely, we assume that an adaptive adversary can arbitrarily overwrite the content of up to δ memory locations, that corrupted locations cannot be detected, and that only O(1) memory locations are safe. In this framework, we call a data structure resilient if it is able to operate correctly (at least) on the set of uncorrupted values. We present a resilient dictionary, implementing search, insert, and delete operations. Our dictionary has O(log n + δ) expected amortized time per operation, and O(n) space complexity, where n denotes the current number of keys in the dictionary. We also describe a deterministic resilient dictionary, with the same amortized cost per operation over a sequence of at least δ
ϵ operations, where ϵ > 0 is an arbitrary constant. Finally, we show that any resilient comparison-based dictionary must take Ω(log n + δ) expected time per search. Our results are achieved by means of simple, new techniques which might be of independent interest for the design of other resilient algorithms. [ABSTRACT FROM AUTHOR]- Published
- 2009
- Full Text
- View/download PDF
13. Optimal resilient sorting and searching in the presence of memory faults
- Author
-
Finocchi, Irene, Grandoni, Fabrizio, and Italiano, Giuseppe F.
- Subjects
- *
COMPUTATIONAL complexity , *SEARCH algorithms , *FAULT tolerance (Engineering) , *COMPUTER storage devices , *MATHEMATICAL models , *INTEGER programming , *COMPUTER science - Abstract
Abstract: We investigate the problem of reliable computation in the presence of faults that may arbitrarily corrupt memory locations. In this framework, we consider the problems of sorting and searching in optimal time while tolerating the largest possible number of memory faults. In particular, we design an time sorting algorithm that can optimally tolerate up to memory faults. In the special case of integer sorting, we present an algorithm with linear expected running time that can tolerate faults. We also present a randomized searching algorithm that can optimally tolerate up to memory faults in expected time, and an almost optimal deterministic searching algorithm that can tolerate faults, for any small positive constant , in worst-case time. All these results improve over previous bounds. [Copyright &y& Elsevier]
- Published
- 2009
- Full Text
- View/download PDF
14. Multilevel Checkpoint/Restart for Large Computational Jobs on Distributed Computing Resources
- Author
-
Florian Schintke and Masoud Gholami Estahbanati
- Subjects
Mean time between failures ,010201 computation theory & mathematics ,Computer science ,Distributed computing ,Memory faults ,0202 electrical engineering, electronic engineering, information engineering ,Graph (abstract data type) ,020201 artificial intelligence & image processing ,Fault tolerance ,0102 computer and information sciences ,02 engineering and technology ,Supercomputer ,01 natural sciences - Abstract
New generations of high-performance computing applications depend on an increasing number of components to satisfy their growing demand for computation. On such large systems, the execution of long-running jobs is more likely affected by component failures. Failure classes vary from frequent transient memory faults to rather rare correlated node errors. Multilevel checkpoint/restart has been introduced to proactively cope with failures at different levels. Writing checkpoints on slower stable devices, which survive fatal failures, causes more overhead than writing them on fast devices (main memory or local SSD), which, however, only protect against light faults. Given a graph of the components of a particular storage hierarchy mapping their fault-domains and their expected mean time to failure (MTTF), we optimize the checkpoint frequencies for each level of the storage hierarchy (multilevel checkpointing) to minimize the overhead and runtime of a given job. We reduce the checkpoint/restart overhead of large data-intensive jobs compared to state-of-the-art solutions on multilevel checkpointing by up to 10 percent in the investigated cases. The improvement increases further with growing checkpoint sizes.
- Published
- 2019
15. Monitoring, Risk Assessment and Actuation for Alzheimer Patients: A Case Study
- Author
-
Pedro Corista, Jorge Calado, Joao Giao, Joao Sarraipa, and Fernando Luis-Ferreira
- Subjects
business.industry ,media_common.quotation_subject ,Memory faults ,medicine.disease ,Quality of life (healthcare) ,Work (electrical) ,Health care ,medicine ,Life expectancy ,Dementia ,Medical emergency ,business ,Risk assessment ,health care economics and organizations ,Autonomy ,media_common - Abstract
Alzheimer is one of the most frequent types of dementia. With the increasing extension of life expectancy, and an increasing incidence above sixty-five years. Near to thirteen million cases are foreseen in 2050 with an estimate cost above two hundred billion dollars in associated care expenses. It becomes important to take measures to ensure quality of life to patients, carers and promote the sustainability of public and personal finances. The major concerns with those patients are memory faults with the tendency for wandering and get lost. The present work proposes a solution for permanent monitoring, risk assessment and reaction, when needed, while extending battery autonomy for the worn device. Data is periodically uploaded to be processed and analysed in a remote infrastructure such as FIWARE. The aim is to establish profiles that better adapt to each citizen of the evergrowing community of dementia patients, including those with Alzheimer disease.Copyright © 2018 by ASME
- Published
- 2018
16. Robust and Adaptive Search
- Author
-
Yann Disser and Stefan Kratsch, Disser, Yann, Kratsch, Stefan, Yann Disser and Stefan Kratsch, Disser, Yann, and Kratsch, Stefan
- Abstract
Binary search finds a given element in a sorted array with an optimal number of log n queries. However, binary search fails even when the array is only slightly disordered or access to its elements is subject to errors. We study the worst-case query complexity of search algorithms that are robust to imprecise queries and that adapt to perturbations of the order of the elements. We give (almost) tight results for various parameters that quantify query errors and that measure array disorder. In particular, we exhibit settings where query complexities of log n + ck, (1+epsilon) log n + ck, and sqrt(cnk)+o(nk) are best-possible for parameter value k, any epsilon > 0, and constant c.
- Published
- 2017
- Full Text
- View/download PDF
17. Resilient Dynamic Programming
- Author
-
Francesco Silvestri, Saverio Caminiti, Emanuele G. Fusco, and Irene Finocchi
- Subjects
Theoretical computer science ,General Computer Science ,Computer science ,cache-oblivious algorithms ,dynamic programming ,gaussian elimination paradigm ,memory faults ,resilient computing ,computer science (all) ,computer science applications1707 ,computer vision and pattern recognition ,applied mathematics ,0102 computer and information sciences ,02 engineering and technology ,Cache-oblivious algorithm ,01 natural sciences ,Fault detection and isolation ,Cache-oblivious algorithms ,Dynamic programming ,Gaussian Elimination Paradigm ,Memory faults ,Resilient computing ,Computer Science (all) ,Computer Science Applications1707 Computer Vision and Pattern Recognition ,Applied Mathematics ,0202 electrical engineering, electronic engineering, information engineering ,Memory hierarchy ,Fingerprint (computing) ,Matrix multiplication ,Computer Science Applications ,010201 computation theory & mathematics ,Theory of computation ,020201 artificial intelligence & image processing ,State (computer science) ,Algorithm - Abstract
We investigate the design of dynamic programming algorithms in unreliable memories, i.e., in the presence of errors that lead the logical state of some bits to be read differently from how they were last written. Assuming that a limited number of memory faults can be inserted at run-time by an adversary with unbounded computational power, we obtain the first resilient algorithms for a broad range of dynamic programming problems, devising a general framework that can be applied to both iterative and recursive implementations. Besides all local dependency problems, where updates to table entries are determined by the contents of neighboring cells, we also settle challenging non-local problems, such as all-pairs shortest paths and matrix multiplication. All our algorithms are correct with high probability and match the running time of their standard non-resilient counterparts while tolerating a polynomial number of faults. The recursive algorithms are also cache-efficient and can tolerate faults at any level of the memory hierarchy. Our results exploit a careful combination of data replication, majority techniques, fingerprint computations, and lazy fault detection. To cope with the complex data access patterns induced by some of our algorithms, we also devise amplified fingerprints, which might be of independent interest in the design of resilient algorithms for different problems.
- Published
- 2017
18. Protecting flash memory areas against memory faults in tiny embedded systems
- Author
-
Patryk Skoncej
- Subjects
010302 applied physics ,Engineering ,Hardware_MEMORYSTRUCTURES ,010308 nuclear & particles physics ,business.industry ,Memory faults ,01 natural sciences ,Maintenance engineering ,Flash memory ,Embedded system ,0103 physical sciences ,Redundancy (engineering) ,business ,Computer hardware ,Test data - Abstract
This paper shows novel combinations of software-based, fault-tolerant techniques for NOR flash memories in tiny embedded systems. The main aim of proposed techniques is to protect from transient and permanent memory faults only those parts of an embedded flash memory which can suffer from the biggest wear. The evaluation of presented approaches was based on test data obtained from real 64KB NOR flash memories.
- Published
- 2016
19. Extending fault periodicity table for testing external memory faults
- Author
-
G. Harutyunyan
- Subjects
Interconnection ,Random access memory ,geography ,geography.geographical_feature_category ,Computer science ,020208 electrical & electronic engineering ,Memory faults ,Hardware_PERFORMANCEANDRELIABILITY ,02 engineering and technology ,Fault (geology) ,020202 computer hardware & architecture ,Built-in self-test ,Test algorithm ,0202 electrical engineering, electronic engineering, information engineering ,Table (database) ,Algorithm ,Auxiliary memory - Abstract
A new solution for building memory BIST infrastructure, based on rules of fault periodicity and regularity in test algorithms was introduced recently. These rules are represented in a form of a Fault Periodicity Table (FPT) considering both known and unknown memory faults in one table. In this paper, application of the proposed methodology for description of external memory faults is shown. Investigation showed that the FPT needs to be extended in order to include the external memory faults. Both interconnect and array faults of external memories are considered in this paper.
- Published
- 2016
20. Study on Historical Memory Fault and Structural Amnesia of Kan Li Siberia Multiple Transmission from Tsinghua Jane qQiyeq
- Author
-
Zhenming Yang
- Subjects
Cognitive science ,Geography ,Cultural anthropology ,Historical memory ,Transmission (telecommunications) ,Memory faults ,medicine ,Amnesia ,medicine.symptom ,Fault (power engineering) ,Cartography - Published
- 2016
21. A Fast Redundancy Analysis Algorithm in ATE for Repairing Faulty Memories
- Author
-
Sungho Kang, Wooheon Kang, and Hyungjun Cho
- Subjects
Repair time ,Automatic test equipment ,General Computer Science ,Computer science ,Memory faults ,Redundancy (engineering) ,Bitmap ,computer.file_format ,Electrical and Electronic Engineering ,computer ,Algorithm ,Electronic, Optical and Magnetic Materials - Abstract
Testing memory and repairing faults have become increasingly important for improving yield. Redundancy analysis (RA) algorithms have been developed to repair memory faults. However, many RA algorithms have low analysis speeds and occupy memory space within automatic test equipment. A fast RA algorithm using simple calculations is proposed in this letter to minimize both the test and repair time. This analysis uses the grouped addresses in the faulty bitmap. Since the fault groups are independent of each other, the time needed to find solutions can be greatly reduced using these fault groups. Also, the proposed algorithm does not need to store searching trees, thereby minimizing the required memory space. Our experiments show that the proposed RA algorithm is very efficient in terms of speed and memory requirements.
- Published
- 2012
22. A Toolkit for Availability Evaluation of a Kind of IA64 System
- Author
-
Zhan Zhang, Jun Qian, Xiaozong Yang, Yi Feng, and Decheng Zuo
- Subjects
File system ,Engineering ,business.industry ,Transaction processing ,Memory faults ,Hardware_PERFORMANCEANDRELIABILITY ,General Medicine ,Fault injection ,IA64 ,computer.software_genre ,fault-injection ,Embedded system ,availability evaluation ,Transaction processing system ,Sensitivity (control systems) ,business ,computer ,Engineering(all) - Abstract
Availability is one of the most important features in transaction processing systems. Fault injection is one of the effective means which can hasten availability tests. An availability evaluation toolkit is designed in this paper for IA64 system which is being widely used in transaction processing business. The availability toolkit consists of fault injection platform and availability analysis platform. A series of fault injection tools are accomplished which covers CPU faults, memory faults, disk faults, IO faults, and file system faults. Availability analysis platform is designed to calculate availability of target system and analyze sensitivity and performability as well. Experiments on a typical IA64 server are described and performed, and the experiment results validate the effectiveness of the toolkit.
- Published
- 2011
23. The Price of Resiliency: a Case Study on Sorting with Memory Faults
- Author
-
Ferraro-Petrillo, Umberto, Finocchi, Irene, and Italiano, Giuseppe F.
- Published
- 2009
- Full Text
- View/download PDF
24. Sorting and Searching in Faulty Memories
- Author
-
Finocchi, Irene and Italiano, Giuseppe F.
- Published
- 2008
- Full Text
- View/download PDF
25. A new redundancy analysis algorithm using one side pivot
- Author
-
Keewon Cho, Woosung Lee, Joo Young Kim, and Sungho Kang
- Subjects
Engineering ,business.industry ,Memory faults ,Redundancy (engineering) ,Algorithm design ,Repair rate ,business ,Algorithm - Abstract
It is important to test the memory and repair faults for improving the memory yield. Many redundancy analysis (RA) algorithms have been developed to repair the memory faults. However, it is difficult to achieve high repair rate and fast analysis speed. The previous RA algorithms do not achieve both high repair rate and fast analysis speed. To overcome this problem, a new RA algorithm called one side pivot (OSP) is proposed. Using the property of pivot fault and its repair priority, the analysis time to find a solution can be reduced. The experimental results show that the proposed algorithm is efficient in terms of repair rate and analysis speed.
- Published
- 2014
26. Extending fault periodicity table for testing faults in memories under 20nm
- Author
-
Samvel Shoukourian, Yervant Zorian, V.A. Vardanian, and G. Harutyunyan
- Subjects
Stuck-at fault ,Built-in self-test ,Computer science ,Test algorithm ,Logic gate ,Memory faults ,Table (database) ,Hardware_PERFORMANCEANDRELIABILITY ,Fault (power engineering) ,Algorithm ,Column (database) - Abstract
A new solution for building memory BIST infrastructure, based on rules of fault periodicity and regularity in test algorithms was introduced recently. These rules are represented in a form of a Fault Periodicity Table (FPT) considering both known and unknown memory faults in one table. Each column of FPT corresponds to a fault nature which can be associated with a variety of different test mechanisms while each row of FPT corresponds to a fault family determined by the complexity of fault sensitization. In this paper, application of the proposed methodology for description of memory faults in technologies below 20nm, including 16/14nm FinFET-based memories, is shown. Specifically, it is shown that all recently discovered FinFET-specific faults successfully fit into FPT.
- Published
- 2014
27. Testing, diagnosis and repair methods for NBTI-induced SRAM faults
- Author
-
Chiung-Hung Chen and Bao Liu
- Subjects
Engineering ,Hardware_MEMORYSTRUCTURES ,Soft error ,business.industry ,Power consumption ,Memory faults ,MOSFET ,Electronic engineering ,Hardware_PERFORMANCEANDRELIABILITY ,Static random-access memory ,business ,Reliability engineering - Abstract
NBTI is a major SRAM aging mechanism, leading to reduced read and hold static noise margins, and increased soft error rate. The existing techniques including guardbanding, on-chip sensor-based detection, and recovery. In this paper, we propose a group of testing, diagnosis, and repair methods for NBTI-induced memory faults. We observe that NBTI leads to SRAM read errors rather than write errors. We propose to identify NBTI-induced memory read errors based on the existing ECC circuitry, differentiate them with soft errors by correction and double checking, and keep them idle for recovery. We further propose an predictive test method for NBTI-induced memory faults by adaptive body biasing. We achieve an adaptive body biasing formula to simulate the NBTI effect. Our experimental results validate the proposed methods and show that they cost little silicon area and power consumption.
- Published
- 2014
28. Transparent testing for intra-word memory faults
- Author
-
I. Voyiatzis, Constantinos Efstathiou, and C. Sgouropoulou
- Subjects
Scheme (programming language) ,Engineering ,Periodic testing ,business.industry ,Memory faults ,Signature (logic) ,Reduction (complexity) ,Built-in self-test ,Embedded system ,Hardware_ARITHMETICANDLOGICSTRUCTURES ,business ,computer ,Word (computer architecture) ,Self test ,computer.programming_language - Abstract
Transparent BIST schemes for RAM modules assure the preservation of the memory contents during periodic testing. Symmetric Transparent Built-in Self Test (BIST) schemes skip the signature prediction phase required in traditional transparent BIST, achieving considerable reduction in test time. In this work we propose a Symmetric transparent BIST scheme that can be utilized to serially apply march tests bit-by-bit to word-organized RAM's, in a transparent manner, in the sense that the initial contents of the RAM are preserved. To the best of our knowledge, this is the first scheme proposed in the open literature to target intraword faults in the concept of transparent BIST for RAMs.
- Published
- 2013
29. [Untitled]
- Author
-
Robin David, Helmut Jürgensen, and Janusz A. Brzozowski
- Subjects
Discrete mathematics ,geography ,Deterministic testing ,geography.geographical_feature_category ,Markov chain ,Memory faults ,Class (philosophy) ,Hardware_PERFORMANCEANDRELIABILITY ,Fault (geology) ,Coupling (probability) ,Computer Science::Hardware Architecture ,Bounded function ,Electrical and Electronic Engineering ,Computer Science::Operating Systems ,Computer Science::Distributed, Parallel, and Cluster Computing ,Mathematics - Abstract
We study the class of “bounded faults” in random-access memories; these are faults that involve a bounded number of cells. This is a very general class of memory faults that includes, for example, the usual stuck-at, coupling, and pattern-sensitive faults, but also many other types of faults. Some bounded faults are known to require deterministic tests of length proportional to n log2 n, where n is the total number of memory cells. The main result of this paper is that, for any bounded fault satisfying certain very mild conditions, the random test length required for a given level of confidence is always O(n).
- Published
- 1997
30. Typing Copyless Message Passing
- Author
-
Bono, Viviana and Padovani, Luca
- Subjects
FOS: Computer and information sciences ,F.1.2, F.3.3, F.3.1, D.4.4 ,Computer Science - Programming Languages ,General Computer Science ,Computer science ,Programming language ,Memory faults ,Message passing ,Linearity ,Process interaction ,Type (model theory) ,computer.software_genre ,Theoretical Computer Science ,Memory leak ,Singularity ,Session (computer science) ,computer ,Programming Languages (cs.PL) - Abstract
We present a calculus that models a form of process interaction based on copyless message passing, in the style of Singularity OS. The calculus is equipped with a type system ensuring that well-typed processes are free from memory faults, memory leaks, and communication errors. The type system is essentially linear, but we show that linearity alone is inadequate, because it leaves room for scenarios where well-typed processes leak significant amounts of memory. We address these problems basing the type system upon an original variant of session types., 50 pages
- Published
- 2012
31. Selection in the Presence of Memory Faults, with Applications to In-place Resilient Sorting
- Author
-
Tsvi Kopelowitz and Nimrod Talmon
- Subjects
Focus (computing) ,Theoretical computer science ,Soft error ,Computer science ,Memory faults ,Sorting ,Element (category theory) ,Program counter ,Selection (genetic algorithm) - Abstract
The selection problem, where one wishes to locate the k th smallest element in an unsorted array of size n, is one of the basic problems studied in computer science. The main focus of this work is designing algorithms for solving the selection problem in the presence of memory faults.
- Published
- 2012
32. Analysis and Test Development for Parasitic Fails in Deep Sub-Micron Memory Devices
- Subjects
memory faults ,Hardware_PERFORMANCEANDRELIABILITY ,sram ,memory tests ,parasitic fails ,tests development - Abstract
Emerging technology trends are gravitating towards extremely high levels of integration at the package and chip levels, and use of deeply scaled technology in nanometer, approaching 10nm CMOS. Challenges will arise due to the ability to design complex systems such as robots that encompass sensors, transducers, communications systems and processors, all of which require memory devices, and are required to be fault-free, and exhibit fault-tolerance, reliability and survivability characteristics. A key area of challenge is in memory testing, since deep scaling and smaller dimensions of semiconductor cell area will exacerbate the presence of complex defects and can induce effects, such as parasitic effects, which necessitate fails in memory devices. In this thesis, parasitic effects induced by spot defects in memory devices have been evaluated. The thesis presents the analysis, evaluation, validation and test remedies for parasitic fails in deep sub-micron memories. On the one hand, it presents analysis for parasitic bit line coupling effects, and the impact of bit line coupling effect on the static random access memory (SRAM) faulty behavior. Thereafter, it determines both the necessary and sufficient detection conditions for memory fault models, and demonstrates the limitations of existing industrial memory tests to adequately detect faults in the presence of bit line coupling. In addition, the thesis presents a systematic approach for test development and optimization, and new memory tests - March SSSc an optimal test that detects all single-cell static faults, and March m-MSS and March BLC that detect all two-cell static faults, in the presence and absence of bit line coupling. On the other hand, this thesis also presents the analysis, evaluation, validation and test remedies for parasitic memory effect in SRAMs. The work presents the impact of the parasitic memory effect on the detection of static faults, and clearly shows that fault detection is influenced by the presence of parasitic node components and not the resistive defect alone; something that must be considered in generating effective memory tests. In addition, the thesis presents the detection conditions and a new memory tests, March SME that targets and detects single-cell static faults, in the presence of the parasitic memory effect.
- Published
- 2011
33. Analysis and Test Development for Parasitic Fails in Deep Sub-Micron Memory Devices
- Author
-
Irobi, I.S. and SIPS, H.J.
- Subjects
memory faults ,Hardware_PERFORMANCEANDRELIABILITY ,sram ,memory tests ,parasitic fails ,tests development - Abstract
Emerging technology trends are gravitating towards extremely high levels of integration at the package and chip levels, and use of deeply scaled technology in nanometer, approaching 10nm CMOS. Challenges will arise due to the ability to design complex systems such as robots that encompass sensors, transducers, communications systems and processors, all of which require memory devices, and are required to be fault-free, and exhibit fault-tolerance, reliability and survivability characteristics. A key area of challenge is in memory testing, since deep scaling and smaller dimensions of semiconductor cell area will exacerbate the presence of complex defects and can induce effects, such as parasitic effects, which necessitate fails in memory devices. In this thesis, parasitic effects induced by spot defects in memory devices have been evaluated. The thesis presents the analysis, evaluation, validation and test remedies for parasitic fails in deep sub-micron memories. On the one hand, it presents analysis for parasitic bit line coupling effects, and the impact of bit line coupling effect on the static random access memory (SRAM) faulty behavior. Thereafter, it determines both the necessary and sufficient detection conditions for memory fault models, and demonstrates the limitations of existing industrial memory tests to adequately detect faults in the presence of bit line coupling. In addition, the thesis presents a systematic approach for test development and optimization, and new memory tests - March SSSc an optimal test that detects all single-cell static faults, and March m-MSS and March BLC that detect all two-cell static faults, in the presence and absence of bit line coupling. On the other hand, this thesis also presents the analysis, evaluation, validation and test remedies for parasitic memory effect in SRAMs. The work presents the impact of the parasitic memory effect on the detection of static faults, and clearly shows that fault detection is influenced by the presence of parasitic node components and not the resistive defect alone; something that must be considered in generating effective memory tests. In addition, the thesis presents the detection conditions and a new memory tests, March SME that targets and detects single-cell static faults, in the presence of the parasitic memory effect.
- Published
- 2011
34. Optimized March Test Flow for Detecting Memory Faults in SRAM Devices Under Bit Line Coupling
- Author
-
Luigi Dilillo, N. Badereddine, Arnaud Virazel, Patrick Girard, Alberto Bosio, L. B. Zordan, Serge Pravossoudovitch, Laboratoire d'Informatique de Robotique et de Microélectronique de Montpellier (LIRMM), Centre National de la Recherche Scientifique (CNRS)-Université de Montpellier (UM), Conception et Test de Systèmes MICroélectroniques (SysMIC), Centre National de la Recherche Scientifique (CNRS)-Université de Montpellier (UM)-Centre National de la Recherche Scientifique (CNRS)-Université de Montpellier (UM), Infineon Technologies France (INFINEON - SOPHIA), Infineon Technologies AG [München], and Peridier, Martine
- Subjects
Coupling ,Engineering ,business.industry ,[SPI.NANO] Engineering Sciences [physics]/Micro and nanotechnologies/Microelectronics ,Memory faults ,Hardware_PERFORMANCEANDRELIABILITY ,02 engineering and technology ,Capacitance ,Test flow ,020204 information systems ,0202 electrical engineering, electronic engineering, information engineering ,Electronic engineering ,Bit line ,Static random-access memory ,Fault model ,[SPI.NANO]Engineering Sciences [physics]/Micro and nanotechnologies/Microelectronics ,business ,ComputingMilieux_MISCELLANEOUS ,Test solution - Abstract
A comprehensive SRAM test must guarantee the correct functioning of each cell of the memory (ability to store and to maintain data), and the corresponding addressing, write and read operations. SRAM testing is mainly based on the concept of fault model used to mimic faulty behaviors. Traditionally, the effects of bit line coupling capacitances have not been considered during the fault analysis. However, recent works show the increasing impact of bit line coupling capacitances on the SRAM behavior. This paper reviews and discusses preview works addressing the issues coming from bit line parasitic capacitances and data contents on SRAM testing, pointing out the impacts of these effects on the existing test solutions. Then, we introduce two optimizations of the state-of-the-art test solution able to take into account the influence of bit line coupling capacitances while reducing the test length of about 60% and 80%, respectively.
- Published
- 2011
35. Detecting memory faults in the presence of bit line coupling in SRAM devices
- Author
-
Said Hamdioui, Sandra Irobi, and Zaid Al-Ars
- Subjects
Coupling ,Engineering ,business.industry ,Transistor ,Memory faults ,Hardware_PERFORMANCEANDRELIABILITY ,law.invention ,Parasitic capacitance ,law ,Fault coverage ,Electronic engineering ,Bit line ,Static random-access memory ,business ,human activities ,Frequency modulation - Abstract
The fault coverage of otherwise efficient memory tests can be dramatically reduced due to the influence of bit line coupling. This paper, analyzes the impact of parasitic bit line coupling and neighborhood coupling data backgrounds on the faulty behavior of SRAMs. It investigates and determines the worst case coupling backgrounds required to induce worst case coupling effects, and validates the analysis through defect injection and circuit simulation of all possible spot defects in the SRAM cell array. The paper clearly demonstrates the inadequacies and limitations of several industrial tests in detecting memory faults in the presence of bit line coupling. Finally, it shows how to detect all single-cell and two-cell faults, both in the absence and in the presence of bit line coupling for any possible spot defect.
- Published
- 2010
36. ENVISAT Prime Memory Faults - Some Lessons Learned
- Author
-
Michel Horblin, Daniel Mesples, and Magnus Nilsson
- Subjects
Geography ,Memory faults ,Real-time computing ,Prime (order theory) ,Remote sensing - Published
- 2010
37. Data Structures Resilient to Memory Faults: An Experimental Study of Dictionaries
- Author
-
Fabrizio Grandoni, Umberto Ferraro-Petrillo, and Giuseppe F. Italiano
- Subjects
computing with unreliable information ,experimental algorithmics ,Theoretical computer science ,fault injection ,Computer science ,Distributed computing ,memory faults ,Testbed ,Sorting ,Context (language use) ,Fault injection ,Terabyte ,Space (commercial competition) ,Data structure ,Search tree ,Theoretical Computer Science ,memory models ,Random search ,Asymptotically optimal algorithm ,sorting ,Focus (optics) ,Settore ING-INF/05 - Sistemi di Elaborazione delle Informazioni - Abstract
We address the problem of implementing data structures resilient to memory faults, which may arbitrarily corrupt memory locations. In this framework, we focus on the implementation of dictionaries and perform a thorough experimental study using a testbed that we designed for this purpose. Our main discovery is that the best-known (asymptotically optimal) resilient data structures have very large space overheads. More precisely, most of the space used by these data structures is not due to key storage. This might not be acceptable in practice, since resilient data structures are meant for applications where a huge amount of data (often of the order of terabytes) has to be stored. Exploiting techniques developed in the context of resilient (static) sorting and searching, in combination with some new ideas, we designed and engineered an alternative implementation, which, while still guaranteeing optimal asymptotic time and space bounds, performs much better in terms of memory without compromising the time efficiency.
- Published
- 2010
38. Resilient dictionaries
- Author
-
Giuseppe F. Italiano, Irene Finocchi, and Fabrizio Grandoni
- Subjects
Sequence ,Amortized analysis ,Current (mathematics) ,Computer science ,memory faults ,software-based fault tolerance ,Binary logarithm ,Data structure ,memory models ,search trees ,unreliable information ,Set (abstract data type) ,Mathematics (miscellaneous) ,Simple (abstract algebra) ,Constant (mathematics) ,Settore ING-INF/05 - Sistemi di Elaborazione delle Informazioni ,Algorithm - Abstract
We address the problem of designing data structures in the presence of faults that may arbitrarily corrupt memory locations. More precisely, we assume that an adaptive adversary can arbitrarily overwrite the content of up to δ memory locations, that corrupted locations cannot be detected, and that only O (1) memory locations are safe. In this framework, we call a data structure resilient if it is able to operate correctly (at least) on the set of uncorrupted values. We present a resilient dictionary, implementing search, insert, and delete operations. Our dictionary has O (log n + δ) expected amortized time per operation, and O ( n ) space complexity, where n denotes the current number of keys in the dictionary. We also describe a deterministic resilient dictionary, with the same amortized cost per operation over a sequence of at least δ ϵ operations, where ϵ > 0 is an arbitrary constant. Finally, we show that any resilient comparison-based dictionary must take Ω(log n + δ) expected time per search. Our results are achieved by means of simple, new techniques which might be of independent interest for the design of other resilient algorithms.
- Published
- 2009
39. Fault Diagnosis Using Test Primitives in Random Access Memories
- Author
-
Zaid Al-Ars and Said Hamdioui
- Subjects
business.industry ,Computer science ,Memory faults ,Diagnostic test ,computer.file_format ,Fault (power engineering) ,Test (assessment) ,Simple (abstract algebra) ,Embedded system ,RDF ,business ,Memory test ,computer ,Random access - Abstract
As diagnostic testing for memory devices increasingly gains in importance, companies are looking for flexible, cost effective methods to perform diagnostics on their failing devices. This paper proposes the new concept of test primitives as a method to diagnose memory faults. Test primitives provide an easy-to-use, extensible, low-cost memory fault diagnosis method that is universally applicable, since it uses simple platform-independent test sequences. The paper defines the concept of test primitives, shows their importance and gives examples to the way they are derived and used in a memory test environment.
- Published
- 2009
40. Counting in the Presence of Memory Faults
- Author
-
Thomas Mølhave, Gerth Stølting Brodal, Allan Grønlund Jørgensen, and Gabriel Moruz
- Subjects
Soft error ,Additive error ,Computer science ,Memory cell ,Memory faults ,Value (computer science) ,Arithmetic ,Algorithm ,Upper and lower bounds - Abstract
The faulty memory RAM presented by Finocchi and Italiano [1] is a variant of the RAM model where the content of any memory cell can get corrupted at any time, and corrupted cells cannot be distinguished from uncorrupted cells. An upper bound, ?, on the number of corruptions and O(1) reliable memory cells are provided. In this paper we investigate the fundamental problem of counting in faulty memory. Keeping many reliable counters in the faulty memory is easily done by replicating the value of each counter ?(?) times and paying ?(?) time every time a counter is queried or incremented. In this paper we decrease the expensive increment cost to o(?) and present upper and lower bound tradeoffs decreasing the increment time at the cost of the accuracy of the counters.
- Published
- 2009
41. Optimal resilient sorting and searching in the presence of dynamic memory faults
- Author
-
Giuseppe F. Italiano, Irene Finocchi, and Fabrizio Grandoni
- Subjects
algorithm ,sorting ,false coins ,Sorting algorithm ,General Computer Science ,Memory faults ,Combinatorial algorithms, SortingSearching, Memory faults, Memory models, Computing with unreliable information ,Theoretical Computer Science ,Computing with unreliable information ,Memory models ,Searching ,Integer ,Search algorithm ,Integer sorting ,Combinatorial algorithms ,Time complexity ,Mathematics ,Discrete mathematics ,Sorting ,Binary logarithm ,Randomized algorithm ,SortingSearching ,Settore ING-INF/05 - Sistemi di Elaborazione delle Informazioni ,Algorithm ,Computer Science(all) - Abstract
We investigate the problem of reliable computation in the presence of faults that may arbitrarily corrupt memory locations. In this framework, we consider the problems of sorting and searching in optimal time while tolerating the largest possible number of memory faults. In particular, we design an O(nlogn) time sorting algorithm that can optimally tolerate up to $O(\sqrt{n\log n}\,)$ memory faults. In the special case of integer sorting, we present an algorithm with linear expected running time that can tolerate $O(\sqrt{n}\,)$ faults. We also present a randomized searching algorithm that can optimally tolerate up to O(logn) memory faults in O(logn) expected time, and an almost optimal deterministic searching algorithm that can tolerate O((logn)1−e) faults, for any small positive constant e, in O(logn) worst-case time. All these results improve over previous bounds.
- Published
- 2009
42. Sorting and searching in faulty memories
- Author
-
Irene Finocchi and Giuseppe F. Italiano
- Subjects
combinatorial algorithms ,computing with unreliable information ,memory faults ,memory models ,searching ,sorting ,Sorting algorithm ,General Computer Science ,Applied Mathematics ,Sorting ,Fault tolerance ,Upper and lower bounds ,Computer Science Applications ,Set (abstract data type) ,Complete information ,Theory of computation ,Settore ING-INF/05 - Sistemi di Elaborazione delle Informazioni ,Algorithm ,Mathematics ,Analysis of algorithms - Abstract
In this paper we investigate the design and analysis of algorithms resilient to memory faults. We focus on algorithms that, despite the corruption of some memory values during their execution, are nevertheless able to produce a correct output at least on the set of uncorrupted values. In this framework, we consider two fundamental problems: sorting and searching. In particular, we prove that any O(nlog n) comparison-based sorting algorithm can tolerate the corruption of at most O((nlog n)1/2) keys. Furthermore, we present one comparison-based sorting algorithm with optimal space and running time that is resilient to O((nlog n)1/3) memory faults. We also prove polylogarithmic lower and upper bounds on resilient searching.
- Published
- 2008
43. An overview of deterministic functional RAM chip testing
- Author
-
A. J. van de Goor and C. A. Verruijt
- Subjects
General Computer Science ,Hierarchy (mathematics) ,Cover (telecommunications) ,Computer science ,Memory faults ,Parallel computing ,Fault (power engineering) ,Chip ,Memory test ,Theoretical Computer Science - Abstract
This paper presents an overview of deterministic functional RAM chip testing. Instead of the traditional ad-hoc approach toward developing memory test algorithms, a hierarchy of functional faults and tests is presented, which is shown to cover all likely functional memory faults. This is done by presenting a novel way of categorizing the faults. All (possible) fault combinations are discussed. Requirements are put forward under which conditions a fault combination can be detected. Finally, memory test algorithms that satisfy the given requirements are presented.
- Published
- 1990
44. Evaluation and design of an ultra-reliable distributed architecture for fault tolerance
- Author
-
Chris J. Walter
- Subjects
Engineering ,business.industry ,Reliability (computer networking) ,Memory faults ,Computer testing ,Fault tolerance ,Avionics ,Reliability engineering ,Embedded system ,Electrical and Electronic Engineering ,Architecture ,Safety, Risk, Reliability and Quality ,business ,Ic devices ,Automatic testing - Abstract
The issues related to the experimental evaluation of an early conceptual prototype of the MAFT (multicomputer architecture for fault tolerance) architecture are discussed. A completely automated testing approach was designed to allow fault-injection experiments to be performed, including stuck-at and memory faults. Over 2000 injection tests were run and the system successfully tolerated all faults. Concurrent with the experimental evaluation, an analytic evaluation was carried out to determine if higher levels of reliability could be achieved. The lessons learned in the evaluation phase culminated in a new design of the MAFT architecture for applications needing ultrareliability. The design uses the concept of redundantly self-checking functions to address the rigid requirements proposed for a future generation of mission-critical avionics. The testing of three subsystems critical to the operation of the new MAFT is presented with close to 50-k test cycles performed over 51 different IC devices to verify the designs. >
- Published
- 1990
45. Precise Identification of Memory Faults Using Electrical Simulation
- Author
-
Zaid Al-Ars, Said Hamdioui, and Georgi Gaydadjiev
- Subjects
Engineering ,business.industry ,Memory faults ,Hardware_PERFORMANCEANDRELIABILITY ,Fault (power engineering) ,Fault indicator ,Reliability engineering ,Stuck-at fault ,Identification (information) ,Fault coverage ,Memory architecture ,Fault model ,business ,Algorithm - Abstract
Recently, a framework describing the space of all fault models has been established. Subsequently, it has been shown that many new faults of that space do exist. Gradually, The number and complexity of observed memory fault models has been gradually increasing. As a result, it has become increasingly difficult to identify the precise functional fault models that a memory suffers from. This paper shows that there are two types of possible imprecision in describing faults: under specification, which leads to tests with insufficient fault coverage, and over specification, which leads to time-inefficient tests. A general method is presented to analyze faulty memory behavior based on electrical simulation and map it precisely onto the corresponding fault models, which makes it possible to generate time-optimal tests with optimal fault coverage.
- Published
- 2007
46. Testing Loaded Programs Using Fault Injection Technique
- Author
-
S. Manaseer, F. A. Masooud, and A. A. Sharieh
- Subjects
Complex software systems ,Error detection ,Process and virtual memory ,Memory faults ,Fault tolerance ,Injection and testing methodology - Abstract
Fault tolerance is critical in many of today's large computer systems. This paper focuses on improving fault tolerance through testing. Moreover, it concentrates on the memory faults: how to access the editable part of a process memory space and how this part is affected. A special Software Fault Injection Technique (SFIT) is proposed for this purpose. This is done by sequentially scanning the memory of the target process, and trying to edit maximum number of bytes inside that memory. The technique was implemented and tested on a group of programs in software packages such as jet-audio, Notepad, Microsoft Word, Microsoft Excel, and Microsoft Outlook. The results from the test sample process indicate that the size of the scanned area depends on several factors. These factors are: process size, process type, and virtual memory size of the machine under test. The results show that increasing the process size will increase the scanned memory space. They also show that input-output processes have more scanned area size than other processes. Increasing the virtual memory size will also affect the size of the scanned area but to a certain limit., {"references":["Broadwell P., Sastry N., and Traupma J., (2001). \"FIG: A Prototype\nTool for Online Verification of Recovery Mechanisms \". ICS SHAMAN Workshop -02 New York, New York USA Copyright 2001\nACM.","Broadwell, P. and Ong E. (2002).\"A Comparison of Static Analysis\nand Fault Injection Techniques for Developing Robust System\nServices\", a project paper, retrieved (January2004), from\n(http://www.cs.berkeley.edu/~pbwell/papers/saswifi.pdf).","Lai M. and Wang S., (1995). \"Software Fault Tolerance\", Wiley & Sons LTd, New York.","Manaseer S. (2004). \"Software Testing Using Software Fault\nInjection\". A Master Thesis in Computer Science , KASIT, University\nof Jordan, Jordan.","Sanders W., (2003).\" Fault Injection Methods and Mechanisms\".\nDepartment of Electrical and Computer Engineering and Coordinated\nScience Laboratory, University of Illinois at Urbana-Champaign.","Silberschatz A. Galvin P., and Gagne G., (2003). Operating System\nConcepts. Wiley & Sons, New York.","Thorhuns R., (2000). \"Software Fault Injection Testing\". A Master\nthesis in Electronic System Design. KTH, Stockholm.","Torres-Pomales W., (2000).\"Software Fault Tolerance: A Tutorial\".\nNASA Center for Aerospace Information (CASI) National Technical\nInformation Service (NTIS).","Voas J., Charron F., and McGraw G., (1997). \"Predicting How Badly\nÔÇÿGood- Software Can Behave\", IEEE Software, 14(4):73-83.\n[10] Voas M., McGraw G., (1998). Software Fault Injection, Inoculating\nPrograms against Errors. Wiley & Sons, New York."]}
- Published
- 2007
- Full Text
- View/download PDF
47. Designing reliable algorithms in unreliable memories
- Author
-
Giuseppe F. Italiano, Irene Finocchi, and Fabrizio Grandoni
- Subjects
sorting networks ,Correctness ,General Computer Science ,high-performance ,Language change ,Computer science ,Computation ,Memory faults ,faults ,information ,Theoretical Computer Science ,Set (abstract data type) ,errors ,Settore ING-INF/05 - Sistemi di Elaborazione delle Informazioni ,Algorithm - Abstract
Some of the present day applications run on computer platforms with large and inexpensive memories, which are also error-prone. Unfortunately, the appearance of even very few memory faults may jeopardize the correctness of the computational results. We say that an algorithm is resilient to memory faults if, despite the corruption of some memory values before or during its execution, it is nevertheless able to get a correct output at least on the set of uncorrupted values (i.e., the algorithm works correctly on uncorrupted data). In this paper we will survey some recent works on resilient algorithms and try to give some insight into the main algorithmic techniques used.
- Published
- 2007
48. Resilient Search Trees
- Author
-
Irene FINOCCHI, Grandoni, F., and Italiano, G. F.
- Subjects
Amortized time ,Memory locations ,Memory faults - Published
- 2007
49. Analysis and Test Development for Parasitic Fails in Deep Sub-Micron Memory Devices
- Author
-
Irobi, I.S. (author) and Irobi, I.S. (author)
- Abstract
Emerging technology trends are gravitating towards extremely high levels of integration at the package and chip levels, and use of deeply scaled technology in nanometer, approaching 10nm CMOS. Challenges will arise due to the ability to design complex systems such as robots that encompass sensors, transducers, communications systems and processors, all of which require memory devices, and are required to be fault-free, and exhibit fault-tolerance, reliability and survivability characteristics. A key area of challenge is in memory testing, since deep scaling and smaller dimensions of semiconductor cell area will exacerbate the presence of complex defects and can induce effects, such as parasitic effects, which necessitate fails in memory devices. In this thesis, parasitic effects induced by spot defects in memory devices have been evaluated. The thesis presents the analysis, evaluation, validation and test remedies for parasitic fails in deep sub-micron memories. On the one hand, it presents analysis for parasitic bit line coupling effects, and the impact of bit line coupling effect on the static random access memory (SRAM) faulty behavior. Thereafter, it determines both the necessary and sufficient detection conditions for memory fault models, and demonstrates the limitations of existing industrial memory tests to adequately detect faults in the presence of bit line coupling. In addition, the thesis presents a systematic approach for test development and optimization, and new memory tests - March SSSc an optimal test that detects all single-cell static faults, and March m-MSS and March BLC that detect all two-cell static faults, in the presence and absence of bit line coupling. On the other hand, this thesis also presents the analysis, evaluation, validation and test remedies for parasitic memory effect in SRAMs. The work presents the impact of the parasitic memory effect on the detection of static faults, and clearly shows that fault detection is influenced by the, Microelectronics & Computer Engineering, Electrical Engineering, Mathematics and Computer Science
- Published
- 2011
50. Automatic March tests generation for multi-port SRAMs
- Author
-
Paolo Prinetto, S. Di Carlo, Alfredo Benso, G. DiNatale, and Alberto Bosio
- Subjects
Random access memory ,Hardware_MEMORYSTRUCTURES ,Computer science ,business.industry ,Data buses ,Memory faults ,DIGITAL SYSTEM DESIGN TEST AND VERIFICATION ,Electronic equipment testing ,Automatic testing ,Digital systems ,Port (computer networking) ,Set (abstract data type) ,Null (SQL) ,Embedded system ,business ,Multi port - Abstract
Testing of Multi-Port (MP) SRAMs requires special tests since the multiple and simultaneous access can sensitize faults that are different from the conventional single-port memory faults. In spite of their growing use, few works have been published on testing MP memories. In addition, most of the published work concentrated only on two ports memories (i.e., 2P memories). This paper presents a methodology to automatically generate march tests for MP memories. It is based on generations of single port memory march test firstly, then extending it to test a generic MP SRAMs. A set of experimental results shows the effectiveness of the proposed solution.
- Published
- 2006
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.