226,864 results on '"Computer Science"'
Search Results
2. Social Information Processing in Children: an ocUlo-pupillometric Tool for Standard Evaluation (SIRCUS)
- Author
-
UMR 1253, iBrain, Université de Tours, Inserm, Tours, France., Laboratory of Fundamental and Applied Computer Science of Tours, EA6300, National Research Agency, France, and Ministry of Health, France
- Published
- 2023
3. The DIALOGUE Study: Swiss-Korean Billateral Collaboration (DIALOGUE)
- Author
-
Yonsei University, University Hospital, Basel, Switzerland, Hopital du Jura, Delemont, Switzerland, Department of Computer Science Yonsei University, Seoul, Korea, National Cancer Center, Korea, Chungnam National University, Ente Ospedaliero Cantonale, Ticino, Switzerland, Swiss National Science Foundation, Insel Gruppe AG, University Hospital Bern, University Hospital, Geneva, and Maria Katapodi, Professor of Nursing Science, Department of Clinical Research
- Published
- 2023
4. Analysis of Factors Determining Increase of Serum Sodium in Hyponatremic Patients
- Author
-
Department of Mathematics and Computer Science, University of Technology Eindhoven and Volker Burst, Prof. MD
- Published
- 2023
5. D-Lung: An Analytics Platform for Lung Cancer Based on Deep Learning Technology
- Author
-
Department of Computer Science & Engineering, CUHK and Professor Winnie W.C. Chu, Professor
- Published
- 2023
6. Towards regression-free and source-free online domain adaptation
- Author
-
Leung, Carson (Computer Science), Mohammed, Noman (Computer Science), Wang, Yang, Wang, Shaowei, Musabe, Taif Al, Leung, Carson (Computer Science), Mohammed, Noman (Computer Science), Wang, Yang, Wang, Shaowei, and Musabe, Taif Al
- Abstract
In real-world applications, inconsistent behavior displayed by AI models before and after updates can have severe consequences, particularly in safety-critical systems like autonomous driving. The updated model may exhibit regression issues, leading to incorrect predictions that were previously accurate. While regression problems have been extensively studied in various contexts, their investigation within the realm of online domain adaptation remains limited. This poses a unique challenge due to the continuous updating of the old model with incoming data streams, potentially worsening regression problems in the source domain. In this work, we address the regression problem in online domain adaptation by mitigating regression rates in the source data while adapting to a streaming target data. Our proposed approach introduces a novel loss incorporated into the Crodobo framework. This ensures that the new model learns from the source domain in a supportive manner, avoiding the acquisition of irrelevant information and preserving privacy by removing data after successful domain adaptation. While the Crodobo framework relies on raw source data, which may contain sensitive information. Hence, it does not guarantee overall privacy protection. Consequently, we also address the privacy concerns by replacing direct access to raw data with synthetic data. Through a generative model, we force the use of synthetic data, transforming online domain adaptation into a source-free process. This guarantees the confidentiality of source data, minimizes storage requirements, and enhances the practicality of the system. Our methodologies undergoe extensive evaluation on datasets such as Visda-C, COVID-DA, and MNIST - USPS. The experimental results validate the effectiveness of our approaches.
- Published
- 2024
7. Bioinformatics, game development and citizen science: connecting the dots
- Author
-
Bunt, Andrea (Computer Science), Domaratzki, Michael (Computer Science), Chamberlain, Jon (University of Essex), Tremblay-Savard, Olivier, de Leon Pereira, Rogerio, Bunt, Andrea (Computer Science), Domaratzki, Michael (Computer Science), Chamberlain, Jon (University of Essex), Tremblay-Savard, Olivier, and de Leon Pereira, Rogerio
- Abstract
Despite the increasing power of computers, there are still complex problems that can benefit from human abilities such as cognition, reasoning, creativity, and problem-solving. These tasks can be broken down into smaller components that humans can solve individually on crowdsourcing platforms, which is a process usually described as human computation. Similarly, citizen science, a form of human computation, invites individuals, known as citizen scientists, to actively participate in scientific research. It is also possible to incorporate gamification elements or turn the tasks into puzzles that can be solved within a video game. This thesis explores using a citizen science game, GeSort, to solve the genome sorting problem. The goal of sorting genomes is to find the smallest set of evolutionary events capable of transforming one genome into another. GeSort represents the sequences of genes of the genomes being compared as series of colored shapes. This allows players to see patterns of matches and mismatches and then use a sequence of operations such as duplications, deletions and inversions to transform one genome into the other. Through the analysis of GeSort, this thesis aims to propose different guidelines and tools to help the citizen science community build improved games and applications. We started our research by evaluating the effects of including educational content in a citizen science game. Next, we developed FORGE, a framework for organizing rewards in gamified environments. Finally, we analyzed data collected through GeSort matches played over a year. Our results suggest that educational content in a citizen science game can increase players' engagement, retention and performance. The use of FORGE can speed up the implementation of a reward system, saving development time and financial resources. Analysis of the data collected through GeSort showed, among other things, that some visual patterns could confuse players, which allowed us to formulate guide
- Published
- 2024
8. Zero-shot synthesis of compilable code for incomplete code snippets using LLMs
- Author
-
Chowdhury, Shaiful (Computer Science), Tabiban, Azadeh (Computer Science), Wang, Shaowei, Kabir, Azmain, Chowdhury, Shaiful (Computer Science), Tabiban, Azadeh (Computer Science), Wang, Shaowei, and Kabir, Azmain
- Abstract
Technical Q&A sites and Large Language Models (LLMs) are valuable resources for software developers seeking knowledge. However, the provided code snippets are often uncompilable and incomplete due to unresolved types, missing libraries, and incorrect import statements. This poses a challenge for users who wish to reuse or analyze these snippets. Existing methods either do not focus on creating compilable code or have low success rates. To address this, we propose ZS4C, a lightweight approach for Zero-shot Synthesis of compilable code For incomplete Code snippets using LLMs. ZS4C operates in two stages: first, it uses an LLM, like GPT-3.5, to identify missing import statements in a snippet; second, it collaborates with a validator (e.g., compiler) to fix compilation errors caused by incorrect imports and syntax issues. We evaluated ZS4C on the StatType-SO benchmark and a new dataset, Python-SO, which includes 539 Python snippets from Stack Overflow across the 20 most popular Python libraries. ZS4C significantly outperforms existing methods, improving the compilation rate from 63% to 95.1% compared to the state-of-the-art SnR, marking a 50.1% improvement. On average, ZS4C can infer more accurate import statements (with an F1 score of 0.98) than SnR, with an improvement of 8.5% in the F1.
- Published
- 2024
9. Active risk management in dynamic teams of heterogeneous robots
- Author
-
Bunt, Andrea (Computer Science), McNeill, Dean (Electrical and Computer Engineering), Baltes, Jacky (Computer Science), Anderson, John, Changizi, Soheil, Bunt, Andrea (Computer Science), McNeill, Dean (Electrical and Computer Engineering), Baltes, Jacky (Computer Science), Anderson, John, and Changizi, Soheil
- Abstract
In Urban Search and Rescue (USAR), hazards such as structural collapse and fire can significantly endanger robots. To mitigate these risks, it is crucial to plan task allocations that adapt to dynamic environments. Most available strategies begin with evaluating the mission in advance and formulating a static plan, which may be inflexible for unforeseen changes. This project seeks to expand our lab’s existing framework by integrating Active Risk Management (ARM) into ongoing missions. The ARM module enhances adaptability by continuously monitoring environmental hazards and initiating risk mitigation tasks. Additionally, a novel method for detecting and escaping local minima allows robots to adjust their navigation patterns, preventing immobilization. The simulation environment now features realistic fire propagation, introducing a dynamic element that rigorously tests the effectiveness of the robots’ risk management strategies. By incorporating these advancements, we aim to increase mission success rates and significantly reduce robot damage in challenging USAR scenarios.
- Published
- 2024
10. An analytical framework to examine and describe people’s expectations of robots
- Author
-
Bunt, Andrea (Computer Science), Gerhard, David (Computer Science), Young, James E., Berzuk, James M., Bunt, Andrea (Computer Science), Gerhard, David (Computer Science), Young, James E., and Berzuk, James M.
- Abstract
We engaged with the problem of expectation discrepancy in human-robot interaction: a known challenge in which the expectations people form when interacting with a social robot may not align with its actual capabilities. This misalignment, an expectation discrepancy, can disappoint users and hinder interaction. While research has proposed ways to mitigate expectation discrepancy, designers lack a systematic approach to analyzing and describing expectations people form of their robot. A more rigorous theoretical framework is a necessary step towards designing robots to purposefully engineering desired expectations. We consulted theories and models from psychology and sociology on expectations between people, and conducted a survey of expectations in human-robot interactions. Through this we developed an analytical framework consisting of a novel model of the cognitive process of human-robot expectation formation, as well as a taxonomy for classifying the types of expectations they form. We finally propose preliminary methods for designers to use this framework as a tool to support systematic analysis of how and why people form expectations of a given robot and what those expectations may be. Such understanding can empower designers with greater control over people’s expectations, enabling them to combat problems of expectation discrepancy.
- Published
- 2024
11. SecureLoop: Design Space Exploration of Secure DNN Accelerators
- Author
-
Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Lee, Kyungmi, Yan, Mengjia, Emer, Joel, Chandrakasan, Anantha, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Lee, Kyungmi, Yan, Mengjia, Emer, Joel, and Chandrakasan, Anantha
- Abstract
Deep neural networks (DNNs) are gaining popularity in a wide range of domains, ranging from speech and video recognition to healthcare. With this increased adoption comes the pressing need for securing DNN execution environments on CPUs, GPUs, and ASICs. While there are active research efforts in supporting a trusted execution environment (TEE) on CPUs, the exploration in supporting TEEs on accelerators is limited, with only a few solutions available [18, 19, 27]. A key limitation along this line of work is that these secure DNN accelerators narrowly consider a few specific architectures. The design choices and the associated cost for securing these architectures do not transfer to other diverse architectures. This paper strives to address this limitation by developing a design space exploration tool for supporting TEEs on diverse DNN accelerators. We target secure DNN accelerators equipped with cryptographic engines where the cryptographic operations are closely coupled with the data movement in the accelerators. These operations significantly complicate the scheduling for DNN accelerators, as the scheduling needs to account for the extra on-chip computation and off-chip memory accesses introduced by these cryptographic operations, and even needs to account for potential interactions across DNN layers. We tackle these challenges in our tool, called SecureLoop, by introducing a scheduling search engine with the following attributes: 1) considers the cryptographic overhead associated with every off-chip data access, 2) uses an efficient modular arithmetic technique to compute the optimal authentication block assignment for each individual layer, and 3) uses a simulated annealing algorithm to perform cross-layer optimizations. Compared to the conventional schedulers, our tool finds the schedule for secure DNN designs with up to 33.2% speedup and 50.2% improvement of energy-delay-product.
- Published
- 2024
12. Multi-color Holograms Improve Brightness in Holographic Displays
- Author
-
Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Kavakl?, Koray, Shi, Liang, Urey, Hakan, Matusik, Wojciech, Ak?it, Kaan, Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Kavakl?, Koray, Shi, Liang, Urey, Hakan, Matusik, Wojciech, and Ak?it, Kaan
- Abstract
Holographic displays generate Three-Dimensional (3D) images by displaying single-color holograms time-sequentially, each lit by a single-color light source. However, representing each color one by one limits brightness in holographic displays. This paper introduces a new driving scheme for realizing brighter images in holographic displays. Unlike the conventional driving scheme, our method utilizes three light sources to illuminate each displayed hologram simultaneously at various intensity levels. In this way, our method reconstructs a multiplanar three-dimensional target scene using consecutive multi-color holograms and persistence of vision. We co-optimize multi-color holograms and required intensity levels from each light source using a gradient descent-based optimizer with a combination of application-specific loss terms. We experimentally demonstrate that our method can increase the intensity levels in holographic displays up to three times, reaching a broader range and unlocking new potentials for perceptual realism in holographic displays.
- Published
- 2024
13. Future Patient - Telerehabilitation of Heart Failure Patients
- Author
-
Aage og Johanne Louis-Hansens Fond, Viewcare A/S, Laboratory of Welfare Technologies - Telehealth & Telerehabilitation, SMI, Department of Health Science and Technology, Aalborg University, Regionshospitalet Viborg, Skive, Technical University of Denmark, University of Aarhus, Danish Heart Foundation, Viborg Healthcare Center, Skive Healthcare Center, Odense University Hospital, Department of Computer Science, AAU, and Birthe Dinesen, Professor
- Published
- 2021
14. Cumulative link models for deep ordinal classification
- Author
-
Vargas, Víctor Manuel, Gutiérrez, Pedro Antonio, and Hervás-Martínez, César
- Published
- 2020
- Full Text
- View/download PDF
15. FashionVLM - Fashion Captioning Using Pretrained Vision Transformer and Large Language Model
- Author
-
Computer Science, Gaurika Gupta, primary and Computer Division, Pritam Shete, additional
- Published
- 2024
- Full Text
- View/download PDF
16. Determination of the pubertal growth spurt by artificial intelligence analysis of cervical vertebrae maturation in lateral cephalometric radiographs
- Author
-
Nogueira-Reis, PhD in Oral Radiology Fernanda, primary, Cascante-Sequeira, PhD student in Oral Radiology Deivi, additional, Farias-Gomes, PhD in Oral Radiology Amanda, additional, de Macedo, PhD in Computer Science Maysa Malfiza Garcia, additional, Watanabe, PhD in Electrical Engineering Renato Naville, additional, Santiago, PhD in Electrical Engineering Anderson Gabriel, additional, Tabchoury, PhD in Dentistry (Pharmacology) Cínthia Pereira Machado, additional, and de Freitas, PhD in Oral Radiology Deborah Queiroz, additional
- Published
- 2024
- Full Text
- View/download PDF
17. Constantin Gaindric -- 80th anniversary
- Author
-
The Editorial Board of Computer Science Journal of Moldova
- Subjects
Electronic computers. Computer science ,QA75.5-76.95 - Abstract
On September 11, 2021 Prof. Constantin Gaindric turns 80! This age does not track in any way with this person when you look at him.
- Published
- 2021
18. Assessing the effect of preprocessing of clinical notes on classification tasks and similarity measures
- Author
-
Katz, Alan (Community Health Sciences), Noman, Mohammed (Computer Science), Lix, Lisa, Moni, Md Moniruzzaman, Katz, Alan (Community Health Sciences), Noman, Mohammed (Computer Science), Lix, Lisa, and Moni, Md Moniruzzaman
- Abstract
Background: Unstructured text data (UTD) in electronic medical records (EMRs) may be challenging to use because of noise including spelling errors, abbreviations, and punctuation symbols. It is expected that preprocessing improves the performance of statistical or machine learning models. Objectives: The research objectives were to assess the effect of the number and order of preprocessing methods (1) on the detection of health conditions from UTD, (2) on clinical and demographic cohort selection criteria in UTD, (3) on the similarity of information contained in pairs of EMR notes for the same patient, and (4) on accurate de-identification of UTD. Method: Study data were from the Informatics for Integrating Biology and the Bedside (i2b2) challenges. The 2008, 2014 and 2018 i2b2 datasets were used for different Objectives. Preprocessing methods included tokenization, removing punctuation symbols, correcting spelling errors, expanding abbreviations, word stemming, and lemmatization. A nested experimental design was adopted, in which order was nested within the number of methods. Balanced random forest, support vector machine, and bidirectional long short-term memory-conditional random field models were used for Objectives 1, 2, and 4, respectively, and model performance was evaluated by accuracy, sensitivity, specificity, F1 score, and precision. For Objective 3, cosine similarity was used to measure the similarity between pairs of notes. Analysis of Variance (ANOVA) and descriptive statistics were used to test research hypotheses. Results: Mean sensitivity, specificity, F1 score, accuracy, and precision were similar across the orders of methods and numbers of methods, for Objectives 1 and 2. Cosine similarity scores were similar across orders of methods for Objective 3. Deep learning models for Objective 4 were not trainable with the preprocessed data. The ANOVA F test results showed no significant effect of the order of methods for different numbers and identical me
- Published
- 2024
19. Developing a comparative framework for machine-learning classifying models based on the Emergency Severity Index triage system
- Author
-
Sherif, Sherif (Electrical and Computer Engineering), Henry, Christopher (Computer Science), Eskicioglu, Rasit, Karajeh, Ala', Sherif, Sherif (Electrical and Computer Engineering), Henry, Christopher (Computer Science), Eskicioglu, Rasit, and Karajeh, Ala'
- Abstract
Emergency departments are among the most crowded facilities in healthcare premises, where they receive a variety of cases, including critical ones and life-threatening conditions. Arranging the order of received patients and providing timely and efficient care is of utmost importance. This procedure is usually carried out by a nurse who considers the patient’s symptoms and vital signs besides ready resources. Existing literature revealed that there is variability in the accuracy of the triage process inside emergencies for a variety of reasons. Therefore, developing an aid tool based on Machine Learning (ML) algorithms would help mitigate this issue and improve the workflow inside such a crucial setting. This work provides a comparison between several ML-based classifying models that were developed from MIMIC-IV-ED and MIMIC-IV databases. Moreover, it presents insights into hidden patterns that explain some outcomes of subgroups in the examined individuals.
- Published
- 2024
20. Zombies and survivors on graph classes
- Author
-
Gunderson, Karen (Mathematics), Durocher, Stephane (Computer Science), Miller, Avery, Liu, Fengyi, Gunderson, Karen (Mathematics), Durocher, Stephane (Computer Science), Miller, Avery, and Liu, Fengyi
- Abstract
Zombies and Survivors is a two-player graph game introduced by Fitzpatrick et al. as a variant of Cops and Robbers. These games can be used to model situations involving pursuit-evasion or search-and-rescue. Two players take turns to move their pieces in a given graph, with the condition that the zombie pieces must move toward the survivor along a shortest path during the zombie player’s turn. The minimum number of zombies that is sufficient to catch the survivor is called the graph's zombie number. First, we give a construction of graphs with zombie number equal to any given k, and extend the result to a construction of graphs that simultaneously have zombie number equal to any given k and cop number equal to any given m. Then, we investigate specific graph classes. We determine the exact zombie number of the Cartesian product of a cycle and a path, and show that this quantity depends on the parity of the cycle. Upper bounds on the zombie number for strong products, Cartesian products and lexicographic products are given. We also show that the throttling numbers of Cartesian products and strong products of two paths are asymptotic to a sublinear polynomial in the length of the path. Finally, the class of graphs formed by a set of cycles with one common center vertex is investigated, and we give a characterization of such graphs that have zombie number 1 or 2.
- Published
- 2024
21. Research on size reduction of magnetic components in grid-connected converters
- Author
-
Filizadeh, Shaahin (Electrical and Computer Engineering), Leung, Carson (Computer Science), Moallem, Mehrdad (Simon Fraser University), Ho, Carl, Dadkhah, Jalal, Filizadeh, Shaahin (Electrical and Computer Engineering), Leung, Carson (Computer Science), Moallem, Mehrdad (Simon Fraser University), Ho, Carl, and Dadkhah, Jalal
- Abstract
The rapid pace of electrification and adoption of renewable energy sources necessitate the development of efficient, cost-effective, and compact power electronic converters. Among the essential components in modern converters, magnetic components play a crucial role, with their size and quantity varying depending on the application, power rating, and converter topology. Reducing the size and number of magnetic components lead to higher converter efficiency, lower cost, and higher power density. This thesis presents two distinct methods for reducing the size and quantity of magnetic components in single-phase and three-phase grid-connected converters. In single-phase converters, an innovative interleaving method is proposed, which effectively increases the power rating of a commercial Gallium Nitride (GaN)-based Power Factor Correction (PFC) converter. The proposed interleaving method along with GaN transistor and an advanced controller operating in Discontinuous Conduction Mode (DCM) lead to substantial inductor size reduction in the PFC converter and Electromagnetic Interference (EMI) filter. For three-phase converters, a novel set of topologies known as "reconfigurable filter" is introduced, minimizing the number of magnetic components. These reconfigurable filters construct LCL filters using only three inductors, a notable reduction compared to the conventional LCL filters requiring six inductors. The key components of the reconfigurable filters are a set of low-frequency bidirectional switches which lead to full utilization of existing magnetic components by changing their roles from the grid side to the converter side. The proposed topologies achieve comparable Total Harmonic Distortion (THD) performance compared to conventional LCL filters while utilizing three fewer inductors. Additionally, the reconfigurable filters suppress leakage current and offer reactive power support, making them versatile for diverse applications. The proposed topologies are applied i
- Published
- 2024
22. Parallelization of hybrid multi-objective evolutionary algorithm on multi-core architectures
- Author
-
Lui, Shaun (Mathematics), Kamali, Shahin (Computer Science), Thulasiraman, Parimala, Sun, Zhuoran, Lui, Shaun (Mathematics), Kamali, Shahin (Computer Science), Thulasiraman, Parimala, and Sun, Zhuoran
- Abstract
Many real world optimization problems involve multiple conflicting objectives, constraints and parameters. Multi-objective optimization (MOO) techniques are used to solve these problems. The goal of MOO is to find a set of optimal solutions, or the Pareto optimal front. Multi-objective evolutionary algorithms are heuristics that evolve a population of candidate solutions to find the Pareto optimal front in a single run. The selection criterion used to select individuals in the population play an important role in determining the quality of the solutions. Pareto-based algorithms use the Pareto selection criterion to evolve different parts of the solution space introducing diverse solutions, but converge slowly to the optimal front. On the other hand, non- Pareto selection Criterion (NPC) algorithms converge faster to the Pareto front, but in the process eliminate other diverse solutions. To compensate for the strengths and weaknesses of PC and NPC, hybrid frameworks such as BCE (bi-criterion evolutionary) have been proposed. In BCE, the PC and NPC algorithms evolve separately, but also co-operate by exchanging information to explore and exploit the objective space. In the literature, two well-known evolutionary algorithms, Non-dominated Sorting Genetic Algorithm II (NSGA-II) (PC) and Multi-objective Evolutionary Algorithm based on Decomposition (MOEA/D) (NPC) have been used as a case study in the BCE framework. However, the individual algorithms are computationally expensive. In this thesis, we study the parallelization of the BCE framework. NSGA-II is highly data parallel, and is well suited for single instruction multiple data architectures. MOEA/D is non-data parallel with some parts of the algorithm being sequential. Therefore, we design the parallel NSGA-II algorithm on the GPU multi-core accelerator and parallel MOEA/D algorithm on multi-core CPU machines using an island model. Using the travelling salesperson benchmark data sets we analyze the performance of t
- Published
- 2024
23. Ant colony optimization with distributed colonies for dynamic environments on multiple GPUs
- Author
-
Mohammed, Noman (Computer Science), Ferens, Ken (Electrical and Computer Engineering), Thulasiraman, Parimala, Wiens, Emanuel, Mohammed, Noman (Computer Science), Ferens, Ken (Electrical and Computer Engineering), Thulasiraman, Parimala, and Wiens, Emanuel
- Abstract
Dynamic environments pose many challenges as the search space is irregular, un- structured, with the data and problem space changing over time. The algorithms executing on these environments should adapt to the varying dynamic conditions. In this research we consider Ant Colony Optimization algorithm (ACO), a technique inspired by real ants in nature and therefore, should be adaptable to dynamic environ- ments. However, some studies in the literature show the contrary. Population-based ACO was introduced, a hybrid technique that combines concepts from Genetic Algo- rithms for solving problems in dynamic environments. In this thesis, we argue and show that ACO is as good as PACO or even better in some cases, by incorporating lo- cal search techniques to exploit the search space, tuning parameters in the algorithm to explore the search space and, using migration between multiple colonies (or island model) for convergence. The multiple colonies are implemented on multiple GPUs for efficiency. We perform various experiments on a dynamic travelling salesperson dataset and compare ACO and PACO with local search and island model. We also show that the parameter tuning has a significant influence on the accuracy.
- Published
- 2024
24. Optimization of geometric measures of sets of moving objects
- Author
-
Li, Ben (Computer Science), Kirkland, Steve (Mathematics), Bose, Prosenjit (Carleton University), Durocher, Stephane, Penha Costa, Ikaro Ruan, Li, Ben (Computer Science), Kirkland, Steve (Mathematics), Bose, Prosenjit (Carleton University), Durocher, Stephane, and Penha Costa, Ikaro Ruan
- Abstract
Given a set S of objects, each moving with linear motion in R^d, consider the diameter D(S, t) of S at time t. In this thesis we explore optimization of extent and proximity measures of S. For instance, one possibility is to identify minimum diameter D(S, t) of S over the domain of time t. D(S, t) is an example of a measure of extent of S. On the basis of this model, other geometric measures could also be explored to be optimized for sets of objects in motion. Let n be the cardinality of S and let M(S, t) be a geometric measure of extent or proximity at time t. Given an integer k, select a subset Q ⊂ S such that |Q| = k and Q has extreme measure M(Q, t) over all possible subsets Q of cardinality k. The present thesis focuses on minimizing and maximizing M(Q, t), in one and two dimensions (d = 1 or d = 2), for which the measure corresponds to set diameter, set width, minimum axis-aligned bounding box, and minimum enclosing disk. For each measure, exact polynomial-time algorithms are proposed for selecting an optimal subset of S and finding the time t of which the subset optimizes the measure.
- Published
- 2024
25. Big data management and mining models and their applications
- Author
-
Wang, Yang (Computer Science), Ho, Carl Ngai Man (Electrical and Computer Engineering), Ezeife, Christiana Ijeoma (University of Windsor), Leung, Carson, Olawoyin, Anifat, Wang, Yang (Computer Science), Ho, Carl Ngai Man (Electrical and Computer Engineering), Ezeife, Christiana Ijeoma (University of Windsor), Leung, Carson, and Olawoyin, Anifat
- Abstract
The world is dynamic, so are big data. The evolving challenges of managing big data volume, variety, veracity, validity, and velocity has resulted in several studies focusing on solving one or more of these perplexing issues. In this Ph.D. research, I focus on the evolving issues arising from big data variety, veracity, privacy, and accessibility. First, I design a conceptual model for capturing and storing variety of big data types including structured, semi-structured and unstructured data types and in addition, design a metadata collection framework for managing the big data in support of machine learning and open data FAIR principle of Findable, Accessibility, Interoperability and Re-usability such that the information about the data are available beyond the life cycle of the data. Second, I design hierarchical spatial-temporal model (HSTM) for managing individual record in big data in the aforementioned open data lake architecture with metadata collection framework. Third, I extend the HSTM and design the resulting hierarchical spatial-temporal privacy preserving model (HSTPPM) for preserving privacy of individual record in big data. Fourth, I extend and design applications of the HSTPPM to big data co-occurrence pattern mining and big data visualization.
- Published
- 2024
26. Machine learning and data science application for financial price prediction and portfolio optimization
- Author
-
Henry, Christopher (Computer Science), Thavaneswaran, Aerambamoorthy (Statistics), Thulasiram, Ruppa K., Dip Das, Joy, Henry, Christopher (Computer Science), Thavaneswaran, Aerambamoorthy (Statistics), Thulasiram, Ruppa K., and Dip Das, Joy
- Abstract
This thesis explores interconnected advanced machine learning (ML) and data science (DS) methodologies for improved predictive accuracy in financial markets and resilient portfolio optimization. Studying the literature on ML/DS methodologies extensively led us to observe a significant lack of application of these advances, such as autoencoder (AE), recurrent neural networks (RNN), etc. in the finance industry. The novelty of this thesis is to study price prediction and portfolio optimization with RNN and AE algorithms. Furthermore, unsupervised ML strategies were studied to introduce robustness in portfolio optimization. For this purpose, two innovative encoder-decoder-based RNN architectures autoencoder-based gated recurrent unit (AE-GRU) and autoencoder-based long short-term memory (AE-LSTM) were proposed, which were shown to be effective in predictive efficacy across diverse asset types and market conditions, showcasing enhanced predictive accuracy for financial assets. Various DS concepts, such as data visualization, Bollinger bands, data-driven volatility estimates, unsupervised ML, etc. were integrated while implementing and experimenting with new architectures for price prediction and portfolio optimization. The proposed models in this thesis showed effectiveness in price prediction and portfolio optimization under varying market conditions. The study also highlights the benefits of diversified portfolios by proposing a novel DL-based model for portfolio construction, especially when coupled with affinity propagation (AP) clustering and appropriate data-driven risk measures based on volatility estimates - with sign correlation (VES) and volatility correlation (VEV). Traditional models optimize portfolio weights using objective functions, while recent innovations emphasize data-driven risk measures for minimum risk weights from random samples. Despite challenges with short-term data featuring negative mean returns, the proposed ML-based diversification approac
- Published
- 2024
27. Improving calibration and multi-frequency inversion in microwave imaging with machine learning
- Author
-
Shafai, Cyrus (Electrical and Computer Engineering), Henry, Christopher (Computer Science), Jeffrey, Ian, Gilmore, Colin, Martin, Ben, Shafai, Cyrus (Electrical and Computer Engineering), Henry, Christopher (Computer Science), Jeffrey, Ian, Gilmore, Colin, and Martin, Ben
- Abstract
This thesis explores two unique ways of using machine learning to improve or facili- tate microwave imaging. The first contribution provides a means of calibration under the assumption of an uncooperative imaging system, a term indicating that it is im- possible or impractical to have any control over the region of interest, which means no known targets are available for calibration. The approach uses a method of style transfer provided by Cycle Generative Adversarial Networks. Cycle Generative Ad- versarial Networks are capable of learning arbitrary transformations between two sets of data in which there exists no paired samples across the two domains. In the case of calibration the two domains are synthetic field data and raw S-parameters from a physical imaging setup. The method is shown to be nearly as good as calibration with a known target for a 2D TM experimental imaging setup. The second contribution of this thesis focuses on how multi-frequency data should best be used in a machine learning model to solve the inverse problem. This work introduces two novel architectures capable of using multi-frequency data and testing the results on experimental data. The first method uses a system of sequential U- Nets referred to as the cascaded multi-frequency approach. This method turned out to be very similar to a recurrent neural network which inspired the creation of a novel LSTM-like architecture - the second method. These two methods were compared i to single frequency inversions and a ‘naive’ multi-frequency inversion which collapse the frequency data into the channels of the single frequency inversion. The models were tested on the same 2D TM experimental imaging setup in which labeled data was acquired through an automated target positioning system. Both multi-frequency approaches show significant improvement over the single frequency counter part and the naive approach.
- Published
- 2024
28. Improving protocols and miner strategies for modern cryptocurrencies
- Author
-
Akcora, Cuneyt (Computer Science), Arora, Sandeep (Marketing), Deters, Ralph (University of Saskatchewan), Thulasiram, Ruppa, Kamali, Shahin, Quinteiro dos Santos, Saulo, Akcora, Cuneyt (Computer Science), Arora, Sandeep (Marketing), Deters, Ralph (University of Saskatchewan), Thulasiram, Ruppa, Kamali, Shahin, and Quinteiro dos Santos, Saulo
- Abstract
Bitcoin, envisioned as a decentralized currency, facilitates secure micro-payments through a distributed consensus and ensures security via its widespread network. However, prioritizing security and decentralization imposes constraints on performance. My thesis explores strategies to boost the blockchain's throughput and performance alongside improving miners' financial incentives. We analyze the effects of increasing block size on throughput and security and how strategic transaction selection by miners before mining can elevate fee collection. We introduce a novel transaction-selection strategy that produces high-quality blocks more efficiently by avoiding traditional sorting and examining the benefits of regularly updating transaction sets to maximize fees. Additionally, we assess the Lightning Network's impact on reducing blockchain load and transaction costs and its potential to lower miners' fee revenues and profits. This study balances the Lightning Network's adoption and integration with third-layer applications to mitigate transaction migration from the blockchain. This thesis contributes to blockchain scalability and efficiency, presenting new transaction selection and block formation strategies that enhance network performance and miners' incentives. It also navigates the economic implications of the Lightning Network, offering a nuanced view of its effects on the blockchain ecosystem.
- Published
- 2024
29. Complexity-based graph attention network for metamorphic malware detection
- Author
-
McLeod, Bob (Electrical and Computer Engineering), Mohammed, Noman (Computer Science), Dansereau, Richard (Carleton University), Ferens, Ken, Brezinski, Kenneth, McLeod, Bob (Electrical and Computer Engineering), Mohammed, Noman (Computer Science), Dansereau, Richard (Carleton University), Ferens, Ken, and Brezinski, Kenneth
- Abstract
This thesis work presents a new approach to malware analysis by creating a specialized sandbox environment for executing and monitoring malware on a host operating system. Over 200 malware samples, along with benignware, were tested in this environment. Application Programming Interface (API) calls were traced of how these executable interacted with the host system, which including registry changes, file system access, and thread activity. Two new methods to measure complexity, Mass Radius and Radius of Gyration Fractal Dimension (FD), were developed and added to a Deep Learning model. These complexity methods helped the model converge faster. Tests showed that the Mass Radius FD measure improved convergence and accuracy over the Radius of Gyration FD in identifying malware. The complexity models performed better than standard models across different datasets. The study also found that shorter sequences of API calls and file events were more likely to indicate malicious behavior. Using GNNExplainer, linked API sequences were linked to specific malware techniques, providing deeper insights into the model’s predictions. The complexity models identified flaws in traditional methods and successfully flagged malware that other commercial sandbox methods were not able to identify, lending credence to the sophistication and applicability of this work.
- Published
- 2024
30. Investigation of building virtual worlds with spatial design domain experts
- Author
-
Gerhard, David (Computer Science), Shields, Jason (Interior Design), Latulipe, Celine, Jung, Min Kyu, Gerhard, David (Computer Science), Shields, Jason (Interior Design), Latulipe, Celine, and Jung, Min Kyu
- Abstract
The popularity of immersive web virtual worlds has increased as a result of the pandemic. Platforms like Gather, that still sees more than a million visitors quarterly post-pandemic, combined real-time video conferencing functionality with virtual worlds: multi-user virtual environments that function as virtual offices or social spaces. This combination has not only incorporated interactivity and engagement in the videoconferencing space but has also brought the capability of virtual world building to the masses, including to people without programming backgrounds. My research investigates the tools used for virtual world building, which are feature-rich creativity support tools. Given their availability, I want to understand how easy it is for non-programmers to use virtual world building tools to build virtual worlds, or more specifically, what the limitations of these virtual world building tools are by having users with backgrounds designing physical world spaces (such as architecture and interior design students) build virtual spaces. In this research, I investigate the technical challenges and creative limitations that such users experience, and how they troubleshoot their virtual world designs. The results show that participants are able to produce interesting virtual worlds with 2D virtual world authoring tools. However, advanced features appear to be too complicated to use consistently. This led to users sticking to basic features that they are familiar with. Additionally, users did not appear to test and debug their work frequently, if not at all. Based on these findings, I explain in what way these virtual world authoring tools frustrate users and produce design suggestions to improve creativity support for them. This thesis contributes results from the user study of non-programmer use of 2D virtual world authoring tools, provides evidence of user difficulty with juggling multiple toolchains and design suggestions to improve virtual world authoring experi
- Published
- 2024
31. SLANG.D: Fast, Modular and Differentiable Shader Programming
- Author
-
Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Bangaru, Sai Praveen, Wu, Lifan, Li, Tzu-Mao, Munkberg, Jacob, Bernstein, Gilbert, Ragan-Kelley, Jonathan, Durand, Fredo, Lefohn, Aaron, He, Yong, Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Bangaru, Sai Praveen, Wu, Lifan, Li, Tzu-Mao, Munkberg, Jacob, Bernstein, Gilbert, Ragan-Kelley, Jonathan, Durand, Fredo, Lefohn, Aaron, and He, Yong
- Abstract
We introduce SLANG.D, an extension to the Slang shading language that incorporates first-class automatic differentiation support. The new shading language allows us to transform a Direct3D-based path tracer to be fully differentiable with minor modifications to existing code. SLANG.D enables a shared ecosystem between machine learning frameworks and pre-existing graphics hardware API-based rendering systems, promoting the interchange of components and ideas across these two domains. Our contributions include a differentiable type system designed to ensure type safety and semantic clarity in codebases that blend differentiable and non-differentiable code, language primitives that automatically generate both forward and reverse gradient propagation methods, and a compiler architecture that generates efficient derivative propagation shader code for graphics pipelines. Our compiler supports differentiating code that involves arbitrary control-flow, dynamic dispatch, generics and higher-order differentiation, while providing developers flexible control of checkpointing and gradient aggregation strategies for best performance. Our system allows us to differentiate an existing real-time path tracer, Falcor, with minimal changes to its shader code. We show that the compiler-generated derivative kernels perform as efficiently as handwritten ones. In several benchmarks, the SLANG.D code achieves significant speedup when compared to prior automatic differentiation systems.
- Published
- 2024
32. GaN Field Emitter Arrays with JA of 10 A/cm2 at VGE = 50 V for Power Applications
- Author
-
Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Shih, P.-C., Zheng, T., Arellano-Jimenez, M. J., Gnade, B., Akinwande, A. I., Palacios, T., Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Shih, P.-C., Zheng, T., Arellano-Jimenez, M. J., Gnade, B., Akinwande, A. I., and Palacios, T.
- Abstract
2022 International Electron Devices Meeting (IEDM), San Francisco, CA, USA, III-Nitrides are attractive as field emission devices for high frequency, high power, and harsh environment applications. A wet-based digital etching and a novel device geometry was used to demonstrate GaN vertical self-alignedgate (SAG) field emitter arrays (FEA) with uniform tips of sub- 10 nm tip radius. The best GaN FEA has a current density (JA) of 10 A/cm2 at VGE = 50 V, which is better than the state-of-the-art Si field emitter arrays at the same bias condition., Department of Defense (DoD)
- Published
- 2024
33. Gamma entrainment using audiovisual stimuli alleviates chemobrain pathology and cognitive impairment induced by chemotherapy in mice
- Author
-
Picower Institute for Learning and Memory, Massachusetts Institute of Technology. Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Broad Institute of MIT and Harvard, Kim, TaeHyun, James, Benjamin T., Kahn, Martin C., Blanco-Duque, Cristina, Abdurrob, Fatema, Islam, Md Rezaul, Lavoie, Nicolas S., Kellis, Manolis, Tsai, Li-Huei, Picower Institute for Learning and Memory, Massachusetts Institute of Technology. Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Broad Institute of MIT and Harvard, Kim, TaeHyun, James, Benjamin T., Kahn, Martin C., Blanco-Duque, Cristina, Abdurrob, Fatema, Islam, Md Rezaul, Lavoie, Nicolas S., Kellis, Manolis, and Tsai, Li-Huei
- Abstract
Patients with cancer undergoing chemotherapy frequently experience a neurological condition known as chemotherapy-related cognitive impairment, or “chemobrain,” which can persist for the remainder of their lives. Despite the growing prevalence of chemobrain, both its underlying mechanisms and treatment strategies remain poorly understood. Recent findings suggest that chemobrain shares several characteristics with neurodegenerative diseases, including chronic neuroinflammation, DNA damage, and synaptic loss. We investigated whether a noninvasive sensory stimulation treatment we term gamma entrainment using sensory stimuli (GENUS), which has been shown to alleviate aberrant immune and synaptic pathologies in mouse models of neurodegeneration, could also mitigate chemobrain phenotypes in mice administered a chemotherapeutic drug. When administered concurrently with the chemotherapeutic agent cisplatin, GENUS alleviated cisplatin-induced brain pathology, promoted oligodendrocyte survival, and improved cognitive function in a mouse model of chemobrain. These effects persisted for up to 105 days after GENUS treatment, suggesting the potential for long-lasting benefits. However, when administered to mice 90 days after chemotherapy, GENUS treatment only provided limited benefits, indicating that it was most effective when used to prevent the progression of chemobrain pathology. Furthermore, we demonstrated that the effects of GENUS in mice were not limited to cisplatin-induced chemobrain but also extended to methotrexate-induced chemobrain. Collectively, these findings suggest that GENUS may represent a versatile approach for treating chemobrain induced by different chemotherapy agents.
- Published
- 2024
34. LiCROM: Linear-Subspace Continuous Reduced Order Modeling with Neural Fields
- Author
-
Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Chang, Yue, Chen, Peter Yichen, Wang, Zhecheng, Chiaramonte, Maurizio M., Carlberg, Kevin, Grinspun, Eitan, Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Chang, Yue, Chen, Peter Yichen, Wang, Zhecheng, Chiaramonte, Maurizio M., Carlberg, Kevin, and Grinspun, Eitan
- Abstract
Linear reduced-order modeling (ROM) simplifies complex simulations by approximating the behavior of a system using a simplified kinematic representation. Typically, ROM is trained on input simulations created with a specific spatial discretization, and then serves to accelerate simulations with the same discretization. This discretization-dependence is restrictive. Becoming independent of a specific discretization would provide flexibility to mix and match mesh resolutions, connectivity, and type (tetrahedral, hexahedral) in training data; to accelerate simulations with novel discretizations unseen during training; and to accelerate adaptive simulations that temporally or parametrically change the discretization. We present a flexible, discretization-independent approach to reduced-order modeling. Like traditional ROM, we represent the configuration as a linear combination of displacement fields. Unlike traditional ROM, our displacement fields are continuous maps from every point on the reference domain to a corresponding displacement vector; these maps are represented as implicit neural fields. With linear continuous ROM (LiCROM), our training set can include multiple geometries undergoing multiple loading conditions, independent of their discretization. This opens the door to novel applications of reduced order modeling. We can now accelerate simulations that modify the geometry at runtime, for instance via cutting, hole punching, and even swapping the entire mesh. We can also accelerate simulations of geometries unseen during training. We demonstrate one-shot generalization, training on a single geometry and subsequently simulating various unseen geometries.
- Published
- 2024
35. Tailors: Accelerating Sparse Tensor Algebra by Overbooking Buffer Capacity
- Author
-
Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Xue, Zi Yu, Wu, Yannan, Emer, Joel, Sze, Vivienne, Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Xue, Zi Yu, Wu, Yannan, Emer, Joel, and Sze, Vivienne
- Abstract
Sparse tensor algebra is a challenging class of workloads to accelerate due to low arithmetic intensity and varying sparsity patterns. Prior sparse tensor algebra accelerators have explored tiling sparse data to increase exploitable data reuse and improve throughput, but typically allocate tile size in a given buffer for the worst-case data occupancy. This severely limits the utilization of available memory resources and reduces data reuse. Other accelerators employ complex tiling during preprocessing or at runtime to determine the exact tile size based on its occupancy. This paper proposes a speculative tensor tiling approach, called overbooking, to improve buffer utilization by taking advantage of the distribution of nonzero elements in sparse tensors to construct larger tiles with greater data reuse. To ensure correctness, we propose a low-overhead hardware mechanism, Tailors, that can tolerate data overflow by design while ensuring reasonable data reuse. We demonstrate that Tailors can be easily integrated into the memory hierarchy of an existing sparse tensor algebra accelerator. To ensure high buffer utilization with minimal tiling overhead, we introduce a statistical approach, Swiftiles, to pick a tile size so that tiles usually fit within the buffer’s capacity, but can potentially overflow, i.e., it overbooks the buffers. Across a suite of 22 sparse tensor algebra workloads, we show that our proposed overbooking strategy introduces an average speedup of 52.7 × and 2.3 × and an average energy reduction of 22.5 × and 2.5 × over ExTensor without and with optimized tiling, respectively.
- Published
- 2024
36. Accelerating RTL Simulation with Hardware-Software Co-Design
- Author
-
Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Elsabbagh, Fares, Sheikhha, Shabnam, Ying, Victor, Nguyen, Quan, Emer, Joel, Sanchez, Daniel, Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Elsabbagh, Fares, Sheikhha, Shabnam, Ying, Victor, Nguyen, Quan, Emer, Joel, and Sanchez, Daniel
- Abstract
Fast simulation of digital circuits is crucial to build modern chips. But RTL (Register-Transfer-Level) simulators are slow, as they cannot exploit multicores well. Slow simulation lengthens chip design time and makes bugs more frequent. We present ASH, a parallel architecture tailored to simulation workloads. ASH consists of a tightly codesigned hardware architecture and compiler for RTL simulation. ASH exploits two key opportunities. First, it performs dataflow execution of small tasks to leverage the fine-grained parallelism in simulation workloads. Second, it performs selective event-driven execution to run only the fraction of the design exercised each cycle, skipping ineffectual tasks. ASH hardware provides a novel combination of dataflow and speculative execution, and ASH’s compiler features several novel techniques to automatically leverage this hardware. We evaluate ASH in simulation using large Verilog designs. An ASH chip with 256 simple cores is gmean 1,485 × faster than 1-core Verilator, and it is 32 × faster than parallel Verilator on a server CPU with 32 complex cores, while using 3 × less area.
- Published
- 2024
37. Integration of an Aptamer-Based Signal-On Probe and a Paper-Based Origami Preconcentrator for Small Molecule Biomarkers Detection
- Author
-
Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Lee, Na E., Hong, Ji H., Lee, Seungmin, Yoo, Yong K., Kim, Kang H., Park, Jeong S., Kim, Cheonjung, Yoon, Junghyo, Yoon, Dae S., Lee, Jeong H., Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Lee, Na E., Hong, Ji H., Lee, Seungmin, Yoo, Yong K., Kim, Kang H., Park, Jeong S., Kim, Cheonjung, Yoon, Junghyo, Yoon, Dae S., and Lee, Jeong H.
- Abstract
Point-of-care testing using paper-based lateral flow assays (LFAs) has emerged as an attractive diagnostic platform. However, detecting small molecules such as cortisol using LFAs is challenging due to limited binding sites and weak signal generation. Here, we report the development of cortisol-specific aptamer-based probes and a paper-based origami preconcentrator (POP) to amplify the probe signal. The cortisol-specific aptamers were conjugated onto gold nanoparticles and hybridized with signal probes to create the cortisol-specific signal-on probe. POP, consisting of patterned layers with convergent wicking zones, induces electrokinetic preconcentration of the released signaling probes. By integrating cortisol-selective aptamer-based probes and POP, we accurately diagnosed cortisol levels within 30 min of signal probe incubation, followed by 10 min of preconcentration. Our sensor was able to detect cortisol levels in the range of 25–1000 ng/mL, with typical cortisol levels in plasma ranging from 40 to 250 ng/mL falling within this range. The successful detection of the wide range of cortisol samples using this approach highlights the potential of this platform as a point-of-care testing tool, particularly for lateral flow assay-based detection of small molecules like cortisol. Our approach offers a convenient and reliable method of cortisol level testing with a portable and accessible diagnosis device.
- Published
- 2024
38. PockEngine: Sparse and Efficient Fine-tuning in a Pocket
- Author
-
Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, MIT-IBM Watson AI Lab, Zhu, Ligeng, Hu, Lanxiang, Lin, Ji, Chen, Wei-Ming, Wang, Wei-Chen, Gan, Chuang, Han, Song, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, MIT-IBM Watson AI Lab, Zhu, Ligeng, Hu, Lanxiang, Lin, Ji, Chen, Wei-Ming, Wang, Wei-Chen, Gan, Chuang, and Han, Song
- Abstract
On-device learning and efficient fine-tuning enable continuous and privacy-preserving customization (e.g., locally fine-tuning large language models on personalized data). However, existing training frameworks are designed for cloud servers with powerful accelerators (e.g., GPUs, TPUs) and lack the optimizations for learning on the edge, which faces challenges of resource limitations and edge hardware diversity. We introduce PockEngine: a tiny, sparse and efficient engine to enable fine-tuning on various edge devices. PockEngine supports sparse backpropagation: it prunes the backward graph and sparsely updates the model with measured memory saving and latency reduction while maintaining the model quality. Secondly, PockEngine is compilation first: the entire training graph (including forward, backward and optimization steps) is derived at compile-time, which reduces the runtime overhead and brings opportunities for graph transformations. PockEngine also integrates a rich set of training graph optimizations, thus can further accelerate the training cost, including operator reordering and backend switching. PockEngine supports diverse applications, frontends and hardware backends: it flexibly compiles and tunes models defined in PyTorch/TensorFlow/Jax and deploys binaries to mobile CPU/GPU/DSPs. We evaluated PockEngine on both vision models and large language models. PockEngine achieves up to 15 × speedup over off-the-shelf TensorFlow (Raspberry Pi), 5.6 × memory saving back-propagation (Jetson AGX Orin). Remarkably, PockEngine enables fine-tuning LLaMav2-7B on NVIDIA Jetson AGX Orin at 550 tokens/s, 7.9 × faster than the PyTorch.
- Published
- 2024
39. TeAAL: A Declarative Framework for Modeling Sparse Tensor Accelerators
- Author
-
Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Nayak, Nandeeka, Odemuyiwa, Toluwanimi O., Ugare, Shubham, Fletcher, Christopher, Pellauer, Michael, Emer, Joel, Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Nayak, Nandeeka, Odemuyiwa, Toluwanimi O., Ugare, Shubham, Fletcher, Christopher, Pellauer, Michael, and Emer, Joel
- Abstract
Over the past few years, the explosion in sparse tensor algebra workloads has led to a corresponding rise in domain-specific accelerators to service them. Due to the irregularity present in sparse tensors, these accelerators employ a wide variety of novel solutions to achieve good performance. At the same time, prior work on design-flexible sparse accelerator modeling does not express this full range of design features, making it difficult to understand the impact of each design choice and compare or extend the state-of-the-art.To address this, we propose TeAAL: a language and simulator generator for the concise and precise specification and evaluation of sparse tensor algebra accelerators. We use TeAAL to represent and evaluate four disparate state-of-the-art accelerators—ExTensor, Gamma, OuterSPACE, and SIGMA—and verify that it reproduces their performance with high accuracy. Finally, we demonstrate the potential of TeAAL as a tool for designing new accelerators by showing how it can be used to speed up vertex-centric programming accelerators—achieving 1.9 × on BFS and 1.2 × on SSSP over GraphDynS.
- Published
- 2024
40. Neural Stress Fields for Reduced-order Elastoplasticity and Fracture
- Author
-
Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Zong, Zeshun, Li, Xuan, Li, Minchen, Chiaramonte, Maurizio M., Matusik, Wojciech, Grinspun, Eitan, Carlberg, Kevin, Jiang, Chenfanfu, Chen, Peter Yichen, Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Zong, Zeshun, Li, Xuan, Li, Minchen, Chiaramonte, Maurizio M., Matusik, Wojciech, Grinspun, Eitan, Carlberg, Kevin, Jiang, Chenfanfu, and Chen, Peter Yichen
- Abstract
We propose a hybrid neural network and physics framework for reduced-order modeling of elastoplasticity and fracture. State-of-the-art scientific computing models like the Material Point Method (MPM) faithfully simulate large-deformation elastoplasticity and fracture mechanics. However, their long runtime and large memory consumption render them unsuitable for applications constrained by computation time and memory usage, e.g., virtual reality. To overcome these barriers, we propose a reduced-order framework. Our key innovation is training a low-dimensional manifold for the Kirchhoff stress field via an implicit neural representation. This low-dimensional neural stress field (NSF) enables efficient evaluations of stress values and, correspondingly, internal forces at arbitrary spatial locations. In addition, we also train neural deformation and affine fields to build low-dimensional manifolds for the deformation and affine momentum fields. These neural stress, deformation, and affine fields share the same low-dimensional latent space, which uniquely embeds the high-dimensional simulation state. After training, we run new simulations by evolving in this single latent space, which drastically reduces the computation time and memory consumption. Our general continuum-mechanics-based reduced-order framework is applicable to any phenomena governed by the elastodynamics equation. To showcase the versatility of our framework, we simulate a wide range of material behaviors, including elastica, sand, metal, non-Newtonian fluids, fracture, contact, and collision. We demonstrate dimension reduction by up to 100,000 × and time savings by up to 10 ×.
- Published
- 2024
41. Growth of Large-Area Single- and Bi-Layer Graphene by Controlled Carbon Precipitation on Polycrystalline Ni Surfaces
- Author
-
Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Massachusetts Institute of Technology. Department of Materials Science and Engineering, Massachusetts Institute of Technology. Department of Physics, Massachusetts Institute of Technology. Research Laboratory of Electronics, Reina, Alfonso, Jia, Xiaoting, Bhaviripudi, Sreekar, Dresselhaus, Mildred, Kong, Jing, Thiele, Stefan, Schaefer, Juergen A., Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Massachusetts Institute of Technology. Department of Materials Science and Engineering, Massachusetts Institute of Technology. Department of Physics, Massachusetts Institute of Technology. Research Laboratory of Electronics, Reina, Alfonso, Jia, Xiaoting, Bhaviripudi, Sreekar, Dresselhaus, Mildred, Kong, Jing, Thiele, Stefan, and Schaefer, Juergen A.
- Abstract
We report graphene films composed mostly of one or two layers of graphene grown by controlled carbon precipitation on the surface of polycrystalline Ni thin films during atmospheric chemical vapor deposition (CVD). Controlling both the methane concentration during CVD and the substrate cooling rate during graphene growth can significantly improve the thickness uniformity. As a result, one- or two- layer graphene regions occupy up to 87% of the film area. Single layer coverage accounts for 5%–11% of the overall film. These regions expand across multiple grain boundaries of the underlying polycrystalline Ni film. The number density of sites with multilayer graphene/graphite (>2 layers) is reduced as the cooling rate decreases. These films can also be transferred to other substrates and their sizes are only limited by the sizes of the Ni film and the CVD chamber. Here, we demonstrate the formation of films as large as 1 in2. These findings represent an important step towards the fabrication of large-scale high-quality graphene samples., National Science Foundation (U.S.) (CTS 05-06830), National Science Foundation (U.S.) (DMR07-04197)
- Published
- 2024
42. A Secure Digital In-Memory Compute (IMC) Macro with Protections for Side-Channel and Bus Probing Attacks
- Author
-
Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Ashok, Maitreyi, Maji, Saurav, Zhang, Xin, Cohn, John, Chandrakasan, Anantha P., Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Ashok, Maitreyi, Maji, Saurav, Zhang, Xin, Cohn, John, and Chandrakasan, Anantha P.
- Abstract
2024 IEEE Custom Integrated Circuits Conference April 21st – 24th, 2024 Denver, CO U.S., Machine learning (ML) accelerators provide energy efficient neural network (NN) implementations for applications such as speech recognition and image processing. Recently, digital IMC has been proposed to reduce data transfer energy, while still allowing for higher bitwidths and accuracies necessary for many workloads, especially with technology scaling [1,2]. Privacy of ML workloads can be exploited with physical side-channel attacks (SCAs) or bus probing attacks (BPAs) [3] (Fig. 1). While SCAs correlate IC power consumption or EM emissions to data or operations, BPAs directly tap traces between the IC and off-chip memory. The inputs reflect private data collected on IoT devices, such as images of faces. The weights, typically stored off-chip, reveal information about proprietary private training datasets. This work presents the first IMC macro protected against SCAs and BPAs to mitigate these risks., National Science Foundation (NSF), MIT-IBM Watson AI Lab, MathWorks Engineering Fellowship
- Published
- 2024
43. Quantum Techniques and Technologies for Cybersecurity in Healthcare
- Author
-
Florence D. Hudson, Founder and CEO, FDHint, LLC, Executive Director, Northeast Big Data Innovation Hub at Columbia University, IEEE Engineering in Medicine and Biology Society Standards Committee, Former IBM VP & CTO, and Special Advisor – NSF Cybersecur and Shantanu Chakrabortty, Founder, Free Dynamics, Clifford Murphy Professor, Preston M. Green Department of Electrical and Systems Engineering, Department of Computer Science and Engineering, Department of Biomedical Engineering, Washington University, St. L
- Subjects
blockchain trusted platform modules ,low resource platforms ,ip protection ,exploiting quantum primitives for healthcare ,blockchain in healthcare today ,low computational footprint for healthcare ,end to end security solutions for healthcare ,Computer applications to medicine. Medical informatics ,R858-859.7 - Abstract
Quantum based security solutions for low resource platforms. Learn and prepare for breaches with a pre-emptive stance and not just when there is an imminent threat.
- Published
- 2022
- Full Text
- View/download PDF
44. Shaping the Future of Healthcare Through Blockchain-Powered Technology
- Author
-
Cathy Mulligan, PhD, Professor of Computer Science, Director of DCentral Lab, Instituto Superior Técnico, UK, Nadia Hewett, Blockchain and Emerging Technology Expert, University of California and Project Lead, Data for Global Common Good, WEF, Susan Somerville, CEO, Chronicled, and Mohan Venkataraman, CTO, Chainyard
- Subjects
blockchain based healthcare delivery systems ,new blockchain-powered technology health solutions ,how does blockchain transform health systems ,blockchain in healthcare ,Computer applications to medicine. Medical informatics ,R858-859.7 - Abstract
Learn the latest developments, solutions and forward-looking approaches with blockchain technology for healthcare. How do panelists view healthcare delivery systems changing through blockchain-powered technology? What solutions are they spearheading along with new models of collaborations with which they are involved? The panel will highlight how blockchain is leveraged to transform health systems.
- Published
- 2022
- Full Text
- View/download PDF
45. Blockchain Applications Presentation Commentry
- Author
-
Dr. Anjum Khurshid, Director Of Data Integration, Department of Population Health, Assistant Professor, Department of Population Health, The University Of Texas at Austin Dell Medical School, USA, Dr. Kayo Fujimoto, Distinguished Professor In Social Determinants of Health, Professor In School Of Public Health At The University Of Texas Health Science Center At Houston, USA, and Dr. Vijayakumar Varadarajan, Adjunct Professor, School of Computer Science and Engineering, The University of New South Wales, Australia
- Subjects
blockchain in healthcare ,blockchain research ,conv2x 2021 ,blockchain academic track ,blockchain applications in healthcare ,blockchain academic scientific review committee ,conv2x blockchain scientific review committee ,Computer applications to medicine. Medical informatics ,R858-859.7 - Abstract
The 2021 ConV2X Annual Symposium featured a scientific program of academic/research presentations in addition to business and industry talks. The research track focused on exploring and sharing developments in blockchain and emerging technologies in health and clinical medicine. Submissions were based on original research, conceptual frameworks, proposed applications, position papers, case studies, and real-world implementation. Selection was based on a peer-review process. Faculty, students, and industry researchers were encouraged to submit abstracts to present ideas before an informed and knowledgeable audience of industry leaders, policy makers, funders, and researchers. All presentations were reviewed by a sub-group of the scientific reveiw committee. This video presentation is an example of the discussions that transpired for each category of submissions, specifically, blockchain applications. Submission Review Committee • Dave Kochalko, CEO of ARTiFACTS • Anjum Khurshid, UT Austin • Carlos Caldas, UT Engineering • Gil Alterovitz, Harvard Medical School • Kayo Fujimoto, UT Health Houston • Lei Zhang, University of Glasglow • Sean Manion, CSciO of ConsenSys Health • Vijayakuman Varadarajan, University of South Wales • Vikram Dhillon, Wayne State University • Yuichi Ikeda, Kyoto University
- Published
- 2022
- Full Text
- View/download PDF
46. Meta learning for point cloud analysis
- Author
-
Livi, Lorenzo (Computer Science), Mohammed, Noman (Computer Science), Wang, Yang, Hatem, Ahmed, Livi, Lorenzo (Computer Science), Mohammed, Noman (Computer Science), Wang, Yang, and Hatem, Ahmed
- Abstract
Point cloud has been highly attracting the attention of the research community, due to their numerous applications in 3D computer vision. While learning-based approaches for point cloud problems have achieved impressive progress, generalization to unknown testing environments remains a major challenge due to the large discrepancies of data captured by different 3D sensors. Existing methods typically train a generic model and the same trained model is applied on each test instance. This could be sub-optimal since it is difficult for the same model to handle all the variations during testing. In this thesis, we propose novel frameworks for point cloud problems that adapt the model in an instance-specific manner during inference. Our model is trained using a meta-learning scheme to provide the model with the ability of fast and effective adaptation at test time. First, we consider the problem of point cloud registration. The objective is to estimate the 3D transformation that aligns a pair of partially overlapped point clouds. Next, we investigate the point cloud upsampling problem. In this setting, the goal is to generate high-resolution point clouds from sparse point clouds. Experimental results demonstrate the effectiveness of our proposed frameworks in improving the performance of state-of-the-art models and achieving superior results.
- Published
- 2023
47. CLIP for point cloud understanding
- Author
-
Livi, Lorenzo (Computer Science), Kai-Sang Leung, Carson (Computer Science), Wang, Yang, Ghose, Shuvozit, Livi, Lorenzo (Computer Science), Kai-Sang Leung, Carson (Computer Science), Wang, Yang, and Ghose, Shuvozit
- Abstract
Contrastive Vision-Language Pre-training (CLIP) based point cloud classification model has added a new direction in the point cloud classification research domain. In this thesis, we propose two novel methods for CLIP-based point cloud classification. First, we propose a Pretrained Point Cloud to Image Translation Network (PPCITNet) that produces generalized colored images along with additional salient visual cues to the point cloud depth maps for CLIP based point cloud classification. In addition, we propose a novel viewpoint adapter that combines the view feature processed by each viewpoint as well as the global intertwined knowledge that exists across the multi-view features. Next, we propose a novel meta-episodic learning framework for CLIP-based point cloud classification. In addition, we introduce dynamic task sampling within the episode based on performance memory. The experimental results demonstrate the superior performance of the proposed model over existing state-of-the-art CLIP-based models on ModelNet10, ModelNet40, and ScanobjectNN datasets.
- Published
- 2023
48. A research study towards the improvement of human security and peace in cyberspace
- Author
-
Mohammed, Noman (Computer Science), Adedayo, Mary O. (Applied Computer Science, The University of Winnipeg), Flaherty, Maureen, Dogbey, Joshua, Mohammed, Noman (Computer Science), Adedayo, Mary O. (Applied Computer Science, The University of Winnipeg), Flaherty, Maureen, and Dogbey, Joshua
- Abstract
The ever-increasing dependence on the internet for the performance of human functions blurs some distinctions between physical and virtual worlds. Tasks like the performance of surgical procedures in hospitals depend on digital tools for efficient and effective health delivery. Still, society is witnessing both new forms of violence and the transposition of violent forms from the physical world to the cyber world. Cyberspace provides an easier option for harm to be caused to individuals because both state and non-state actors can extend their actions beyond their physical reach; commercial spyware is being deployed against political opponents in many countries. Without obtaining express consent, some organizations may be involved in trading the personal data of clients for business gains. During all these, the search for cyber peace has become difficult because of divergent views on what constitutes cyberviolence and the potency of cyber weapons. This research seeks to integrate discussions among scholars and the perspectives of some cybersecurity practitioners on cyber peace and violence. The decision to interrogate cyber peace and violence from the perspectives of cyber-security practitioners will contribute towards building some stability for these evolving concepts within the peace and conflict doctrines. Four cybersecurity professionals were interviewed on the subject. The basic human needs theory, human rights, and social justice theories are used to interrogate the understanding of cyber peace and violence. The results indicate that cyber harms targeting both state and non-state actors and installations should be considered in conflict analysis. This approach helps to enhance the concept of positive and negative cyber peace as possibilities.
- Published
- 2023
49. SnuggleBot: a cuddly companion robot for lonely people to use at home
- Author
-
Latulipe, Celine (Computer Science), Tremblay-Savard, Olivier (Computer Science), Young, James E., Passler Bates, Danika, Latulipe, Celine (Computer Science), Tremblay-Savard, Olivier (Computer Science), Young, James E., and Passler Bates, Danika
- Abstract
We designed, prototyped, and deployed a novel companion robot for private in-home use that aims to promote engagement to support people living with loneliness. We identified and explored three key approaches to encourage engagement and provide comfort and prototyped a novel robot aiming to embody these principles: this resulted in SnuggleBot, a novel robot that is physically comforting, socially engaging, and requires care, to provide structure and increase engagement. We deployed our prototype unsupervised for a minimum of 7 weeks (with optional longer involvement up to 6 months) into the homes of seven people who live alone and self-identified as lonely. We reflect on our specific designs and how they promoted engagement and companionship. Our results indicate that robot designs incorporating physical comfort, social engagement, and requiring care have the potential to promote companionship, with many participants showing signs of bonding with the robot. Further, our design strategies were generally successful in that they promoted the behaviours and reactions that we intended. We found that most participants expressed that the robot is comfortable. The robot also promoted animism and engagement. Additionally, participants reported wellbeing benefits because of caring for the robot. This supports future research into robots developed with our design strategies, which can leverage our results to improve on our implementations of the design strategies.
- Published
- 2023
50. Studying the impact of early test termination due to assertion failure on code coverage and spectrum-based fault localization
- Author
-
Thulasiraman, Parimala (Computer Science), Rouhani, Sara (Computer Science), Wang, Shaowei, Uddin, Md. Ashraf, Thulasiraman, Parimala (Computer Science), Rouhani, Sara (Computer Science), Wang, Shaowei, and Uddin, Md. Ashraf
- Abstract
An assertion is commonly used to validate the expected program’s behavior (e.g., if the returned value of a method equals an expected value) in software testing. Although it is a recommended practice to use only one assertion in a single test to avoid code smell (e.g., Assertion Roulette), we observe that it is common to have multiple assertions in a single test. One issue with tests that have multiple assertions is that when the test fails at an early assertion, the test will terminate at that point, and the remaining testing code will not be executed. This, in turn, can potentially reduce the code coverage and the performance of techniques that rely on code coverage information (e.g., spectrum-based fault localization). We refer to such a scenario as early test termination. Understanding the impact of early test termination on test coverage is important for software testing and debugging, particularly for the techniques that rely on coverage information obtained from the testing. In this study, we investigated 207 versions of 6 open-source projects. We found that a non-negligible portion of the failed tests (19.1%) is early terminated due to assertion failure, which leads to the skipping of 15.3% to 60.5% of the test code on average, and a negative impact on testing coverage. To mitigate early test termination, we propose two approaches, i.e., Trycatch (adding a try-catch block surrounding an assertion) and Slicing (slicing a test into a set of independent sub-tests, in which only one assertion and its dependent code are contained). After applying our approaches, the line/branch coverage get improved in 55% of the studied versions. Moreover, Slicing improves the performance of SBFL by 15.1% and 10.66% in terms of Mean First Rank (MFR) for Ochiai and Tarantula, respectively. We also provide actionable suggestions to prevent early test termination, and approaches to mitigate early test termination if it already exists in their project.
- Published
- 2023
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.