4,191 results on '"ELECTRONIC file management"'
Search Results
2. Tarazu: An Adaptive End-to-end I/O Load-balancing Framework for Large-scale Parallel File Systems.
- Author
-
Paul, Arnab K., Neuwirth, Sarah, Wadhwa, Bharti, Wang, Feiyi, Oral, Sarp, and Butt, Ali R.
- Subjects
MATHEMATICAL optimization ,MODEL airplanes ,OPENFLOW (Computer network protocol) ,SCALABILITY ,ELECTRONIC file management ,SUPERCOMPUTERS - Abstract
The imbalanced I/O load on large parallel file systems affects the parallel I/O performance of high-performance computing (HPC) applications. One of the main reasons for I/O imbalances is the lack of a global view of system-wide resource consumption. While approaches to address the problem already exist, the diversity of HPC workloads combined with different file striping patterns prevents widespread adoption of these approaches. In addition, load-balancing techniques should be transparent to client applications. To address these issues, we propose Tarazu, an end-to-end control plane where clients transparently and adaptively write to a set of selected I/O servers to achieve balanced data placement. Our control plane leverages real-time load statistics for global data placement on distributed storage servers, while our design model employs trace-based optimization techniques to minimize latency for I/O load requests between clients and servers and to handle multiple striping patterns in files. We evaluate our proposed system on an experimental cluster for two common use cases: the synthetic I/O benchmark IOR and the scientific application I/O kernel HACC-I/O. We also use a discrete-time simulator with real HPC application traces from emerging workloads running on the Summit supercomputer to validate the effectiveness and scalability of Tarazu in large-scale storage environments. The results show improvements in load balancing and read performance of up to 33% and 43%, respectively, compared to the state-of-the-art. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
3. An effective secondary personalization file system driven by FileForge module.
- Author
-
Gaojian Liu, Yufei Hu, and Ngai Cheong
- Subjects
ACADEMIC departments ,STUDENT affairs services ,ADMINISTRATIVE procedure ,DATABASES ,DIGITAL technology ,TRAUMA registries ,ELECTRONIC file management - Abstract
Digital service platforms provided by academic support departments in Macao assist academic staff and students in various areas such as registry, student affairs, academic activities, and research. As the number of undergraduate students increases and new departments are established, academic staff often face the challenge of dealing with paperwork that contains similar content but different formats. This situation results in redundancies and a waste of time. This paper presents our endeavors to simplify administrative procedures in higher education by automating restructured documentation and developing secondary file systems. The paper presents two case studies: Scenario One focuses on streamlining the publication system for academic staff who submit papers in different formats. At the same time, Scenario Two aims to simplify the daily paperwork process for academic staff. Both cases involve transforming the distribution of administrative documents, transitioning from a standardized form with guidelines to a customized form with concise tips. This approach allows academic staff to handle only the necessary information, which may not be available in the database or requires verification. The case studies serve to demonstrate the effectiveness of this administrative simplification. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
4. A Lightweight File System Design for Unikernel.
- Author
-
Cho, Kyungwoon and Bahn, Hyokyung
- Subjects
ELECTRONIC file management ,SYSTEMS design - Abstract
Unikernels are specialized operating system (OS) kernels optimized for a single application or service, offering advantages such as rapid boot times, high performance, minimal memory usage, and enhanced security compared to general-purpose OS kernels. Unikernel applications must remain compatible with the runtime environment of general-purpose kernels, either through binary or source compatibility. As a result, many Unikernel projects have prioritized system call compatibility over performance enhancements. In this paper, we explore the design principles of Unikernel file systems and introduce a new file system tailored for Unikernels named ULFS (Ultra Lightweight File System). ULFS provides system call services akin to those of general-purpose OS kernels but achieves superior performance and security with significantly fewer system resources. Specifically, ULFS is developed as a lightweight file system embracing Unikernel design principles. It streamlines system calls, removes unnecessary locks, and omits permission checks for multiple users, utilizing a non-hypervisor architecture. This approach significantly reduces the memory footprint of the file system and enhances performance. Through measurement studies, we assess the performance and memory requirements of various file systems from major Unikernel projects. Our findings demonstrate that ULFS surpasses several existing Unikernel file systems, including Rumpvfs, Ramfs-u, Ramfs-q, 9pfs, and Hcfs. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
5. Research for Practice: Crash Consistency: Keeping data safe in the presence of crashes is a fundamental problem.
- Author
-
ALAGAPPAN, RAMNATTHAN
- Subjects
- *
COMPUTER system failures , *COMPUTER system failure prevention , *COMPUTER storage capacity , *APPLICATION software , *ELECTRONIC file management , *COMPUTERS - Abstract
This article discusses system crashes and how each level of the system needs to be implemented correctly and the system component interfaces need to be used correctly by applications in order to prevent crashes. Several papers are cited within this article exploring the file system, a lower level component within the system, an exploration of interface-level guarantees with bug-finders, crash-consistent programs, and how the newer concept of persistent memory interacts with system crashes.
- Published
- 2023
- Full Text
- View/download PDF
6. "Digesting the abundance of idol matter" key factors in personal information management experiences of selected social science faculty.
- Author
-
Nwagwu, Williams
- Subjects
PERSONAL information management ,INFORMATION literacy ,COMPUTER literacy ,SOCIAL scientists ,INFORMATION literacy education ,ELECTRONIC file management - Abstract
Purpose: The purpose of this paper is to examine the personal information management (PIM) behaviours of social science faculty in Africa. The study examined the experiences and encounters of selected social scientists in Africa in organising and finding and re-finding of the information they previously created or stored. More specifically, the study sought to examine how faculty keep and refind information, the files and folders in which they store the information. Also, the study examined the nature and characteristics of faculty information spaces with particular respect to electronic documents including emails and paper documents. Design/methodology/approach: Sample survey research design and a mixed methods approach consisting of qualitative and quantitative were used. Data was collected using a discursive technique, an interview schedule and a questionnaire. Data analysis was conducted using factorial analysis of mixed data design, guided by a combination of category and codes identification using NVivo and Statistical Package for the Social Sciences (SPSS) 17. Principal component analysis (PCA) of factor analysis was executed to identify key components. Findings: Eleven issues, namely, time, infrastructure, importance of the information, folder/file management, document characteristics and organisational context played significant roles in the PIM behaviours of the respondents. Others were importance of the information, document overload, memory, workload and computer literacy. PCA extracted four major components, namely, document overload, time, computer literacy and importance of the information. Research limitations/implications: An expansion in the number of faculty involved in this study would probably yield a more reliable outcome. Extending the study to cover Africa would also yield a more applicable result. Practical implications: The key PIM issues identified in this study, namely, document overload, time, computer literacy and importance of the information should constitute the focus of continuous information literacy education aimed at improving PIM social scientists' faculty in Africa. Social implications: Improved PIM of social science faculty will result to improved research productivity and good health. Originality/value: PIM of social scientists has not been examined in the literature, and yet it is crucial for further understanding their learning and information behaviours, and improving their productivity. The design and administration of a questionnaire constructed based on codes extracted from qualitative and discursive sessions to the same respondents from whom the qualitative data was collected makes the findings very strong. A further deployment of factorial analysis of mixed data design to handle qualitative data makes the contribution of the study very significant. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
7. New Spam Filtering Method with Hadoop Tuning-Based MapReduce Naïve Bayes.
- Author
-
Keungyeup Ji and Youngmi Kwon
- Subjects
ELECTRONIC file management ,EMAIL ,MACHINE learning ,COMPUTER programming ,ACCURACY - Abstract
As the importance of email increases, the amount of malicious email is also increasing, so the need for malicious email filtering is growing. Since it is more economical to combine commodity hardware consisting of a medium server or PC with a virtual environment to use as a single server resource and filter malicious email using machine learning techniques, we used a Hadoop MapReduce framework and Naïve Bayes among machine learning methods for malicious email filtering. Naïve Bayes was selected because it is one of the top machine learning methods(Support Vector Machine (SVM), Naïve Bayes, K-Nearest Neighbor(KNN), and Decision Tree) in terms of execution time and accuracy. Malicious email was filtered with MapReduce programming using the Naïve Bayes technique, which is a supervised machine learning method, in a Hadoop framework with optimized performance and also with the Python program technique with the Naïve Bayes technique applied in a bare metal server environment with the Hadoop environment not applied. According to the results of a comparison of the accuracy and predictive error rates of the two methods, the Hadoop MapReduce Naïve Bayes method improved the accuracy of spam and ham email identification 1.11 times and the prediction error rate 14.13 times compared to the non-Hadoop Python Naïve Bayes method. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
8. A Novel Distributed File System Using Blockchain Metadata.
- Author
-
Kumar, Deepa S., Dija, S., Sumithra, M. D., Rahman, M. Abdul, and Nair, Praseeda B.
- Subjects
COMPUTER workstation clusters ,BLOCKCHAINS ,CLOUD computing ,ELECTRONIC data processing ,ONLINE data processing ,ELECTRONIC file management ,METADATA ,SOCIAL media - Abstract
Cluster computing has become an inevitable part of data processing as the huge volume of data being produced from different sources like online social media, IoT, mobiledata, sensor data, black box data and so on increases in an exponentially fast manner. Distributed File System defines different methods to distribute, read and eliminate the files among different cluster computing nodes. It is found that popular distributed file systems such as Google File System and Hadoop Distributed File System store metadata centrally. This creates a chance for a Single Point of Failure that arises the need for backup and alternative solutions to recover the metadata on the failure of the metadata server. Also, the name node server is built using expensive and reliable hardware. For small and medium clusters, it is not cost effective to maintain expensive name node server. Even though cheap commodity hardware may substitute the name node functionality, they are prone to hardware failure. This paper proposes a novel distributed file system to distribute files over a cluster of machines connected in a Peer-to-Peer network. The most significant feature of the file system is its capability to distribute the metadata using distributed consensus, using hash values. Although the distributed metadata is visible to the public, the methodology ensures that it is immutable and irrefutable. As part of the in-depth research, the proposed file system has been successfully tested in the Google Cloud Platform. Also, the basic operations like read, write, and delete on Distributed File System with distributed metadata are compared with that of Hadoop Distributed File System based on distribution time on the same cluster setup. The novel distributed file system provides better results compared to the existing methodologies. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
9. Efficient I/O Merging Scheme for Distributed File Systems.
- Author
-
An, Byoung Chul and Sung, Hanul
- Subjects
- *
SCALABILITY , *METADATA , *ELECTRONIC file management - Abstract
Recently, decentralized file systems are widely used to overcome centralized file systems' load asymmetry between nodes and the scalability problem. Due to the lack of a metadata server, decentralized systems require more RPC requests to control metadata processing between clients and servers, which adversely impacts the I/O performance and traffic imbalance by increasing RPC latency. In this paper, we propose an efficient I/O scheme to reduce the RPC overhead in decentralized file systems. Instead of sending a single RPC request at a time, we enqueued the RPCs in the global queue and merged them into larger RPC requests, thus avoiding excessive RPC latency overheads. The experimental results showed that our scheme improves write and read performance by up to 13% and 16%, respectively, compared with those of the original. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
10. 铜基粉末冶金刹车闸片磨损形貌演变研究.
- Author
-
申长慧, 米雪, 彭金方, 唐攀, 杨文贤, 方婧婷, 黄银, and 朱旻昊
- Subjects
MECHANICAL wear ,FRETTING corrosion ,POWDER metallurgy ,SURFACE roughness ,SURFACE morphology ,ELECTRONIC file management - Abstract
Copyright of Lubrication Engineering (0254-0150) is the property of Editorial Office of LUBRICATION ENGINEERING and its content may not be copied or emailed to multiple sites or posted to a listserv without the copyright holder's express written permission. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract. (Copyright applies to all Abstracts.)
- Published
- 2023
- Full Text
- View/download PDF
11. An Empirical Configuration Study of a Common Document Clustering Pipeline.
- Author
-
Eklund, Anton, Forsman, Mona, and Drewes, Frank
- Subjects
DOCUMENT clustering ,NATURAL language processing ,DIMENSIONS ,EMBEDDINGS (Mathematics) ,ELECTRONIC file management - Abstract
Document clustering is frequently used in applications of natural language processing, e.g. to classify news articles or create topic models. In this paper, we study document clustering with the common clustering pipeline that includes vectorization with BERT or Doc2Vec, dimension reduction with PCA or UMAP, and clustering with K-Means or HDBSCAN.We discuss the interactions of the different components in the pipeline, parameter settings, and how to determine an appropriate number of dimensions. The results suggest that BERT embeddings combined with UMAP dimension reduction to no less than 15 dimensions provides a good basis for clustering, regardless of the specific clustering algorithm used. Moreover, while UMAP performed better than PCA in our experiments, tuning the UMAP settings showed little impact on the overall performance. Hence, we recommend configuring UMAP so as to optimize its time efficiency. According to our topic model evaluation, the combination of BERT and UMAP, also used in BERTopic, performs best. A topic model based on this pipeline typically benefits from a large number of clusters. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
12. Fast and Low Overhead Metadata Operations for NVM-Based File System Using Slotted Paging.
- Author
-
Lin, Fangzhu, Xiao, Chunhua, Liu, Weichen, Wu, Lin, Shi, Chen, and Ning, Kun
- Subjects
- *
ELECTRONIC file management , *METADATA , *NONVOLATILE memory , *DYNAMIC random access memory , *EDIBLE fats & oils , *JOURNAL writing , *ELECTRIC breakdown - Abstract
Existing nonvolatile memory (NVM)-based file systems can fully leverage the characteristics of NVM to obtain better performance than traditional disk-based file systems. It has the potential capacity to efficiently manage metadata and perform fast metadata operations. However, most NVM-based file systems mainly focus on managing file metadata (inode), while pay little attention to directory metadata (dentry), which also has a noticeable impact on the file system performance. Besides, the traditional journaling technique that guarantees metadata consistency may not yield satisfactory performance on NVM-based file systems. To solve these problems, in this article we propose a fast and low overhead metadata operation mechanism, called FLOMO. It first adopts a novel slotted-paging structure in NVM to reorganize dentry for efficiently performing dentry operations, and utilizes the red–black tree in DRAM to accelerate dentry lookup and the search process of dentry deletion. Moreover, FLOMO presents a selective journaling scheme for metadata updates, which partially logs the changes related to dentry in the proposed slotted page, thereby, mitigating the redundant journaling overhead. To verify FLOMO, we implement it in a typical NVM-based file system, the persistent memory file system (PMFS). Experimental results show that FLOMO accelerates the metadata operations in PMFS by 34.4% $\sim 59$ %, and notably reduces the journaling overhead for metadata, shortening the latency by 59% on average. For real-world applications, FLOMO has higher throughput compared with PMFS, PMFS without journal, and NOVA, achieving up to $2.1\times $ , $1.1\times $ , and $1.3\times $ performance improvement, respectively. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
13. A Research on the Sharing Platform of Wild Bird Data in Yunnan Province Based on Blockchain and Interstellar File System.
- Author
-
Yang, Huaiyuan, Li, Yucheng, Zhou, Hua, Zhao, Yili, and Song, Lei
- Subjects
- *
BLOCKCHAINS , *DATA warehousing , *ELECTRONIC data processing , *ACCESS control , *ELECTRONIC file management , *INFORMATION sharing - Abstract
Sharing scientific data is an effective means to rationally exploit scientific data and is vital to promote the development of the industrial chain and improve the level of science and technology. In recent years, the popularity of the open data platform has increased, but problems remain, including imperfect system architecture, unsound privacy and security, and non-standardized interaction data. To address these problems, the blockchain's decentralization, smart contracts, distributed storage, and other features can be used as the core technology for open data systems. This paper addresses the problems of opening, allocation-right confirmation, sharing, and rational use of wild-bird data from Yunnan Province, China. A data storage model is proposed based on the blockchain and interstellar file system and is applied to wild-bird data to overcome the mutual distrust between ornithology institutions in the collaborative processing and data storage of bird data. The model provides secure storage and secure access control of bird data in the cloud, thereby ensuring the decentralized and secure storage of wild-bird data for multiple research institutions. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
14. The HMML Authority File: Current Status and Future Directions.
- Author
-
Soufi, Denise and Gullo, Daniel K.
- Subjects
ELECTRONIC file management ,ANNOTATIONS ,MANUSCRIPTS ,MUSEUMS - Abstract
This annotation describes the HMML Authority File (HAF), an open access database of authority records used by the Hill Museum & Manuscript Library that currently focuses on name authorities related to the Eastern Christian and Islamic manuscript traditions, but will eventually include all of HMML's authority files. It explains the project's history, describes how the file is populated, and discusses the best methods for searching and accessing the rich data stored in the file. As the file is a Beta project, the annotation also discusses its shortcomings and their intended solutions, as well as other future directions. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
15. HR document and file management for managers.
- Subjects
ELECTRONIC file management ,EXECUTIVES - Abstract
The article presents suggestion for HR document and file management for managers including recommendations to check with HR before pitching any documents related to Hiring, promotions, demotions and terminations, Employee performance, and Safety and health.
- Published
- 2022
16. Activity-Centric Computing Systems.
- Author
-
BARDRAM, JAKOB E., JEURIS, STEVEN, TELL, PAOLO, HOUBEN, STEVEN, and VOIDA, STEPHEN
- Subjects
- *
COMPUTER science , *RESEARCH & development , *CONCEPTUAL models , *TEAMS in the workplace , *ELECTRONIC file management - Abstract
The article discusses the notion of Activity-Centric Computing (ACC) in relation to application-centric computing. Topics include the history of ACC systems in relation to research and development, the relation of ACC to conceptual models, and the relation of ACC to collaborative work and film or resource management.
- Published
- 2019
- Full Text
- View/download PDF
17. Editorial: On My Way In III: It's Not All About Me: Writing a Cover Letter for an Academic Position; In This Issue; In This Issue – Reviews.
- Subjects
- *
COVER letters , *TEACHERS , *JOB vacancies , *JOB applications , *ELECTRONIC file management - Abstract
In this article, the author offers suggestions for writing a cover letter for an academic position. It mentions cover letter as a genre of academic writing means recognizing that it presents an opportunity to make an argument and cover letters show that the applicant has considered the needs of the hiring institution. It also mentions tempting to take a cover letter off the shelf and out of the electronic files.
- Published
- 2022
- Full Text
- View/download PDF
18. Security Research in Personnel Electronic File Management Based on Blockchain Technology.
- Author
-
Wang, Hongbing and Zhang, Jian
- Subjects
ELECTRONIC file management ,SECURITY personnel ,BLOCKCHAINS ,ARCHIVES collection management ,PROFESSIONAL employees ,INFORMATION resources management - Abstract
Compared with traditional files, electronic personnel files have the characteristics of the economy, environmental protection, convenience, and sharing and are gradually replacing traditional paper files. However, the development of electronic archives is still in its infancy, and there are still many problems, including the professional quality of personnel, information management, and the security of electronic archives storage. As an emerging technology, blockchain technology has the characteristics of decentralization, immutability, and traceability. This paper applies blockchain technology in the management of electronic archives, overcomes the internal distortion and insecurity of electronic archives, and designs an electronic archives management module based on blockchain technology. Through the application in some schools, the effectiveness and practicability of the algorithm are proved. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
19. Cross‐Analysis of Researchers' Different Shared File Management Activities in Cloud Storage.
- Subjects
- *
ELECTRONIC file management , *CLOUD storage , *INTERNET surveys , *INFORMATION resources management , *PERSONAL information management - Abstract
Using cloud storage for collaborative projects has become common practice among researchers. However, despite its popularity, we lack an understanding of researchers' shared file management practices in cloud storage. More importantly, although researchers engage in various shared file management activities (i.e., create, update, organize, find/re‐find, keep, archive, and delete) in cloud storage, no studies have comparatively investigated these activities. Based on 475 responses from researchers collected through an online survey, this study examined researchers' different shared file management activities by focusing on the differences in the frequency, perceived ease of use, and satisfaction for each activity, and identified well‐supported activities as well as activities that require increased support. This study deepens our understanding of researchers' shared file management practice in cloud storage by cross‐examining an array of shared file management activities, and contributes to the development of tools and applications that better support researchers' shared file management in cloud storage. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
20. Jellyfi sh in the Hammerspace.
- Author
-
Beier, Bela
- Subjects
FILE Transfer Protocol (Computer network protocol) ,ELECTRONIC file management ,ANIMATION (Cinematography) - Abstract
The article presents an interview with Jeremy Smith, chief technology officer (CTO) at animation firm Jellyfish Pictures. He discusses global file system that is able to move data from different point depending on one's pipeline. He mentions use of Object storage at Jellyfish. He also mentions shot made up of the pre-existing shot while looking at animation.
- Published
- 2022
21. Wie bekämpft der Bund die Wirtschaftskriminalität?
- Author
-
Steiner, Yves
- Subjects
PREVENTION of money laundering ,ELECTRONIC file management ,ECONOMIC crime ,COMMODITY futures ,LAW enforcement ,ASSET-liability management ,MONEY laundering - Abstract
Copyright of Volkswirtschaft is the property of State Secretariat for Economic Affairs (SECO) and its content may not be copied or emailed to multiple sites or posted to a listserv without the copyright holder's express written permission. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract. (Copyright applies to all Abstracts.)
- Published
- 2022
22. Survey on Various Cloud Security Approaches.
- Author
-
Sagar, K. V. Daya, Sri, P. S. G. Aruna, Krishna, Chinta Venkata Murali, BRAHMMESWARA, BALA, and Sakhamuri, Sridevi
- Subjects
- *
CLOUD computing , *DATA plans , *ELECTRONIC file management , *CLOUD storage , *AUDITING - Abstract
Cloud computing plays a significant role in effective data handling based on the increase in data usage in various realtime applications. Data auditing is performed on certain files and the authenticator with deduplication. It addresses the problem of key management to deduce the file content based on the malicious activities performed on the cloud. So, based on effective auditing of the integrity of the data and authenticator, the data is checked properly and minimizes the overhead of cloud storage overhead. In this work, the cloud audit and authenticator approach is proposed based on a certain file system, which makes the malicious user get authenticate the data auditing verification as the existing algorithm has low security based on entropy. They propose a data auditing approach that integrates with file management and the authenticator of data deduplication. The proposed approach performs the authenticator process and new form of file tag, which helps guarantee effective security based on the random generation of message key. The proposed approach achieves minimum computational overhead based on the authenticator and data block generation in the performance analysis. Then the security verification is performed on various attacks, such as brute force attack, a man-in-the-middle attack, etc., to check whether the approach is safe or unsafe against the attacks. [ABSTRACT FROM AUTHOR]
- Published
- 2022
23. Aggregated Traffic Anomaly Detection Using Time Series Forecasting on Call Detail Records.
- Author
-
Mokhtari, Arian, Ghorbani, Niloofar, and Bahrak, Behnam
- Subjects
ANOMALY detection (Computer security) ,TRAFFIC monitoring ,TIME series analysis ,TIME management ,METADATA ,CYBERTERRORISM ,ELECTRONIC file management - Abstract
Mobile network operators store an enormous amount of information like log files that describe various events and users' activities. Analysis of these logs might be used in many critical applications such as detecting cyber attacks, finding behavioral patterns of users, security incident response, and network forensics. In a cellular network, call detail records (CDRs) is one type of such logs containing metadata of calls and usually includes valuable information about contacts such as the phone numbers of originating and receiving subscribers, call duration, the area of activity, type of call (SMS or voice call), and a timestamp. With anomaly detection, it is possible to determine abnormal reduction or increment of network traffic in an area or for a particular person. This paper's primary goal is to study subscribers' behavior in a cellular network, mainly predicting the number of calls in a region and detecting anomalies in the network traffic. In this paper, a new hybrid method is proposed based on various anomaly detection methods such as GARCH, K-means, and neural network to determine the anomalous data. Moreover, we have discussed the possible causes of such anomalies. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
24. How Big Are Peoples' Computer Files? File Size Distributions Among User‐managed Collections.
- Author
-
Dinneen, Jesse David and Nguyen, Ba Xuan
- Subjects
- *
ELECTRONIC file management , *PERSONAL information management , *INFORMATION resources management , *HUMAN-computer interaction , *DEMOGRAPHIC surveys - Abstract
Improving file management interfaces and optimising system performance requires current data about users' digital collections and particularly about the file size distributions of such collections. However, prior works have examined only the sizes of system files and users' work files in varied contexts, and there has been no such study since 2013; it therefore remains unclear how today's file sizes are distributed, particularly personal files, and further if distributions differ among the major operating systems or common occupations. Here we examine such differences among 49 million files in 348 user collections. We find that the average file size has grown more than ten‐fold since the mid‐2000s, though most files are still under 8 MB, and that there are demographic and technological influences in the size distributions. We discuss the implications for user interfaces, system optimisation, and PIM research. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
25. montage: NVM-based scalable synchronization framework for crash-consistent file systems.
- Author
-
Sul, Woong, Yeom, Heon Y., and Han, Hyuck
- Subjects
- *
DESIGN exhibitions , *SYNCHRONIZATION , *DATA warehousing , *METADATA , *SCALABILITY , *ELECTRONIC file management , *ATOMIC displacements - Abstract
In file systems, a single write system call can make multiple modifications to data and metadata, but such changes are not flushed in an atomic way. To retain the consistency of file systems, conventional approaches guarantee crash consistency in exchange for sacrificing system performance. To mitigate the performance penalty, non-volatile memory (NVM) technologies are believed to be a good candidate for this purpose owing to their low latency and byte-addressibility. However, none of the prior proposals that exploit NVM provide both scalability and strict ordering of modifications. In this article, we propose montage, a crash consistency framework for file systems that consists of two parts. First, montage splits NVM space into multiple staging areas and synchronizes the flushing of modifications these to the storage device. Second, montage uses the pipelined architecture to speed up data flushing to storage. We apply montage to two journaling file systems (ext4 and JFS) and evaluate them on a multicore server with high-performance storage. The evaluation results demonstrate that our design exhibits better performance by a wide margin than recent NVM-based journaling file systems. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
26. Fast and Powerful Hashing Using Tabulation.
- Author
-
Thorup, Mikkel
- Subjects
- *
HASHING , *MATHEMATICAL bounds , *INDEPENDENCE (Mathematics) , *ELECTRONIC file management , *COMPUTER logic - Abstract
Randomized algorithms are often enjoyed for their simplicity, but the hash functions employed to yield the desired probabilistic guarantees are often too complicated to be practical. Here, we survey recent results on how simple hashing schemes based on tabulation provide unexpectedly strong guarantees. Simple tabulation hashing dates back to Zobrist (A new hashing method with application for game playing. Technical Report 88, Computer Sciences Department, University of Wisconsin). Keys are viewed as consisting of c characters and we have precomputed character tables h1, . . ., hc mapping characters to random hash values. A key x = (x1, . . ., xc) is hashed to h1[x1] ⊕ h2[x2]. . . . . ⊕ hc[xc]. This schemes is very fast with character tables in cache. Although simple tabulation is not even four-independent, it does provide many of the guarantees that are normally obtained via higher independence, for example, linear probing and Cuckoo hashing. Next, we consider twisted tabulation where one input character is "twisted" in a simple way. The resulting hash function has powerful distributional properties: Chernoffstyle tail bounds and a very small bias for minwise hashing. This is also yields an extremely fast pseudorandom number generator that is provably good for many classic randomized algorithms and data-structures. Finally, we consider double tabulation where we compose two simple tabulation functions, applying one to the output of the other, and show that this yields very high independence in the classic framework of Wegman and Carter.26 In fact, w.h.p., for a given set of size proportional to that of the space consumed, double tabulation gives fully random hashing. We also mention some more elaborate tabulation schemes getting near-optimal independence for given time and space. Although these tabulation schemes are all easy to implement and use, their analysis is not. [ABSTRACT FROM AUTHOR]
- Published
- 2017
- Full Text
- View/download PDF
27. HBIM-GIS INTEGRATION WITH AN IFC-TO-SHAPEFILE APPROACH: THE PALAZZO TROTTI VIMERCATE PILOT CASE STUDY.
- Author
-
Gabriele, M. and Previtali, M.
- Subjects
DATA conversion ,DATA integration ,PILOT projects ,GEOGRAPHIC information systems ,ELECTRONIC file management - Abstract
The proprietary software investments in the data integration field are incrementing, and the progresses are visible in the possibility to directly open in a GIS environment a 3D software data format. Still, this is limited to the integration between the proprietary data formats and standards, ArcGIS environment shapefile multipatch and Revit 3D model, by using a proprietary software (ArcGIS). This study takes advantage of the lesson-learnt results in the proprietary data integration field, wanting to replicate a similar result using the IFC open standard, which is not directly openable by a GIS interface and needs to overcome a conversion that in most of the cases leads to semantic and geometric losses. So, an IFC-to-shapefile data conversion was performed, stressing (i) the way information is stored in the attribute table to query the geometries and perform geoprocessing, by (ii) implementing workarounds to keep the Revit instances' shared parameters in the IFC file, (iii) meanwhile having a high Level of Detail of the HBIM. The research performed the IFC-to-shapefile data conversion through FME (Feature Manipulation Engine), benefitting of the flexibility of the shapefile format and of the IFC' possibility to keep a high LOD in the export phase. Both allowed to properly query and manage the elements of an HBIM in a GIS (ArcGIS environment), and, using relational attributes table, retrieve the information contained in each Revit instance' property panel, as the shared parameters that implement the BIM Level of Information (LOI). [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
28. ENABLING AIR QUALITY MONITORING WITH THE OPEN DATA CUBE: IMPLEMENTATION FOR SENTINEL-5P AND GROUND SENSOR OBSERVATIONS.
- Author
-
Cedeno Jimenez, J. R., Oxoli, D., and Brovelli, M. A.
- Subjects
AIR quality monitoring ,METADATA ,AIR pollution measurement ,PYTHON programming language ,CUBES ,ELECTRONIC file management ,OPEN source software ,DETECTORS - Published
- 2021
- Full Text
- View/download PDF
29. A functional BCI model by the IEEE P2731 working group: data storage and sharing.
- Author
-
Bianchi, Luigi, Antonietti, Alberto, Bajwa, Garima, Ferrante, Raffaele, Mahmud, Mufti, and Balachandran, Pradeep
- Subjects
- *
INFORMATION sharing , *GROUP process , *DATA warehousing , *ELECTRONIC file management , *OPTICAL disks - Abstract
The IEEE P2731 working group is in the process of developing a functional model so that virtually any BCI system could be described according to it. In this paper, we want to stimulate the discussion for the definition and selection of the information that should be stored in a file for effectively allowing the sharing of BCI data and tools amongst researchers. Establishing these requirements and procedures will accelerate BCI development and lay the foundation for accessible and scalable BCI technology as well as provide the foundations for the definition of a standard file format. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
30. Image Steganography-Based GUI Design to Hide Agricultural Data.
- Author
-
SOLAK, Serdar and ALTINISIK, Umut
- Subjects
- *
COMPUTER engineering , *SIGNAL-to-noise ratio , *DATA transmission systems , *CRYPTOGRAPHY , *ELECTRONIC file management , *GRAPHICAL user interfaces , *ALGORITHMS - Abstract
Throughout the ages, safely preserving and transmitting data that have extraordinary importance for humanity has increased its importance with rapid advances in computer technology. Steganography stores hidden data within the files, which are unnoticed by third parties, so it provides secure transmission of data to the receiver. In this study, a steganography-based GUI design has been carried out, which ensures that the agricultural data is safely stored and communicated to the other party. We used LSB one-bit, two-bit, three-bit substitution and PVD algorithms with GUI for stages of agricultural data hiding and extracting at cover images. We also provided extra security using the embedded key and shifting operations on the hidden data before hiding data the cover image. In short, we confused the hidden data in the cover image so that malicious people can't understand. In experimental studies, performance analysis was evaluated by comparing various criteria as similarity ratio (Structural Similarity Index Measure, SSIM), stego image quality (Peak Signal-to-Noise Ratio, PSNR) and data hiding capacity (Payload). [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
31. Review of Mitchell-Schuitevoerder (2020): A Project-Based Approach to Translation Technology.
- Author
-
Orrego-Carmona, David
- Subjects
COMPUTER adaptive testing ,COMPUTER file management ,TRANSLATING & interpreting ,SELF-efficacy ,SPEECH perception ,VOICE recognition software ,ELECTRONIC file management - Abstract
Mitchell-Schuitevoerder, Rosemary 2020 A Project-Based Approach to Translation Technology London and New York Routledge xxxv, 164 9780367138844 I A Project-Based Approach to Translation Technology i is probably the most updated, adaptable, and accessible resource for translators to effectively implement translation technologies in every step of the translation workflow. Chapter 1 covers the basic principles of CAT tools and discusses how translators, language service providers (LSPs) and clients interact with these tools. Mitchell-Schuitevoerder not only mentions computer literacy skills as essential for learning about CAT tools but specifically highlights how these skills can ensure a successful CAT tool operation. Chapter 3 deals with the integration of machine translation (MT) into CAT tools and focuses on the efficient use of MT in the translation workflow. [Extracted from the article]
- Published
- 2021
- Full Text
- View/download PDF
32. Enhancing HDFS with a full-text search system for massive small files.
- Author
-
Xu, Wentao, Zhao, Xin, Lao, Bin, and Nong, Ge
- Subjects
- *
TEXT files , *RECORDS management , *METADATA , *ELECTRONIC file management - Abstract
HDFS is a popular open-source system for scalable and reliable file management, which is designed as a general-purpose solution for distributed file storage. While it works well for medium or large files, it will suffer heavy performance degradations in case of lots of small files. To overcome this drawback, we propose here a system to enhance HDFS with a distributed true full-text search system SAES of 100% recall and precision ratios. By indexing the meta data of each file, e.g., name, size, date and description, files can be quickly accessed by efficient searches over metadata. Moreover, by merging many small files into a large file to be stored with better space and I/O efficiencies, the negative performance impacts caused by directly storing each small file individually are avoided. An experimental study is conducted for function and performance tests on both realistic and artificial data. The experimental results show that the system works well for file operations such as uploading, downloading and deleting. Moreover, the RAM consumption for managing massive small files is dramatically reduced, which is critical for good system performance. The proposed system could be a potential storage solution for massive small files. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
33. ROOA: CloudIDE framework for extension development.
- Author
-
Javidpanah, Mohsen, Javadpour, Amir, and Rezaei, Samira
- Subjects
TEXT editors (Computer programs) ,ELECTRONIC file management ,PROGRAMMING languages ,INFORMATION resources management ,COMPILERS (Computer programs) - Abstract
The Program is a mediator as all IDEs are mediator programs, in the sense that IDEs like Eclipse, Eric IDE, and net beans are considered as a mediator between a programmer and Interpreter or the programming language compiler. In fact, in this paper we can say that all things the programs do can be done with a simple text editor and we can see the Program output by direct contact with interpreter or compiler. The only difference between the Program provided by me and other IDEs is that the environment is special for building the extension. It means that instead of being an interface between user and interpreter or compiler, standards provided by other applications are usually different. The program tries to fill in the absence of an Efficient IDE in the field of building an extension. On the other hand, it reduces the cost of extension standard providers for presenting an IDE specific to that standard. In fact, in the second target the program provided will be a platform that has the ability to mount the next standards. Standard providers should only conduct a Part of coding. A constant part is common in all standards. We should find a solution for them in all IDEs, such as file management and information management that have been implemented in constant part of the IDE. The variable and non-similar part is implemented as classes. If needed to implement, it should be inherited from them and some of their functions must be overridden. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
34. Migrating Engineering Windows HPC applications to Linux HTCondor and Slurm Clusters.
- Author
-
Doglioni, C., Kim, D., Stewart, G.A., Silvestris, L., Jackson, P., Kamleh, W., Alandes Pradillo, Maria, Høimyr, Nils, Llopis Sanmillan, Pablo, and Tapani Jylhänkangas, Markus
- Subjects
- *
MESSAGE passing (Computer science) , *HIGH performance computing , *QUANTUM chromodynamics , *USER interfaces , *DATA analysis , *ELECTRONIC file management - Abstract
The CERN IT department has been maintaining different High Performance Computing (HPC) services over the past five years. While the bulk of computing facilities at CERN are running under Linux, a Windows cluster was dedicated for engineering simulations and analysis related to accelerator technology development. The Windows cluster consisted of machines with powerful CPUs, big memory, and a low-latency interconnect. The Linux cluster resources are accessible through HTCondor, and are used for general purpose parallel but single-node type jobs, providing computing power to the CERN experiments and departments for tasks such as physics event reconstruction, data analysis, and simulation. For HPC workloads that require multi-node parallel environments for Message Passing Interface (MPI) based programs, there is another Linux-based HPC service that is comprised of several clusters running under the Slurm batch system, and consist of powerful hardware with low-latency interconnects. In 2018, it was decided to consolidate compute intensive jobs in Linux to make a better use of the existing resources. Moreover, this was also in line with CERN IT strategy to reduce its dependencies on Microsoft products. This paper focuses on the migration of Ansys [1], COMSOL [2] and CST [3] users from Windows HPC to Linux clusters. Ansys, COMSOL and CST are three engineering applications used at CERN for different domains, like multiphysics simulations and electromagnetic field problems. Users of these applications are in different departments, with different needs and levels of expertise. In most cases, the users have no prior knowledge of Linux. The paper will present the technical strategy to allow the engineering users to submit their simulations to the appropriate Linux cluster, depending on their simulation requirements. We also describe the technical solution to integrate their Windows workstations in order from them to be able to submit to Linux clusters. Finally, we discuss the challenges and lessons learnt during the migration. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
35. Improvements in utilisation of the Czech national HPC center.
- Author
-
Doglioni, C., Kim, D., Stewart, G.A., Silvestris, L., Jackson, P., Kamleh, W., Svatoš, Michal, Chudoba, Jiří, and Vokáč, Petr
- Subjects
- *
DISTRIBUTED computing , *ELECTRONIC file management , *COMPUTER software installation , *LARGE Hadron Collider , *WORKFLOW management systems - Abstract
The distributed computing system of the ATLAS experiment at LHC is allowed to opportunistically use resources at the Czech national HPC center IT4Innovations in Ostrava. The jobs are submitted via an ARC Compute Element (ARC-CE) installed at the grid site in Prague. Scripts and input files are shared between the ARC-CE and a shared file system located at the HPC centre via sshfs. This basic submission system has worked there since the end of 2017. Several improvements were made to increase the amount of resource that ATLAS can use. The most significant change was the migration of the submission system to enable pre-emptable jobs, to adapt to the HPC management's decision to start pre-empting opportunistic jobs. Another improvement of the submission system was related to the sshfs connection which seemed to be a limiting factor of the system. Now, the submission system consists of several ARC-CE machines. Also, various parameters of sshfs were tested in an attempt to increase throughput. As a result of the improvements, the utilisation of the Czech national HPC center by the ATLAS distributed computing increased. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
36. Physics Data Production on HPC: Experience to be efficiently running at scale.
- Author
-
Doglioni, C., Kim, D., Stewart, G.A., Silvestris, L., Jackson, P., Kamleh, W., Poat, M D, Lauret, J, Porter, J, and Balewski, J
- Subjects
- *
HIGH performance computing , *PHYSICS experiments , *BIG data , *ELECTRONIC file management , *SCALABILITY , *SOFTWARE configuration management - Abstract
The Solenoidal Tracker at RHIC (STAR) is a multi-national supported experiment located at the Brookhaven National Lab and is currently the only remaining running experiment at RHIC. The raw physics data captured from the detector is on the order of tens of PBytes per data acquisition campaign, making STAR fit well within the definition of a big data science experiment. The production of the data has typically run using a High Throughput Computing (HTC) approach either done on a local farm or via Grid computing resources. Especially, all embedding simulations (complex workflow mixing real and simulated events) have been run on standard Linux resources at NERSC's Parallel Distributed Systems Facility (PDSF). However, as per April 2019 PDSF has been retired and High Performance Computing (HPC) resources such as the Cray XC-40 Supercomputer known as "Cori" have become available for STAR's data production as well as embedding. STAR has been the very first experiment to show feasibility of running a sustainable data production campaign on this computing resource. In this contribution, we hope to share with the community the best practices for using such resource efficiently. The use of Docker containers with Shifter is the standard approach to run on HPC at NERSC – this approach encapsulates the environment in which a standard STAR workflow runs. From the deployment of a tailored Scientific Linux environment (with the set of libraries and special configurations required for STAR to run) to the deployment of third-party software and the STAR specific software stack, we've learned it has become impractical to rely on a set of containers comprising each specific software release. To this extent, a solution based on the CernVM File System (CVMFS) for the deployment of software and services has been deployed but it doesn't stop there. One needs to make careful scalability considerations when using a resource like Cori, such as avoiding metadata lookups, scalability of distributed filesystems, and real limitations of containerized environments on HPC. Additionally, CVMFS clients are not compatible on Cori nodes and one needs to rely on an indirect NFS mount scheme using custom services known as DVS servers designed to forward data to worker nodes. In our contribution, we will discuss our strategies from the past and our current solution based on CVMFS. The second focus of our presentation will be to discuss strategies to find the most efficient use of database Shifter containers serving our data production (a near "database as a service" approach) and the best methods to test and scale your workflow efficiently. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
37. Evolution of the CERNBox platform to support the Malt project.
- Author
-
Doglioni, C., Kim, D., Stewart, G.A., Silvestris, L., Jackson, P., Kamleh, W., González Labrador, Hugo, Bippus, Vincent Nicolas, Bukowiec, Sebastian, Castro, Diogo, Dellabella, Sebastien, Kwiatek, Michal, Lo Presti, Giuseppe, Mascetti, Luca, Mościcki, Jakub T., Puentes, Esteban, Seweryn, Piotr Jan, and Smyrnakis, Apostolos
- Subjects
- *
CLOUD storage , *SYNCHRONIZATION software , *ACCESS to information , *DATA visualization , *ELECTRONIC file management , *GATEWAYS (Computer networks) - Abstract
CERNBox is the CERN cloud storage hub for more than 25,000 users at CERN. It allows synchronising and sharing files on all major desktop and mobile platforms (Linux, Windows, MacOSX, Android, iOS) providing universal, ubiquitous, online- and offline access to any data stored in the CERN EOS infrastructure. CERNBox also provides integration with other CERN services for big science: visualisation tools, interactive data analysis and real-time collaborative editing. Over the last two years, CERNBox has evolved from a pure cloud sync and share platform into a collaborative service, to support new applications such as DrawIO for diagrams and organigrams sketching, OnlyOffice and Collabora Online for documents editing, and DXHTML Gantt for project management, as alternatives to traditional desktop applications. Moving to open source applications has the advantage to reduce licensing costs and enables easier integration within the CERN infrastructure. This move from commercial software to open source solutions is part of the MALT project, led by the IT department at CERN to reduce the dependencies on commercial solutions. As part of the MALT project, CERNBox is the chosen solution to replace Home directories of the Windows DFS file system. Access to storage from Windows managed devices for end-users is largely covered by synchronization clients. However, online access using standard CIFS/SMB protocol is required for shared use-cases, such as central login services (Terminal Services) and visitor desktop computers. We present recent work to introduce a set of Samba gateways running in High Availability cluster mode to enable direct access to the CERNBox backend storage (EOS). [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
38. Machine Learning-based Anomaly Detection of Ganglia Monitoring Data in HEP Data Center.
- Author
-
Doglioni, C., Kim, D., Stewart, G.A., Silvestris, L., Jackson, P., Kamleh, W., Chen, Juan, Wang, Lu, and Hu, Qingbao
- Subjects
- *
ANOMALY detection (Computer security) , *MACHINE learning , *DATA libraries , *DATA visualization , *ELECTRONIC file management - Abstract
This paper introduces a generic and scalable anomaly detection framework. Anomaly detection can improve operation and maintenance efficiency and assure experiments can be carried out effectively. The framework facilitates common tasks such as data sample building, retagging and visualization, deviation measurement and performance measurement for machine learning-based anomaly detection methods. The samples we used are sourced from Ganglia monitoring data. There are several anomaly detection methods to handle spatial and temporal anomalies within the framework. Finally, we show the rudimental application of the framework on Lustre distributed file systems in daily operation and maintenance. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
39. Certifying a File System Using Crash Hoare Logic: Correctness in the Presence of Crashes.
- Author
-
Chajed, Tej, Haogang Chen, Chlipala, Adam, Kaashoek, M. Frans, Zeldovich, Nickolai, and Ziegler, Daniel
- Subjects
- *
ELECTRONIC file management , *HOARE logic , *COMPUTER system failures , *INFORMATION storage & retrieval systems , *AUTOMATION , *ERRORS - Abstract
FSCQ is the first file system with a machine-checkable proof that its implementation meets a specification, even in the presence of fail-stop crashes. FSCQ provably avoids bugs that have plagued previous file systems, such as performing disk writes without sufficient barriers or forgetting to zero out directory blocks. If a crash happens at an inopportune time, these bugs can lead to data loss. FSCQ's theorems prove that, under any sequence of crashes followed by reboots, FSCQ will recover its state correctly without losing data. To state FSCQ's theorems, this paper introduces the Crash Hoare logic (CHL), which extends traditional Hoare logic with a crash condition, a recovery procedure, and logical address spaces for specifying disk states at different abstraction levels. CHL also reduces the proof effort for developers through proof automation. Using CHL, we developed, specified, and proved the correctness of the FSCQ file system. Although FSCQ's design is relatively simple, experiments with FSCQ as a user-level file system show that it is sufficient to run Unix applications with usable performance. FSCQ's specifications and proofs required significantly more work than the implementation, but the work was manageable even for a small team of a few researchers. [ABSTRACT FROM AUTHOR]
- Published
- 2017
- Full Text
- View/download PDF
40. Erasure-Coding-Based Storage and Recovery for Distributed Exascale Storage Systems.
- Author
-
Kim, Jeong-Joon and Armada, Manuel
- Subjects
SYSTEMS availability ,DATA replication ,STORAGE ,ELECTRONIC file management - Abstract
Various techniques have been used in distributed file systems for data availability and stability. Typically, a method for storing data in a replication technique-based distributed file system is used, but due to the problem of space efficiency, an erasure-coding (EC) technique has been utilized more recently. The EC technique improves the space efficiency problem more than the replication technique does. However, the EC technique has various performance degradation factors, such as encoding and decoding and input and output (I/O) degradation. Thus, this study proposes a buffering and combining technique in which various I/O requests that occurred during encoding in an EC-based distributed file system are combined into one and processed. In addition, it proposes four recovery measures (disk input/output load distribution, random block layout, multi-thread-based parallel recovery, and matrix recycle technique) to distribute the disk input/output loads generated during decoding. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
41. IDDS:一种双链结构传染病数据共享区块链模型.
- Author
-
刘炜, 李阳, 田钊, 彭宇飞, and 佘维
- Subjects
- *
INFORMATION sharing , *PROBLEM solving , *DATA warehousing , *DATA security , *COMMUNICABLE diseases , *ELECTRONIC file management - Abstract
The infectious disease prevention information system has the problems that data is difficult to circulate and share between monitoring systems.To solve these problems,this paper proposed an infectious disease data sharing model based on delegated proof of stake consensus algorithm combined with blockchain technology.By virtue of the characteristics of the blockchain,such as decentralization,non-tampering and collective maintenance,it used double-chain as the blockchain architecture,which improved the work efficiency.Combined with inter-planetary file system,it obtained large capacity storage space,which solved the space problem faced by block data storage and ensured the stability and sharing security of data storage.This paper proposed a consensus algorithm for disease prevention and control to realize the efficient operation.In addition,by comparing with other data sharing models,the advantages of this model in data storage and sharing security are highlighted. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
42. K‐PSO: An improved PSO‐based container scheduling algorithm for big data applications.
- Author
-
Liu, Bo, Li, Jiawei, Lin, Weiwei, Bai, Weihua, Li, Pengfei, and Gao, Qian
- Subjects
BIG data ,PARTICLE swarm optimization ,ALGORITHMS ,COMPUTING platforms ,ELECTRONIC file management ,DATA warehousing - Abstract
Summary: In recent years, Docker container technology is being applied in the field of cloud computing at an explosive speed. The scheduling of Docker container resources has gradually become a research hotspot. Existing big data computing and storage platforms apply with traditional virtual machine technology, which often results in low resource utilization, a long time for flexible scaling and expanding clusters. In this paper, we propose an improved container scheduling algorithm for big data applications named Kubernetes‐based particle swarm optimization(K‐PSO). Experimental results show that the proposed K‐PSO algorithm converges faster than the basic PSO algorithm, and the running time of the algorithm is cut in about half. The K‐PSO container scheduling algorithm and algorithm experiment for big data applications are implemented in the Kubernetes container cloud system. Our experimental results show that the node resource utilization rate of the improved scheduling strategy based on K‐PSO algorithm is about 20% higher than that of the Kube‐scheduler default strategy, balanced QoS priority strategy, ESS strategy, and PSO strategy, while the average I/O performance and average computing performance of Hadoop cluster are not degraded. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
43. Tax Court holds firm to deadline for petition filed after midnight Eastern time.
- Author
-
Beavers, James A.
- Subjects
ELECTRONIC file management ,ELECTRONIC data processing ,TAX returns ,LAW clerks - Abstract
The article discusses that U.S. Tax Court held that a Tax Court petition filed electronically by the taxpayers. It mentions Tax Court held that a document that a taxpayer files electronically with the court is filed when it is received by the court as determined in reference to where the court is located. It also mentions Tax Court found that, as with Federal Rule 6(a), Rule 22(d) is in accord with the idea that the court's electronic filing system serves as a substitute for clerk of court.
- Published
- 2023
44. AB+DM: TWINNING.
- Author
-
TSOMONDO, DZANA
- Subjects
- *
PHOTOGRAPHY , *WIT & humor , *ELECTRONIC file management , *PALETTE (Color range) , *ARTISTS - Published
- 2021
45. Handling conditional queries and data storage on Hyperledger Fabric efficiently.
- Author
-
Yan, Tianlu, Chen, Wei, Zhao, Pengpeng, Li, Zhixu, Liu, An, and Zhao, Lei
- Subjects
- *
DATA warehousing , *BLOCKCHAINS , *ELECTRONIC file management - Abstract
As a popular consortium blockchain platform, Hyperledger Fabric has received increasing attention recently. When executing transactions on such platform, it usually costs a lot of time and hardly to achieve high efficiency. Although efficiently handling transactions can be leveraged to support various use-cases, it presents significant challenges as data on Hyperledger Fabric is organized on file-system and exposed via limited API. We tackle the problem in two ways: conditional queries and data storage. In this paper, we propose the following novel methods. To improve the performance of conditional queries on Hyperledger Fabric, we use all attributes of the query to create composite keys before executing it. In order to achieve further performance improvements, we build an index called AUP in the second method, where we also study the update of AUP during transactions. To speed up data storage on Hyperledger Fabric, We create a cache for the data in the block header. The extensive experiments conducted on the real-world dataset demonstrate that the proposed methods can achieve high performance in terms of efficiency and memory cost. Finally, We implement a prototype system. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
46. IDENTIFICATION OF MARINE EMERGENCY RESPONSE OF ELECTRONIC NAVIGATION OPERATOR.
- Author
-
P. S., Nosov, V. V., Cherniavskyi, S. M., Zinchenko, I. S., Popovych, Ya. А., Nahrybelnyi, and H. V., Nosova
- Subjects
ELECTRONIC navigation ,PROBLEM solving ,MENTAL work ,ELECTRONIC file management ,EXPERIMENTAL films ,P-adic analysis - Abstract
Copyright of Radio Electronics, Computer Science, Control is the property of Zaporizhzhia National Technical University and its content may not be copied or emailed to multiple sites or posted to a listserv without the copyright holder's express written permission. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract. (Copyright applies to all Abstracts.)
- Published
- 2021
- Full Text
- View/download PDF
47. The biglasso Package: A Memory- and Computation-Efficient Solver for Lasso Model Fitting with Big Data in R.
- Author
-
Yaohui Zeng and Breheny, Patrick
- Subjects
- *
R (Computer program language) , *BIG data , *ELECTRONIC file management , *REGRESSION analysis - Abstract
Penalized regression models such as the lasso have been extensively applied to analyzing high-dimensional data sets. However, due to memory limitations, existing R packages like glmnet and ncvreg are not capable of fitting lasso-type models for ultrahigh-dimensional, multi-gigabyte data sets that are increasingly seen in many areas such as genetics, genomics, biomedical imaging, and high-frequency finance. In this research, we implement an R package called biglasso that tackles this challenge. biglasso utilizes memory-mapped files to store the massive data on the disk, only reading data into memory when necessary during model fitting, and is thus able to handle out-of-core computation seamlessly. Moreover, it's equipped with newly proposed, more efficient feature screening rules, which substantially accelerate the computation. Benchmarking experiments show that our biglasso package, as compared to existing popular ones like glmnet, is much more memory-and computation-efficient. We further analyze a 36 GB simulated GWAS data set on a laptop with only 16 GB RAM to demonstrate the out-of-core computation capability of biglasso in analyzing massive data sets that cannot be accommodated by existing R packages. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
48. LOFFS: A Low-Overhead File System for Large Flash Memory on Embedded Devices.
- Author
-
Runyu Zhang, Duo Liu, Xianzhang Chen, Xiongxiong She, Chaoshu Yang, Yujuan Tan, Zhaoyan Shen, and Zili Shao
- Subjects
ELECTRONIC file management ,MACHINE learning ,RESOURCE allocation ,FLASH memory ,INTERNET of things - Abstract
Emerging applications like machine learning in embedded devices (e.g., satellite and vehicles) require huge storage space, which recently stimulates the widespread deployment of large-capacity flash memory in IoT devices. However, existing embedded file systems fall short in managing large-capacity storage efficiently for excessive memory consumption and poor booting performance. In this paper, we propose a novel embedded file system, LOFFS, to tackle the above issues and manage large-capacity NAND flash on resource-limited embedded devices. We redesign the space management mechanisms and construct hybrid file structures to achieve high performance with minimum resource occupation. We have implemented LOFFS in Linux, and the experimental results show that LOFFS outperforms YAFFS by 55.8% on average with orders of magnitude reductions on memory footprint. [ABSTRACT FROM AUTHOR]
- Published
- 2020
49. Efficient Multi-Grained Wear Leveling for Inodes of Persistent Memory File Systems.
- Author
-
Chaoshu Yang, Duo Liu, Runyu Zhang, Xianzhang Chen, Shun Nie, Fengshun Wang, Qingfeng Zhuge, and Sha, Edwin H.-M.
- Subjects
ELECTRONIC file management ,LINUX operating systems ,DATA analysis ,PROBLEM solving ,ACCURACY - Abstract
Existing persistent memory file systems usually store inodes in fixed locations, which ignores the external and internal imbalanced wears of inodes on the persistent memory (PM). Therefore, the PM for storing inodes can be easily damaged. Existing solutions achieve low accuracy of wear-leveling with high-overhead data migrations. In this paper, we propose a Lightweight and Multi-grained Wear-leveling Mechanism, called LMWM, to solve these problems. We implement the proposed LMWM in Linux kernel based on NOVA, a typical persistent memory file system. Compared with MARCH, the state-of-the-art wear-leveling mechanism for inode table, experimental results show that LMWM can improve 2.5x lifetime of PM and 1.12x performance, respectively. [ABSTRACT FROM AUTHOR]
- Published
- 2020
50. LHCb Continuous Integration and Deployment system: a message based approach.
- Author
-
Chitic, Stefan-Gabriel, Couturier, Ben, Clemencic, Marco, Closier, Joel, Forti, A., Betev, L., Litmaath, M., Smirnova, O., and Hristov, P.
- Subjects
- *
PYTHON programming language , *COMPUTER software , *ELECTRONIC file management , *MACHINE learning , *DISTRIBUTED computing - Abstract
A continuous integration system is crucial to maintain the quality of the 6 millions lines of C++ and Python source code of the LHCb software in order to ensure consistent builds of the software as well as to run the unit and integration tests. Jenkins automation server is used for this purpose. It builds and tests around 100 configurations and produces in the order of 1500 built artifacts per day which are installed on the CVMFS file system or potentially on the developers' machines. Faced with a large and growing number of configurations built every day, and in order to ease inter-operation between the continuous integration system and the developers, we decided to put in place a flexible messaging system. As soon as the built artifacts have been produced, the distributed system allows their deployment based on the priority of the configurations. We will describe the architecture of the new system, which is based on RabbitMQ messaging system (and the pika Python client library), and uses priority queues to start the LHCb software integration tests and to drive the installation of the nightly builds on the CVMFS file system. We will also show how the introduction of an event based system can help with the communication of results to developers. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.