368 results on '"data allocation"'
Search Results
2. An effective and efficient parallel large-scale cross-media retrieval in mobile cloud network.
- Author
-
Jiang, Nan, Zhuang, Yi, and Chiu, Dickson K.W.
- Abstract
With the rapid growth of multimedia data (e.g., text, image, video, audio and 3D model, etc) in the web, there are a large number of media objects with different modalities in the multimedia documents such as webpages, which exhibit latent semantic correlation. As a new type of multimedia retrieval method, cross-media retrieval is becoming increasingly attractive, through which users can get the results with various media types with the same semantic information by submitting a retrieval of any media type. The explosive increasing of the number of media objects, however, makes it difficult for the traditional local standalone mode to process efficiently. So the powerful parallel processing capability of cloud computing is accommodated to facilitate the efficient large-scale cross-media retrieval. In this paper, based on a Multi-Layer-Cross-Reference-Graph(MLCRG) model, we propose an efficient parallel cross-media retrieval (PCMR) method in which two enabling techniques (i.e., 1) the adaptive cross-media data allocation algorithm and 2) the PCIndex scheme) are accommodated to effectively speedup the retrieval performance. To the best of our knowledge, there is little research on the parallel retrieval processing of the large-scale cross-media databases in the mobile cloud network. Extensive experiments are conducted to testify that our proposed PCIndex method outperform the three competitors (e.g., the PFAR (Mao et al, 22), the MBSR (Retrieval 4(2):153-164, 42) and the SPECH (Knowl Based Syst 251(5):1-13, 40)) in terms of the effectiveness and efficiency, respectively. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
3. Automatic Parallelization of Iterative Loops Nests on Distributed Memory Computing Systems
- Author
-
Bagliy, A. P., Metelitsa, E. A., Steinberg, B. Ya., Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, and Malyshkin, Victor, editor
- Published
- 2023
- Full Text
- View/download PDF
4. FSIMR: File-system-aware Data Management for Interlaced Magnetic Recording.
- Author
-
YI-HAN LIEN, YEN-TING CHEN, YUAN-HAO CHANG, YU-PEI LIANG, and WEI-KUAN SHIH
- Subjects
DATA management ,UNITS of time ,OPTICAL disks ,DIRECTORIES - Abstract
Interlaced Magnetic Recording (IMR) is an emerging recording technology for hard-disk drives (HDDs) that provides larger storage capacity at a lower cost. By partially overlapping (interlacing) each bottom track with two adjacent top tracks, IMR-based HDDs successfully increase the data density while incurring some hardware write constraints. To update each bottom track, the data on two adjacent top tracks must be read and rewritten to avoid losing their valid data, resulting in additional overhead for performing read-modify-write (RMW) operations. Therefore, researchers have proposed various data management schemes to mitigate such overhead in recent years, aiming at improving the write performance. However, these designs have not taken into account the data characteristics of the file system, which is a crucial layer of operating systems for storing/ retrieving data into/from HDDs. Consequently, the write performance improvement is limited due to the unawareness of spatial locality and hotness of data. This paper proposes a file-system-aware data management scheme called FSIMR to improve system write performance. Noticing that data of the same directory may have higher spatial locality and are mostly updated at the same time, FSIMR logically partitions the IMR-based HDD into fixed-sized zones; data belonging to the same directory will be arranged to one zone to reduce the time of seeking to-be-updated data (seek time). Furthermore, cold data within a zone are arranged to bottom tracks and updated in an out-of-place manner to eliminate RMW operations. Our experimental results show that the proposed FSIMR could reduce the seek time by up to 14% without introducing additional RMW operations, compared to existing designs. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
5. A novel hybrid meta‐heuristic‐oriented latency sensitive cloud object storage system.
- Author
-
Nataraj, N. and Nataraj, R. V.
- Subjects
CLOUD storage ,METAHEURISTIC algorithms ,CLOUD computing ,NP-hard problems ,DATA transmission systems - Abstract
Summary: Cloud providers must find out how to properly arrange data in a limited count of servers while ensuring latency assurances to reduce total storage expenses. Timeout is also important to consider because it has a substantial impact on response latency. The core aim of this task is to implement a new cloud object storage system strategy that handles challenges like "latency‐sensitive data allocation, latency‐sensitive data re‐allocation, and latency‐sensitive workload consolidation." The main contribution here is that distributing the latency of the cloud object storage system allows for better data allocation, data reallocation, and workload consolidation. The primary aim is to use the fewest number of servers feasible to fulfill all requests while maintaining their latency requirements, lowering the overall data transmission cost. As a consequence, Whale Butterfly Optimization Method (WBOA) is a novel hybrid meta‐heuristic algorithm that solves NP‐hard problems by combining baseline advanced algorithms. The simulation outcomes reveal that the offered paradigm consistently provides the greatest outcomes regarding throughput utilization, lower latency, higher storage, and number of used nodes when compared to competing techniques. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
6. Data Allocation with Neural Similarity Estimation for Data-Intensive Computing
- Author
-
Vamosi, Ralf, Schikuta, Erich, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Groen, Derek, editor, de Mulatier, Clélia, editor, Paszynski, Maciej, editor, Krzhizhanovskaya, Valeria V., editor, Dongarra, Jack J., editor, and Sloot, Peter M. A., editor
- Published
- 2022
- Full Text
- View/download PDF
7. Auction-Based Data Transaction in Mobile Networks: Data Allocation Design and Performance Analysis
- Author
-
Bharathi, B., HemanthChowdary, G., Girish NagaKumar, G., Kacprzyk, Janusz, Series Editor, Pal, Nikhil R., Advisory Editor, Bello Perez, Rafael, Advisory Editor, Corchado, Emilio S., Advisory Editor, Hagras, Hani, Advisory Editor, Kóczy, László T., Advisory Editor, Kreinovich, Vladik, Advisory Editor, Lin, Chin-Teng, Advisory Editor, Lu, Jie, Advisory Editor, Melin, Patricia, Advisory Editor, Nedjah, Nadia, Advisory Editor, Nguyen, Ngoc Thanh, Advisory Editor, Wang, Jun, Advisory Editor, Mallick, Pradeep Kumar, editor, Bhoi, Akash Kumar, editor, Marques, Gonçalo, editor, and Hugo C. de Albuquerque, Victor, editor
- Published
- 2021
- Full Text
- View/download PDF
8. An optimized cost-based data allocation model for heterogeneous distributed computing systems.
- Author
-
Tarun, Sashi, Dubey, Mithilesh Kumar, Batth, Ranbir Singh, and Kaur, Sukhpreet
- Subjects
HETEROGENEOUS distributed computing ,SWARM intelligence ,DIRECTED acyclic graphs ,ARCHITECTURAL models - Abstract
Continuous attempts have been made to improve the flexibility and effectiveness of distributed computing systems. Extensive effort in the fields of connectivity technologies, network programs, high processing components, and storage helps to improvise results. However, concerns such as slowness in response, long execution time, and long completion time have been identified as stumbling blocks that hinder performance and require additional attention. These defects increased the total system cost and made the data allocation procedure for a geographically dispersed setup difficult. The load-based architectural model has been strengthened to improve data allocation performance. To do this, an abstract job model is employed, and a data query file containing input data is processed on a directed acyclic graph. The jobs are executed on the processing engine with the lowest execution cost, and the system's total cost is calculated. The total cost is computed by summing the costs of communication, computation, and network. The total cost of the system will be reduced using a Swarm intelligence algorithm. In heterogeneous distributed computing systems, the suggested approach attempts to reduce the system's total cost and improve data distribution. According to simulation results, the technique efficiently lowers total system cost and optimizes partitioned data allocation. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
9. Distributing Data in Real Time Spatial Data Warehouse
- Author
-
Hamdi, Wael, Faiz, Sami, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Woeginger, Gerhard, Editorial Board Member, Yung, Moti, Editorial Board Member, and Qiu, Meikang, editor
- Published
- 2020
- Full Text
- View/download PDF
10. Preliminary Experience with OpenMP Memory Management Implementation
- Author
-
Roussel, Adrien, Carribault, Patrick, Jaeger, Julien, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Woeginger, Gerhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Milfeld, Kent, editor, de Supinski, Bronis R., editor, Koesterke, Lars, editor, and Klinkenberg, Jannis, editor
- Published
- 2020
- Full Text
- View/download PDF
11. Optimization of data allocation in hierarchical memory for blocked shortest paths algorithms
- Author
-
A. A. Prihozhy
- Subjects
shortest paths algorithm ,hierarchical memory ,direct mapped cache ,performance ,block conflict graph ,data allocation ,equitable coloring ,defective coloring ,Information technology ,T58.5-58.64 - Abstract
This paper is devoted to the reduction of data transfer between the main memory and direct mapped cache for blocked shortest paths algorithms (BSPA), which represent data by a D[M×M] matrix of blocks. For large graphs, the cache size S = δ×M2, δ < 1 is smaller than the matrix size. The cache assigns a group of main memory blocks to a single cache block. BSPA performs multiple recalculations of a block over one or two other blocks and may access up to three blocks simultaneously. If the blocks are assigned to the same cache block, conflicts occur among the blocks, which imply active transfer of data between memory levels. The distribution of blocks on groups and the block conflict count strongly depends on the allocation and ordering of the matrix blocks in main memory. To solve the problem of optimal block allocation, the paper introduces a block conflict weighted graph and recognizes two cases of block mapping: non-conflict and minimum-conflict. In first case, it formulates an equitable color-class-size constrained coloring problem on the conflict graph and solves it by developing deterministic and random algorithms. In second case, the paper formulates a problem of weighted defective color-count constrained coloring of the conflict graph and solves it by developing a random algorithm. Experimental results show that the equitable random algorithm provides an upper bound of the cache size that is very close to the lower bound estimated over the size of a complete subgraph, and show that a non-conflict matrix allocation is possible at δ = 0.5 for M = 4 and at δ = 0.1 for M = 20. For a low cache size, the weighted defective algorithm gives the number of remaining conflicts that is up to 8.8 times less than the original BSPA gives. The proposed model and algorithms are applicable to set-associative cache as well.
- Published
- 2021
- Full Text
- View/download PDF
12. Recommender systems and market approaches for industrial data management
- Author
-
Jess, Torben and McFarlane, Duncan
- Subjects
025.04 ,Data allocation ,Data management ,Recommender systems ,Market-based algorithms ,Data overload - Abstract
Industrial companies are dealing with an increasing data overload problem in all aspects of their business: vast amounts of data are generated in and outside each company. Determining which data is relevant and how to get it to the right users is becoming increasingly difficult. There are a large number of datasets to be considered, and an even higher number of combinations of datasets that each user could be using. Current techniques to address this data overload problem necessitate detailed analysis. These techniques have limited scalability due to their manual effort and their complexity, which makes them unpractical for a large number of datasets. Search, the alternative used by many users, is limited by the user’s knowledge about the available data and does not consider the relevance or costs of providing these datasets. Recommender systems and so-called market approaches have previously been used to solve this type of resource allocation problem, as shown for example in allocation of equipment for production processes in manufacturing or for spare part supplier selection. They can therefore also be seen as a potential application for the problem of data overload. This thesis introduces the so-called RecorDa approach: an architecture using market approaches and recommender systems on their own or by combining them into one system. Its purpose is to identify which data is more relevant for a user’s decision and improve allocation of relevant data to users. Using a combination of case studies and experiments, this thesis develops and tests the approach. It further compares RecorDa to search and other mechanisms. The results indicate that RecorDa can provide significant benefit to users with easier and more flexible access to relevant datasets compared to other techniques, such as search in these databases. It is able to provide a fast increase in precision and recall of relevant datasets while still keeping high novelty and coverage of a large variety of datasets.
- Published
- 2017
- Full Text
- View/download PDF
13. Data, User and Power Allocations for Caching in Multi-Access Edge Computing.
- Author
-
Xia, Xiaoyu, Chen, Feifei, He, Qiang, Cui, Guangming, Grundy, John C., Abdelrazek, Mohamed, Xu, Xiaolong, and Jin, Hai
- Subjects
- *
EDGE computing , *NASH equilibrium , *INFORMATION retrieval , *ALLOCATION (Accounting) , *VENDOR-managed inventory - Abstract
In the multi-access edge computing (MEC) environment, app vendors’ data can be cached on edge servers to ensure low-latency data retrieval. Massive users can simultaneously access edge servers with high data rates through flexible allocations of transmit power. The ability to manage networking resources offers unique opportunities to app vendors but also raises unprecedented challenges. To ensure fast data retrieval for users in the MEC environment, edge data caching must take into account the allocations of data, users, and transmit power jointly. We make the first attempt to study the Data, User, and Power Allocation (DUPA $^3$ 3 ) problem, aiming to serve the most users and maximize their overall data rate. First, we formulate the DUPA $^3$ 3 problem and prove its $\mathcal {NP}$ NP -completeness. Then, we model the DUPA $^3$ 3 problem as a potential DUPA $^3$ 3 game admitting at least one Nash equilibrium and propose a two-phase game-theoretic decentralized algorithm named DUPA $^3$ 3 Game to achieve the Nash equilibrium as the solution to the DUPA $^3$ 3 problem. To evaluate DUPA $^3$ 3 Game, we analyze its theoretical performance and conduct extensive experiments to demonstrate its effectiveness and efficiency. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
14. Achieving Robust and Transferable Performance for Conservation‐Based Models of Dynamical Physical Systems.
- Author
-
Zheng, Feifei, Chen, Junyi, Maier, Holger R., and Gupta, Hoshin
- Subjects
DYNAMICAL systems ,CONSERVATION laws (Physics) ,CONCEPTUAL models ,DATA distribution ,CALIBRATION - Abstract
Because physics‐based models of dynamical systems are constrained to obey conservation laws, they must typically be fed long sequences of temporally consecutive (TC) data during model calibration and evaluation. When memory time scales are long (as in many physical systems), this requirement makes it difficult to ensure distributional similarity when partitioning the data into independent, TC, calibration and evaluation subsets. The consequence can be poor and/or uncertain model performance when applied to new situations. To address this issue, we propose a novel strategy for achieving robust and transferable model performance. Instead of partitioning the data into TC calibration and evaluation periods, the model is run in continuous simulation mode for the entire period, and specific time steps are assigned (via a deterministic data‐allocation approach) for use in computing the calibration and evaluation metrics. Generative adversarial testing shows that this approach results in consistent calibration and evaluation data subset distributions. When tested using three conceptual rainfall‐runoff models applied to 163 catchments representing a wide range of hydro‐climatic conditions, the proposed "distributionally consistent (DC)" strategy consistently resulted in better overall performance than achieved using the traditional "TC" strategy. Testing on independent data periods confirmed superior robustness and transferability of the DC‐calibrated models, particularly under conditions of larger runoff skewness. Because the approach is generally applicable to physics‐based models of dynamical systems, it has the potential to significantly improve the confidence associated with prediction and uncertainty estimates generated using such models. Plain Language Summary: When developing models (whether physical or conceptual) it is common practice to partition the available data into separate calibration and evaluation subsets. Further, these subsets need to consist of temporally consecutive data due to the long residence times (memory) of the system state. Such an approach can result in low robustness and poor generalization ability, due to significant variation in the hydro‐climatic conditions represented by the two subsets. Here, we show that by discarding the idea of partitioning historical data into time‐consecutive subsets and instead running model simulations in a continuous manner through the entire data set, it becomes possible to ensure that the data used for calibration and evaluation come from statistically similar distributions, thereby improving the quality of the calibrated model. An additional benefit is that the model states need only be initialized once at the outset of the simulation. An important feature of our strategy is that, by removing the requirement for that calibration and evaluation data to consist of temporally consecutive observations, a very large number of data partitions can be achieved, which significantly improves the possibility of ensuring distributional similarity of the two subsets, thereby improving robustness and generalization ability of the model. Key Points: A novel strategy is proposed to calibrate and validate CRR models by discarding the use of time‐consecutive dataThe proposed approach achieves robust and transferable CRR model performance based on results from 163 catchmentsThe proposed method is generally applicable to all physics‐based models with potential to significantly improve model prediction confidence [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
15. Real-Time Distribution Algorithm for Fully Comparison Data Based on Storm.
- Author
-
Dong, Chang-qing, Chen, Chen, Ren, Nver, and Cai, Jian-jun
- Subjects
- *
DISTRIBUTION (Probability theory) , *REDUNDANCY in engineering , *TRANSACTION costs , *STATISTICS , *COST allocation , *SETTLEMENT costs , *SPANNING trees - Abstract
Current data allocation algorithms neglect the problems of unsatisfactory allocation results and long execution time caused by the redundancy of full comparative data and the complexity of data types. To solve these problems, a real-time allocation algorithm of full comparison data based on storm is proposed. Firstly, the phase unwrapping algorithm of minimum spanning tree is used to remove redundant data in full comparison data; then, the distributed data clustering algorithm and storm framework are used to realize the full comparison data clustering after redundancy removal. Several main factors affecting the selection of statistical information are summarized according to the clustering results. Then the communication cost of data loading and transaction processing is determined, and the trade-off between read-only transaction and update transaction cost is achieved. By judging whether the total cost of read-only transaction and update transaction is reduced or not, the replica is eliminated, and a full comparison data allocation algorithm with minimum total cost of read-only transaction and update transaction is proposed to realize real-time allocation of full-comparative data. The example analysis shows that the proposed algorithm can meet the user's needs in terms of execution time, acceleration ratio, storage efficiency and cost. Compared with the reference algorithm, the proposed algorithm has the lowest execution time, the highest acceleration ratio and the closest allocation cost to the ideal overhead. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
16. On K-means clustering-based approach for DDBSs design
- Author
-
Ali A. Amer
- Subjects
DDBS ,Data allocation ,Data replication ,Query clustering ,K-means algorithm ,Computer engineering. Computer hardware ,TK7885-7895 ,Information technology ,T58.5-58.64 ,Electronic computers. Computer science ,QA75.5-76.95 - Abstract
Abstract In Distributed Database Systems (DDBS), communication costs and response time have long been open-ended challenges. Nevertheless, when DDBS is carefully designed, the desired reduction in communication costs will be achieved. Data fragmentation (data clustering) and data allocation are on popularity as the prime strategies in constant use to design DDBS. Based on these strategies, on the other hand, several design techniques have been presented in the literature to improve DDBS performance using either empirical results or data statistics, making most of them imperfect or invalid particularly, at least, at the initial stage of DDBSs design. In this paper, thus, a heuristic k-means approach for vertical fragmentation and allocation is introduced. This approach is primarily focused on DDBS design at the initial stage. Many techniques are being joined in a step to make a promising work. A brief yet effective experimental study, on both artificially-created and real datasets, has been conducted to demonstrate the optimality of the proposed approach, comparing with its counterparts, as the obtained results has been shown encouraging.
- Published
- 2020
- Full Text
- View/download PDF
17. Data Allocation Based on Evolutionary Data Popularity Clustering
- Author
-
Vamosi, Ralf, Lassnig, Mario, Schikuta, Erich, Hutchison, David, Series Editor, Kanade, Takeo, Series Editor, Kittler, Josef, Series Editor, Kleinberg, Jon M., Series Editor, Mattern, Friedemann, Series Editor, Mitchell, John C., Series Editor, Naor, Moni, Series Editor, Pandu Rangan, C., Series Editor, Steffen, Bernhard, Series Editor, Terzopoulos, Demetri, Series Editor, Tygar, Doug, Series Editor, Weikum, Gerhard, Series Editor, Shi, Yong, editor, Fu, Haohuan, editor, Tian, Yingjie, editor, Krzhizhanovskaya, Valeria V., editor, Lees, Michael Harold, editor, Dongarra, Jack, editor, and Sloot, Peter M. A., editor
- Published
- 2018
- Full Text
- View/download PDF
18. Empirical Study of Data Allocation in Heterogeneous Memory
- Author
-
Zhao, Hui, Qiu, Meikang, Gai, Keke, Hutchison, David, Series editor, Kanade, Takeo, Series editor, Kittler, Josef, Series editor, Kleinberg, Jon M., Series editor, Mattern, Friedemann, Series editor, Mitchell, John C., Series editor, Naor, Moni, Series editor, Pandu Rangan, C., Series editor, Steffen, Bernhard, Series editor, Terzopoulos, Demetri, Series editor, Tygar, Doug, Series editor, Weikum, Gerhard, Series editor, and Qiu, Meikang, editor
- Published
- 2018
- Full Text
- View/download PDF
19. Dynamic Data Allocation and Task Scheduling on Multiprocessor Systems With NVM-Based SPM
- Author
-
Yan Wang, Kenli Li, and Keqin Li
- Subjects
Data allocation ,endurance ,execution cost ,nonvolatile memory ,wear-leveling ,Electrical engineering. Electronics. Nuclear engineering ,TK1-9971 - Abstract
Low-power and short-latency memory access is critical to the performance of chip multiprocessor (CMP) system devices, especially to bridge the performance gap between memory and CPU. Together with increased demand for low-energy consumption and high-speed memory, scratch-pad memory (SPM) has been widely adopted in multiprocessor systems. In this paper, we employ a hybrid SPM, composed of a static random-access memory and a nonvolatile memory (NVM), to replace the cache in CMP. However, there are several challenges related to the CMP that need to be addressed, including how to dynamically assign processors to application tasks and dynamically allocate data to memories. To solve these problems based on this architecture, we propose a novel dynamic data allocation and task scheduling algorithm, i.e., dynamic greedy data allocation and task scheduling (DGDATS). Experiments on DSP benchmarks demonstrate the effectiveness and efficiency of our proposed algorithms; namely, our proposed algorithm can generate a highly efficient dynamic data allocation and task scheduling approach to minimize the total execution cost and produce the least amount of write operations on NVMs. Our extensive simulation study demonstrates that our proposed algorithm exhibits an excellent performance compared with the heuristic allocation (HA) and adaptive genetic algorithm for data allocation (AGADA) algorithms. Based on the CMP systems with hybrid SPMs, DGDATS reduces the total execution cost by 22.18% and 51.37% compared with those of the HA and AGADA algorithms, respectively. Additionally, the average number of write operations on NVM is 19.82% lower than that of HA.
- Published
- 2019
- Full Text
- View/download PDF
20. Cost Reduction for Data Allocation in Heterogenous Cloud Computing Using Dynamic Programming
- Author
-
Zhao, Hui, Qiu, Meikang, Gai, Keke, Li, Jie, He, Xin, Hutchison, David, Series editor, Kanade, Takeo, Series editor, Kittler, Josef, Series editor, Kleinberg, Jon M., Series editor, Mattern, Friedemann, Series editor, Mitchell, John C., Series editor, Naor, Moni, Series editor, Pandu Rangan, C., Series editor, Steffen, Bernhard, Series editor, Terzopoulos, Demetri, Series editor, Tygar, Doug, Series editor, Weikum, Gerhard, Series editor, and Qiu, Meikang, editor
- Published
- 2017
- Full Text
- View/download PDF
21. Spatial Network Big Databases: An Introduction
- Author
-
Yang, KwangSoo, Shekhar, Shashi, Yang, KwangSoo, and Shekhar, Shashi
- Published
- 2017
- Full Text
- View/download PDF
22. Management and Analysis of Big Graph Data: Current Systems and Open Challenges
- Author
-
Junghanns, Martin, Petermann, André, Neumann, Martin, Rahm, Erhard, Zomaya, Albert Y., editor, and Sakr, Sherif, editor
- Published
- 2017
- Full Text
- View/download PDF
23. Task scheduling on heterogeneous multiprocessor systems through coherent data allocation.
- Author
-
Deng, Zexi, Shen, Hong, Cao, Dunqian, Yan, Zihan, and Huang, Huimin
- Subjects
MULTIPROCESSORS ,SCHEDULING ,ENERGY consumption ,TASKS ,DATABASES ,ASSIGNMENT problems (Programming) - Abstract
Energy consumption has become one of the main bottlenecks that limit the performance improvement of heterogeneous multiprocessor systems. In a heterogeneous distributed shared‐memory multiprocessor system (HDSMS), each processor can access all the memories, and each data can be stored in different memories. This article aims at addressing the problem of task scheduling and data allocation (TSDA) on HDSMS. To minimize the total energy consumption under a time constraint for TSDA, we propose two algorithms: the extended tree assignment for task scheduling incorporating data allocation (ETATS‐DA) and critical path task scheduling and data allocation (CPTSDA). The ETATS‐DA algorithm first utilizes the extended tree assignment to search the near optimal solution for task assignment, and then allocates data to memory based on the result of assignment. The CPTSDA algorithm considers TSDA jointly on a critical path simultaneously. Our proposed algorithms perform coherent data allocation under the consideration of best task scheduling by running two different heuristic strategies, respectively, and taking the best result as the final result. We conduct a large number of simulation experiments to test the performance of our algorithms, and the results validate the higher performance of our methods compared with the state‐of‐the‐art algorithms. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
24. Marking of Electrode Sheets in the Production of Lithium-Ion Cells as an Enabler for Tracking and Tracing.
- Author
-
Sommer, Alessandro, Leeb, Matthias, Haghi, Sajedeh, Günter, Florian J., and Reinhart, Gunther
- Abstract
The production of lithium-ion batteries is highly complex and characterized by continuous as well as discrete material flows and processes. A first step towards controlling the complexity of battery production is to create transparency through data collection. Electrodes as one of the key elements in a battery cell play a decisive role for the battery performance. The allocation of production data to electrodes enables a detailed digital twin and an individual grading system. This paper presents a concept for the marking of electrode sheets and requirements on markers as well as on marking technologies due to boundary conditions. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
25. DDNN based data allocation method for IIoT.
- Author
-
Wang, Chuting, Guo, Ruifeng, Hu, Yi, Yu, Haoyu, and Su, Wenju
- Subjects
DATABASES ,MANUFACTURING processes ,ARTIFICIAL intelligence ,MACHINE learning ,INDUSTRIAL sites ,DEEP learning - Abstract
With the complete application of artificial intelligence in the field of industrial production and manufacturing and the rapid development of edge computing, industrial processing sites often need to deploy machine learning tasks at edges and terminals. We propose a data allocation method based on Distributed Deep Neural Networks (DDNN) framework, which allocates data to edge servers or stays locally for processing. DDNN divides deep learning tasks and deploys pre-trained shallow neural networks and deep neural networks at local or edges, respectively. However, all data is processed locally, and the failure is sent to the edge server or the cloud. It will lead to excessive pressure on local terminal equipment and long-term idle edge servers, which cannot meet industrial production's real-time requirements on user privacy and time-sensitive tasks. In this paper, the complexity and inference error rate of machine learning model, the data processing speed of local equipment and edge server, and the transmission time are comprehensively considered to establish the system model. A joint optimization problem is proposed to minimize the total data processing delay. The optimal solution is derived analytically, and the optimal data allocation methhod is given. Simulation experiments are designed to verify the method's effectiveness and study the influence of key parameters on the allocation method. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
26. A data distribution model for RDF.
- Author
-
Schroeder, Rebeca, Penteado, Raqueline R. M., and Hara, Carmem S.
- Subjects
DATA distribution ,MESSAGE passing (Computer science) ,DATA modeling ,RDF (Document markup language) ,ELECTRONIC data processing ,COMBINED sewer overflows ,DATABASES - Abstract
The ever-increasing amount of RDF data made available requires data to be partitioned across multiple servers. We have witnessed some research progress made towards scaling RDF query processing based on suitable data distribution methods. In general, they work well for queries matching simple triple patterns, but they are not efficient for queries involving more complex patterns. In this paper, we present an RDF data distribution method which overcomes the shortcomings of the current approaches in order to scale RDF storage both on the volume of data and query processing. We apply a method that identifies frequent patterns accessed by queries in order to keep related data in the same partition. We deploy our reasoning on a summarized view of data in order to avoid exhaustive analysis on large datasets. As result, partitioning templates are obtained from data items in an RDF structure. In addition, we provide an approach for dynamic data insertions even if new data do not conform to the original RDF structure. Apart from the repartitioning approaches, we use an overflow repository to store data which may not follow the original schema. Our study shows that our method scales well and is effective to improve the overall performance by decreasing the amount of message passing among servers, compared to alternative data distribution approaches for RDF. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
27. SmartHeating: On the Performance and Lifetime Improvement of Self-Healing SSDs.
- Author
-
Cui, Jinhua, Liu, Junwei, Huang, Jianhang, and Yang, Laurence T.
- Subjects
- *
FLASH memory , *SOLID state drives , *RECORDS management , *HIGH temperatures , *THRESHOLD voltage - Abstract
In NAND flash memory-based solid-state drives (SSDs), during the idle time between the consecutive program/erase cycles (dwell time), the dielectric damage of flash cell can be partially repaired, also known as the self-recovery effect. As the effectiveness of the self-recovery effect can be improved under high temperature, self-healing SSDs are proven feasible to extend the flash endurance significantly. However, current self-healing SSDs perform the heating operations on all the worn-out blocks without considering the data retention requirement, and measures the lifetime of flash memory based on the worst-case self-recovery effect, leading to some unnecessary heating operations and the degraded performance. We propose SmartHeating, a smart heating scheme that exploits the dwell time variation and the write hotness variation to improve the I/O performance and the lifetime of self-healing SSDs. SmartHeating tracks the dwell time of all worn-out flash blocks, predicts their self-recovery effect and reliability, and avoids performing heating operations on the worn-out flash blocks that still have strong flash reliability. In addition, by exploiting the data hotness variation, SmartHeating only heats the worn-out flash blocks that store write-cold data, while allocating write-hot data to a small portion of worn-out flash blocks with negligible refresh overhead. The experimental results show that SmartHeating reduces the number of heating operations by 12.5% on average, boosts I/O performance of flash storage systems by 21.0%, and improves the lifetime of flash memory by $1.20\times $ compared with conventional heating scheme. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
28. ASGOP: An aggregated similarity-based greedy-oriented approach for relational DDBSs design
- Author
-
Ali A. Amer, Marghny H. Mohamed, and Khaled Al_Asri
- Subjects
Information science ,Computer science ,Vertical fragmentation ,Clustering ,Data allocation ,Data replication ,Science (General) ,Q1-390 ,Social sciences (General) ,H1-99 - Abstract
In the literature of distributed database system (DDBS), several methods sought to meet the satisfactory reduction on transmission cost (TC) and were seen substantially effective. Data Fragmentation, site clustering, and data distribution have been considered the major leading TC-mitigating influencers. Sites clustering, on one hand, aims at grouping sites appropriately according to certain similarity metrics. On the other hand, data distribution seeks to allocate the fragmented data into clusters/sites properly. The combination of these methods, however, has been shown fruitful concerning TC reduction along with network overheads. In this work, hence, a heuristic clustering-based approach for vertical fragmentation and data allocation is meticulously designed. The focus is directed on proposing an influential solution for improving relational DDBS throughputs across an aggregated similarity-based fragmentation procedure, an effective site clustering and a greedy algorithm-driven data allocation model. Moreover, the data replication is also considered so TC is further minimized. Through the delineated-below evaluation, the findings of experimental implementation have been observed to be promising.
- Published
- 2020
- Full Text
- View/download PDF
29. TTEC: Data Allocation Optimization for Morphable Scratchpad Memory in Embedded Systems
- Author
-
Linbo Long, Qing Ai, Xiaotong Cui, and Jun Liu
- Subjects
Data allocation ,scratchpad memory ,morphable NVM ,embedded systems ,Electrical engineering. Electronics. Nuclear engineering ,TK1-9971 - Abstract
Scratchpad memory (SPM) is widely utilized in many embedded systems as a software-controlled on-chip memory to replace the traditional cache. New non-volatile memory (NVM) has emerged as a promising candidate to replace SRAM in SPM, due to its significant benefits, such as low-power consumption and high performance. In particular, several representative NVMs, such as PCM, ReRAM, and STT-RAM can build multiple-level cells (MLC) to achieve even higher density. Nevertheless, this triggers off higher energy overhead and longer access latency compared with its single-level cell (SLC) counterpart. To address this issue, this paper first proposes a specific SPM with morphable NVM, in which the memory cell can be dynamically programmed to the MLC mode or SLC mode. Considering the benefits of high-density MLC and low-energy SLC, a simple and novel optimization technique, named theory of thermal expansion and contraction, is presented to minimize the energy consumption and access latency in embedded systems. The basic idea is to dynamically adjust the size configure of SLC/MLC in SPM according to the different workloads of program and allocate the optimal storage medium for each data. Therefore, an integer linear programming formulation is first built to produce an optimal SLC/MLC SPM partition and data allocation. In addition, a corresponding approximation algorithm is proposed to achieve near-optimal results in polynomial time. Finally, the experimental results show that the proposed technique can effectively improve the system performance and reduce the energy consumption.
- Published
- 2018
- Full Text
- View/download PDF
30. Latency-Sensitive Data Allocation and Workload Consolidation for Cloud Storage
- Author
-
Song Yang, Philipp Wieder, Muzzamil Aziz, Ramin Yahyapour, Xiaoming Fu, and Xu Chen
- Subjects
Cloud Storage ,data allocation ,latency ,workload consolidation ,Electrical engineering. Electronics. Nuclear engineering ,TK1-9971 - Abstract
Customers often suffer from the variability of data access time in (edge) cloud storage service, caused by network congestion, load dynamics, and so on. One efficient solution to guarantee a reliable latency-sensitive service (e.g., for industrial Internet of Things application) is to issue requests with multiple download/upload sessions which access the required data (replicas) stored in one or more servers, and use the earliest response from those sessions. In order to minimize the total storage costs, how to optimally allocate data in a minimum number of servers without violating latency guarantees remains to be a crucial issue for the cloud provider to deal with. In this paper, we study the latency-sensitive data allocation problem, the latency-sensitive data reallocation problem and the latency-sensitive workload consolidation problem for cloud storage. We model the data access time as a given distribution whose cumulative density function is known, and prove that these three problems are NP-hard. To solve them, we propose an exact integer nonlinear program (INLP) and a Tabu Search-based heuristic. The simulation results reveal that the INLP can always achieve the best performance in terms of lower number of used nodes and higher storage and throughput utilization, but this comes at the expense of much higher running time. The Tabu Search based heuristic, on the other hand, can obtain close-to-optimal performance, but in a much lower running time.
- Published
- 2018
- Full Text
- View/download PDF
31. Tracking and Tracing for Data Mining Application in the Lithium-ion Battery Production.
- Author
-
Wessel, Jacob, Turetskyy, Artem, Wojahn, Olaf, Herrmann, Christoph, and Thiede, Sebastian
- Abstract
The production of Lithium-Ion Battery (LIB) cells is characterized by the interlinking of different production processes with a manifold of intermediate products. To be able to ensure high quality and enable a traceability of different production and product characteristics (e.g. energy consumption, material), a tracking and tracing concept is required. In this paper, a practical tracking and tracing concept throughout the production of LIB cells, enabling inline data-driven applications, is introduced. As a part of this concept an intelligent tracking and tracing platform is shown, which allows the generation of a pre-clustered data sets to facilitate future data-driven applications. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
32. On K-means clustering-based approach for DDBSs design.
- Author
-
Amer, Ali A.
- Subjects
DISTRIBUTED databases ,K-means clustering ,DESIGN techniques - Abstract
In Distributed Database Systems (DDBS), communication costs and response time have long been open-ended challenges. Nevertheless, when DDBS is carefully designed, the desired reduction in communication costs will be achieved. Data fragmentation (data clustering) and data allocation are on popularity as the prime strategies in constant use to design DDBS. Based on these strategies, on the other hand, several design techniques have been presented in the literature to improve DDBS performance using either empirical results or data statistics, making most of them imperfect or invalid particularly, at least, at the initial stage of DDBSs design. In this paper, thus, a heuristic k-means approach for vertical fragmentation and allocation is introduced. This approach is primarily focused on DDBS design at the initial stage. Many techniques are being joined in a step to make a promising work. A brief yet effective experimental study, on both artificially-created and real datasets, has been conducted to demonstrate the optimality of the proposed approach, comparing with its counterparts, as the obtained results has been shown encouraging. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
33. 减轻 CMT 中乱序程度的发送端数据分配方案.
- Author
-
王振朝, 李海潇, and 侯欢欢
- Subjects
- *
BANDWIDTHS , *ALGORITHMS , *DISEASES , *DATA , *TCP/IP - Abstract
In order to reduce the effect of the packet disorder on the transmission performance at the receiver in the CMT system, this paper developed a new data allocation scheme in the transmission terminal. The scheme predicted the packet forward transmission delay based on the bandwidth, round trip transmission delay and congestion window of the path, and used it as a measure factor of the path transmission priority in the CMT system. The sender assigned packets in the queue for each path based on the path transmission priority and sender cache status, which did not cause the receiver to be out of order. Simulation results show that compared with the round-robin and ATLB algorithm, the proposed data allocation scheme can reduce the number of disorder data apparently. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
34. Delineation of Traffic Analysis Zone for Public Transportation OD Matrix Estimation Based on Socio-spatial Practices
- Author
-
Moghaddam, S. M. Hassan Mahdavi, Ameli, Mostafa, Rao, K. Ramachandra, Tiwari, Geetam, and Technische Universität Dresden
- Subjects
ddc:360 ,public transportation delineation ,travel demand ,traffic analysis zones ,data allocation ,empirical data ,Abgrenzung des öffentlichen Verkehrs, Verkehrsnachfrage, Verkehrsanalysezonen, Daten Zuweisung, empirische Daten - Abstract
This paper aims to develop and validate an efficient method for delineation of public transit analysis zones (PTTAZ), particularly for origin-destination (OD) matrix prediction for transit operation planning. Existing methods have a problem in reflecting the level of spatial precision, travel characteristics, travel demand growth, access to transit stations, and most importantly, the direction of transit routes. This study proposes a new methodology to redelineate existing traffic analysis zones (TAZ) to create PTTAZ in order to allocate travel demand to transit stops. We aim to achieve an accurate prediction of the OD matrix for public transportation (PT). The matrix should reflect the passenger accessibility in the socioeconomic and socio-spatial characterization of PTTAZ and minimize intrazonal trips. The proposed methodology transforms TAZ-based to PTTAZ-based data with sequential steps through multiple statistical methods. In short, the generation of PTTAZ establishes homogeneous sub-zones representing the relationship between passenger flow, network structure, land use, population, socio-economic characteristics, and, most importantly, existing bus transit infrastructure. To validate the proposed scheme, we implement the framework for India’s Vishakhapatnam bus network and compare the results with the household survey. The results show that the PTTAZ-based OD matrix represents a realistic scenario for PT demand.
- Published
- 2023
35. Designing Parallel Relational Data Warehouses: A Global, Comprehensive Approach
- Author
-
Benkrid, Soumia, Bellatreche, Ladjel, Cuzzocrea, Alfredo, Kacprzyk, Janusz, Series editor, Catania, Barbara, editor, Cerquitelli, Tania, editor, Chiusano, Silvia, editor, Guerrini, Giovanna, editor, Kämpf, Mirko, editor, Kemper, Alfons, editor, Novikov, Boris, editor, Palpanas, Themis, editor, Pokorný, Jaroslav, editor, and Vakali, Athena, editor
- Published
- 2014
- Full Text
- View/download PDF
36. Adaptive Distributed RDF Graph Fragmentation and Allocation based on Query Workload.
- Author
-
Peng, Peng, Zou, Lei, Chen, Lei, and Zhao, Dongyan
- Subjects
- *
DATA integrity , *RESOURCE allocation , *ELECTRONIC data processing , *QUERYING (Computer science) , *RDF (Document markup language) - Abstract
As massive volumes of Resource Description Framework (RDF) data are growing, designing a distributed RDF database system to manage them is necessary. In designing this system, it is very common to partition the RDF data into some parts, called fragments, which are then distributed. Thus, the distribution design comprises two steps: fragmentation and allocation. In this study, we explore the workload for fragmentation and allocation, which aims to reduce the communication cost during SPARQL query processing. Specifically, we adaptively maintain some frequent access patterns (FAPs) to reflect the characteristics of the workload while ensuring the data integrity and approximation ratio. Based on these frequent access patterns, we propose three fragmentation strategies, namely vertical, horizontal, and mixed fragmentation, to divide RDF graphs while meeting different types of query processing objectives. After fragmentation, we discuss how to allocate these fragments to various sites while balancing the fragments. Finally, we discuss how to process queries based on the results of fragmentation and allocation. Experiments over large RDF datasets confirm the superior performance of our proposed solutions. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
37. Cooperative Data Caching for Cloud Data Servers
- Author
-
Mingcong Yang, Kai Guo, and Yongbing Zhang
- Subjects
cloud data centers ,data allocation ,mixed integer programming ,Management information systems ,T58.6-58.62 - Abstract
Thanks to the advance of cloud computing technologies, users can access the data stored at cloud data centers at any time and from any where. However, the data centers are usually sparsely distributed over the Internet and are far away from end users. In this paper, we consider to construct a cache network by a large number of cache nodes close to the end users in order to minimize the data access delay.We firstly formulate the problem of placing the replicas of data items to cache nodes as a mixed integer programming (MIP) problem. Then, we proposed an efficient heuristic algorithm that allocates at least one replica of each data item in the cache network and attempt to allocate more data items so as to minimize the total data access cost. The simulation results show that our proposed algorithm behaves much better than a well-known LRU algorithm and the computation complexity is limited.
- Published
- 2016
- Full Text
- View/download PDF
38. Joint Power and Data Allocation in Multi-Carrier Full-Duplex Relaying Networks Operating With Finite Blocklength Codes
- Author
-
Eduard A. Jorswieck, Bo Li, Hao Jiang, Anke Schmeink, Yulin Hu, and Xiaopeng Yuan
- Subjects
Computer science ,Applied Mathematics ,Electronic engineering ,Multi carrier ,Electrical and Electronic Engineering ,Data allocation ,Joint (audio engineering) ,Computer Science Applications ,Power (physics) - Published
- 2022
- Full Text
- View/download PDF
39. A Novel Storage Management in Embedded Environment
- Author
-
Wei, Lin, Yan-yuan, Zhang, and Zhu, Min, editor
- Published
- 2011
- Full Text
- View/download PDF
40. A Data Allocation Method in Multi-processors Task Scheduling Procedure
- Author
-
Wang, Chao, Liu, Wei, Yuan, Pei-yuan, and Zheng, Dehuai, editor
- Published
- 2011
- Full Text
- View/download PDF
41. Verification of Partitioning and Allocation Techniques on Teradata DBMS
- Author
-
Bellatreche, Ladjel, Benkrid, Soumia, Ghazal, Ahmad, Crolotte, Alain, Cuzzocrea, Alfredo, Hutchison, David, Series editor, Kanade, Takeo, Series editor, Kittler, Josef, Series editor, Kleinberg, Jon M., Series editor, Mattern, Friedemann, Series editor, Mitchell, John C., Series editor, Naor, Moni, Series editor, Nierstrasz, Oscar, Series editor, Pandu Rangan, C., Series editor, Steffen, Bernhard, Series editor, Sudan, Madhu, Series editor, Terzopoulos, Demetri, Series editor, Tygar, Doug, Series editor, Vardi, Moshe Y., Series editor, Weikum, Gerhard, Series editor, Xiang, Yang, editor, Cuzzocrea, Alfredo, editor, Hobbs, Michael, editor, and Zhou, Wanlei, editor
- Published
- 2011
- Full Text
- View/download PDF
42. An optimized cost-based data allocation model for heterogeneous distributed computing systems
- Author
-
Sashi Tarun, Mithilesh Kumar Dubey, Ranbir Singh Batth, and Sukhpreet Kaur
- Subjects
Total cost ,Execution time ,General Computer Science ,Data allocation ,Computation cost ,Electrical and Electronic Engineering ,Communication cost ,Network cost - Abstract
Continuous attempts have been made to improve the flexibility and effectiveness of distributed computing systems. Extensive effort in the fields of connectivity technologies, network programs, high processing components, and storage helps to improvise results. However, concerns such as slowness in response, long execution time, and long completion time have been identified as stumbling blocks that hinder performance and require additional attention. These defects increased the total system cost and made the data allocation procedure for a geographically dispersed setup difficult. The load-based architectural model has been strengthened to improve data allocation performance. To do this, an abstract job model is employed, and a data query file containing input data is processed on a directed acyclic graph. The jobs are executed on the processing engine with the lowest execution cost, and the system's total cost is calculated. The total cost is computed by summing the costs of communication, computation, and network. The total cost of the system will be reduced using a Swarm intelligence algorithm. In heterogeneous distributed computing systems, the suggested approach attempts to reduce the system's total cost and improve data distribution. According to simulation results, the technique efficiently lowers total system cost and optimizes partitioned data allocation.
- Published
- 2022
43. : A Methodology for Effectively and Efficiently Designing Parallel Relational Data Warehouses on Heterogenous Database Clusters
- Author
-
Bellatreche, Ladjel, Cuzzocrea, Alfredo, Benkrid, Soumia, Hutchison, David, Series editor, Kanade, Takeo, Series editor, Kittler, Josef, Series editor, Kleinberg, Jon M., Series editor, Mattern, Friedemann, Series editor, Mitchell, John C., Series editor, Naor, Moni, Series editor, Nierstrasz, Oscar, Series editor, Pandu Rangan, C., Series editor, Steffen, Bernhard, Series editor, Sudan, Madhu, Series editor, Terzopoulos, Demetri, Series editor, Tygar, Doug, Series editor, Vardi, Moshe Y., Series editor, Weikum, Gerhard, Series editor, Bach Pedersen, Torben, editor, Mohania, Mukesh K., editor, and Tjoa, A Min, editor
- Published
- 2010
- Full Text
- View/download PDF
44. Query Optimization over Parallel Relational Data Warehouses in Distributed Environments by Simultaneous Fragmentation and Allocation
- Author
-
Bellatreche, Ladjel, Cuzzocrea, Alfredo, Benkrid, Soumia, Hutchison, David, editor, Kanade, Takeo, editor, Kittler, Josef, editor, Kleinberg, Jon M., editor, Mattern, Friedemann, editor, Mitchell, John C., editor, Naor, Moni, editor, Nierstrasz, Oscar, editor, Pandu Rangan, C., editor, Steffen, Bernhard, editor, Sudan, Madhu, editor, Terzopoulos, Demetri, editor, Tygar, Doug, editor, Vardi, Moshe Y., editor, Weikum, Gerhard, editor, Hsu, Ching-Hsien, editor, Yang, Laurence T., editor, Park, Jong Hyuk, editor, and Yeo, Sang-Soo, editor
- Published
- 2010
- Full Text
- View/download PDF
45. A Distributed Intrusion Detection Model via Nondestructive Partitioning and Balanced Allocation for Big Data.
- Author
-
Xiaonian Wu, Chuyun Zhang, Runlian Zhang, Yujue Wang, and Jinhua Cui
- Subjects
BIG data ,DETECTORS ,DATA integrity ,COMPUTER simulation ,PARALLEL algorithms - Abstract
There are two key issues in distributed intrusion detection system, that is, maintaining load balance of system and protecting data integrity. To address these issues, this paper proposes a new distributed intrusion detection model for big data based on nondestructive partitioning and balanced allocation. A data allocation strategy based on capacity and workload is introduced to achieve local load balance, and a dynamic load adjustment strategy is adopted to maintain global load balance of cluster. Moreover, data integrity is protected by using session reassemble and session partitioning. The simulation results show that the new model enjoys favorable advantages such as good load balance, higher detection rate and detection efficiency. [ABSTRACT FROM AUTHOR]
- Published
- 2018
- Full Text
- View/download PDF
46. Data allocation optimization for query processing in graph databases using Lucene.
- Author
-
Mathew, Anita Brigit
- Subjects
- *
NONRELATIONAL databases , *SOCIAL networks , *WIRELESS sensor nodes , *METAHEURISTIC algorithms , *GRAPHIC methods - Abstract
Abstract Methodological handling of queries is a crucial requirement in social networks connected to a graph NoSQL database that incorporates massive amounts of data. The massive data need to be partitioned across numerous nodes so that the queries when executed can be retrieved from a parallel structure. A novel storage mechanism for effective query processing must to be established in graph databases for minimizing time overhead. This paper proposes a metaheuristic algorithm for partitioning of graph database across nodes by placement of all related information on same or adjacent nodes. The graph database allocation problem is proved to be NP-Hard. A metaheuristic algorithm comprising of Best Fit Decreasing with Ant Colony Optimization is proposed for data allocation in a distributed architecture of graph NoSQL databases. Lucene index is applied on proposed allocation for faster query processing. The proposed algorithm with Lucene is evaluated based on simulation results obtained from different heuristics available in literature. [ABSTRACT FROM AUTHOR]
- Published
- 2018
- Full Text
- View/download PDF
47. A Novel Bitrate Adaptation Method for Heterogeneous Wireless Body Area Networks.
- Author
-
Cwalina, Krzysztof K., Ambroziak, Slawomir J., Rajchowski, Piotr, Sadowski, Jaroslaw, and Stefanski, Jacek
- Subjects
BODY area networks ,BIT rate - Abstract
In the article, a novel bitrate adaptation method for data streams allocation in heterogeneous Wireless Body Area Networks (WBANs) is presented. The efficiency of the proposed algorithm was compared with other known algorithms of data stream allocation using computer simulation. A dedicated simulator has been developed using results of measurements in the real environment. The usage of the proposed adaptive data streams allocation method by transmission rate adaptation based on radio channel parameters can increase the efficiency of resources’ usage in a heterogeneous WBANs, in relation to fixed bitrates transmissions and the use of well-known algorithms. This increase of efficiency has been shown regardless of the mobile node placement on the human body. [ABSTRACT FROM AUTHOR]
- Published
- 2018
- Full Text
- View/download PDF
48. Efficient query retrieval in Neo4jHA using metaheuristic social data allocation scheme.
- Author
-
Mathew, Anita Brigit, Madhu Kumar, S.D., Krishnan, K. Murali, and Salam, Sameera M.
- Subjects
- *
BIG data , *QUERY (Information retrieval system) , *ONLINE data processing , *ANT algorithms , *LINEAR programming - Abstract
Large amount of data from social networks needs to be shared, distributed and indexed in a parallel structure to be able to make best use of the data. Neo4j High Availability (Neo4jHA) is a popular open-source graph database used for query handling on large social data. This paper analyses how storing and indexing of social data across machines can be carried out by placing all the related information on the same or adjacent machines, with replication. The social graph data allocation problem referred to as Neo4jHA allocation has proved to be NP-Hard in this paper. An integration of Best Fit Decreasing algorithm with Ant Colony Optimization based metaheuristics is proposed for data allocation in a distributed architecture of Neo4jHA. The evaluation of the algorithm is carried out by simulation. The query processing efficiency is compared with other heuristic algorithms like First Fit, Best Fit, First Fit Decreasing and Best Fit Decreasing found in literature. A Skip List index was constructed on Neo4jHA of every machine after the implementation of the proposed allocation strategy for enhancing the query processing efficiency. The results illustrate how the proposed algorithm outperforms other data allocation approaches in query execution with and without an index. [ABSTRACT FROM AUTHOR]
- Published
- 2018
- Full Text
- View/download PDF
49. On Lack of Robustness in Hydrological Model Development Due to Absence of Guidelines for Selecting Calibration and Evaluation Data: Demonstration for Data‐Driven Models.
- Author
-
Zheng, Feifei, Maier, Holger R., Wu, Wenyan, Dandy, Graeme C., Gupta, Hoshin V., and Zhang, Tuqiao
- Subjects
FLOOD risk ,ROBUST statistics ,HYDROLOGY - Abstract
Abstract: Hydrological models are used for a wide variety of engineering purposes, including streamflow forecasting and flood‐risk estimation. To develop such models, it is common to allocate the available data to calibration and evaluation data subsets. Surprisingly, the issue of how this allocation can affect model evaluation performance has been largely ignored in the research literature. This paper discusses the evaluation performance bias that can arise from how available data are allocated to calibration and evaluation subsets. As a first step to assessing this issue in a statistically rigorous fashion, we present a comprehensive investigation of the influence of data allocation on the development of data‐driven artificial neural network (ANN) models of streamflow. Four well‐known formal data splitting methods are applied to 754 catchments from Australia and the U.S. to develop 902,483 ANN models. Results clearly show that the choice of the method used for data allocation has a significant impact on model performance, particularly for runoff data that are more highly skewed, highlighting the importance of considering the impact of data splitting when developing hydrological models. The statistical behavior of the data splitting methods investigated is discussed and guidance is offered on the selection of the most appropriate data splitting methods to achieve representative evaluation performance for streamflow data with different statistical properties. Although our results are obtained for data‐driven models, they highlight the fact that this issue is likely to have a significant impact on all types of hydrological models, especially conceptual rainfall‐runoff models. [ABSTRACT FROM AUTHOR]
- Published
- 2018
- Full Text
- View/download PDF
50. Fuzzy Neighborhood Allocation (FNA): A Fuzzy Approach to Improve Near Neighborhood Allocation in DDB
- Author
-
Basseda, Reza, Rahgozar, Maseud, Lucas, Caro, Sarbazi-Azad, Hamid, editor, Parhami, Behrooz, editor, Miremadi, Seyed-Ghassem, editor, and Hessabi, Shaahin, editor
- Published
- 2009
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.