31 results
Search Results
2. Virtual Resource Allocation Based on Link Interference in Cayley Wireless Data Centers.
- Author
-
Luo, Juan, Guo, Yaling, Fu, Shan, Li, Keqin, and He, Wenfeng
- Subjects
INTERFERENCE (Telecommunication) ,DATA libraries ,RESOURCE allocation ,VIRTUAL networks ,MATHEMATICAL mappings ,COMPUTER algorithms - Abstract
Cayley data centers are well known patterns of completely wireless data centers (WDCs). However, low link reliability and link interference will affect the construction of virtual networks. This paper proposes a virtual resource mapping algorithm on the basis of Cayley structures. First, we analyze the characteristics of Cayley WDCs and model networks in WDCs, where a virtual network is modeled as a traditional undirected graph, while a physical topology is modeled as a directed graph. Second, we propose a virtual resource mapping and coloring algorithm based on link interference called VRMCA-LI. We build a connection interference matrix for each node and use a coloring method to avoid interference. VRMCA-LI uses the same color for the nodes that are within the transmitting angle of a sending node and whose signal-to-noise ratio is less than a threshold. These nodes with the same color cannot be allocated to virtual nodes at the same time. The allocation of nodes and links are concurrent, which performs dynamic adjustment to save mapping time. Third, our experimental results show that VRMCA-LI outperforms WVNEA-LR and PG-VNE in terms of mapping time of virtual nodes, acceptance rate of virtual networks, and average node utilization rate. [ABSTRACT FROM AUTHOR]
- Published
- 2015
- Full Text
- View/download PDF
3. Analysis of Backward Congestion Notification with Delay for Enhanced Ethernet Networks.
- Author
-
Jiang, Wanchun, Ren, Fengyuan, Wu, Yongwei, Lin, Chuang, and Stojmenovic, Ivan
- Subjects
INTERNET traffic ,ETHERNET ,COMPUTER networks ,PERFORMANCE evaluation ,DATA libraries ,MATHEMATICAL models - Abstract
At present, companies and standards organizations are enhancing Ethernet as the unified switch fabric for all of the TCP/IP traffic, the storage traffic and the high performance computing traffic in data centers. Backward congestion notification (BCN) is the basic mechanism for the end-to-end congestion management enhancement of Ethernet. To fulfill the special requirements of the unified switch fabric, i.e., losslessness and low transmission delay, BCN should hold the buffer occupancy around a target point tightly. Thus, the stability of the control loop and the buffer size are critical to BCN. Currently, the impacts of delay on the performance of BCN are unidentified. When the speed of Ethernet increases to 40 Gbps or 100 Gbps in the near future, the number of on-the-fly packets becomes the same order with the buffer size of switch. Accordingly, the impacts of delay will become significant. In this paper, we analyze BCN, paying special attention on the delay. We model the BCN system with a set of segmented delayed differential equations, and then deduce sufficient condition for the uniformly asymptotic stability of BCN. Subsequently, the bounds of buffer occupancy are estimated, which provides direct guidelines on setting buffer size. Finally, numerical analysis and experiments on the NetFPGA platform verify our theoretical analysis. [ABSTRACT FROM AUTHOR]
- Published
- 2014
- Full Text
- View/download PDF
4. Disaggregation and Sharing of I/O Devices in Cloud Data Centers.
- Author
-
Suzuki, Jun, Hidaka, Yoichi, Higuchi, Junichi, Hayashi, Yuki, Kan, Masaki, and Yoshikawa, Takashi
- Subjects
INFORMATION sharing ,INPUT-output analysis ,CLOUD computing ,DATA libraries ,GRAPHICS processing units ,DATA transmission systems - Abstract
Input/output (I/O) devices such as a graphics processing unit and a solid-state drive are inserted into I/O slots of a host in data center platforms. With this sort of configuration the I/O devices are used exclusively by the host with resultant inefficient resource usage. In addition, the maximum number of I/O devices that can be assigned to each host is limited by the number of its I/O slots. This paper proposes a method to solve the resource usage inefficiency and assignment limitation by disaggregating I/O devices from hosts. An I/O device accommodated in a device pool is interconnected with multiple hosts by a standard Ethernet and can be flexibly assigned to one of them. Introducing assignment flexibility of I/O devices among hosts improves the resource usage efficiency and enables as many I/O devices as desired being assigned to each host. The authors further propose a method that enables an I/O device to be simultaneously shared among multiple hosts and multiple I/O packets to be aggregated into an Ethernet frame to suppress the performance overhead of the device disaggregation. Evaluation results showed that I/O devices that are compliant to PCI Express are shared with the overhead up to 27 percent. [ABSTRACT FROM AUTHOR]
- Published
- 2016
- Full Text
- View/download PDF
5. Online Energy Estimation of Relational Operations in Database Systems.
- Author
-
Xu, Zichen, Tu, Yi-Cheng, and Wang, Xiaorui
- Subjects
DATABASE management ,DATA libraries ,ENERGY consumption ,SEARCH algorithms ,CENTRAL processing units - Abstract
Data centers are well known to consume a large amount of energy. As databases are one of the major applications in a data center, building energy-aware database systems has become an active research topic recently. The quantification of the energy cost of database systems is an important task in design. In this paper, we report our recent efforts on this issue, with a focus on the energy cost estimation of query plans during query optimization. We start from building a series of physical models for energy estimation of individual relational operators based on their resource consumption patterns. As the execution of a query plan is a combination of multiple relational operators, we use the physical models as a basis for a comprehensive energy model for the entire query. To address the challenge of maintaining accuracy under system and workload dynamics, we develop an online scheme that dynamically adjusts model parameters based on statistical signal modeling. Our models are implemented in a real database management system and evaluated on a physical test bed. The results show that our solution achieves a high accuracy (worst-case error 13.7 percent) despite noises. Our models also help identify query plans with significantly higher energy efficiency. [ABSTRACT FROM PUBLISHER]
- Published
- 2015
- Full Text
- View/download PDF
6. On the Network Power Effectiveness of Data Center Architectures.
- Author
-
Shang, Yunfei, Li, Dan, Zhu, Jing, and Xu, Mingwei
- Subjects
COMPUTER networks ,COMPUTER architecture ,DATA libraries ,CLOUD computing ,BANDWIDTHS - Abstract
Cloud computing not only requires high-capacity data center networks to accelerate bandwidth-hungry computations, but also causes considerable power expenses to cloud providers. In recent years many advanced data center network architectures have been proposed to increase the network throughput, such as Fat-Tree
[1] and BCube[2] , but little attention has been paid to the power efficiency of these network architectures. This paper makes the first comprehensive comparison study for typical data center networks with regard to their Network Power Effectiveness(NPE), which indicates the end-to-end bps per watt in data transmission and reflects the tradeoff between power consumption and network throughput. We take switches, server NICs and server CPU cores into account when evaluating the network power consumption. We measure NPE under both regular routing and power-aware routing, and investigate the impacts of topology size, traffic load, throughput threshold in power-aware routing, network power parameter as well as traffic pattern. The results show that in most cases Flattened Butterfly possesses the highest NPE among the architectures under study, and server-centric architectures usually have higher NPEs than Fat-Tree and VL2 architectures. In addition, the sleep-on-idle technique and power-aware routing can significantly improve the NPEs for all the data center architectures, especially when the traffic load is low. We believe that the results are useful for cloud providers, when they design/upgrade data center networks or employ network power management. [ABSTRACT FROM PUBLISHER]- Published
- 2015
- Full Text
- View/download PDF
7. Power Budgeting Techniques for Data Centers.
- Author
-
Zhan, Xin and Reda, Sherief
- Subjects
CLOUD computing ,DATA libraries ,DATA science ,DYNAMIC programming ,ALGORITHMS ,SERVER farms (Computer network management) - Abstract
The development of cloud computing and data science result in rapid increases of number and scale of data centers. Because of cost and sustainability concerns, energy efficiency has been a major goal for data center architects. Focusing on reducing the cooling power and making full use of available computing power, power budgeting is an increasingly important requirement for data center operations. In this paper, we present a framework of power budgeting, considering both computing power and cooling power, in data centers to maximize the system normalized performance (SNP) of the entire center under a total power budget. Maximizing the SNP for a given power budget is equivalent to maximizing the energy efficiency. We propose a method to partition the total power budget among the cooling and computing infrastructure in a self-consistent way, where the cooling power is sufficient to extract the heat of the computing power. Intertwinedly, we devise an optimal computing power budgeting technique based on dynamic programming algorithm to determine the optimal power caps for the individual servers such that the available power could be efficiently translated to performance improvements. The optimal computing budgeting technique leverages a proposed online throughput predictor based on performance counter measurements to estimate the change in throughput of heterogeneous workloads as a function of allocated server power caps. We demonstrate that our proposed power budgeting method outperforms previous methods by 3-4 percent in terms of SNP using our data center simulation environment. While maintaining the improvement of SNP, our method improve fairness at best by 57 percent. We also evaluate the performance of our method in power saving scenario and dynamic power budgeting case. [ABSTRACT FROM AUTHOR]
- Published
- 2015
- Full Text
- View/download PDF
8. Cost-Aware Cooperative Resource Provisioning for Heterogeneous Workloads in Data Centers.
- Author
-
Zhan, Jianfeng, Wang, Lei, Li, Xiaona, Shi, Weisong, Weng, Chuliang, Zhang, Wenyao, and Zang, Xiutao
- Subjects
WORKLOAD of computer networks ,DATA libraries ,COST analysis ,CLOUD computing ,INTERNET servers ,MULTIPLEXING ,SEARCH engines - Abstract
Recent cost analysis shows that the server cost still dominates the total cost of high-scale data centers or cloud systems. In this paper, we argue for a new twist on the classical resource provisioning problem: heterogeneous workloads are a fact of life in large-scale data centers, and current resource provisioning solutions do not act upon this heterogeneity. Our contributions are threefold: first, we propose a cooperative resource provisioning solution, and take advantage of differences of heterogeneous workloads so as to decrease their peak resources consumption under competitive conditions; second, for four typical heterogeneous workloads: parallel batch jobs, web servers, search engines, and MapReduce jobs, we build an agile system PhoenixCloud that enables cooperative resource provisioning; and third, we perform a comprehensive evaluation for both real and synthetic workload traces. Our experiments show that our solution could save the server cost aggressively with respect to the noncooperative solutions that are widely used in state-of-the-practice hosting data centers or cloud systems: for example, EC2, which leverages the statistical multiplexing technique, or RightScale, which roughly implements the elastic resource provisioning technique proposed in related state-of-the-art work. [ABSTRACT FROM AUTHOR]
- Published
- 2013
- Full Text
- View/download PDF
9. Cost-Aware Streaming Workflow Allocation on Geo-Distributed Data Centers.
- Author
-
Chen, Wuhui, Paik, Incheon, and Li, Zhenni
- Subjects
WORKFLOW software ,DATA libraries ,VIRTUAL machine systems ,CLOUD computing ,BIG data ,ELECTRONIC data processing - Abstract
The virtual machine (VM) allocation problem in cloud computing has been widely studied in recent years, and many algorithms have been proposed in the literature. Most of them have been successfully applied to batch processing models such as MapReduce; however, none of them can be applied to streaming workflow well because of the following weaknesses: 1) failure to capture the characteristics of tasks in streaming workflow for the short life cycle of data streams; 2) most algorithms are based on the assumptions that the price of VMs and traffic among data centers (DCs) are static and fixed. In this paper, we propose a streaming workflow allocation algorithm that takes into consideration the characteristics of streaming work and the price diversity among geo-distributed DCs, to further achieve the goal of cost minimization for streaming big data processing. First, we construct an extended streaming workflow graph (ESWG) based on the task semantics of streaming workflow and the price diversity of geo-distributed DCs, and the streaming workflow allocation problem is formulated into mixed integer linear programming based on the ESWG. Second, we propose two heuristic algorithms to reduce the computational space based on task combination and DC combination in order to meet the strict latency requirement. Finally, our experimental results demonstrate significant performance gains with lower total cost and execution time. [ABSTRACT FROM PUBLISHER]
- Published
- 2017
- Full Text
- View/download PDF
10. NIPD: Non-Intrusive Power Disaggregation in Legacy Datacenters.
- Author
-
Tang, Guoming, Jiang, Weixiang, Xu, Zhifeng, Liu, Fangming, and Wu, Kui
- Subjects
ELECTRIC power failures ,DATA libraries ,CLIENT/SERVER computing ,ELECTRIC power conservation ,NONLINEAR statistical models ,FEATURE extraction - Abstract
Fine-grained power monitoring, which refers to power monitoring at the server level, is critical to the efficient operation and energy saving of datacenters. Fined-grained power monitoring, however, is extremely challenging in legacy datacenters that host server systems not equipped with power monitoring sensors. Installing power monitoring hardware at the server level not only incurs high costs but also complicates the maintenance of high-density server clusters and enclosures. In this paper, we present a zero-cost, purely software-based solution to this challenging problem. We use a novel technique of non-intrusive power disaggregation (NIPD) that establishes power mapping functions (PMFs) between the states of servers and their power consumption, and infer the power consumption of each server with the aggregated power of the entire datacenter. The PMFs that we have developed can support both linear and nonlinear power models via the state feature transformation. To reduce the training overhead, we further develop adaptive PMFs update strategies and ensure that the training data and state features are appropriately selected. We implement and evaluate NIPD over a real-world datacenter with $326$
nodes. The results show that our solution can provide high precision power estimation at both rack level and server level. In specific, with PMFs including only two nonlinear terms, our power estimation i) at rack level has mean relative error of $2.18$ percent corresponding to the idle and peak power, respectively. [ABSTRACT FROM AUTHOR]- Published
- 2017
- Full Text
- View/download PDF
11. On Soft Error Reliability of Virtualization Infrastructure.
- Author
-
Xu, Xin and Huang, H. Howie
- Subjects
SOFT errors ,RELIABILITY in engineering ,VIRTUAL machine systems ,DATA libraries ,CLOUD computing - Abstract
Hardware errors are no longer exceptions in modern cloud data centers. Although virtualization provides software failure isolation among different virtual machines (VM), the virtualization infrastructure including the hypervisor and privileged VMs remains vulnerable to hardware errors. What makes matters worse is that such errors are unlikely bounded by the virtualization boundary and may lead to loss of work in multiple guest VMs due to unexpected and/or mishandled failures. To understand reliability implication of hardware errors in virtualized systems, in this paper we develop a simulation-based framework that enables a comprehensive fault injection study on the hypervisor with a wide range of configurations. Our analysis shows that, in current systems, many hardware errors can propagate through various paths for an extended time before an observed failure (e.g., whole system crash). We further discuss the challenges of designing error tolerance techniques for the hypervisor. [ABSTRACT FROM PUBLISHER]
- Published
- 2016
- Full Text
- View/download PDF
12. TE-Shave: Reducing Data Center Capital and Operating Expenses with Thermal Energy Storage.
- Author
-
Zheng, Wenli, Ma, Kai, and Wang, Xiaorui
- Subjects
HEAT storage ,ENERGY conservation ,DATA libraries ,CAPITAL investments ,OPERATING costs ,ELECTRIC batteries - Abstract
Power shaving has recently been proposed to dynamically shave the power peaks of a data center with energy storage devices (ESD), such that more servers can be safely hosted. In addition to the reduction of capital investment (cap-ex), power shaving also helps cut the electricity bills (op-ex) of a data center by reducing the high utility tariffs related to peak power. However, existing work on power shaving focuses exclusively on electrical ESDs (e.g., UPS batteries) to shave the server-side power demand. In this paper, we propose TE-Shave, a generalized power shaving framework that exploits both UPS batteries and a new knob, thermal energy storage (TES) tanks equipped in many data centers. Specifically, TE-Shave utilizes stored cold water or ice to manipulate the cooling power, which accounts for 30-40 percent of the total power cost of a data center. Our extensive evaluation with real-world workload traces shows that TE-Shave saves cap-ex and op-ex up to $2,668/day and $825/day, respectively, for a data center with 17,920 servers. Even for future data centers that are projected to have more efficient cooling and thus a smaller portion of cooling power, e.g., a quarter of today’s level, TE-Shave still leads to 28 percent more savings than existing work that focuses only on the server-side power. TE-Shave is also coordinated with traditional TES solutions for further reduced op-ex, and integrated with processor throttling to cap the power draw (i.e., power capping). Our hardware testbed results show that TE-Shave can improve the system performance up to 23 percent. [ABSTRACT FROM PUBLISHER]
- Published
- 2015
- Full Text
- View/download PDF
13. EXR: Greening Data Center Network with Software Defined Exclusive Routing.
- Author
-
Li, Dan, Shang, Yunfei, He, Wu, and Chen, Congjie
- Subjects
DATA libraries ,ROUTING (Computer network management) ,SERVER farms (Computer network management) ,ECOLOGICAL impact ,BANDWIDTHS ,COMPUTER simulation - Abstract
The explosive expansion of data center sizes aggravates the power consumption and carbon footprint, which has restricted the sustainable growth of cloud services and seriously troubled data center operators. In recent years, plenty of advanced data center network architectures have been proposed. They usually employ richly-connected topologies and multi-path routing to provide high network capacity. Unfortunately, they also undergo inefficient network energy usage during the traffic valley time. To address the problem, many energy-aware flow scheduling algorithms are proposed recently, primarily considering how to aggregate traffic by flexibly choosing the routing paths, with flows fairly sharing the link bandwidths. In this paper, we leverage software defined network (SDN) technique and explore a new solution to energy-aware flow scheduling, i.e., scheduling flows in the time dimension and using exclusive routing (EXR) for each flow, i.e., a flow always exclusively utilizes the links of its routing path. The key insight is that exclusive occupation of link resources usually results in higher link utilization in high-radix data center networks, since each flow does not need to compete for the link bandwidths with others. When scheduling the flows, EXR leaves flexibility to operators to define the priorities of flows, e.g., based on flow size, flow deadline, etc. Extensive simulations and testbed experiments both show that EXR can effectively save network energy compared with the regular fair-sharing routing (FSR), and significantly reduce the average flow completion time if assigning higher scheduling priorities to smaller flows. [ABSTRACT FROM PUBLISHER]
- Published
- 2015
- Full Text
- View/download PDF
14. Proactive Data Migration for Improved Storage Availability in Large-Scale Data Centers.
- Author
-
Wu, Suzhen, Jiang, Hong, and Mao, Bo
- Subjects
DATA libraries ,BACK up systems ,SERVER farms (Computer network management) ,INFORMATION retrieval ,MATHEMATICAL optimization ,ARTIFICIAL intelligence - Abstract
In face of high partial and complete disk failure rates and untimely system crashes, the executions of low-priority background tasks become increasingly frequent in large-scale data centers. However, the existing algorithms are all reactive optimizations and only exploit the temporal locality of workloads to reduce the user I/O requests during the low-priority background tasks. To address the problem, this paper proposes Intelligent Data Outsourcing (IDO), a zone-based and proactive data migration optimization, to significantly improve the efficiency of the low-priority background tasks. The main idea of IDO is to proactively identify the hot data zones of RAID-structured storage systems in the normal operational state. By leveraging the prediction tools to identify the upcoming events, IDO proactively migrates the data blocks belonging to the hot data zones on the degraded device to a surrogate RAID set in the large-scale data centers. Upon a disk failure or crash reboot, most user I/O requests addressed to the degraded RAID set can be serviced directly by the surrogate RAID set rather than the much slower degraded RAID set. Consequently, the performance of the background tasks and user I/O performance during the background tasks are improved simultaneously. Our lightweight prototype implementation of IDO and extensive trace-driven experiments on two case studies demonstrate that, compared with the existing state-of-the-art approaches, IDO effectively improves the performance of the low-priority background tasks. Moreover, IDO is portable and can be easily incorporated into any existing algorithms for RAID-structured storage systems. [ABSTRACT FROM PUBLISHER]
- Published
- 2015
- Full Text
- View/download PDF
15. Exploring Server Redundancy in Nonblocking Multicast Data Center Networks.
- Author
-
Guo, Zhiyang and Yang, Yuanyuan
- Subjects
CLIENT/SERVER computing ,MULTICASTING (Computer networks) ,DATA libraries ,COMPUTER networks ,TOPOLOGY - Abstract
Clos networks and their variations such as folded-Clos networks (fat-trees) have been widely adopted as network topologies in data center networks. Since multicast is an essential communication pattern in many cloud services, nonblocking multicast communication can ensure the high performance of such services. However, nonblocking multicast Clos networks are costly due to the large number of middle stage switches required. On the other hand, server redundancy is ubiquitous in today’s data centers to provide high availability of services. In this paper, we explore such server redundancy in data centers to reduce the cost of nonblocking multicast Clos data center networks (DCNs). To facilitate our analysis, we first consider an ideal fault-free data center with no server failure. We give an algorithm to assign active servers evenly among input stage switches in a multicast Clos DCN where each server has one or more redundant backups depending on the availability requirements of services they provide. We show that the sufficient nonblocking condition on the number of middle stage switches for a multicast Clos DCN can be significantly reduced by exploring server redundancy. Then, to complete our analysis, we consider a practical faulty data center, where one or more active servers may fail at any time. We give a strategy to re-balance the active servers among input stage switches after server failures so that the same nonblocking condition still holds. Finally, we provide a multicast routing algorithm with linear time complexity to configure multicast connections in Clos DCNs. [ABSTRACT FROM AUTHOR]
- Published
- 2015
- Full Text
- View/download PDF
16. PowerTracer: Tracing Requests in Multi-Tier Services to Reduce Energy Inefficiency.
- Author
-
Lu, Gang, Zhan, Jianfeng, Wang, Haining, Yuan, Lin, Gao, Yunwei, Weng, Chuliang, and Qi, Yong
- Subjects
ENERGY consumption ,DATA libraries ,ELECTRIC potential ,COMPUTER architecture ,PULSE frequency modulation ,ENERGY conservation - Abstract
As energy has become one of the key operating costs in running a data center and power waste commonly exists, it is essential to reduce energy inefficiency inside data centers. In this paper, we develop an innovative framework, called PowerTracer, for diagnosing energy inefficiency and saving power. Inside the framework, we first present a resource tracing method based on request tracing in multi-tier services of black boxes. Then, we propose a generalized methodology of applying a request tracing approach for energy inefficiency diagnosis and power saving in multi-tier service systems. With insights into service performance and resource consumption of individual requests, we develop (1) a bottleneck diagnosis tool that pinpoints the root causes of energy inefficiency, and (2) a power saving method that enables dynamic voltage and frequency scaling (DVFS) with online request tracing. We implement a prototype of PowerTracer, and conduct extensive experiments to validate its effectiveness. Our tool analyzes several state-of-the-practice and state-of-the-art DVFS control policies and uncovers existing energy inefficiencies. Meanwhile, the experimental results demonstrate that PowerTracer outperforms its peers in power saving. [ABSTRACT FROM PUBLISHER]
- Published
- 2015
- Full Text
- View/download PDF
17. On Nonblocking Multicast Fat-Tree Data Center Networks with Server Redundancy.
- Author
-
Guo, Zhiyang and Yang, Yuanyuan
- Subjects
DATA libraries ,SERVER farms (Computer network management) ,PERFORMANCE evaluation ,UBIQUITOUS computing ,ROUTING (Computer network management) ,COMPUTATIONAL complexity - Abstract
Fat-tree networks have been widely adopted as network topologies in data center networks (DCNs). However, it is costly for fat-tree DCNs to support nonblocking multicast communication, due to the large number of core switches required. Since multicast is an essential communication pattern in many cloud services and nonblocking multicast communication can ensure the high performance of such services, reducing the cost of nonblocking multicast fat-tree DCNs is very important. On the other hand, server redundancy is ubiquitous in today’s data centers to provide high availability of services. In this paper, we explore server redundancy in data centers to reduce the cost of nonblocking multicast fat-tree data center networks (DCNs). First, we present a multirate network model that accurately describes the communication environment of the fat-tree DCNs. We then show that the sufficient condition on the number of core switches required for nonblocking multicast communication under the multirate model can be significantly reduced when the fat-tree DCNs are 2-redundant, i.e., each server in the data center has exactly one redundant backup. We also study the general redundant fat-tree DCNs where servers may have different numbers of redundant backups depending on the availability requirements of services they provide, and show that a higher redundancy level further reduces the cost of nonblocking multicast fat-tree DCNs. Then, to complete our analysis, we consider a practical faulty data center, where one or more active servers may fail at any time. We give a strategy to re-balance the active servers among edge switches after server failures so that the same nonblocking condition still holds. Finally, we give a multicast routing algorithm with linear time complexity to configure multicast connections in fat-tree DCNs. [ABSTRACT FROM PUBLISHER]
- Published
- 2015
- Full Text
- View/download PDF
18. Efficient Server Provisioning and Offloading Policies for Internet Data Centers with Dynamic Load-Demand.
- Author
-
Xu, Dan, Liu, Xin, and Fan, Bin
- Subjects
INTERNET traffic ,CLIENT/SERVER computing ,DATA libraries ,ENERGY conservation ,QUALITY of service ,PROBABILITY theory - Abstract
In data centers, traffic demand varies in both large and small time scales. A data center with dynamic traffic often needs to over-provision active servers to meet the peak demand, which incurs significant energy cost. In this paper, our goal is to reduce energy cost of a set of distributed Internet data centers (IDCs) while maintaining the quality of service of the dynamic traffic. In particular, we consider the outage probability as the QoS metric, where outage is defined as service demand exceeding the capacity. We require the outage probability at each IDC to be smaller than a predefined threshold. Our goal is thus to minimize total energy cost over all IDCs, subject to the outage probability constraint. We achieve the goal by dynamically adjusting server capacity and performing load shifting in different time scales. We propose three different load-shifting and joint capacity allocation schemes with different complexity and performance. Our schemes leverage both stochastic multiplexing gain and electricity-price diversity. Thus, improving over prior work, our schemes reduce energy consumption/cost even when all IDCs have the same electricity price. We use both simulated load traces and real traffic traces to evaluate the performance of the proposed schemes. Results show that our proposed schemes are efficient in reducing energy cost, and robust in QoS provisioning. [ABSTRACT FROM AUTHOR]
- Published
- 2015
- Full Text
- View/download PDF
19. Power and Thermal-Aware Workload Allocation in Heterogeneous Data Centers.
- Author
-
Al-Qawasmeh, Abdulla M., Pasricha, Sudeep, Maciejewski, Anthony A., and Siegel, Howard Jay
- Subjects
HETEROGENEOUS computing ,DATA libraries ,ENERGY consumption ,MATHEMATICAL optimization ,SIMULATION methods & models - Abstract
Many of today’s data centers experience physical limitations on the power needed to run the data center. The first problem that we study is maximizing the performance (quantified by the reward collected for completing tasks by their individual deadlines) of a data center that is subject to total power consumption (of compute nodes and CRAC units) and thermal constraints. The second problem that we study is how to minimize the power consumption in a data center while guaranteeing that the overall performance does not drop below a specified threshold. For both problems, we develop novel optimization techniques for assigning the performance states of cores at the data center level to optimize the operation of the data center. The resource allocation (assignment) techniques in this paper are thermal aware as they consider effects of performance state assignments on temperature and power consumption by the CRAC units. Our simulation studies show that in some cases our assignment technique achieves about 17% average improvement in the reward collected, and about 9% reduction in power consumption compared to an assignment technique that only considers putting a core in the performance state with the highest performance or turning the core off. [ABSTRACT FROM AUTHOR]
- Published
- 2015
- Full Text
- View/download PDF
20. From the Cloud to the Atmosphere: Running MapReduce across Data Centers.
- Author
-
Jayalath, Chamikara, Stephen, Julian, and Eugster, Patrick
- Subjects
BIG data ,DATA mining ,DATA libraries ,SERVER farms (Computer network management) ,ECONOMIC change ,CLOUD computing - Abstract
Efficiently analyzing big data is a major issue in our current era. Examples of analysis tasks include identification or detection of global weather patterns, economic changes, social phenomena, or epidemics. The cloud computing paradigm along with software tools such as implementations of the popular MapReduce framework offer a response to the problem by distributing computations among large sets of nodes. In many scenarios, input data are, however, geographically distributed (geodistributed) across data centers, and straightforwardly moving all data to a single data center before processing it can be prohibitively expensive. Above-mentioned tools are designed to work within a single cluster or data center and perform poorly or not at all when deployed across data centers. This paper deals with executing sequences of MapReduce jobs on geo-distributed data sets. We analyze possible ways of executing such jobs, and propose data transformation graphs that can be used to determine schedules for job sequences which are optimized either with respect to execution time or monetary cost. We introduce G-MR, a system for executing such job sequences, which implements our optimization framework. We present empirical evidence in Amazon EC2 and VICCI of the benefits of G-MR over common, naïve deployments for processing geodistributed data sets. Our evaluations show that using G-MR significantly improves processing time and cost for geodistributed data sets. [ABSTRACT FROM AUTHOR]
- Published
- 2014
- Full Text
- View/download PDF
21. On the Optimal Allocation of Virtual Resources in Cloud Computing Networks.
- Author
-
Papagianni, Chrysa, Leivadeas, Aris, Papavassiliou, Symeon, Maglaris, Vasilis, Cervelló-Pastor, Cristina, and Monje, Álvaro
- Subjects
CLOUD computing ,COMPUTER networks ,DISTRIBUTED computing ,COST effectiveness ,DATA libraries ,COMPUTER simulation ,QUALITY of service - Abstract
Cloud computing builds upon advances on virtualization and distributed computing to support cost-efficient usage of computing resources, emphasizing on resource scalability and on demand services. Moving away from traditional data-center oriented models, distributed clouds extend over a loosely coupled federated substrate, offering enhanced communication and computational services to target end-users with quality of service (QoS) requirements, as dictated by the future Internet vision. Toward facilitating the efficient realization of such networked computing environments, computing and networking resources need to be jointly treated and optimized. This requires delivery of user-driven sets of virtual resources, dynamically allocated to actual substrate resources within networked clouds, creating the need to revisit resource mapping algorithms and tailor them to a composite virtual resource mapping problem. In this paper, toward providing a unified resource allocation framework for networked clouds, we first formulate the optimal networked cloud mapping problem as a mixed integer programming (MIP) problem, indicating objectives related to cost efficiency of the resource mapping procedure, while abiding by user requests for QoS-aware virtual resources. We subsequently propose a method for the efficient mapping of resource requests onto a shared substrate interconnecting various islands of computing resources, and adopt a heuristic methodology to address the problem. The efficiency of the proposed approach is illustrated in a simulation/emulation environment, that allows for a flexible, structured, and comparative performance evaluation. We conclude by outlining a proof-of-concept realization of our proposed schema, mounted over the European future Internet test-bed FEDERICA, a resource virtualization platform augmented with network and computing facilities. [ABSTRACT FROM AUTHOR]
- Published
- 2013
- Full Text
- View/download PDF
22. “Cool” Load Balancing for High Performance Computing Data Centers.
- Author
-
Sarood, Osman, Miller, Phil, Totoni, Ehsan, and Kalé, Laxmikant V.
- Subjects
LOAD balancing (Computer networks) ,PERFORMANCE evaluation ,DATA libraries ,ELECTRIC power consumption ,MICROPROCESSORS ,ELECTRIC potential ,TEMPERATURE effect - Abstract
As we move to exascale machines, both peak power demand and total energy consumption have become prominent challenges. A significant portion of that power and energy consumption is devoted to cooling, which we strive to minimize in this work. We propose a scheme based on a combination of limiting processor temperatures using dynamic voltage and frequency scaling (DVFS) and frequency-aware load balancing that reduces cooling energy consumption and prevents hot spot formation. Our approach is particularly designed for parallel applications, which are typically tightly coupled, and tries to minimize the timing penalty associated with temperature control. This paper describes results from experiments using five different Charm++ and MPI applications with a range of power and utilization profiles. They were run on a 32-node (128-core) cluster with a dedicated air conditioning unit. The scheme is assessed based on three metrics: the ability to control processors' temperature and hence avoid hot spots, minimization of timing penalty, and cooling energy savings. Our results show cooling energy savings of up to 63 percent, with a timing penalty of only 2-23 percent. [ABSTRACT FROM AUTHOR]
- Published
- 2012
- Full Text
- View/download PDF
23. eRAID: Conserving Energy in Conventional Disk-Based RAID System.
- Author
-
Wang, Jun, Huijun Zhu, and Dong Li
- Subjects
ENERGY consumption ,CLIENT/SERVER computing ,DATA libraries ,ENERGY conservation ,REDUNDANCY in engineering ,OPTICAL disks ,ALGORITHMS ,SIMULATION methods & models ,INFORMATION resources management - Abstract
Recently, high-energy consumption has become a serious concern for both storage servers and data centers. Recent research studies have utilized the short transition times of multispeed disks to decrease energy consumption. Manufacturing challenges and costs have so far prevented commercial deployment of multispeed disks. In this paper, we propose an energy saving policy, eRAID (energy-efficient RAID), for conventional disk-based mirrored and parity redundant disk array architectures. eRAID saves energy by spinning down partial or the entire mirror disk group with constraints of acceptable performance degradation. We first develop a multiconstraint energy-saving model for the RAID environment by considering both disk characteristics and workload features. Then, we develop a performance (response time and throughput) control scheme for eRAID based on the analytical model. Experimental results show that eRAID can save up to 32 percent energy while satisfying the predefined performance requirement. [ABSTRACT FROM AUTHOR]
- Published
- 2008
- Full Text
- View/download PDF
24. Sliding Mode Congestion Control for Data Center Ethernet Networks.
- Author
-
Jiang, Wanchun, Ren, Fengyuan, Shu, Ran, Wu, Yongwei, and Lin, Chuang
- Subjects
SLIDING mode control ,DATA libraries ,SERVER farms (Computer network management) ,ETHERNET ,OSCILLATIONS - Abstract
Recently, Ethernet is enhanced as the unified switch fabric of data centers, called data center Ethernet. One of the indispensable enhancements is end-to-end congestion management, and currently quantized congestion notification (QCN) has been ratified as the corresponding standard. However, our experiments show that QCN suffers from large oscillations of the queue length at the bottleneck link such that the buffer is emptied frequently and accordingly the link utilization degrades, with certain system parameters and network configurations. This phenomenon is corresponding to our theoretical analysis result that QCN fails to enter into the sliding mode motion (SMM) pattern with certain system parameters and network configurations. Knowing the drawbacks of QCN and realizing the advantage that congestion management system is insensitive to the changes of parameters and network configurations in the SMM pattern, we present sliding mode congestion control (SMCC), which can enter into the SMM pattern under any conditions. SMCC is simple, stable, fair, has short response time, and can be easily used to replace QCN because both of them follow the framework developed by the IEEE 802.1Qau work group. Experiments on the NetFPGA platform show that SMCC is superior to QCN, especially when traffic pattern and network states are variable. [ABSTRACT FROM PUBLISHER]
- Published
- 2015
- Full Text
- View/download PDF
25. Analysis of Reliability Dynamics of SSD RAID.
- Author
-
Li, Yongkun, Lee, Patrick P.C., and Lui, John C.S.
- Subjects
BIT error rate ,DATA libraries ,MARKOV processes ,RELIABILITY in engineering ,DESKTOP environments (Computer interfaces) ,COMPUTER simulation - Abstract
Solid-state drives (SSDs) have been widely deployed in desktops and data centers. However, SSDs suffer from bit errors, and the bit error rate is time dependent since it increases as an SSD wears down. Traditional storage systems mainly use parity-based RAID to provide reliability guarantees by striping redundancy across multiple devices, but the effectiveness of traditional RAID schemes in SSDs remains debatable. In particular, an open problem is how different parity distributions over multiple devices influence the reliability of an SSD RAID array. That is, should we evenly distribute patsaverties as suggested by conventional wisdom, or unevenly distribute parties as recently proposed for SSD RAID? To address this fundamental problem, we propose the first analytical model to quantify the reliability dynamics of an SSD RAID array as it ages. Specifically, we develop a “non-homogeneous” continuous time Markov chain model, and derive the transient reliability solution. We validate our model via trace-driven simulation and conduct numerical analysis to analyze the reliability dynamics of SSD RAID arrays subject to different parity distributions, error rates, and SSD array configurations. Our model enables system practitioners to decide the appropriate parity distribution based on their reliability requirements. [ABSTRACT FROM PUBLISHER]
- Published
- 2016
- Full Text
- View/download PDF
26. Optimal Task Placement with QoS Constraints in Geo-Distributed Data Centers Using DVFS.
- Author
-
Gu, Lin, Zeng, Deze, Barnawi, Ahmed, Guo, Song, and Stojmenovic, Ivan
- Subjects
OPTIMAL control theory ,QUALITY of service ,ELECTRIC power consumption ,DATA libraries ,CLIENT/SERVER computing - Abstract
With the rising demands on cloud services, the electricity consumption has been increasing drastically as the main operational expenditure (OPEX) to data center providers. The geographical heterogeneity of electricity prices motivates us to study the task placement problem over geo-distributed data centers. We exploit the dynamic frequency scaling technique and formulate an optimization problem that minimizes OPEX while guaranteeing the quality-of-service, i.e., the expected response time of tasks. Furthermore, an optimal solution is discovered for this formulated problem. The experimental results show that our proposal achieves much higher cost-efficiency than the traditional resizing scheme, i.e., by activating/deactivating certain servers in data centers. [ABSTRACT FROM AUTHOR]
- Published
- 2015
- Full Text
- View/download PDF
27. Adaptive-Acceleration Data Center TCP.
- Author
-
Zhang, Tao, Wang, Jianxin, Huang, Jiawei, Huang, Yi, Chen, Jianer, and Pan, Yi
- Subjects
TCP/IP ,BANDWIDTHS ,COMPUTER simulation ,DATA libraries ,ARTIFICIAL intelligence - Abstract
Providing deadline-sensitive services is a challenge in data centers. Because of the conservativeness in additive increase congestion avoidance, current transmission control protocols are inefficient in utilizing the super high bandwidth of data centers. This may cause many deadline-sensitive flows to miss their deadlines before achieving their available bandwidths. We propose an Adaptive-Acceleration Data Center TCP, A $\!^2$
DTCP, which takes into account both network congestion and latency requirement of application service. By using congestion avoidance with an adaptive increase rate that varies between additive and multiplicative, A $\!^2$ DTCP accelerates bandwidth detection thus achieving high bandwidth utilization efficiency. At-scale simulations and real testbed implementations show that A $\!^2$ DTCP significantly reduces the missed deadline ratio compared to D $\!^2$ TCP and DCTCP. In addition, A$\!^2$ DTCP can co-exist with conventional TCP as well without requiring more changes in switch hardware than D $\!^2$ TCP and DCTCP. [ABSTRACT FROM PUBLISHER]- Published
- 2015
- Full Text
- View/download PDF
28. Endurance-Aware Flash-Cache Management for Storage Servers.
- Author
-
Suei, Pei-Lun, Yeh, Mi-Yen, and Kuo, Tei-Wei
- Subjects
FLASH memory ,CACHE memory ,ENERGY consumption ,COMPUTER storage devices ,DATA libraries ,CLIENT/SERVER computing - Abstract
As flash memory emerges as a high-performance and energy-efficient alternative for storage devices, how to accommodate disk-based storage servers with a flash-memory cache might provide a promising solution to resolve the energy-efficiency concerns of storage servers and their data centers. In this work, we propose a cache design method over flash memory without any additional hardware support for storage servers. In particular, a scalable set-associative flash-cache management design is proposed to significantly improve the flash endurance with excellent caching response time and hit rates. The capability of the proposed scheme was evaluated by extensive experiments over Microsoft’s server traces, in which significant improvement on endurance was achieved, compared to existing work. [ABSTRACT FROM AUTHOR]
- Published
- 2014
- Full Text
- View/download PDF
29. ANTELOPE: A Semantic-Aware Data Cube Scheme for Cloud Data Center Networks.
- Author
-
Hua, Yu, Liu, Xue, and Jiang, Hong
- Subjects
SEMANTIC computing ,DATA analysis ,SCHEME programming language ,CLOUD computing ,DATA libraries ,BANDWIDTHS ,SCALABILITY - Abstract
Today’s cloud data centers contain more than millions of servers and offer high bandwidth. A fundamental problem is how to significantly improve the large-scale system’s scalability to interconnect a large number of servers and meanwhile support various online services in cloud computing. One way is to deal with the challenge of potential mismatching between the network architecture and the data placement. To address this challenge, we present ANTELOPE, a scalable distributed data-centric scheme in cloud data centers, in which we systematically take into account both the property of network architecture and the optimization of data placement. The basic idea behind ANTELOPE is to leverage precomputation based data cube to support online cloud services. Since the construction of data cube suffers from the high costs of full materialization, we use a semantic-aware partial materialization solution to significantly reduce the operation and space overheads. Extensive experiments on real system implementations demonstrate the efficacy and efficiency of our proposed scheme. [ABSTRACT FROM PUBLISHER]
- Published
- 2014
- Full Text
- View/download PDF
30. Integrated Approach to Data Center Power Management.
- Author
-
Ganesh, Lakshmi, Weatherspoon, Hakim, Marian, Tudor, and Birman, Ken
- Subjects
DATA libraries ,PERFORMANCE evaluation ,GREEN products ,COMPUTER storage devices ,CLIENT/SERVER computing ,ENERGY consumption of computers ,ONLINE information services - Abstract
Energy accounts for a significant fraction of the operational costs of a data center, and data center operators are increasingly interested in moving toward low-power designs. Two distinct approaches have emerged toward achieving this end: the power-proportional approach focuses on reducing disk and server power consumption, while the green data center approach focuses on reducing power consumed by support-infrastructure like cooling equipment, power distribution units, and power backup equipment. We propose an integrated approach, which combines the benefits of both. Our solution enforces power-proportionality at the granularity of a rack or even an entire containerized data center; thus, we power down not only idle IT equipment, but also their associated support-infrastructure. We show that it is practical today to design data centers to power down idle racks or containers—and in fact, current online service trends strongly enable this model. Finally, we show that our approach combines the energy savings of power-proportional and green data center approaches, while performance remains unaffected. [ABSTRACT FROM AUTHOR]
- Published
- 2013
- Full Text
- View/download PDF
31. Power-Aware Storage Cache Management.
- Author
-
Zhu, Qingbo and Zhou, Yuanyuan
- Subjects
ENERGY consumption ,DATA libraries ,POWER resources ,CONSUMPTION (Economics) ,CACHE memory ,DYNAMIC programming - Abstract
Reducing energy consumption is an important issue for data centers. Among the various components of a data center, storage is one of the biggest energy consumers. Previous studies have shown that the average idle period for a server disk in a data center is very small compared to the time taken to spin down and spin up. This significantly limits the effectiveness of disk power management schemes. This article proposes several power-aware storage cache management algorithms that provide more opportunities for the underlying disk power management schemes to save energy. More specifically, we present an offline energy- optimal cache replacement algorithm using dynamic programming, which minimizes the disk energy consumption. We also present an offline power-aware greedy algorithm that is more energy-efficient than Belady's offline algorithm (which minimizes cache misses only). We also propose two online power-aware algorithms, PA-LRU and PB-LRU. Simulation results with both a real system and synthetic workloads show that, compared to LRU, our online algorithms can save up to 22 percent more disk energy and provide up to 64 percent better average response time. We have also investigated the effects of four storage cache write policies on disk energy consumption. [ABSTRACT FROM AUTHOR]
- Published
- 2005
- Full Text
- View/download PDF
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.