4,785 results on '"Service level objective"'
Search Results
2. FAULT TOLERANCE USING SELF-HEALING SLA AND LOAD BALANCED DYNAMIC RESOURCE PROVISIONING IN CLOUD COMPUTING
- Author
-
Mayank Sohani and Dr. S. C. Jain
- Subjects
quality of service ,cloud service provider ,service level agreement ,service level objective ,and predictive cloud computing system ,Information technology ,T58.5-58.64 ,Electronic computers. Computer science ,QA75.5-76.95 - Abstract
In cloud computing, heterogeneous resources located in different data centers are used to provide pay-per-use performance. Resource provisioning encounters problems due to uncertainty and dispersion of resources in cloud computing environments. According to our research, there are various techniques, but the existing methods and frameworks cannot handle the behaviors of these applications, environment, and resources. The cloud provider wants a framework to execute the task for achieving cost-effective delivery and reliable cloud services. Cloud workload's essential requirement is Quality of Service (QoS) based on adequate resource provisioning as per cloud workload needs. The resource discovery and best paring of workload-resource as per the cloud user's QoS need for workload requirements is an optimization problem. The resource provisioning is appropriate and can be acceptable if offered as per QoS needs in a controlled way. A mutual contract of Service Level Agreement (SLA) is signed between Cloud Service Provider (CSP) and cloud users, which determines the Service Level Objectives (SLO). The SLA contains the QoS parameters and notifies the dependent SLA's recent status. The user will get Quality of Service (QoS), based on prediction, where the resource would manage efficiently and lead to perform loads and deliver cost-effective and reliable cloud services. This paper presents a prediction-based resource management technique called Predictive Cloud Computing Systems (PCCSs). Focus on the self-healing-based prediction that handles unexpected failures and self-configuration-based prediction of resources and applications. Predictive Cloud Computing Systems (PCCSs) performance evaluated in the cloud simulator. The simulation results reveal that Predictive Cloud Computing Systems (PCCSs) achieve better results than existing techniques as execution time, cost-effective, resource conflict, SLA breach while delivering reliable services. [JJCIT 2021; 7(2.000): 206-222]
- Published
- 2021
- Full Text
- View/download PDF
3. Latency‐aware adaptive micro‐batching techniques for streamed data compression on graphics processing units.
- Author
-
Stein, Charles M., Rockenbach, Dinei A., Griebler, Dalvan, Torquati, Massimo, Mencagli, Gabriele, Danelutto, Marco, and Fernandes, Luiz G.
- Subjects
DATA compression ,GRAPHICS processing units - Abstract
Summary: Stream processing is a parallel paradigm used in many application domains. With the advance of graphics processing units (GPUs), their usage in stream processing applications has increased as well. The efficient utilization of GPU accelerators in streaming scenarios requires to batch input elements in microbatches, whose computation is offloaded on the GPU leveraging data parallelism within the same batch of data. Since data elements are continuously received based on the input speed, the bigger the microbatch size the higher the latency to completely buffer it and to start the processing on the device. Unfortunately, stream processing applications often have strict latency requirements that need to find the best size of the microbatches and to adapt it dynamically based on the workload conditions as well as according to the characteristics of the underlying device and network. In this work, we aim at implementing latency‐aware adaptive microbatching techniques and algorithms for streaming compression applications targeting GPUs. The evaluation is conducted using the Lempel‐Ziv‐Storer‐Szymanski compression application considering different input workloads. As a general result of our work, we noticed that algorithms with elastic adaptation factors respond better for stable workloads, while algorithms with narrower targets respond better for highly unbalanced workloads. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
4. Less Provisioning: A Hybrid Resource Scaling Engine for Long-Running Services With Tail Latency Guarantees
- Author
-
Zhao Laiping, Keqiu Li, Rongqi Zhang, and Binlei Cai
- Subjects
Computer Networks and Communications ,business.industry ,Computer science ,Distributed computing ,Testbed ,Service level objective ,Provisioning ,Workload ,Cloud computing ,Computer Science Applications ,Hardware and Architecture ,Collaborative filtering ,Time series ,Latency (engineering) ,business ,Software ,Information Systems - Abstract
Modern resource management frameworks guarantee low tail latency for long-running services using the resource over-provisioning method, resulting in serious waste of resources and increasing the service costs greatly. To reduce the over-provisioning cost, we present HRSE, a hybrid resource scaling engine that enables much more efficient resource provisioning for both periodic and non-periodic workloads of long-running services while guaranteeing the tail latency Service Level Objective (SLO). HRSE employs a convolution-based time series analysis to identify periodic patterns in workloads. If periodic patterns are discovered, HRSE estimates the just-right amount of resources based on the periodic features through a top-K based collaborative filtering approach. Otherwise, it leverages wavelet-clustering to capture the short-term patterns in non-periodic workloads and predict the resource demands for the near future. To further enforce the tail latency SLO, HRSE uses an online reprovisioning mechanism that dynamically adjusts the resources to mitigate the performance uncertainty due to workload burstinesses. We fully implement HRSE on top of Docker and conduct extensive experiments using traces from production systems. Testbed experiments show that HRSE is able to increase the average resource utilization to 43% and 45% for periodic and non-periodic workloads respectively while guaranteeing the same tail latency objective.
- Published
- 2022
5. Autonomous Lifecycle Management for Resource-Efficient Workload Orchestration for Green Edge Computing
- Author
-
Francesc Guim, Thijs Metsch, Hassnaa Moustafa, Timothy Verrall, David Carrera, Nicola Cadenelli, Jiang Chen, David Doria, Chadie Ghadie, and Raul Gonzalez Prats
- Subjects
Service (systems architecture) ,Computer Networks and Communications ,Renewable Energy, Sustainability and the Environment ,business.industry ,Computer science ,Distributed computing ,Service level objective ,Cloud computing ,Workload ,Product life-cycle management ,Enhanced Data Rates for GSM Evolution ,Orchestration (computing) ,business ,Edge computing - Abstract
Edge computing is an important pillar for green computation by bringing the Cloud resources to the Edge, serving real-time applications, and reducing the computing and network resources required to transfer data for processing in the Cloud. 5G brings network densification and enables massive IoT and V2X applications, which triggers the need for edge computing to host network functions and user-facing services in a converged edge platform(s). Several edge computing deployments are being observed by ecosystem players (telco, ISVs, chip vendors, CSPs,... etc.) for IoT or V2X services, however, focusing on converged network functions and services. The point that is still in its early stages is the dynamic workload orchestration across the converged edge platforms running network functions and multi-tenant IoT services with different compute requirements and different Service Level Objectives (SLOs). This paper focuses on autonomous life cycle management for converged edge platform(s) to enable resource-efficient workload orchestration, contributing to the green goal. We present a solution for intelligent dynamic resources configuration on edge computing platforms hosting multi-tenant services while guaranteeing the SLO for each service and helping green communication goal. The presented solution has been deployed in a trial, and we present results on efficient resources configuration.
- Published
- 2022
6. DAVINCI: online and Dynamic Adaptation of eVolvable vIrtual Network services over Cloud Infrastructures
- Author
-
Claes Edstrom, Laaziz Lahlou, and Nadjia Kara
- Subjects
Service (systems architecture) ,Downtime ,Computer Networks and Communications ,Computer science ,business.industry ,Distributed computing ,Service level objective ,Cloud computing ,Elasticity (cloud computing) ,Hardware and Architecture ,Overhead (computing) ,Minimum-cost flow problem ,business ,Virtual network ,Software - Abstract
Service function chains, or more generally, virtual network services, evolve throughout their life-cycle due to traffic fluctuations, resource usage, and the stringent requirements of the SLO (Service Level Objective) that must be fulfilled. Network operators resort to elasticity mechanisms (e.g., vertical and horizontal scaling) and migration procedures to adapt the service function chains dynamically to meet their requirements constantly. However, most state-of-the-art solutions do not evaluate the overall impact of such mechanisms on the service function chains or the resulting penalty costs (e.g., migration downtime and SLO violation). To bridge this gap, we propose DAVINCI, a decision-making tool with different adaptation policies that allows the network operator to adapt the service function chains to minimize network changes (e.g., reallocating paths and VNFs) while keeping penalty costs at their lowest level. Moreover, DAVINCI allows for the adaptations (e.g., migration, vertical and horizontal scaling) to be expressed as a set of decisions and leverages the Min Cost Flow problem to estimate migration downtime. The analytical evaluation shows that DAVINCI outperforms the existing state-of-the-art solution (NFV-PEAR) in terms of migration, traffic costs, and the overhead induced by the migration and elasticity mechanisms while significantly reducing penalty costs.
- Published
- 2022
7. Migrating From Legacy to Software Defined Networks: A Network Reliability Perspective
- Author
-
Anwar Haque, Yaser Al Mtawa, and Hanan Lutfiyya
- Subjects
Service (systems architecture) ,Computer science ,Reachability ,Distributed computing ,Reliability (computer networking) ,Metric (mathematics) ,Service level objective ,Electrical and Electronic Engineering ,Safety, Risk, Reliability and Quality ,Network topology ,Software-defined networking ,Telecommunications network - Abstract
Designing survivable communication networks to achieve carrier-grade five-nines reliability is of paramount importance for the network operators. This article addresses service reliability and its related aspects such as nodal reachability, network connectivity, and edge-disjoint routing in both traditional networks and software defined networks (SDNs). The proposed roadmap is based on two phases: Fundamental analytical phase and performance evaluation phase. In the first phase, a graph operator is defined to analyze the characteristics of the reliability metric and its associated reachability feature. This phase will focus on both the macro- and micro-level properties of reliability. In the second phase, we exploit the analysis in the former phase to get an insight into the performance evaluation of traditional and SDN-based networks against the reliability metric, and then calculate the statistical significance of the mean difference of their reliability values. Reliability under edge-disjoint paths to avoid resource competition is also investigated. Various types of topologies are utilized to test the service reliability of both architecture designs. Extensive simulation results show that SDN-based networks have comparable performance to its legacy counterpart against the operational reliability metric. Our findings not only shed light on enhancing reliability using edge-disjoint paths under link failure scenarios but also expected to benefit the operators to achieve their service level objectives while migrating from legacy to SDN-based platform.
- Published
- 2021
8. iAgree Studio: A Platform to Edit and Validate WS–Agreement Documents
- Author
-
Müller, Carlos, Gutiérrez, Antonio Manuel, Resinas, Manuel, Fernández, Pablo, Ruiz-Cortés, Antonio, Hutchison, David, editor, Kanade, Takeo, editor, Kittler, Josef, editor, Kleinberg, Jon M., editor, Mattern, Friedemann, editor, Mitchell, John C., editor, Naor, Moni, editor, Nierstrasz, Oscar, editor, Pandu Rangan, C., editor, Steffen, Bernhard, editor, Sudan, Madhu, editor, Terzopoulos, Demetri, editor, Tygar, Doug, editor, Vardi, Moshe Y., editor, Weikum, Gerhard, editor, Basu, Samik, editor, Pautasso, Cesare, editor, Zhang, Liang, editor, and Fu, Xiang, editor
- Published
- 2013
- Full Text
- View/download PDF
9. Applying Complex Event Processing towards Monitoring of Multi-party Contracts and Services for Logistics – A Discussion
- Author
-
Roth, Martin, Donath, Steffi, van der Aalst, Wil, Series editor, Mylopoulos, John, Series editor, Rosemann, Michael, Series editor, Shaw, Michael J., Series editor, Szyperski, Clemens, Series editor, Daniel, Florian, editor, Barkaoui, Kamel, editor, and Dustdar, Schahram, editor
- Published
- 2012
- Full Text
- View/download PDF
10. Web Engineering for Cloud Computing : (Web Engineering Forecast: Cloudy with a Chance of Opportunities)
- Author
-
Toffetti, Giovanni, Hutchison, David, editor, Kanade, Takeo, editor, Kittler, Josef, editor, Kleinberg, Jon M., editor, Mattern, Friedemann, editor, Mitchell, John C., editor, Naor, Moni, editor, Nierstrasz, Oscar, editor, Pandu Rangan, C., editor, Steffen, Bernhard, editor, Sudan, Madhu, editor, Terzopoulos, Demetri, editor, Tygar, Doug, editor, Vardi, Moshe Y., editor, Weikum, Gerhard, editor, Grossniklaus, Michael, editor, and Wimmer, Manuel, editor
- Published
- 2012
- Full Text
- View/download PDF
11. Who Do You Call? Problem Resolution through Social Compute Units
- Author
-
Sengupta, Bikram, Jain, Anshu, Bhattacharya, Kamal, Truong, Hong-Linh, Dustdar, Schahram, Hutchison, David, editor, Kanade, Takeo, editor, Kittler, Josef, editor, Kleinberg, Jon M., editor, Mattern, Friedemann, editor, Mitchell, John C., editor, Naor, Moni, editor, Nierstrasz, Oscar, editor, Pandu Rangan, C., editor, Steffen, Bernhard, editor, Sudan, Madhu, editor, Terzopoulos, Demetri, editor, Tygar, Doug, editor, Vardi, Moshe Y., editor, Weikum, Gerhard, editor, Liu, Chengfei, editor, Ludwig, Heiko, editor, Toumani, Farouk, editor, and Yu, Qi, editor
- Published
- 2012
- Full Text
- View/download PDF
12. QoS Contract-Aware Reconfiguration of Component Architectures Using E-Graphs
- Author
-
Tamura, Gabriel, Casallas, Rubby, Cleve, Anthony, Duchien, Laurence, Hutchison, David, Series editor, Kanade, Takeo, Series editor, Kittler, Josef, Series editor, Kleinberg, Jon M., Series editor, Mattern, Friedemann, Series editor, Mitchell, John C., Series editor, Naor, Moni, Series editor, Nierstrasz, Oscar, Series editor, Pandu Rangan, C., Series editor, Steffen, Bernhard, Series editor, Sudan, Madhu, Series editor, Terzopoulos, Demetri, Series editor, Tygar, Doug, Series editor, Vardi, Moshe Y., Series editor, Weikum, Gerhard, Series editor, Barbosa, Luís Soares, editor, and Lumpe, Markus, editor
- Published
- 2012
- Full Text
- View/download PDF
13. Blockchain-enabled real-time SLA monitoring for cloud-hosted services
- Author
-
Waheed Iqbal, Junaid Arshad, Hassan Zaib, Sidrah Abdullah, and Kashif Mehboob Khan
- Subjects
Service (business) ,Blockchain ,Computer Networks and Communications ,business.industry ,Computer science ,Service level objective ,Cloud computing ,Computer security ,computer.software_genre ,Service level ,Scalability ,Key (cryptography) ,The Internet ,business ,computer ,Software - Abstract
Cloud computing is an important technology for businesses and individual users to obtain computing resources over the Internet on-demand and flexibly. Although cloud computing has been adopted across diverse applications, the owners of time-and-performance critical applications require cloud service providers’ guarantees about their services, such as availability and response times. Service Level Agreements (SLAs) are a mechanism to communicate and enforce such guarantees typically represented as service level objectives (SLOs), and financial penalties are imposed on SLO violations. Due to delays and inaccuracies caused by manual processing, an automatic method to periodically verify SLA terms in a transparent and trustworthy manner is fundamental to effective SLA monitoring, leading to the acceptance and credibility of such service to the customers of cloud services. This paper presents a blockchain-based distributed infrastructure that leverages fundamental blockchain properties to achieve immutable and trustworthy SLA monitoring within cloud services. The paper carries out an in-depth empirical investigation for the scalability of the proposed system in order to address the challenge of transparently enforcing real-time monitoring of cloud-hosted services leveraging blockchain technology. This will enable all the stakeholders to enforce accurate execution of SLA without any imprecisions and delays by maintaining an immutable ledger publicly across blockchain network. The experimentation takes into consideration several attributes of blockchain which are critical in achieving optimum performance. The paper also investigates key characteristics of these factors and their impact to the behaviour of the system for further scaling it up under various cases for increased service utilization.
- Published
- 2021
14. Automated Validation of Compensable SLAs
- Author
-
Antonio Manuel Gutiérrez, Carlos Müller, Octavio Martín-Díaz, Antonio Ruiz-Cortés, Manuel Resinas, and Pablo Fernandez
- Subjects
Service (systems architecture) ,Information Systems and Management ,Computer Networks and Communications ,Computer science ,business.industry ,Reliability (computer networking) ,Service management ,Service level objective ,020207 software engineering ,Cloud computing ,Provisioning ,02 engineering and technology ,Computer Science Applications ,Service-level agreement ,Hardware and Architecture ,020204 information systems ,0202 electrical engineering, electronic engineering, information engineering ,business ,Software engineering ,Constraint satisfaction problem - Abstract
A Service Level Agreement (SLA) regulates the provisioning of a service by defining a set of guarantees. Each guarantee sets a Service Level Objective (SLO) on some service metrics, and optionally a compensation that is applied when the SLO is unfulfilled or overfulfilled. Currently, there are software tools and research proposals that use the information about compensations to automate and optimise certain parts of the service management. However, they assume that compensations are well defined, which is too optimistic in some circumstances and can lead to undesirable situations. In this article we discuss about the notion of validity of guarantees with a compensation, which we refer to as compensable guarantees (CG). We describe an abstract model of CGs and we provide a technique that leverages constraint satisfaction problem solvers to automatically validate them. We also present a materialisation of the model of CGs in iAgree, a language to specify SLAs and a tooling support that implements our whole approach. An assessment over 319 CGs taken from 24 real-world SLAs suggests that the expressiveness and effectiveness of our proposal can pave the way for using CGs in a safer and more reliable way.
- Published
- 2021
15. SLA as a Complementary Currency in Peer-2-Peer Markets
- Author
-
Petri, Ioan, Rana, Omer, Cosmin Silaghi, Gheorghe, Hutchison, David, Series editor, Kanade, Takeo, Series editor, Kittler, Josef, Series editor, Kleinberg, Jon M., Series editor, Mattern, Friedemann, Series editor, Mitchell, John C., Series editor, Naor, Moni, Series editor, Nierstrasz, Oscar, Series editor, Pandu Rangan, C., Series editor, Steffen, Bernhard, Series editor, Sudan, Madhu, Series editor, Terzopoulos, Demetri, Series editor, Tygar, Doug, Series editor, Vardi, Moshe Y., Series editor, Weikum, Gerhard, Series editor, Altmann, Jörn, editor, and Rana, Omer F., editor
- Published
- 2010
- Full Text
- View/download PDF
16. Extending WS-Agreement with Multi-round Negotiation Capability
- Author
-
Rumpl, Angela, Wäldrich, Oliver, Ziegler, Wolfgang, Wieder, Philipp, editor, Yahyapour, Ramin, editor, and Ziegler, Wolfgang, editor
- Published
- 2010
- Full Text
- View/download PDF
17. Lessons Learned from Implementing WS-Agreement
- Author
-
Battré, Dominic, Hovestadt, Matthias, Wäldrich, Oliver, Wieder, Philipp, editor, Yahyapour, Ramin, editor, and Ziegler, Wolfgang, editor
- Published
- 2010
- Full Text
- View/download PDF
18. The PLASTIC Framework and Tools for Testing Service-Oriented Applications
- Author
-
Bertolino, Antonia, De Angelis, Guglielmo, Frantzen, Lars, Polini, Andrea, Hutchison, David, Series editor, Kanade, Takeo, Series editor, Kittler, Josef, Series editor, Kleinberg, Jon M., Series editor, Mattern, Friedemann, Series editor, Mitchell, John C., Series editor, Naor, Moni, Series editor, Nierstrasz, Oscar, Series editor, Pandu Rangan, C., Series editor, Steffen, Bernhard, Series editor, Sudan, Madhu, Series editor, Terzopoulos, Demetri, Series editor, Tygar, Doug, Series editor, Vardi, Moshe Y., Series editor, Weikum, Gerhard, Series editor, De Lucia, Andrea, editor, and Ferrucci, Filomena, editor
- Published
- 2009
- Full Text
- View/download PDF
19. Managing the Alignment between Business and Software Services Requirements from a Capability Model Perspective
- Author
-
Grandry, Eric, Dubois, Eric, Picard, Michel, Rifaut, André, Hutchison, David, Series editor, Kanade, Takeo, Series editor, Kittler, Josef, Series editor, Kleinberg, Jon M., Series editor, Mattern, Friedemann, Series editor, Mitchell, John C., Series editor, Naor, Moni, Series editor, Nierstrasz, Oscar, Series editor, Pandu Rangan, C., Series editor, Steffen, Bernhard, Series editor, Sudan, Madhu, Series editor, Terzopoulos, Demetri, Series editor, Tygar, Doug, Series editor, Vardi, Moshe Y., Series editor, Weikum, Gerhard, Series editor, Mähönen, Petri, editor, Pohl, Klaus, editor, and Priol, Thierry, editor
- Published
- 2008
- Full Text
- View/download PDF
20. Managing User Expectations with Component Performance Contracts
- Author
-
Coppola, Massimo, Laforenza, Domenico, Tonellotto, Nicola, Danelutto, Marco, Vanneschi, Marco, Zoccolo, Corrado, Talia, Domenico, Yahyapour, Ramin, and Ziegler, Wolfgang
- Published
- 2008
- Full Text
- View/download PDF
21. On QoE-Oriented Cloud Service Orchestration for Application Providers
- Author
-
Kannappan Palaniappan, Dmitrii Chemodanov, Huy Trinh, Prasad Calyam, Jon Patman, and Samaikya Valluripally
- Subjects
Information Systems and Management ,Computer Networks and Communications ,business.industry ,Computer science ,Quality of service ,Service level objective ,Cloud computing ,Application service provider ,Virtualization ,computer.software_genre ,Computer Science Applications ,Hardware and Architecture ,Forwarding plane ,Orchestration (computing) ,Quality of experience ,business ,computer ,Computer network - Abstract
New virtualization technologies allow Infrastructure Providers (InPs) to lease their resources to Application Service Providers (ASPs) for highly scalable delivery of cloud services to end-users. However, existing literature lacks knowledge on Quality of Experience (QoE)-oriented cloud service orchestration algorithms that can guide ASPs on how to plan their budget to enhance satisfactory QoE delivery to end-users. In contrast to the InP’s cloud service orchestration, the ASP’s orchestration should not rely on expensive infrastructure control mechanisms such as Software-Defined Networking (SDN), or require apriori knowledge on the number of services to be instantiated and their anticipated placement location within InP’s infrastructure. In this paper, we address this issue of delivering satisfactory user QoE by synergistically optimizing both ASP’s management and data planes . The optimization within the ASP management plane first maximizes Service Level Objective (SLO) coverage of users when application services are being deployed, and are not yet operational. The optimization of the ASP data plane then enhances satisfactory user QoE delivery when applications services are operational with real user access. Our evaluation of QoE-oriented algorithms using realistic numerical simulations, real-world cloud testbed experiments with actual users and ASP case studies show notably improved performance over existing cloud service orchestration solutions.
- Published
- 2021
22. Towards Deadline Guaranteed Cloud Storage Services
- Author
-
Guoxin Liu, Haiying Shen, Haoyu Wang, and Lei Yu
- Subjects
Information Systems and Management ,Computer Networks and Communications ,business.industry ,Computer science ,Service level objective ,020206 networking & telecommunications ,Cloud computing ,02 engineering and technology ,Load balancing (computing) ,020202 computer hardware & architecture ,Computer Science Applications ,Load management ,Data access ,Hardware and Architecture ,Server ,0202 electrical engineering, electronic engineering, information engineering ,business ,Cloud storage ,Queue ,Computer network - Abstract
More and more organizations move their data and workload to commercial cloud storage systems. However, the multiplexing and sharing of the resources in a cloud storage system present unpredictable data access latency to tenants, which may make online data-intensive applications unable to satisfy their deadline requirements. Thus, it is important for cloud storage systems to provide deadline guaranteed services. In this paper, to meet a current form of service level objective (SLO) that constrains the percentage of each tenant’s data access requests failing to meet its required deadline below a given threshold, we build a mathematical model to derive the upper bound of acceptable request arrival rate on each server. We then propose a Deadline Guaranteed storage service (called DGCloud ) that incorporates three basic algorithms. Its deadline-aware load balancing scheme redirects requests and creates replicas to release the excess load of each server beyond the derived upper bound. Its workload consolidation algorithm tries to maximally reduce servers while still satisfying the SLO to maximize the resource utilization. Its data placement optimization algorithm re-schedules the data placement to minimize the transmission cost of data replication. We further propose three enhancement methods to further improve the performance of DGCloud . A dynamic load balancing method allows an overloaded server to quickly offload its excess workload. A data request queue improvement method sets different priorities to the data responses in a server’s queue so that more requests can satisfy the SLO requirement. A wakeup server selection method selects a sleeping server that stores more popular data to wake up, which allows it to handle more data requests. Our trace-driven experiments in simulation and Amazon EC2 show the superior performance of DGCloud compared with previous methods in terms of deadline guarantees and system resource utilization, and the effectiveness of its individual algorithms.
- Published
- 2021
23. A QoS Test-Bed Generator for Web Services
- Author
-
Bertolino, Antonia, De Angelis, Guglielmo, Polini, Andrea, Hutchison, David, editor, Kanade, Takeo, editor, Kittler, Josef, editor, Kleinberg, Jon M., editor, Mattern, Friedemann, editor, Mitchell, John C., editor, Naor, Moni, editor, Nierstrasz, Oscar, editor, Pandu Rangan, C., editor, Steffen, Bernhard, editor, Sudan, Madhu, editor, Terzopoulos, Demetri, editor, Tygar, Doug, editor, Vardi, Moshe Y., editor, Weikum, Gerhard, editor, Baresi, Luciano, editor, Fraternali, Piero, editor, and Houben, Geert-Jan, editor
- Published
- 2007
- Full Text
- View/download PDF
24. Reliable Orchestration of Resources Using WS-Agreement
- Author
-
Ludwig, Heiko, Nakata, Toshiyuki, Wäldrich, Oliver, Wieder, Philipp, Ziegler, Wolfgang, Hutchison, David, editor, Kanade, Takeo, editor, Kittler, Josef, editor, Kleinberg, Jon M., editor, Mattern, Friedemann, editor, Mitchell, John C., editor, Naor, Moni, editor, Nierstrasz, Oscar, editor, Pandu Rangan, C., editor, Steffen, Bernhard, editor, Sudan, Madhu, editor, Terzopoulos, Demetri, editor, Tygar, Dough, editor, Vardi, Moshe Y., editor, Weikum, Gerhard, editor, Gerndt, Michael, editor, and Kranzlmüller, Dieter, editor
- Published
- 2006
- Full Text
- View/download PDF
25. Burst-aware predictive autoscaling for containerized microservices
- Author
-
David Carrera, Waheed Iqbal, Muhammad Imran Abdullah, Josep Lluis Berral, Jordà Polo, Universitat Politècnica de Catalunya. Departament d'Arquitectura de Computadors, Barcelona Supercomputing Center, and Universitat Politècnica de Catalunya. CAP - Grup de Computació d'Altes Prestacions
- Subjects
Information Systems and Management ,Computació en núvol ,Computer Networks and Communications ,Computer science ,Distributed computing ,Cloud computing ,Dynamic priority scheduling ,Microservices ,computer.software_genre ,Autoscaling ,Containers ,Burstiness ,Resource allocation ,Informàtica::Arquitectura de computadors [Àrees temàtiques de la UPC] ,business.industry ,Quality of service ,Response time guarantees ,Service level objective ,Virtualization ,Computer Science Applications ,Assignació de recursos ,Hardware and Architecture ,Service-level objectives ,business ,computer - Abstract
Autoscaling methods are used for cloud-hosted applications to dynamically scale the allocated resources for guaranteeing Quality-of-Service (QoS). The public-facing application serves dynamic workloads, which contain bursts and pose challenges for autoscaling methods to ensure application performance. Existing State-of-the-art autoscaling methods are burst-oblivious to determine and provision the appropriate resources. For dynamic workloads, it is hard to detect and handle bursts online for maintaining application performance. In this article, we propose a novel burst-aware autoscaling method which detects burst in dynamic workloads using workload forecasting, resource prediction, and scaling decision making while minimizing response time service-level objectives (SLO) violations. We evaluated our approach through a trace-driven simulation, using multiple synthetic and realistic bursty workloads for containerized microservices, improving performance when comparing against existing state-of-the-art autoscaling methods. Such experiments show an increase of × 1.09 in total processed requests, a reduction of × 5.17 for SLO violations, and an increase of × 0.767 cost as compared to the baseline method. This work was partially supported by the European Research Council (ERC) under the EU Horizon 2020 programme (GA 639595), the Spanish Ministry of Economy, Industry and Competitiveness (TIN2015-65316-P and IJCI2016-27485) and the Generalitat de Catalunya (2014-SGR-1051).
- Published
- 2022
26. Utility Computing: A Better Model for Outsourcing Success?
- Author
-
Kaplan, Jeff and Brudenall, Peter, editor
- Published
- 2005
- Full Text
- View/download PDF
27. Deadline-Aware Cost Optimization for Spark
- Author
-
Wojciech Golab, Supratik Mukhopadhyay, and Subhajit Sidhanta
- Subjects
Mathematical optimization ,Information Systems and Management ,business.industry ,Computer science ,Big data ,Service level objective ,020207 software engineering ,Cloud computing ,02 engineering and technology ,Cluster (spacecraft) ,Cost optimization ,Parallel processing (DSP implementation) ,Approximation error ,Spark (mathematics) ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,business ,Information Systems - Abstract
We present OptEx, a closed-form model of job execution on Apache Spark, a popular parallel processing engine. To the best of our knowledge, OptEx is the first work that analytically models job completion time on Spark. The model can be used to estimate the completion time of a given Spark job on a cloud, with respect to the size of the input dataset, the number of iterations, and the number of nodes comprising the underlying cluster. Experimental results demonstrate that OptEx yields a mean relative error of 6 percent in estimating the job completion time. Furthermore, the model can be applied for estimating the cost-optimal cluster composition for running a given Spark job on a cloud under a completion deadline specified in the SLO (i.e., Service Level Objective). We show experimentally that OptEx is able to correctly estimate the required cluster composition for running a given Spark job under a given SLO deadline with an accuracy of 98 percent. We also provide a tool which can classify Spark jobs into job categories based on bisimilarity analysis on lineage graphs collected from the given jobs.
- Published
- 2021
28. AI-Driven Provisioning in the 5G Core
- Author
-
Puneet Sharma, Sonia Fahmy, Lianjie Cao, and Amit Sheoran
- Subjects
Computer Networks and Communications ,Computer science ,business.industry ,Mobile broadband ,Reliability (computer networking) ,Service level objective ,020206 networking & telecommunications ,Provisioning ,02 engineering and technology ,Service provider ,Communications system ,0202 electrical engineering, electronic engineering, information engineering ,Resource management ,business ,5G ,Computer network - Abstract
Network slicing enables communication service providers to partition physical infrastructure into logically independent networks. Network slices must be provisioned to meet the service-level objectives (SLOs) of disparate offerings, such as enhanced mobile broadband, ultrareliable low-latency communications, and massive machine-type communications. Network orchestrators must customize service placement and scaling to achieve the SLO of each network slice. In this article, we discuss the challenges encountered by network orchestrators in allocating resources to disparate 5G network slices, and propose the use of artificial intelligence to make core placement and scaling decisions that meet the requirements of network slices deployed on shared infrastructure. We explore how artificial intelligence-driven scaling algorithms, coupled with functionality-aware placement, can enable providers to design closed-loop solutions to meet the disparate SLOs of future network slices.
- Published
- 2021
29. CEDULE+: Resource Management for Burstable Cloud Instances Using Predictive Analytics
- Author
-
Feng Yan, Ahsan Ali, Riccardo Pinciroli, and Evgenia Smirni
- Subjects
Profiling (computer programming) ,Computer Networks and Communications ,Computer science ,business.industry ,Distributed computing ,Process (computing) ,Service level objective ,020206 networking & telecommunications ,Workload ,Cloud computing ,02 engineering and technology ,Predictive analytics ,Resource (project management) ,0202 electrical engineering, electronic engineering, information engineering ,Resource management ,Electrical and Electronic Engineering ,business - Abstract
Nearly all principal cloud providers now provide burstable instances in their offerings. The main attraction of this type of instance is that it can boost its performance for a limited time to cope with workload variations. Although burstable instances are widely adopted, it is not clear how to efficiently manage them to avoid waste of resources. In this article, we use predictive data analytics to optimize the management of burstable instances. We design CEDULE+, a data-driven framework that enables efficient resource management for burstable cloud instances by analyzing the system workload and latency data. CEDULE+ selects the most profitable instance type to process incoming requests and controls CPU, I/O, and network usage to minimize the resource waste without violating Service Level Objectives (SLOs). CEDULE+ uses lightweight profiling and quantile regression to build a data-driven prediction model that estimates system performance for all combinations of instance type, resource type, and system workload. CEDULE+ is evaluated on Amazon EC2, and its efficiency and high accuracy are assessed through real-case scenarios. CEDULE+ predicts application latency with errors less than 10%, extends the maximum performance period of a burstable instance up to 2.4 times, and decreases deployment costs by more than 50%.
- Published
- 2021
30. Moving Co-Branding to the Web: Service-Level Agreement Implications
- Author
-
Diaz, Oscar, Trujillo, Salvador, Camarinha-Matos, Luis M., editor, and Afsarmanesh, Hamideh, editor
- Published
- 2004
- Full Text
- View/download PDF
31. Dynamic Surge Protection: An Approach to Handling Unexpected Workload Surges with Resource Actions that Have Lead Times
- Author
-
Lassettre, E., Coleman, D. W., Diao, Y., Froehlich, S., Hellerstein, J. L., Hsiung, L., Mummert, T., Raghavachari, M., Parker, G., Russell, L., Surendra, M., Tseng, V., Wadia, N., Ye, P., Goos, Gerhard, editor, Hartmanis, Juris, editor, van Leeuwen, Jan, editor, Brunner, Marcus, editor, and Keller, Alexander, editor
- Published
- 2003
- Full Text
- View/download PDF
32. Alleviating I/O Interference in Virtualized Systems With VM-Aware Persistency Control
- Author
-
Young Ik Eom, Taehyung Lee, and Minho Lee
- Subjects
Hardware_MEMORYSTRUCTURES ,Disk cache flush operation ,General Computer Science ,Computer science ,business.industry ,General Engineering ,sync ,Service level objective ,Cloud computing ,Throughput ,Disk buffer ,computer.software_genre ,map table ,TK1-9971 ,write buffer ,Virtual machine ,Embedded system ,Server ,virtual machine ,General Materials Science ,Electrical engineering. Electronics. Nuclear engineering ,Latency (engineering) ,business ,computer - Abstract
Consolidating multiple servers into a physical machine is now a commonplace in cloud infrastructures. The virtualized systems often arrange virtual disks of multiple virtual machines (VMs) on the same underlying storage device while striving to guarantee the service level objective (SLO) of the performance of each VM. Unfortunately, sync operations called by a VM may make it hard to satisfy the performance SLO by disturbing I/O activities of other VMs. In this paper, we experimentally uncover that the disk cache flush operation incurs significant I/O interference among VMs, and revisit the internal architecture and flush mechanism of the flash memory-based SSD. Then, we present vFLUSH, a novel VM-aware flush mechanism, that supports VM-based persistency control for the disk cache flush operation. We also discuss the long-tail latency issue in vFLUSH and an efficient scheme for mitigating the problem. Our evaluation with various micro- and macro-benchmarks shows that vFLUSH reduces the average latency of disk cache flush operations by up to 58.5%, thereby producing improvements in throughput by up to $1.93\times $ . The method for alleviating the long-tail latency problem, which is applied to vFLUSH, achieves a significant reduction in tail latency by up to 75.9%, with a modest throughput degradation by 2.9–7.2%.
- Published
- 2021
33. A Deep Neural Network-Based Multi-Label Classifier for SLA Violation Prediction in a Latency Sensitive NFV Application
- Author
-
Alan Davy, Nikita Jalodia, and Mohit Taneja
- Subjects
Artificial neural network ,Network function virtualization ,business.industry ,Computer science ,Quality of service ,Reliability (computer networking) ,Distributed computing ,Deep learning ,Service level objective ,deep learning ,Cloud computing ,TK5101-6720 ,neural networks ,Telecommunications network ,machine learning ,classification algorithms ,Scalability ,Telecommunication ,Artificial intelligence ,business ,Transportation and communications ,multi-label classification ,HE1-9990 - Abstract
Recent advancements in the domain of Network Function Virtualization (NFV), and rollout of next-generation networks have necessitated the requirement for the upkeep of latency-critical application architectures in future networks and communications. While Cloud service providers recognize the evolving mission-critical requirements in latency sensitive verticals such as autonomous driving, multimedia, gaming, telecommunications, and virtual reality, there is a wide gap to bridge the Quality of Service (QoS) constraints for the end-user experience. Most latency-critical services are over-provisioned on all fronts to offer reliability, which is inefficient towards scalability in the long run. To address this, we propose a strategy to model frequent violations on the application level as a multi-output target to enable more complex decision-making in the management of virtualised communication networks. In this work, we utilize data from a real-world deployment to configure and draft a realistic set of Service Level Objectives (SLOs) for a voice based NFV application, and develop a deep neural network based multi-label classification methodology to identify and predict multiple categories of SLO breaches associated with an application state. With this, we aim to gain granular SLA and SLO violation insights, enabling us to study and mitigate their impact and inform precision in drafting proactive scaling policies. We further compare the performance against a set of multi-label compatible machine learning classifiers, and address class imbalance in a multi-label setup. We perform a comprehensive evaluation to assess the performance on example-based, label-based and ranking-based measures, and demonstrate the suitability of deep learning in such a use-case.
- Published
- 2021
34. Market-based autonomous resource and application management in private clouds.
- Author
-
Costache, Stefania, Kortas, Samuel, Morin, Christine, and Parlavantzas, Nikos
- Subjects
- *
HIGH performance computing , *CLOUD computing , *MARKETS , *APPLICATION software management , *VIRTUAL machine systems - Abstract
High Performance Computing (HPC) clouds need to be efficiently shared between selfish tenants having applications with different resource requirements and Service Level Objectives (SLOs). The main difficulty relies on providing concurrent resource access to such tenants while maximizing the resource utilization. To overcome this challenge, we propose Merkat, a market-based SLO-driven cloud platform. Merkat relies on a market-based model specifically designed for on-demand fine-grain resource allocation to maximize resource utilization and it uses a combination of currency distribution and dynamic resource pricing to ensure proper resource distribution among tenants. To meet the tenant’s SLO, Merkat uses autonomous controllers, which apply adaptation policies that: (i) dynamically tune the application’s provisioned CPU and memory per virtual machine in contention periods, or (ii) dynamically change the number of virtual machines. Our evaluation with simulation and on the Grid’5000 testbed shows that Merkat provides flexible support for different application types and SLOs and good tenant satisfaction compared to existing centralized systems, while the infrastructure resource utilization is improved. [ABSTRACT FROM AUTHOR]
- Published
- 2017
- Full Text
- View/download PDF
35. SLOC: Service Level Objectives for Next Generation Cloud Computing
- Author
-
Andrea Morichetta, Xiaoning Ding, Schahram Dustdar, Ying Xiong, Thomas Pusztai, Deepak Vij, and Stefan Nastic
- Subjects
Service (systems architecture) ,Computer Networks and Communications ,business.industry ,Computer science ,Quality of service ,Distributed computing ,Service level objective ,020206 networking & telecommunications ,Cloud computing ,02 engineering and technology ,Resource (project management) ,Elasticity (cloud computing) ,Service level ,0202 electrical engineering, electronic engineering, information engineering ,Resource management ,business - Abstract
Since the emergence of cloud computing service level objectives (SLOs) and service level agreements (SLAs) have put themselves forward as one of the key enablers for cloud's on-demand, pay-as-you-go service consumption model. To date, the vast majority of cloud platforms provide support for SLAs only in terms of statically predefined SLOs, e.g., service availability, and low-level resource capacity guarantees, e.g., CPU usage. Unfortunately, there is only limited support to clearly map workload performance requirements to the resource capacity guarantees. In this article, we introduce SLOC— a novel elasticity framework, which promotes a novel performance-driven, SLO-native approach to cloud computing. We outline the main research challenges, vision, and approach of our SLOC framework toward the SLO-native paradigm in next generation cloud computing.
- Published
- 2020
36. Profit model for admission control and resource scheduling for optimum utilization of resources in cloud computing environment
- Author
-
Priyesh Kanungo, Sumant Katiyal, and Nikky Ahuja
- Subjects
Risk analysis (engineering) ,business.industry ,Computer science ,Software as a service ,Service level ,Resource allocation ,Service level objective ,Cloud computing ,Customer satisfaction ,Profit model ,Service provider ,business - Abstract
Globalization has brought many changes in modern business world. Problems of enterprise-oriented software applications like distribution and configuration of resources presents challenge to the traditional software sales model. Cloud computing presents the solution for such problems. With the help of cloud-based model one can earn good return by making the QoS demands of their users (customers) satisfied. In order to fulfill the infrastructure demands (like network, storage, etc.) to their users, service providers have to maintain their own hardware or they lease it from the Iaa S providers. For maintaining their own hardware SaaS providers will have to incur an extra cost, but if taken on lease, zero maintenance cost involved. Moreover, if provider wants to optimize cost and gain customer satisfaction, they will have to satisfy their customers/users by maintaining service level objectives. This paper presents a model for resource allocation in such a way as to minimize cost and maximize profit by satisfying the service level needs of the customers. The model is designed to enable the service providers with an ability to cope with the changing needs of customers, mapping customer requests to infrastructure level parameters and better handling of heterogeneous VMs.
- Published
- 2020
37. PEMC: Power Efficiency Measurement Calculator to Compute Power Efficiency and CO₂ Emissions in Cloud Data Centers
- Author
-
Abdullah Alghamdi, Mohamed A. Elmagzoub, Asadullah Shaikh, and Mueen Uddin
- Subjects
General Computer Science ,business.industry ,Computer science ,020209 energy ,Big data ,General Engineering ,Service level objective ,Cloud computing ,02 engineering and technology ,law.invention ,Reliability engineering ,Calculator ,law ,Power usage effectiveness ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,General Materials Science ,Data center ,Electrical and Electronic Engineering ,business ,Electrical efficiency ,Efficient energy use - Abstract
The power consumption of cloud data centers has a considerable impact on the environment and climate change nowadays. Researchers are seeking to find practical solutions to reduce power consumption in data centers while guaranteeing the desired level of services and service level objectives. With the establishment of the data center industry, the demand for computation and data storage has been continually rising. Energy efficiency is one of the most significant issues faced by these big data centers to meet such high computational requirements. There are many industry acceptable metrics available such as PUE, DCiE, DCP, etc. Power Usage Effectiveness (PUE) metric has proven to be the most popular in measuring energy efficiency; however, it measures the power efficiency alone with no consideration for CO2 emissions and the costs involved in total power usage across data centers. In this article, we proposed a novel Power Efficiency Measurement Calculator (PEMC) that combines and calculates the power efficiency, CO2 emissions, and the total annual costs incurred. The pseudocode and algorithm to perform these specific PUE, DCiE, and CO2 emission functions are given to explain the working of proposed work. Finally, the proposed PEMC calculator was tested and validated through a case study performed in one of the tier-level data centers in Malaysia and the results demonstrate its effectiveness compared with Power Usage Effectiveness (PUE) and other known calculators.
- Published
- 2020
38. Using Service Providers
- Author
-
Adam Freeman
- Subjects
Process management ,Work (electrical) ,Service delivery framework ,Computer science ,Table (database) ,Service level objective ,Dependency injection ,Context (language use) ,Service level requirement ,Service provider - Abstract
In the previous chapter, I introduced services and explained how they are distributed using dependency injection. When using dependency injection, the objects that are used to resolve dependencies are created by service providers, known more commonly as providers. In this chapter, I explain how providers work, describe the different types of provider that are available, and demonstrate how providers can be created in different parts of the application to change the way that services behave. Table 20-1 puts providers in context.
- Published
- 2022
39. Latency-aware failover strategies for containerized web applications in distributed clouds
- Author
-
Yasser Aldwyan and Richard O. Sinnott
- Subjects
Computer Networks and Communications ,business.industry ,Computer science ,Distributed computing ,Interoperability ,Service level objective ,020206 networking & telecommunications ,Cloud computing ,02 engineering and technology ,Failover ,Bottleneck ,Hardware and Architecture ,Backup ,High availability ,0202 electrical engineering, electronic engineering, information engineering ,Web application ,020201 artificial intelligence & image processing ,business ,Software - Abstract
Despite advances in Cloud computing, ensuring high availability (HA) remains a challenge due to varying loads and the potential for Cloud outages. Deploying applications in distributed Clouds can help overcome this challenge by geo-replicating applications across multiple Cloud data centers (DCs). However, this distributed deployment can be a performance bottleneck due to network latencies between users and DCs as well as inter-DC latencies incurred during the geo-replication process. For most web applications, both HA and Performance (HAP) are essential and need to meet pre-agreed Service Level Objectives (SLOs). Efficiently placing and managing primary and backup replicas of applications in distributed Clouds to achieve HAP is a challenging task. Existing solutions consider either HA or performance but not both. In this paper we propose an approach for automating the process of providing a latency-aware failover strategy through a server placement algorithm leveraging genetic algorithms that factor in the proximity of users and inter-DC latencies. To facilitate the distributed deployment of applications and avoid the overheads of Clouds, we utilize container technologies. To evaluate our proposed approach, we conduct experiments on the Australia-wide National eResearch Collaboration Tools and Resources (NeCTAR - www.nectar.org.au) Research Cloud. Our results show at least a 23.3% and 22.6% improvement in response times under normal and failover conditions respectively compared to traditional, latency-unaware approaches. Also, the 95th percentile of response times in our approach are at most1.5 ms above the SLO compared to 11–32 ms using other approaches.
- Published
- 2019
40. Tritium
- Author
-
Fábio Oliveira, Sadie L. Allen, Ayse K. Coskun, Mert Toslali, and Srinivasan Parthasarathy
- Subjects
business.industry ,Computer science ,Analytics ,Reliability (computer networking) ,Distributed computing ,Metric (mathematics) ,Service level objective ,Cloud computing ,Microservices ,business ,Modularity ,Data type - Abstract
Microservice architectures are widely used in cloud-native applications as their modularity allows for independent development and deployment of components. With the many complex interactions occurring in between components, it is difficult to determine the effects of a particular microservice rollout. Site Reliability Engineers must be able to determine with confidence whether a new rollout is at fault for a concurrent or subsequent performance problem in the system so they can quickly mitigate the issue. We present Tritium, a cross-layer analytics system that synthesizes several types of data to suggest possible causes for Service Level Objective (SLO) violations in microservice applications. It uses event data to identify new version rollouts, tracing data to build a topology graph for the cluster and determine services potentially affected by the rollout, and causal impact analysis applied to metric time-series to determine if the rollout is at fault. Tritium works based on the principle that if a rollout is not responsible for a change in an upstream or neighboring SLO metric, then the rollout's telemetry data will do a poor job predicting the behavior of that SLO metric. In this paper, we experimentally demonstrate that Tritium can accurately attribute SLO violations to downstream rollouts and outline the steps necessary to fully realize Tritium.
- Published
- 2021
41. Parslo
- Author
-
Amirhossein Mirhosseini, Thomas F. Wenisch, and Sameh Elnikety
- Subjects
Computer science ,Total cost ,business.industry ,Distributed computing ,Reliability (computer networking) ,Scalability ,Service level objective ,Cloud computing ,Microservices ,Latency (engineering) ,Gradient descent ,business - Abstract
Modern cloud services are implemented as graphs of loosely-coupled microservices to improve programmability, reliability, and scalability. Service Level Objectives (SLOs) define end-to-end latency targets for the entire service to ensure user satisfaction. In such environments, each microservice is independently deployed and (auto-)scaled. However, it is unclear how to optimally scale individual microservices when end-to-end SLOs are violated or underutilized, and how to size each microservice to meet the end-to-end SLO at minimal total cost. In this paper, we propose Parslo---a Gradient Descent-based approach to assign partial SLOs among nodes in a microservice graph under an end-to-end latency SLO. At a high level, the Parslo algorithm breaks the end-to-end SLO budget into small incremental "SLO units", and iteratively allocates one marginal SLO unit to the best candidate microservice to achieve the highest total cost savings until the entire end-to-end SLO budget is exhausted. Parslo achieves a near-optimal solution, seeking to minimize the total cost for the entire service deployment, and is applicable to general microservice graphs that comprise patterns like dynamic branching, parallel fan-out, and microservice dependencies. Parslo reduces service deployment costs by more than 6x in real microservice-based applications, compared to a state-of-the-art partial SLO assignment scheme.
- Published
- 2021
42. OneEdge
- Author
-
Enrique Saurez, Umakishore Ramachandran, Alexandros Daglis, and Harshit Gupta
- Subjects
Situation awareness ,Control theory ,Computer science ,Distributed computing ,Service level objective ,Control reconfiguration ,Resource management ,Enhanced Data Rates for GSM Evolution ,Edge computing ,Scheduling (computing) - Abstract
Resource management for geo-distributed infrastructures is challenging due to the scarcity and non-uniformity of edge resources, as well as the high client mobility and workload surges inherent to situation awareness applications. Due to their centralized nature, state-of-the-art schedulers that work well in datacenters lack the performance and feature requirements of such applications. We present OneEdge, a hybrid control plane that enables autonomous decision-making at edge sites for localized, rapid single-site application deployment. Edge sites handle mobility, churn, and load spikes, by cooperating with a centralized controller that allows coordinated multi-site scheduling and dynamic reconfiguration. OneEdge's scheduling decisions are driven by each application's end-to-end service level objective (E2E SLO) as well as the specific requirements of situation awareness applications. OneEdge's novel distributed state management combines autonomous decision-making at the edge sites for rapid localized resource allocations with decision-making at the central controller when multi-site application deployment is needed. Using a mix of applications on multi-region Azure instances, we show that, in contrast to centralized or fully distributed control planes, OneEdge caters to the unique requirements of situation awareness applications. Compared to a centralized control plane, OneEdge reduces deployment latency by 66% for single-site applications, without compromising E2E SLOs.
- Published
- 2021
43. FSA
- Author
-
Tulika Mitra, Pravein Govindan Kannan, Raj Joshi, Mun Choon Chan, and Nishant Budhdev
- Subjects
Schedule ,Network packet ,business.industry ,Computer science ,ComputerSystemsOrganization_COMPUTER-COMMUNICATIONNETWORKS ,Testbed ,Scalability ,Wireless ,Service level objective ,business ,Multiplexing ,5G ,Computer network - Abstract
5G networks are gaining pace in development and deployment in recent years. One of 5G's key objective is to support a variety of use cases with different Service Level Objectives (SLOs). Slicing is a key part of 5G that allows operators to provide a tailored set of resources to different use cases in order to meet their SLOs. Existing works focus on slicing in the frontend or the C-RAN. However, slicing is missing in the fronthaul network that connects the frontend to the C-RAN. This leads to over-provisioning in the fronthaul and the C-RAN, and also limits the scalability of the network. In this paper, we design and implement Fronthaul Slicing Architecture (FSA), which to the best of our knowledge, is the first slicing architecture for the fronthaul network. FSA runs in the switch dataplane and uses information from the wireless schedule to identify the slice of a fronthaul data packet at line-rate. It enables multipoint-to-multipoint routing as well as packet prioritization to provide multiplexing gains in the fronthaul and the C-RAN, making the system more scalable. Our testbed evaluation using scaled-up LTE traces shows that FSA can support accurate multipoint-to-multipoint routing for 80 Gbps of fronthaul traffic. Further, the slice-aware packet scheduling enabled by FSA's packet prioritization reduces the 95th percentile Flowlet Completion Times (FCT) of latency-sensitive traffic by up to 4 times.
- Published
- 2021
44. Combining task scheduling and data replication for SLA compliance and enhancement of provider profit in clouds
- Author
-
Tarek Hamrouni, Riad Mokadem, Amel Khelifa, Faouzi Ben Charrada, Laboratoire d'Informatique, Programmation, Algorithmique et Heuristique (LIPAH), Faculté des Sciences Mathématiques, Physiques et Naturelles de Tunis (FST), Université de Tunis El Manar (UTM)-Université de Tunis El Manar (UTM), Optimisation Dynamique de Requêtes Réparties à grande échelle (IRIT-PYRAMIDE), Institut de recherche en informatique de Toulouse (IRIT), Université Toulouse 1 Capitole (UT1), Université Fédérale Toulouse Midi-Pyrénées-Université Fédérale Toulouse Midi-Pyrénées-Université Toulouse - Jean Jaurès (UT2J)-Université Toulouse III - Paul Sabatier (UT3), Université Fédérale Toulouse Midi-Pyrénées-Centre National de la Recherche Scientifique (CNRS)-Institut National Polytechnique (Toulouse) (Toulouse INP), Université Fédérale Toulouse Midi-Pyrénées-Université Toulouse 1 Capitole (UT1), Université Fédérale Toulouse Midi-Pyrénées, and Université de Tunis El Manar (UTM)
- Subjects
Data replication ,Operations research ,Computer science ,Cloud computing ,02 engineering and technology ,Bottleneck ,Scheduling (computing) ,Service-level agreement ,Artificial Intelligence ,0202 electrical engineering, electronic engineering, information engineering ,Economic model ,Resource management ,[INFO]Computer Science [cs] ,business.industry ,Service level agreement ,Service level objective ,Triadic concept analysis ,Replication (computing) ,Correlation ,Task scheduling ,[INFO.INFO-PF]Computer Science [cs]/Performance [cs.PF] ,Data access ,Cloud provider ,Profit ,020201 artificial intelligence & image processing ,business - Abstract
International audience; Task scheduling and data replication are highly coupled resource management techniques that are widely used by cloud providers to improve the overall system performance and ensure service level agreement (SLA) compliance while preserving their own economic profit. However, balancing the trade-off between system performance and provider profit is very challenging. In this paper, we propose a novel scheduling algorithm called Bottleneck and Cost Value Scheduling (BCVS) algorithm coupled with a novel dynamic data replication strategy called Correlation and Economic Model-based Replication (CEMR). The main goal is to improve data access effectiveness in order to meet service level objectives in terms of response time SLORT and minimum availability SLOMA, while preserving the provider profit. The BCVS algorithm focuses on reducing system bottleneck situations caused by data transfer when the CEMR focuses on preventing future SLA violations and guaranteeing a minimum availability. An economic model is also proposed to estimate the cloud provider profit. Simulation results indicate that the proposed combination of scheduling and replication algorithms offers higher monetary profit for the cloud provider by up to 30% compared to existing strategies. Moreover, it allows better performance.
- Published
- 2021
45. Memory at Your Service: Fast Memory Allocation for Latency-critical Services
- Author
-
Junxian Zhao, Aidi Pi, Xiaobo Zhou, and Shaoqi Wang
- Subjects
Service (business) ,FOS: Computer and information sciences ,Hardware_MEMORYSTRUCTURES ,Memory sharing ,Computer science ,Service level objective ,computer.software_genre ,Memory management ,Computer Science - Distributed, Parallel, and Cluster Computing ,Benchmark (computing) ,User space ,Operating system ,Batch processing ,Distributed, Parallel, and Cluster Computing (cs.DC) ,Latency (engineering) ,computer - Abstract
Co-location and memory sharing between latency-critical services, such as key-value store and web search, and best-effort batch jobs is an appealing approach to improving memory utilization in multi-tenant datacenter systems. However, we find that the very diverse goals of job co-location and the GNU/Linux system stack can lead to severe performance degradation of latency-critical services under memory pressure in a multi-tenant system. We address memory pressure for latency-critical services via fast memory allocation and proactive reclamation. We find that memory allocation latency dominates the overall query latency, especially under memory pressure. We analyze the default memory management mechanism provided by GNU/Linux system stack and identify the reasons why it is inefficient for latency-critical services in a multi-tenant system. We present Hermes, a fast memory allocation mechanism in user space that adaptively reserves memory for latency-critical services. It advises Linux OS to proactively reclaim memory of batch jobs. We implement Hermes in GNU C Library. Experimental result shows that Hermes reduces the average and the $99^{th}$ percentile memory allocation latency by up to 54.4% and 62.4% for a micro benchmark, respectively. For two real-world latency-critical services, Hermes reduces both the average and the $99^{th}$ percentile tail query latency by up to 40.3%. Compared to the default Glibc, jemalloc and TCMalloc, Hermes reduces Service Level Objective violation by up to 84.3% under memory pressure.
- Published
- 2021
46. On improving service provision through the use of customer-centric semantic service models
- Author
-
Nikolaos Loutas
- Subjects
Service (business) ,Service quality ,Knowledge management ,Customer advocacy ,Customer Service Assurance ,business.industry ,Service design ,Service level objective ,Service level requirement ,Differentiated service ,business - Abstract
Οι τεχνολογίες πληροφοριών και επικοινωνίας (ΤΠΕ) έφεραν επανάσταση στην παροχή υπηρεσιών. Σε αυτή την εργασία, θα εστιάσουμε την προσοχή μας στις υπηρεσίες των οποίων η παροχή υποστηρίζεται από ΤΠΕ, π.χ. μέσω της χρήσης των πληροφοριακών συστημάτων των επιχειρήσεων ή μέσω του διαδικτύου. Οι υπηρεσιοστρεφείς αρχιτεκτονικές (Service Oriented Architecture – SOA) είναι σήμερα το κυρίαρχο πρότυπο για το σχεδιασμό, την ανάπτυξη και την εφαρμογή τέτοιων πληροφοριακών συστημάτων. Παραδοσιακοί κλάδοι υπηρεσιών, όπως ο τουρισμός, η υγεία και η δημόσια διοίκηση, είναι πλέον σε θέση να παρέχουν τις υπηρεσίες τους πιο αποτελεσματικά, μέσω διαφόρων διαύλων, προσαρμοσμένων στις ανάγκες των πελατών τους. Κατά συνέπεια, ένας αυξανόμενος αριθμός υπηρεσιών παρέχεται ηλεκτρονικά μέσω του διαδικτύου, διαμορφώνοντας αυτό που αναφέρεται στη βιβλιογραφία ως διαδίκτυο των υπηρεσιών (Web of Services). Η παροχή υπηρεσιών έχει προχωρήσει αναμφίβολα σε μεγάλο βαθμό. Ωστόσο, εντοπίστηκαν δύο προβλήματα που εμποδίζουν την εξατομικευμένη παροχή υπηρεσιών προσανατολισμένων στον πελάτη, ιδίως στο πλαίσιο του διαδικτύου των υπηρεσιών: i.Την περιορισμένη συμμετοχή των πελατών, ιδιαίτερα στην μοντελοποίηση και στην περιγραφή των υπηρεσιών, και ii.Την ύπαρξη πολλών ανταγωνιστικών ή επικαλυπτόμενων προτύπων και μοντέλων για υπηρεσίες, τα οποία εντέλει δεν είναι πλήρη. Επιστήμες, όπως το μάρκετινγκ και το μάνατζμεντ, ευαγγελίζονται τη σημασία της συμμετοχής του πελάτη καθ' όλη τη διάρκεια του κύκλου ζωής των υπηρεσιών και δίνουν ιδιαίτερη σημασία στην αξιολόγηση αυτών από τους πελάτες. Παρατηρήσαμε όμως ότι οι αρχές αυτές δεν τηρούνται στην παροχή υπηρεσιών στο διαδίκτυο. Οι ερευνητές στο διαδίκτυο των υπηρεσιών υποθέτουν ότι ο πελάτης έχει συνήθως έναν παθητικό ρόλο, που περιορίζεται στην κατανάλωση των υπηρεσιών και όχι απαραίτητα και στο σχεδιασμό τους. Θεωρούν ότι η μοντελοποίηση και η περιγραφή της υπηρεσίας δημιουργείται αποκλειστικά από το πάροχο. Έτσι χάνονται ή δεν εκμεταλλεύονται επαρκώς πολύτιμες πληροφορίες, γνώμες και ανατροφοδότηση που προέρχονται από τους πελάτες. Παράλληλα, διαφορετικές προσπάθειες προτυποποίησης και μοντελοποίησης των υπηρεσιών δημιούργησαν επικαλυπτόμενα, και συχνά διαφορετικά (μη-διαλειτουργικά), μοντέλα για την περιγραφή τους, κανένα από τα οποία δεν λαμβάνει υπόψη την οπτική των πελατών. Στόχος της παρούσας εργασίας είναι να αναπτυχθεί ένα ολιστικό μοντέλο δεδομένων για την υπηρεσίας που θα συνδυάζει τόσο την οπτική του παρόχου όσο και αυτή των πελατών. Πιο συγκεκριμένα, οι στόχοι της εργασίας ήταν: . Να μελετηθεί και να αναλυθεί η έννοια της υπηρεσίας, προκειμένου να δημιουργηθεί ένα κοινό βασικό μοντέλο δεδομένων. II. Να επεκταθεί το κοινό βασικό μοντέλο δεδομένων, προκειμένου να συμπεριλάβει την μοντελοποίηση της οπτικής του πελάτη για την υπηρεσία, δηλαδή πληροφορίες σχετικά με το πώς ο πελάτης την αντιλαμβάνεται και τη βιώνει. III. Να αξιολογηθεί η πληρότητα, η χρηστικότητα και η χρησιμότητα του προαναφερθέντος μοντέλου υπηρεσιών μέσω την ανάπτυξης πιλοτικών εφαρμογών που το χρησιμοποιούν προκειμένου να ικανοποιήσουν τις απαιτήσεις αναφορικά με την αναζήτηση και την εξατομίκευση υπηρεσιών στην περιοχή εφαρμογής της δημόσιας διοίκησης. Τα βασικά θεωρητικά αποτελέσματα αυτής της εργασίας περιλαμβάνουν: 1.Το μοντέλο του κύκλου ζωής της υπηρεσίας από την οπτική του πελάτη. Απεικονίζει τη συμμετοχή των πελατών κατά τις διάφορες φάσεις μιας υπηρεσίας. Περιλαμβάνει πέντε φάσεις, ήτοι: συνειδητοποίηση αναγκών, αναζήτηση, εξατομίκευση, παροχή και αξιολόγηση υπηρεσίας. 2.Το κοινό βασικό μοντέλο δεδομένων για την υπηρεσία (Unified Service Model). Αποτελεί τον κοινό τόπο μεταξύ ενός πλήθους διαφορετικών αλλά και συχνά επικαλυπτόμενων μοντέλων τα οποία επιχειρούν την αναπαράσταση των υπηρεσιών από τη οπτική του παρόχου. 3.Το πελατοκεντρικό μοντέλο υπηρεσιών (Customer Service Model). Αποτελεί μια επέκταση του κοινού βασικού μοντέλου, προκειμένου να συμπεριλάβει την μοντελοποίηση της οπτικής του πελάτη για την υπηρεσία. Το μοντέλο εισάγει έννοιες όπως η θυσία, η προσδοκία του πελάτη, η ανατροφοδότηση και η αντιλαμβανόμενη αξία. Οι έννοιες του μοντέλου μελετώνται και οργανώνονται βάσει των οπτικών του πλαισίου Zachman. 4.Το παράδειγμα των κοινωνικών περιγραφών υπηρεσιών (social descriptions of services). Μια νέα έκφανση για την περιγραφή υπηρεσιών προερχόμενη από τον πελάτη. Οι περιγραφές αυτές εκφράζουν τις προσδοκίες και την ανατροφοδότηση των πελατών, και ενσωματώνουν πληροφορίες από την αξιολόγησή των υπηρεσιών από τους πελάτες. Οι κοινωνικές περιγραφές των υπηρεσιών συμπληρώνουν τις περιγραφές που δημιουργούνται και διατηρούνται από τους παρόχους. Οι βασικές θεωρητικές συμβολές της εργασίας συνοψίζονται στα εξής: •Συμβάλλει στην μοντελοποίηση της υπηρεσίας, η οποία εξακολουθεί να αποτελεί ένα ενεργό ερευνητικό πεδίο και ένα απαιτητικό πεδίο εφαρμογής, μέσω της ανάπτυξης ενός ολιστικού, πελατοκεντρικού μοντέλου, καλύπτοντας την μοντελοποίηση τόσο της ίδιας της υπηρεσίας όσο και του κύκλου ζωής της. •Συμβάλλει στην ικανοποίηση της απαίτησης για πελατοκεντρικές υπηρεσίες, η οποία είναι μια από τις βασικές αρχές που διέπουν το διαδίκτυο των υπηρεσιών. •Τοποθετημένη στην περιοχή της Επιστήμης των Υπηρεσιών (Service Science) συνδυάζει τη μελέτη της υπηρεσίας τόσο από την τεχνική σκοπιά (επιστήμη υπολογιστών) όσο και από την επιχειρηματική (μάρκετινγκ, μάνατζμεντ…), μεταφέροντας έτσι γνώση από το ένα πεδίο στο άλλο. •Με την εισαγωγή ενός κοινού βασικού μοντέλου για την υπηρεσία, το οποίο καλύπτει όλες τις υφιστάμενες προσπάθειες, μειώνει τα σημασιολογικά εμπόδια διαλειτουργικότητας και συμβάλλει στην σημασιολογική διασύνδεση των υπηρεσιών, καθώς και των υφιστάμενων περιγραφών τους που κατασκευάζονται χρησιμοποιώντας διαφορετικά σημασιολογικά μοντέλα.
- Published
- 2021
47. Causal Modeling based Fault Localization in Cloud Systems using Golden Signals
- Author
-
Ajay Gupta, Amit Paradkar, Larisa Shwartz, Qing Wang, Pooja Aggarwal, Prateeti Mohapatra, Seema Nagar, and Atri Mandal
- Subjects
Computer science ,Reliability (computer networking) ,Distributed computing ,Service level objective ,Domain knowledge ,Microservices ,Centrality ,Fault (power engineering) ,Throughput (business) ,Causal model - Abstract
In cloud-native applications, a large fraction of operational failures, known as outages, result in violations of Service Level Objectives (SLOs). SLOs are defined around specific measurable characteristics: availability, throughput, frequency, response time, and quality. Four metrics, latency, traffic, errors, and saturation, ensure coverage for most outages of an application. These are often called golden signals. The dynamicity and complexity of cloud-native applications complicate Site Reliability Engineers’ (SREs) efforts in problem determination, in particular in its fault localization. The fault localization is often a try-and-error process in which SREs rely on their domain knowledge and experience. It is laborious and frequently results in long Mean Time To Resolution (MTTR) for outages. This paper describes a lightweight fault localization system, that establishes causal relationships among the golden signal service errors and error logs, and further leverages PageRank centrality of the derived causal graph for generating a ranked list of faulty microservices.
- Published
- 2021
48. SLO Script: A Novel Language for Implementing Complex Cloud-Native Elasticity-Driven SLOs
- Author
-
Andrea Morichetta, Stefan Nastic, Victor Casamayor Pujol, Thomas Pusztai, Deepak Vij, Schahram Dustdar, Ying Xiong, and Xiaoning Ding
- Subjects
Service (systems architecture) ,Computer science ,business.industry ,Service level objective ,Cloud computing ,Service provider ,computer.software_genre ,Elasticity (cloud computing) ,Type safety ,Object model ,Web service ,Software engineering ,business ,computer - Abstract
Service Level Objectives (SLOs) allow defining expected performance of cloud services, such that cloud service providers know what they guarantee and service consumers know what to expect. Most approaches focus on low-level SLOs, closely related to resources, e.g., average CPU or memory usage, and are usually bound to specific elasticity controllers. We present SLO Script, a language and accompanying framework, motivated by real-world, industrial needs to allow service providers to define complex, high-level SLOs in an orchestrator-independent manner. The main features of SLO Script include: i) novel abstractions (StronglyTypedSLO) with type safety features, ensuring compatibility between SLOs and elasticity strategies, ii) abstractions that enable decoupling of SLOs from elasticity strategies, iii) a strongly typed metrics API, and iv) an orchestrator-independent object model that enables language extensibility. We present a case study about a real-world, cloud-native application and evaluate our language while implementing a realistic Cost Efficiency SLO.
- Published
- 2021
49. Impact of Distributed Rate Limiting on Load Distribution in a Latency-sensitive Messaging Service
- Author
-
Roch Guerin, Chong Li, Christopher Gill, Jiangnan Liu, and Chenyang Lu
- Subjects
Networking and Internet Architecture (cs.NI) ,FOS: Computer and information sciences ,Flexibility (engineering) ,Service (systems architecture) ,Computer science ,business.industry ,Distributed computing ,Control (management) ,Service level objective ,Cloud computing ,Workload ,Computer Science - Networking and Internet Architecture ,Resource (project management) ,Server ,business - Abstract
The cloud's flexibility and promise of seamless auto-scaling notwithstanding, its ability to meet service level objectives (SLOs) typically calls for some form of control in resource usage. This seemingly traditional problem gives rise to new challenges in a cloud setting, and in particular a subtle yet significant trade-off involving load-distribution decisions (the distribution of workload across available cloud resources to optimize performance), and rate limiting (the capping of individual workloads to prevent global over-commitment). This paper investigates that trade-off through the design and implementation of a real-time messaging system motivated by Internet-of- Things (IoT) applications, and demonstrates a solution capable of realizing an effective compromise. The paper's contributions are in both explicating the source of this trade-off, and in demonstrating a possible solution.
- Published
- 2021
50. A Novel Middleware for Efficiently Implementing Complex Cloud-Native SLOs
- Author
-
Schahram Dustdar, Stefan Nastic, Xiaoning Ding, Ying Xiong, Deepak Vij, Victor Casamayor Pujol, Andrea Morichetta, and Thomas Pusztai
- Subjects
Flexibility (engineering) ,Elasticity (cloud computing) ,Cost efficiency ,Control theory ,Computer science ,business.industry ,Middleware ,Distributed computing ,Service level objective ,Provisioning ,Cloud computing ,business - Abstract
Service Level Objectives (SLOs) guide the elasticity of cloud applications, e.g., by deciding when and how much the resources provisioned to an application should be changed. Evaluating SLOs requires metrics, which can be directly measured on the application or system, or, more elaborately, be composed from multiple low-level metrics. The implementation of such metrics and SLOs, the triggering of elasticity strategies, and allowing configurability by the user deploying an application, requires a flexible middleware. In this paper, we present a middleware that provides an orchestrator-independent SLO controller for periodically evaluating SLOs and triggering elasticity strategies, while decoupling SLOs from the elasticity strategies to increase flexibility, and provider-independent services for obtaining low-level metrics and composing them into higher-level metrics. We evaluate our middleware by implementing a motivating use case, featuring a cost efficiency SLO for an application deployed on Kubernetes.
- Published
- 2021
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.