Session B-3


10:00 AM — 11:30 AM EDT
May 4 Wed, 10:00 AM — 11:30 AM EDT

Cutting Tail Latency in Commodity Datacenters with Cloudburst

Gaoxiong Zeng (Hong Kong University of Science and Technology, China); Li Chen (Huawei, China); Bairen Yi (Bytedance, China); Kai Chen (Hong Kong University of Science and Technology, China)

Long tail latency of short flows (or messages) greatly affects user-facing applications in datacenters. Prior solutions to the problem introduce significant implementation complexities, such as global state monitoring, complex network control, or non-trivial switch modifications. While promising superior performance, they are hard to implement in practice.

This paper presents Cloudburst, a simple, effective yet readily deployable solution achieving similar or even better results without introducing the above complexities. At its core, Cloudburst explores forward error correction (FEC) over multipath - it proactively spreads FEC-coded packets generated from messages over multipath in parallel, and recovers them with the first few arriving ones. As a result, Cloudburst is able to obliviously exploit underutilized paths, thus achieving low tail latency. We have implemented Cloudburst as a user-space library, and deployed it on a testbed with commodity switches. Our testbed and simulation experiments show the superior performance of Cloudburst. For example, Cloudburst achieves 63.69% and 60.06% reduction in 99th percentile message/flow completion time (FCT) compared to DCTCP and PIAS, respectively.

EdgeMatrix: A Resources Redefined Edge-Cloud System for Prioritized Services

Yuanming Ren, Shihao Shen, Yanli Ju and Xiaofei Wang (Tianjin University, China); Wenyu Wang (Shanghai Zhuichu Networking Technologies Co., Ltd., China); Victor C.M. Leung (Shenzhen University, China & The University of British Columbia, Canada)

The edge-cloud system has the potential to combine the advantages of heterogeneous devices and truly realize ubiquitous computing. However, for service providers to meet the Service-Level-Agreement (SLA) requirements, the complex networked environment brings inherent challenges such as multi-resource heterogeneity, resource competition, and networked system dynamics. In this paper, we design a framework for the edge-cloud system, namely EdgeMatrix, to maximize the throughput while meeting various SLAs for quality requirements. First, EdgeMatrix introduces Networked Multi-agent Actor-Critic (NMAC) algorithm to redefines physical resources as logically isolated resource combinations, i.e., resource cells. Then, we use a clustering algorithm to group the cells with similar characteristics into various sets, i.e., resource channels, for different channels can offer different SLA guarantees. Besides, we design a multi-task mechanism to solve the problem of joint service orchestration and request dispatch (JSORD) among edge-cloud clusters, significantly reducing the time complexity than traditional methods. To ensure stability, EdgeMatrix adopts a two-time-scale framework, i.e., coordinating resources and services at the large time scale and dispatching requests at the small time scale. The real trace-based experimental results verify that EdgeMatrix can improve system throughput in complex networked environments, reduce SLA violations, and significantly reduce the time complexity than traditional methods.

TRUST: Real-Time Request Updating with Elastic Resource Provisioning in Clouds

Jingzhou Wang, Gongming Zhao, Hongli Xu and Yangming Zhao (University of Science and Technology of China, China); Xuwei Yang (Huawei Technologies, China); He Huang (Soochow University, China)

In a commercial cloud, service providers (e.g., video streaming service provider) rent resources from cloud vendors (e.g., Google Cloud Platform) and provide services to cloud users, making a profit from the price gap. Cloud users acquire services by forwarding their requests to corresponding servers. In practice, as a common scenario, traffic dynamics will cause server overload or load-unbalancing. Existing works mainly deal with the problem by two methods: elastic resource provisioning and request updating. Elastic resource provisioning is a fast and agile solution but may cost too much since service providers need to buy extra resources from cloud vendors. Though request updating is a free solution, it will cause a significant delay, resulting in a bad users' QoS. In this paper, we present a new scheme, called real-time request updating with elastic resource provisioning (TRUST), to help service providers pay less cost with users' QoS guarantee in clouds. In addition, we propose an efficient algorithm for TRUST with a bounded approximation factor based on progressive-rounding. Both small-scale experiment results and large-scale simulation results show the superior performance of our proposed algorithm compared with state-of-the-art benchmarks.

VITA: Virtual Network Topology-aware Southbound Message Delivery in Clouds

Luyao Luo, Gongming Zhao and Hongli Xu (University of Science and Technology of China, China); Liguang Xie and Ying Xiong (Futurewei Technologies, USA)

Southbound message delivery from the control plane to the data plane is one of the essential issues in multi-tenant clouds. A natural method of southbound message delivery is that the control plane directly communicates with compute nodes in the data plane. However, due to the large number of compute nodes, this method may result in massive control overhead. The Message Queue (MQ) model can solve this challenge by aggregating and distributing messages to queues. Existing MQ-based solutions often perform message aggregation based on the physical network topology, which do not align with the fundamental requirements of southbound message delivery, leading to high message redundancy on compute nodes. To address this issue, we design and implement VITA, the first-of-its-kind work on virtual network topology-aware southbound message delivery. However, it is intractable to optimally deliver southbound messages according to the virtual attributes of messages. Thus, we design two algorithms, submodular-based approximation algorithm and simulated annealing-based algorithm, to solve different scenarios of the problem. Both experiment and simulation results show that VITA can reduce the total traffic amount of redundant messages by 45%-75% and reduce the control overhead by 33%-80% compared with state-of-the-art solutions.

Session Chair

Hong Xu (The Chinese University of Hong Kong)

Session B-6

Edge Computing

4:30 PM — 6:00 PM EDT
May 4 Wed, 4:30 PM — 6:00 PM EDT

MoDEMS: Optimizing Edge Computing Migrations For User Mobility

Taejin Kim (Carnegie Mellon University, USA); Sandesh Dhawaskar sathyanarayana (Energy Sciences Network, Lawrence Berkeley National Laboratory & University of Colorado Boulder, USA); Siqi Chen (University of Colorado Boulder, USA); Youngbin Im (Ulsan National Institute of Science and Technology, Korea (South)); Xiaoxi Zhang (Sun Yat-sen University, China); Sangtae Ha (University of Colorado Boulder, USA); Carlee Joe-Wong (Carnegie Mellon University, USA)

Edge computing capabilities in 5G wireless networks promise to benefit mobile users: computing tasks can be offloaded from user devices to nearby edge servers, reducing users' experienced latencies. Few works have addressed how this offloading should handle long-term user mobility: as devices move, they will need to offload to different edge servers, which may require migrating data or state information from one edge server to another. In this paper, we introduce MoDEMS, a system model and architecture that provides a rigorous theoretical framework and studies the challenges of such migrations to minimize the service provider cost and user latency. We show that this cost minimization problem can be expressed as an integer linear programming problem, which is hard to solve due to resource constraints at the servers and unknown user mobility patterns. We show that finding the optimal migration plan is in general NP-hard, and we propose alternative heuristic solution algorithms that perform well in both theory and practice. We finally validate our results with real user mobility traces, ns-3 simulations, and an LTE testbed experiment. MoDEMS based migrations reduce the latency experienced by users of edge applications by 33% compared to previously proposed migration approaches.

Optimal Admission Control Mechanism Design for Time-Sensitive Services in Edge Computing

Shutong Chen (Huazhong University of Science and Technology, China); Lin Wang (VU Amsterdam & TU Darmstadt, The Netherlands); Fangming Liu (Huazhong University of Science and Technology, China)

Edge computing is a promising solution for reducing service latency by provisioning time-sensitive services directly from the network edge. However, upon workload peaks at the resource-limited edge, an edge service has to queue service requests, incurring high waiting time. Such quality of service (QoS) degradation ruins the reputation and reduces the long-term revenue of the service provider.
To address this issue, we propose an admission control mechanism for time-sensitive edge services. Specifically, we allow the service provider to offer admission advice to arriving requests regarding whether to join for service or balk to seek alternatives. Our goal is twofold: maximizing revenue of the service provider and ensuring QoS if the provided admission advice is followed. To this end, we propose a threshold structure that estimates the highest length of the request queue. Leveraging such a threshold structure, we propose a mechanism to balance the trade-off between increasing revenue from accepting more requests and guaranteeing QoS by advising requests to balk. Rigorous analysis shows that our mechanism achieves the goal and that the provided admission advice is optimal for end-users to follow. We further validate our mechanism through trace-driven simulations with both synthetic and real-world service request traces.

Towards Online Privacy-preserving Computation Offloading in Mobile Edge Computing

Xiaoyi Pang (Wuhan University, China); Zhibo Wang (Zhejiang University, China); Jingxin Li and Ruiting Zhou (Wuhan University, China); Ju Ren (Tsinghua University, China); Zhetao Li (Xiangtan University, China)

Mobile Edge Computing (MEC) is a new paradigm where mobile users can offload computation tasks to the nearby MEC server. Some works have pointed out that the true amount of offloaded tasks may reveal the sensitive information of users, and proposed several privacy-preserving offloading mechanisms. However, to the best of our knowledge, none of them can provide strict privacy guarantee. In this paper, we propose a novel online privacy-preserving computation offloading mechanism, called OffloadingGuard, to generate efficient offloading strategies for users in real time, which provide strict user privacy guarantee while minimizing the total cost of task computation. To this end, we design a deep reinforcement learning-based offloading model which allows each user to adaptively determine the satisfactory perturbed offloading ratio according to the time-varying channel state at each time slot to achieve trade-off between user privacy and computation cost. In particular, to strictly protect the true amount of offloaded tasks and prevent the untrusted MEC server from revealing mobile users' privacy, a range-constrained Laplace distribution is designed to obfuscate the original offloading ratio of each user and restrict the perturbed offloading ratio in a rational range. OffloadingGuard is proved to satisfy \epsilon-differential privacy, and extensive experiments demonstrate its effectiveness.

Two Time-Scale Joint Service Caching and Task Offloading for UAV-assisted Mobile Edge Computing

Ruiting Zhou and Xiaoyi Wu (Wuhan University, China); Haisheng Tan (University of Science and Technology of China, China); Renli Zhang (Wuhan University, China)

The emergence of unmanned aerial vehicles (UAVs) extends the mobile edge computing (MEC) services in broader coverage to offer new flexible and low-latency computing services for user equipment (UE) in the era of 5G and beyond. One of the fundamental requirements in UAV-assisted MEC is the low latency, which can be jointly optimized with service caching and task offloading. However, this is challenged by the communication overhead involved with service caching and constrained by limited energy capacity. In this work, we present a comprehensive optimization framework with the objective of minimizing the service latency while incorporating the unique features of UAVs. Specifically, to reduce the caching overhead, we make caching placement decision every T slots (specified by service providers), and adjust UAV trajectory, UE-UAV association, and task offloading decisions at each time slot under the constraints of UAV's energy. By leveraging Lyapunov optimization approach and dependent rounding technique, we design an alternating optimization-based algorithm, named TJSO, which iteratively optimizes caching and offloading decisions. Theoretical analysis proves that TJSO converges to the near-optimal solution in polynomial time. Extensive simulations verify that our solution can reduce the service delay for UEs while maintaining low energy consumption when compared to the three baselines.

Session Chair

Jianli Pan (University of Missouri, St. Louis)

Made with in Toronto · Privacy Policy · INFOCOM 2020 · INFOCOM 2021 · © 2022 Duetone Corp.