Technical Sessions

All sessions will be held in Washington AB unless otherwise noted.

Proceedings Front Matter: 
Cover Page | Title Page and List of Organizers | Table of Contents | Message from the Program Co-Chairs

Full Proceedings PDFs (Download available for all conference attendees)
 ICAC '14 Full Proceedings (PDF)
 ICAC '14 Proceedings Interior (PDF, best for mobile devices)
 Errata Slip (PDF)

Full Proceedings ePub (for iPad and most eReaders)
 ICAC '14 Full Proceedings (ePub)
 Errata Slip (ePub)

Full Proceedings Mobi (for Kindle)
 ICAC '14 Full Proceedings (Mobi)
 Errata Slip (Mobi)

Download Proceedings Archive (Conference Attendees Only)

Attendee Files 
ICAC '14 Proceedings Archive (ZIP)

 

Wednesday, June 18, 2014

8:00 a.m.–8:45 a.m. Wednesday

Continental Breakfast

Washington Foyer

8:45 a.m.–9:00 a.m. Wednesday

Opening Remarks

Program Co-Chairs: Giuliano Casale, Imperial College London, and Xiaohui (Helen) Gu, North Carolina State University

9:00 a.m.–10:30 a.m. Wednesday

Keynote Address I

Session Chair: Xiaoyun Zhu, VMware

Machine Learning Solves Only Half of the Puzzle

Yuanyuan Zhou, University of California, San Diego

As computer systems become ever-so complex to manage and optimize, various machine learning or data mining techniques have become popular in analyzing a large amount of system data. In this talk, I will share my limited experience and challenges we have encountered when exploring these techniques to solve system problems in our research projects and also commercial products.

As computer systems become ever-so complex to manage and optimize, various machine learning or data mining techniques have become popular in analyzing a large amount of system data. In this talk, I will share my limited experience and challenges we have encountered when exploring these techniques to solve system problems in our research projects and also commercial products.

Yuanyuan Zhou is currently a Qualcomm Chair Professor at the University of California, San Diego. Before UCSD, she was a tenured associate professor at University of Illinois at Urbana Champaign. Her research interests span the areas of operating systems, software engineering, system reliability and maintainability. She has co-founded three startups. Her recent startup, PatternInsight, has successfully deployed software quality assurance tools in many companies. In 2012, its Log Insight business line was acquired by VMware and now Log Insight is a VMware Product offering to its many data center customers for data center management. Dr. Zhou is an ACM Fellow and obtained her Ph.D. from Princeton. She has the great fortune of working with many talented students and colleagues.

10:30 a.m.–11:00 a.m. Wednesday

Break with Refreshments

Washington Foyer

11:00 a.m.–12:15 p.m. Wednesday

Model-Driven Management and Self-Adaptation

Session Chair: Dilma Da Silva, Qualcomm Research

Storage Workload Isolation via Tier Warming: How Models Can Help

Ji Xue and Feng Yan, College of William and Mary; Alma Riska, EMC Corporation; Evgenia Smirni, College of William and Mary

Storage systems are often deployed in a tiered form to enable high performance and availability. These tiers utilize all possible volatile and non-volatile storage technologies, including DRAM, SSD, and HDD. The tradeoffs among their cost, features, and capabilities can make their effective integration into a single storage entity complex. Here, we propose an autonomic technique that learns user traffic patterns in a storage system over long time-scales to optimize user performance but also volume of completed system work. Our purpose is to multiplex as best as possible user workload with storage system features (e.g., voluminous internal system work) such that the latter is not starved but rather completed with minimal impact on user performance. Key to achieving the above is to use an autonomic learning engine to predict when the user workload intensity increases/ decreases and then proactively stop/start bulky internal system work. Being proactive allows the system to effectively bring into the fast tier the active user working set just-in-time and right before it is needed most, i.e., when user traffic suddenly peaks. We illustrate the effectiveness of this mechanism by using both trace driven simulations from production systems as well experiments on a real testbed.

Available Media

Model-driven Elasticity and DoS Attack Mitigation in Cloud Environments

Cornel Barna and Mark Shtern, York University; Michael Smit, Dalhousie University; Hamoun Ghanbari and Marin Litoiu, York University

Workloads for web applications can change rapidly. When the change is an increase in customers, a common adaptive approach to maintain SLAs is elasticity, the on-demand allocation of computing resources. However, application-level denial-of-service (DoS) attacks can also cause changes in workload, and require an entirely different response. These two issues are often addressed separately (in both research and application). This paper presents a model-driven adaptive management mechanism which can correctly scale a web application, mitigate a DoS attack, or both, based on an assessment of the business value of workload. This approach is enabled by modifying a layered queuing network model previously used to model data centers to also accurately predict short-term cloud behavior, despite cloud variability over time. We evaluate our approach on Amazon EC2 and demonstrate the ability to horizontally scale a sample web application in response to an increase in legitimate traffic while mitigating multiple DoS attacks, achieving the established performance goal.

Available Media

Integrating Adaptation Mechanisms Using Control Theory Centric Architecture Models: A Case Study

Filip Křikava, University of Lille 1 and Inria; Philippe Collet, Université Nice Sophia Antipolis; Romain Rouvoy, University of Lille 1 and Inria

Control theory provides solid foundations for developing reliable and scalable feedback control for software systems. Although, feedback controllers have been acknowledged to efficiently solve common classes of problems, their adoption by state-of-the-art approaches for designing self-adaptation in legacy software systems remains limited and at best consists in ad hoc integrations, which are usually engineered manually.

In this paper, we revisit the Znn.com case study and we present an alternative implementation based on classical feedback controllers. We show how these controllers can be easily integrated into software systems through control theory centric architecture models and domainspecific modeling support. We also provide an assessment of the resulting properties, quality attributes and limitations.

Available Media
12:15 p.m.–2:00 p.m. Wednesday

FCW '14 Luncheon

Grand Ballroom ABC

2:00 p.m.–3:30 p.m. Wednesday

Cloud Resource Management

Session Chair: Christopher Stewart, The Ohio State University

ShuttleDB: Database-Aware Elasticity in the Cloud

3:15 pm

Sean Barker, University of Massachusetts Amherst; Yun Chi, Square Inc.; Hakan Hacιgumuş, NEC Laboratories America; Prashant Shenoy and Emmanuel Cecchet, University of Massachusetts Amherst

Motivated by the growing popularity of database-as-aservice clouds, this paper presents ShuttleDB, a holistic approach enabling flexible, automated elasticity of database tenants in the cloud. We first propose a database-aware live migration and replication method designed to work with off-the-shelf databases without any database engine modifications. We then combine these database-aware techniques with VM-level mechanisms to implement a flexible elasticity approach that can achieve efficient scale up, scale out, or scale back for diverse tenants with fluctuating workloads. Our experimental evaluation of the ShuttleDB prototype shows that by applying migration and replication techniques at the tenant level, automated elasticity can be achieved both intra- and inter-datacenter in a database agnostic way. We further show that ShuttleDB can reduce the time and data transfer needed for elasticity by 80% or more compared to tenant-oblivious approaches.

Available Media

Matrix: Achieving Predictable Virtual Machine Performance in the Clouds

2:15 pm

Ron C. Chiang, The George Washington University; Jinho Hwang, IBM T. J. Watson Research Center; H. Howie Huang and Timothy Wood, The George Washington University

The success of cloud computing builds largely upon on-demand supply of virtual machines (VMs) that provide the abstraction of a physical machine on shared resources. Unfortunately, despite recent advances in virtualization technology, there still exists an unpredictable performance gap between the real and desired performance. The main contributing factors include contention to the shared physical resources among co-located VMs, limited control of VM allocation, as well as lack of knowledge on the performance of a specific VM out of tens of VM types offered by public cloud providers. In this work, we propose Matrix, a novel performance and resource management system that ensures the desired performance of an application achieved on a VM. To this end, Matrix utilizes machine learning methods - clustering models with probability estimates - to predict the performance of new workloads in a virtualized environment, choose a suitable VM type, and dynamically adjust the resource configuration of a virtual machine on the fly. The evaluations on a private cloud, and two public clouds (Rackspace and Amazon EC2) show that for an extensive set of cloud applications, Matrix is able to estimate application performance with average 90% accuracy. In addition, Matrix can deliver the target performance within 3% variance, and do so with the best cost-efficiency in most cases.

Available Media

Adaptive, Model-driven Autoscaling for Cloud Applications

Anshul Gandhi, Parijat Dube, Alexei Karve, Andrzej Kochut, and Li Zhang, IBM Research

Applications with a dynamic workload demand need access to a flexible infrastructure to meet performance guarantees and minimize resource costs. While cloud computing provides the elasticity to scale the infrastructure on demand, cloud service providers lack control and visibility of user space applications, making it difficult to accurately scale the underlying infrastructure. Thus, the burden of scaling falls on the user.

In this paper, we propose a new cloud service, Dependable Compute Cloud (DC2), that automatically scales the infrastructure to meet the user-specified performance requirements. DC2 employs Kalman filtering to automatically learn the (possibly changing) system parameters for each application, allowing it to proactively scale the infrastructure to meet performance guarantees. DC2 is designed for the cloud - it is application-agnostic and does not require any offline application profiling or benchmarking. Our implementation results on OpenStack using a multi-tier application under a range of workload traces demonstrate the robustness and superiority of DC2 over existing rule-based approaches.

Available Media

Exploring Graph Analytics for Cloud Troubleshooting

Chengwei Wang, Karsten Schwan, Brian Laub, Mukil Kesavan, and Ada Gavrilovska, Georgia Institute of Technology

We propose VFocus, a platform which uses streaming graph analytics to narrow down the search space for troubleshooting and management in large scale data centers. This paper describes useful guidance operations which are realized with graph analytics and validated with representative use cases. The first case is based on real data center traces to measure the performance of troubleshooting operations supported by VFocus. In the second use case, the utility of VFocus is demonstrated by detecting data hotspots in a big data stream processing application. Experimental results show that VFocus guidance operations can troubleshoot Virtual Machine (VM) migration failures with accuracy of 83% and with delays of only hundreds of milliseconds when tracking migrations on 256 servers hosing 1024 VMs. Such successes are achieved with negligible runtime overheads and low perturbation for applications, in comparison to brute-force approaches.

Available Media
3:30 p.m.–4:00 p.m. Wednesday

Break with Refreshments

Washington Foyer

4:00 p.m.–5:30 p.m. Wednesday

Network and System Management

Session Chair: Cristian Lumezanu, NEC Labs

Inferring Origin Flow Patterns in Wi-Fi with Deep Learning

3:15 pm

Youngjune L. Gwon and H. T. Kung, Harvard University

We present a novel application of deep learning in networking. The envisioned system can learn the original flow characteristics such as a burst size and inter-burst gaps conceived at the source from packet sampling done at a receiverWi-Fi node. This problem is challenging because CSMA introduces complex, irregular alterations to the origin pattern of the flow in the presence of competing flows. Our approach is semi-supervised learning. We first work through multiple layers of feature extraction and subsampling from unlabeled flow measurements.We use a feature extractor based on sparse coding and dictionary learning, and our subsampler performs overlapping max pooling. Given the layers of learned feature mapping, we train SVM classifiers with deep feature representation resulted at the top layer. The proposed scheme has been evaluated empirically in a custom wireless simulator and OPNET. The results are promising that we achieve superior classification performance over ARMAX, Naïve Bayes classifiers, and Gaussian mixture models optimized by the EM algorithm.

Available Media

Guarded Modules: Adaptively Extending the VMM's Privilege Into the Guest

Kyle C. Hale and Peter A. Dinda, Northwestern University

When a virtual machine monitor (VMM) provides code that executes in the context of a guest operating system, allowing that code to have privileged access to specific hardware and VMM resources can enable new mechanisms to enhance functionality, performance, and adaptability. We present a software technique, guarded execution of privileged code in the guest, that allows the VMM to provide this capability, as well as an implementation for Linux guests in the Palacios VMM. Our system, which combines compile-time, link-time, and runtime techniques, provides the module developer with the following guarantees: (1) A kernel module will remain unmodified and it will acquire privilege only when untrusted code invokes it through developer-chosen, valid entry points with a valid stack. (2) Any execution path leaving the module will trigger a revocation of privilege. (3) The module has access to private memory. The system also provides the administrator with a secure method to bind a specific module with particular privileges implemented by the VMM. This lays the basis for guaranteeing that only trusted code in the guest can utilize special privileges. We give two examples of guarded Linux kernel modules: a network interface driver with direct access to the physical NIC and an idle loop that uses instructions not usually permitted in a guest, but which can be adaptively selected when no other virtual core shares the physical core. In both cases only the guarded module has these privileges.

Available Media

Active Control of Memory for Java Virtual Machines and Applications

2:30 pm

Norman Bobroff, Peter Westerink, and Liana Fong, IBM T. J. Watson Research Center

A controller is implemented to manage memory as an elastic resource similar to computing cycles for Java applications. The controller actively arbitrates constrained memory between collocated JVMs in response to demand. A key aspect of the work is that JVM metrics are used as proxies for application KPIs so that application performance instrumentation and modeling are not required. A metric corresponding to the allocation rate of memory is derived from the JVM metrics and established as the measure of application performance and is used as the effective feedback mechanism to the controller. The controller is based on a fair share policy in which memory is distributed to equalize the marginal performance value to all JVMs. The design is tested for effectiveness and stability using the suite of SPECjvm2008 and SPECjbb2005 benchmarks.

Available Media

Is Your Web Server Suffering from Undue Stress due to Duplicate Requests?

Fahad A. Arshad, Amiya K. Maji, Sidharth Mudgal, and Saurabh Bagchi, Purdue University

An important, if not very well known, problem that afflicts many web servers is duplicate client browser requests due to server-side problems. A legitimate request is followed by a redundant request, thus increasing the load on the server and corrupting state at the server end (such as, the hit count for the page) and at the client end (such as, state maintained through a cookie). This problem has been reported in many developer blogs and has been found to afflict even popular web sites, such as CNN and YouTube. However, to date, there has not been a scientific, technical solution to this problem that is browser vendor neutral. In this paper, we provide such a solution which we call GRIFFIN. We identify that the two root causes of the problem are missing resource at the server end or duplicated Javascripts embedded in the page. We have the insight that dynamic tracing of the function call sequence creates a signature that can be used to differentiate between legitimate and duplicate requests. We apply our technique to find unreported problems in a large production scientific collaboration web service called HUBzero, which are fixed upon reporting the problems. Our experiments show an average overhead of 1.29X for tracing the PHP-runtime on HUBzero across 60 unique HTTP transactions. GRIFFIN has zero false-positives (when run across HTTP transaction of size one and two) and an average detection accuracy of 78% across 60 HTTP transactions.

Available Media
6:30 p.m.–8:00 p.m. Wednesday

Wednesday Reception

Grand Ballroom AB

 

Thursday, June 19, 2014

8:00 a.m.–9:00 a.m. Thursday

Continental Breakfast

Washington Foyer

9:00 a.m.–10:30 a.m. Thursday

Keynote Address II

Session Chair: Giuliano Casale, Imperial College London

The Enterprise and Big Data Systems: Yesterday, Today, and Tomorrow

Lucy Cherkasova, HP Labs

Processing ever-increasing amounts of information and providing a meaningful analysis of large datasets (Big Data) has become a significant computing challenge in the Enterprise environment. New tools, frameworks, and systems have been proposed for Big Data processing. They target a variety of data (everything from business transactions to sensor data to tweets) and aim to offer new useful insights via advanced real-time analytics and/or batch-driven data analysis. The common theme of these underlying systems is that they represent a scale-out approach on commodity machines. Using a MapReduce framework I will present and analyze challenges in performance management of such systems. I will talk about the community and enterprise efforts to design unified and/or integrated data processing frameworks that aim to simplify application development and enhance data analytics.

Processing ever-increasing amounts of information and providing a meaningful analysis of large datasets (Big Data) has become a significant computing challenge in the Enterprise environment. New tools, frameworks, and systems have been proposed for Big Data processing. They target a variety of data (everything from business transactions to sensor data to tweets) and aim to offer new useful insights via advanced real-time analytics and/or batch-driven data analysis. The common theme of these underlying systems is that they represent a scale-out approach on commodity machines. Using a MapReduce framework I will present and analyze challenges in performance management of such systems. I will talk about the community and enterprise efforts to design unified and/or integrated data processing frameworks that aim to simplify application development and enhance data analytics. Finally, I will discuss hardware and resource usage patterns imposed by modern and emerging scale-out applications and their possible impact on the future system design.

Dr. Ludmila Cherkasova is a principal scientist in the Systems Research Lab, at Hewlett-Packard Labs, Palo Alto, USA. Her research interests include the analysis, design, and management of concurrent and distributed systems (such as emerging systems for Big Data processing, internet and media applications, virtualized environments, and next generation data centers). She has authored over 100 referred publications and more than 70 patent applications. Over the years she has mentored and co-advised more than 15 PhD students/interns. She has served as a Track/PC co-chair of 6 International conferences. She is an ACM Distinguished Scientist and recognized by the Certificate of Appreciation from the IEEE Computer Society. She has earned 6 Best Paper awards. Her most recent works were on the design of performance analysis and optimization techniques for MapReduce environments.

10:30 a.m.–11:00 a.m. Thursday

Break with Refreshments

Washington Foyer

11:00 a.m.–12:30 p.m. Thursday

MBDS Panel

Session Chair: Karsten Schwan, Georgia Institute of Technology

Panelists: Lucy Cherkasova, HP Labs; Hubertus Franke, IBM T. J. Watson Research Center; Alex Moundalexis, Cloudera; Christopher Stewart, The Ohio State University


12:30 p.m.–2:00 p.m. Thursday

FCW '14 Luncheon

Grand Ballroom ABC

2:00 p.m.–3:30 p.m. Thursday

MBDS Track

Session Chair: Karsten Schwan, Georgia Institute of Technology

A Model-Based Namespace Metadata Benchmark for HDFS

Cristina L. Abad, Escuela Superior Politécnica del Litoral; Yi Lu and Roy H. Campbell, University of Illinois at Urbana-Champaign; Nathan Roberts, Yahoo, Inc.

Efficient namespace metadata management is increasingly important as next-generation storage systems are designed for peta and exascales. New schemes have been proposed; however, their evaluation has been insufficient due to a lack of an appropriate namespace metadata benchmark. We describe MimesisBench, a novel namespace metadata benchmark for next-generation storage systems, and demonstrate its usefulness through a study of the scalability and performance of the Hadoop Distributed File System (HDFS).

Available Media

Towards Combining Online & Offline Management for Big Data Applications

3:15 pm

Brian Laub, Chengwei Wang, Karsten Schwan, and Chad Huneycutt, Georgia Institute of Technology

Traditional data center monitoring systems focus on collecting basic metrics such as CPU and memory usage, in a centralized location, giving administrators a summary of global system health via a database of observations. Conversely, emerging research systems are focusing on scalable, distributed monitoring capable of quickly detecting and alerting administrators to anomalies. This paper outlines VStore, a system that seeks to combine fast online anomaly detection with offline storage and analysis of monitoring data. VStore can be used as a historical reference to help guide administrators towards quickly classifying and fixing anomalous behavior once a problem has been detected. We demonstrate this idea with a distributed big streaming data application, and explore three common fault scenarios in this application. We show that each scenario exhibits a slightly different monitoring history, which may be undetectable by online algorithms that are resource-constrained. We also offer a discussion of how historical data captured by VStore can be combined with online monitoring tools to improve troubleshooting efforts in the data center.

Available Media

An Enterprise Dynamic Thresholding System

Mazda A. Marvasti, Arnak V. Poghosyan, Ashot N. Harutyunyan, and Naira M. Grigoryan, VMware, Inc.

We demonstrate an enterprise Dynamic Thresholding System for data-agnostic management of monitoring flows. The dynamic thresholding based on data historical behavior enables adaptive and more accurate control of business environments compared to static thresholding. We manifest the main blocks of a complex analytical engine that is implemented in VMware vCenter Operations Manager as a principal foundation of the company’s data-driven anomaly detection.

Available Media

User-Centric Heterogeneity-Aware MapReduce Job Provisioning in the Public Cloud

Eric Pettijohn and Yanfei Guo, University of Colorado, Colorado Springs; Palden Lama, University of Texas at San Antonio; Xiaobo Zhou, University of Colorado, Colorado Springs

Cloud datacenters are becoming increasingly heterogeneous with respect to the hardware on which virtual machine (VM) instances are hosted. As a result, ostensibly identical instances in the cloud show significant performance variability depending on the physical machines that host them. In our case study on Amazon’s EC2 public cloud, we observe that the average execution time of Hadoop MapReduce jobs vary by up to 30% in spite of using identical VM instances for the Hadoop cluster. In this paper, we propose and develop U-CHAMPION, a user-centric middleware that automates job provisioning and configuration of the Hadoop MapReduce framework in a public cloud to improve job performance and reduce the cost of leasing VM instances. It addresses the unique challenges of hardware heterogeneity-aware job provisioning in the public cloud through a novel selective-instance-reacquisition technique. It applies a collaborative filtering technique based on UV Decomposition for online estimation of ad-hoc job execution time. We have implemented U-CHAMPION on Amazon EC2 and compared it with a representative automated MapReduce job provisioning system. Experimental results with the PUMA benchmarks show that U-CHAMPION improves MapReduce job performance and reduces the cost of leasing VM instances by as much as 21%.

Available Media
3:30 p.m.–4:00 p.m. Thursday

Break with Refreshments

Washington Foyer

4:00 p.m.–5:30 p.m. Thursday

SCPS Track

Session Chair: Sokwoo Rhee, NIST

Exploiting Temporal Diversity of Water Efficiency to Make Data Center Less "Thirsty"

3:30 pm

Mohammad A. Islam, Kishwar Ahmed, Shaolei Ren, and Gang Quan, Florida International University

Data centers, which include both cyber (e.g., servers) and physical (e.g., cooling units) assets, are notorious for their energy consumption and carbon footprint. Nonetheless, a less-known fact about data centers is that they are extremely “thirsty” (for cooling), consuming millions of gallons water each day and raising serious concerns amid extended droughts. To curtail the surging water footprint, we adopt a holistic cyber-physical approach and incorporate the inherent physical characteristic of data center— time-varying water efficiency — into server provisioning and workload management. Specifically, we propose an online batch job scheduling algorithm, called WACE (minimization of WAter, Carbon and Electricity cost), which dynamically adjusts server provisioning to reduce the water consumption by deferring delay-tolerant batch jobs to water-efficient time periods. We demonstrate the effectiveness of WACE via trace-based simulations, showing that WACE reduces 27% water consumption compared to state-of-the-art scheduling algorithms.

Available Media

Real-time Edge Analytics for Cyber Physical Systems using Compression Rates

2:30 pm

Sokratis Kartakis and Julie A. McCann, Imperial College London

There is a movement in many practical applications of Cyber-Physical Systems to push processing to the edge. This is particularly important were the CPS is carrying out monitoring and control, where the latency between the decision making and control message reception should be minimal. However, CPS are limited by the capabilities of the typically battery powered low resourced devices. In this paper we present a self-adaptive scheme that both reduces the amount of resources required to store high sample rate data at the edge and at the same time carries out initial data analytics. Using out Smart Water datasets, plus a selection from other real world CPS applications, we show that our algorithm reduces computation by 98%; data volumes by 55%; while requiring only 11KB of memory at runtime (including the compression algorithm). In addition we show that our system supports self-tuning and automatic reconfiguration which means that manual tuning is alleviated and the scheme can be both applied to any kind of raw data automatically and is able self-optimize as the nature of the incoming data changes over time.

Available Media

Self-Optimizing Citizen-Centric Mobile Urban Sensing Systems

Usman Adeel, Shusen Yang, and Julie A. McCann, Imperial College London

In this paper, we develop a novel networking scheme that supports both real-time and delay-tolerant urban sensing applications. This maintains optimality through self-adapting its communications strategy using either inexpensive short-range opportunistic transmissions or reliable long-range cellular radios. Core to this scheme is the trading of mobile sensor data in a virtual market where we demonstrate that our scheme can incentivize phone users to participate. We show that the scheme can optimise network throughput while minimising total phone costs, in terms of 3G and battery costs.

Available Media

Gait Recognition using Encodings with Flexible Similarity Metrics

Michael B. Crouse, Kevin Chen, and H.T. Kung, Harvard University

Gait signals detectable by sensors on ubiquitous personal devices such as smartphones can reveal characteristics unique to each individual, and thereby offer a new approach to recognizing users. Conventional pattern matching approaches use inner-product based distance measures which are not robust to common variations in time-series analysis (e.g., shifts and stretching). This is unfortunate given that it is well understood that capturing such variations is paramount for model performance. This work shows how machine learning methods which encode gait signals into a feature space based on a dictionary can use convolution and Dynamic TimeWarping (DTW) similarity measures to improve classification accuracy in a variety of situations common to gait recognition. We also show that data augmentation is crucial in gait recognition, as diverse training data in practical applications is very limited. We validate the effectiveness of these methods empirically, and demonstrate the identification of user gait patterns where shift and stretch variations in measurements are substantial. We present a new gait dataset that contains a complete representation of the variations that can be expected in real-world recognition scenarios. We compare our techniques against the current state of the art gait period detection and normalization schemes on our dataset and show improved classification accuracy under all experimental scenarios.

Available Media
6:30 p.m.–8:00 p.m. Thursday

USENIX ATC '14 and ICAC '14 Poster Session and Reception

Grand Ballroom AB

View the list of accepted posters.

 

Friday, June 20, 2014

8:00 a.m.–9:00 a.m. Friday

Continental Breakfast

Washington Foyer

9:00 a.m.–10:30 a.m. Friday

Keynote Address III

Session Chair: Xiaohui (Helen) Gu, North Carolina State University

Conquering Big Data with Spark and BDAS

Ion Stoica, University of California, Berkeley

Ion Stoica is a Professor in the EECS Department at University of California at Berkeley. He received his PhD from Carnegie Mellon University in 2000. He does research on cloud computing and networked computer systems. Past work includes the Dynamic Packet State (DPS), Chord DHT, Internet Indirection Infrastructure (i3), declarative networks, replay-debugging, and multi-layer tracing in distributed systems. His current research focuses on resource management and scheduling for data centers, cluster computing frameworks, and network architectures. He is an ACM Fellow and has received numerous awards, including the SIGCOMM Test of Time Award (2011), and the ACM doctoral dissertation award (2001). In 2006, he co-founded Conviva, a startup to commercialize technologies for large scale video distribution, and in 2013, he co-founded Databricks as startup to commercialize technologies for Big Data processing.

Today, big and small organizations alike collect huge amounts of data, and they do so with one goal in mind: extract "value" through sophisticated exploratory analysis, and use it as the basis to make decisions as varied as personalized treatment and ad targeting. Unfortunately, existing data analytics tools are slow in answering queries, as they typically require to sift through huge amounts of data stored on disk, and are even less suitable for complex computations, such as machine learning algorithms. These limitations leave the potential of extracting value of big data unfulfilled.

Today, big and small organizations alike collect huge amounts of data, and they do so with one goal in mind: extract "value" through sophisticated exploratory analysis, and use it as the basis to make decisions as varied as personalized treatment and ad targeting. Unfortunately, existing data analytics tools are slow in answering queries, as they typically require to sift through huge amounts of data stored on disk, and are even less suitable for complex computations, such as machine learning algorithms. These limitations leave the potential of extracting value of big data unfulfilled.

To address this challenge, we are developing Berkeley Data Analytics Stack (BDAS), an open source data analytics stack that provides interactive response times for complex computations on massive data. To achieve this goal, BDAS supports efficient, large-scale in-memory data processing, and allows users and applications to trade between query accuracy, time, and cost. In this talk, I'll present the architecture, challenges, results, and our experience with developing BDAS, with a focus on Apache Spark, an in-memory cluster computing engine that provides support for a variety of workloads, including batch, streaming, and iterative computations. In a relatively short time, Spark has become the most active big data project in the open source community, and is already being used by over one hundred of companies and research institutions.

Ion Stoica is a Professor in the EECS Department at University of California at Berkeley. He received his PhD from Carnegie Mellon University in 2000. He does research on cloud computing and networked computer systems. Past work includes the Dynamic Packet State (DPS), Chord DHT, Internet Indirection Infrastructure (i3), declarative networks, replay-debugging, and multi-layer tracing in distributed systems. His current research focuses on resource management and scheduling for data centers, cluster computing frameworks, and network architectures. He is an ACM Fellow and has received numerous awards, including the SIGCOMM Test of Time Award (2011), and the ACM doctoral dissertation award (2001). In 2006, he co-founded Conviva, a startup to commercialize technologies for large scale video distribution, and in 2013, he co-founded Databricks as startup to commercialize technologies for Big Data processing.

10:30 a.m.–11:00 a.m. Friday

Break with Refreshments

Washington Foyer

11:00 a.m.–12:15 p.m. Friday

Scheduling, Pricing, and Incentive

Session Chair: Evangelia Kalyvianaki, City University London

On-demand, Spot, or Both: Dynamic Resource Allocation for Executing Batch Jobs in the Cloud

Ishai Menache, Microsoft Research; Ohad Shamir, Weizmann Institute; Navendu Jain, Microsoft Research

Cloud computing provides an attractive computing paradigm in which computational resources are rented on-demand to users with zero capital and maintenance costs. Cloud providers offer different pricing options to meet computing requirements of a wide variety of applications. An attractive option for batch computing is spot-instances, which allows users to place bids for spare computing instances and rent them at a (often) substantially lower price compared to the fixed on-demand price. However, this raises three main challenges for users: how many instances to rent at any time? what type (on-demand, spot, or both)? and what bid value to use for spot instances? In particular, renting on-demand risks high costs while renting spot instances risks job interruption and delayed completion when the spot market price exceeds the bid. This paper introduces an online learning algorithm for resource allocation to address this fundamental tradeoff between computation cost and performance. Our algorithm dynamically adapts resource allocation by learning from its performance on prior job executions while incorporating history of spot prices and workload characteristics. We provide theoretical bounds on its performance and prove that the average regret of our approach (compared to the best policy in hindsight) vanishes to zero with time. Evaluation on traces from a large datacenter cluster shows that our algorithm outperforms greedy allocation heuristics and quickly converges to a small set of best performing policies.

Available Media

Real-Time Scheduling of Skewed MapReduce Jobs in Heterogeneous Environments

2:45 pm

Nikos Zacheilas and Vana Kalogeraki, Athens University of Economics and Business

Supporting real-time jobs on MapReduce systems is particularly challenging due to the heterogeneity of the environment, the load imbalance caused by skewed data blocks, as well as real-time response demands imposed by the applications. In this paper we describe our approach for scheduling real-time, skewed MapReduce jobs in heterogeneous systems. Our approach comprises the following components: (i) a distributed scheduling algorithm for scheduling real-time MapReduce jobs endto- end, and (ii) techniques for handling the data skewness that frequently arises in MapReduce environments and can lead to significant load imbalances. Our detailed experimental results using real datasets on a truly heterogeneous environment, Planetlab, illustrate that our approach is practical, exhibits good performance and consistently outperforms its competitors.

Available Media

Colocation Demand Response: Why Do I Turn Off My Servers?

3:30 pm

Shaolei Ren and Mohammad A. Islam, Florida International University

Data centers are promising participants in demand response programs (i.e., reducing a large electricity demand upon utility’s request), making power grid more stable and sustainable. In this paper, we focus on enabling colocation data center demand response. Colocation is an integral yet unique segment of data center industry, where multiple tenants house their servers in one shared facility. Nonetheless, differing from owner-operated data centers (e.g., Google), colocation data center suffers from “split incentive”: colocation operator desires demand response for financial incentives but has no control over tenants’ servers, while tenants who own the servers may not desire demand response due to lack of incentives. To break “split incentive”, we propose a first-of-its-kind incentive mechanism, called iCODE (incentivizing COlocation tenants for DEmand response), based on reverse auction: tenants, who voluntarily submit energy reduction bids to colocation operator, will be financially rewarded if their bids are accepted. We formally model how each tenant decides its bids and how colocation operator decides winning bids. We perform a trace-based simulation to evaluate iCODE. We show that iCODE can reduce colocation energy consumption by over 50% during demand response periods, unleashing the potential of colocation demand response.

Available Media
12:15 p.m.–2:00 p.m. Friday

FCW '14 Luncheon

Grand Ballroom ABC

2:00 p.m.–3:30 p.m. Friday

Resource and Workload Management

Session Chair: Howie Huang, The George Washington University

Self-Tuning Intel Transactional Synchronization Extensions

Nuno Diegues and Paolo Romano, INESC-ID and Instituto Superior Técnico, University of Lisbon

Awarded Best Paper!

Transactional Memory was recently integrated in Intel processors under the name TSX.We show that its performance can be significantly affected by the configuration of its interplay with the software-based fallback: in fact, there does not seem to exist a single configuration that can perform best independently of the application and workload. We address this challenge by introducing an innovative self-tuning approach that exploits lightweight reinforcement learning techniques to identify the optimal TSX configuration in a workload-oblivious manner, i.e. not requiring any off-line/a-priori sampling of the application’s workload. To achieve total transparency for the programmer, we integrated the proposed algorithm in the GCC compiler. Our evaluation shows improvements up to 2× over state of the art approaches, while remaining within 5% from the performance achievable using optimal static configurations.

Available Media

CloudPowerCap: Integrating Power Budget and Resource Management across a Virtualized Server Cluster

3:30 pm

Yong Fu, Washington University in St. Louis; Anne Holler, VMware, Inc.; Chenyang Lu, Washington University in St. Louis

In many data centers, server racks are highly underutilized due to maintaining the sum of the server nameplate power below the power provisioned to the rack. The root cause of this rack underutilization is that the server nameplate power is often much higher than can be reached in practice. Although statically setting per-host power caps can ensure the sum of the servers’ maximum power draw does not exceed the rack’s provisioned power, it burdens the data center operator with managing the rack power budget across the hosts. In this paper we present Cloud- PowerCap, a practical and scalable solution for power cap management in a virtualized cluster. CloudPowerCap, closely integrated with a cloud resource management system, dynamically adjusts the per-host power caps for hosts in the cluster to respect not only the rack power budget but also the resource management system’s constraints and objectives. Evaluation based on an industrial cloud simulator demonstrates effectiveness and efficacy of CloudPowerCap.

Available Media

A Comprehensive Resource Management Solution for Web-based Systems

2:30 pm

Filippo Seracini, Massimiliano Menarini, and Ingolf Krueger, University of California, San Diego; Luciano Baresi, Sam Guinea, and Giovanni Quattrocchi, Politecnico di Milano

This paper presents an autonomic resource management solution that looks at the non-functional qualities of a Web-based system, as well as at the characteristics of the infrastructural resources it uses. It exploits a detailed performance model of both these aspects to increase the efficiency of resource allocation. The solution is evaluated on an auction and shopping benchmark web site, and compared to a baseline approach and to an existing solution from literature. Results show that, by jointly taking into account the different software and hardware facets of our application, we can reduce the amount of resources allocated by up to 42.5% compared with an existing work from the literature.

Available Media

PCP: A Generalized Approach to Optimizing Performance Under Power Constraints through Resource Management

Henry Hoffmann, University of Chicago; Martina Maggio, Lund University

Many computing systems are constrained by power budgets. While they could temporarily draw more power, doing so creates unsustainable temperatures and unwanted electricity consumption. Developing systems that operate within power budgets is a constrained optimization problem: configuring the components within the system to maximize performance while maintaining sustainable power consumption. This is a challenging problem because many different components within a system affect power/performance tradeoffs and they interact in complex ways. Prior approaches address these challenges by fixing a set of components and designing a power budgeting framework that manages only that one set of components. If new components become available, then this framework must be redesigned and reimplemented. This paper presents PCP, a general solution to the power budgeting problem that works with arbitrary sets of components, even if they are not known at design time or change during runtime. To demonstrate PCP, we implement it in software and deploy it on a Linux/x86 platform.

Available Media
3:30 p.m.–4:00 p.m. Friday

Break with Refreshments

Washington Foyer

4:00 p.m.–5:30 p.m. Friday

Energy in Data Centers

Session Chair: Paolo Costa, Microsoft Research Cambridge

Coordinating Liquid and Free Air Cooling with Workload Allocation for Data Center Power Minimization

3:00 pm

Li Li, Wenli Zheng, Xiaodong Wang, and Xiaorui Wang, The Ohio State University

Data centers are seeking more efficient cooling techniques to reduce their operating expenses, because cooling can account for 30-40% of the power consumption of a data center. Recently, liquid cooling has emerged as a promising alternative to traditional air cooling, because it can help eliminate undesired air recirculation. Another emerging technology is free air cooling, which saves chiller power by utilizing outside cold air for cooling. Some existing data centers have already started to adopt both liquid and free air cooling techniques for significantly improved cooling efficiency and more data centers are expected to follow.

In this paper, we propose SmartCool, a power optimization scheme that effectively coordinates different cooling techniques and dynamically manages workload allocation for jointly optimized cooling and server power. In sharp contrast to the existing work that addresses different cooling techniques in an isolated manner, SmartCool systematically formulates the integration of different cooling systems as a constrained optimization problem. Furthermore, since geo-distributed data centers have different ambient temperatures, SmartCool dynamically dispatches the incoming requests among a network of data centers with heterogeneous cooling systems to best leverage the high efficiency of free cooling. A light-weight heuristic algorithm is proposed to achieve a near-optimal solution with a low run-time overhead. We evaluate SmartCool both in simulation and on a hardware testbed. The results show that SmartCool outperforms two state-of-the-art baselines by having a 38% more power savings.

Available Media

Managing Green Datacenters Powered by Hybrid Renewable Energy Systems

Chao Li, University of Florida; Rui Wang, Beihang University; Tao Li, University of Florida; Depei Qian, Beihang University; Jingling Yuan, Wuhan University of Technology

The rapidly growing server energy expenditure and the warning of climate change have forced the IT industry to look at datacenters powered by renewable energy. Existing proposals on this issue yield sub-optimal performance as they typically assume certain specific type of renewable energy sources and overlook the benefits of cross-source coordination. This paper takes the first step toward exploring green datacenters powered by hybrid renewable energy systems that include baseload power supply, intermittent power supply, and backup energy storage. We propose GreenWorks, a power management framework for green high-performance computing datacenters powered by renewable energy mix. Specifically, GreenWorks features a hierarchical power coordination scheme tailored to the timing and capacity of different renewable energy sources. Using real datacenter workload traces and renewable power generation data, we show that our scheme allows green datacenters to achieve better design trade-offs.

Available Media

WattValet: Heterogenous Energy Storage Management in Data Centers for Improved Power Capping

Shen Li, Shaohan Hu, Shiguang Wang, Siyu Gu, Chenji Pan, and Tarek Abdelzaher, University of Illinois at Urbana-Champaign

This paper presents WattValet, an efficient solution to reduce data center peak power consumption by using heterogeneous energy storage. We henceforth call an energy storage device, a battery, with the understanding that the discussion applies to other devices as well such as pumped hydraulic and thermal systems. Previous work on energy storage management in data centers often ignores or underestimates their degree of heterogeneity. Even if batteries used in a data center are of the same model and purchased at the same time, differences in storing temperature and humidity, as well as discharging cycles and depth, gradually drive their characteristics apart. We show that differences in battery characteristics, such as discharge rates, if not fully accounted for, can lead to significantly suboptimal power caps. A new algorithm, called WattValet, is described that reduces peak power consumption by efficiently exploiting heterogeneity. Evaluation using Wikipedia traces shows that the power cap generated by WattValet is within 2% of the optimal solution, whereas WattValet finishes the computation orders of magnitude faster than the optimal solution even in small-scale experiments.

Available Media