8:00 am–9:00 am |
Monday |
Continental Breakfast
Texas Ballroom Foyer
|
9:00 am–9:15 am |
Monday |
Program Co-Chairs: Eric Eide, University of Utah, and Mathias Payer, Purdue University
|
9:15 am–10:30 am |
Monday |
Session Chair: Mathias Payer, Purdue University
Tim Leek, MIT Lincoln Laboratory
(joint work with New York University and Northeastern University) Work on automating vulnerability discovery has long been hampered by a shortage of ground-truth corpora with which to evaluate tools and techniques. This lack of ground truth prevents authors and users of tools alike from being able to measure such fundamental quantities as miss and false alarm rates. In this talk, we detail LAVA, a novel dynamic taint analysis-based technique for producing ground-truth corpora by quickly and automatically injecting large numbers of realistic bugs into program source code. Every LAVA bug is accompanied by an input that triggers it whereas normal inputs are extremely unlikely to do so. These vulnerabilities are synthetic but, we argue, still realistic, in the sense that they are embedded deep within programs and are triggered by real inputs. LAVA has already been used to inject thousands of bugs into programs of between 10K and 2M LOC, and we have begun to use the resulting corpora to evaluate bug finding tools. Work on automating vulnerability discovery has long been hampered by a shortage of ground-truth corpora with which to evaluate tools and techniques. This lack of ground truth prevents authors and users of tools alike from being able to measure such fundamental quantities as miss and false alarm rates. In this talk, we detail LAVA, a novel dynamic taint analysis-based technique for producing ground-truth corpora by quickly and automatically injecting large numbers of realistic bugs into program source code. Every LAVA bug is accompanied by an input that triggers it whereas normal inputs are extremely unlikely to do so. These vulnerabilities are synthetic but, we argue, still realistic, in the sense that they are embedded deep within programs and are triggered by real inputs. LAVA has already been used to inject thousands of bugs into programs of between 10K and 2M LOC, and we have begun to use the resulting corpora to evaluate bug finding tools. Our vision is to scale up the LAVA infrastructure to enable frequent online self-evaluation. Developers and evaluators of bug finding tools and techniques will be able to obtain fresh corpora seeded with unknown vulnerabilities on demand, submit their results to be graded automatically, and receive feedback in a tight iterative loop. It is our hope that this will encourage lively and healthy competition that is informed by meaningful performance measures.
|
10:30 am–11:00 am |
Monday |
Break with Refreshments
Texas Ballroom Foyer
|
11:00 am–12:30 pm |
Monday |
Session Chair: Stephen Schwab, USC Information Sciences Institute (ISI)
Robert L. Nord, Ipek Ozkaya, Edward J. Schwartz, Forrest Shull, and Rick Kazman, Carnegie Mellon University Software vulnerabilities originating from design decisions are hard to find early and time consuming to fix later. We investigated whether the problematic design decisions themselves might be relatively easier to find, based on the concept of “technical debt,” i.e., design or implementation constructs that are expedient in the short term but make future changes and fixes more costly. If so, can knowing which components contain technical debt help developers identify and manage certain classes of vulnerabilities? This paper provides our approach for using knowledge of technical debt to identify software vulnerabilities that are difficult to find using only static analysis of the code. We present initial findings from a study of the Chromium open source project that motivates the need to examine a combination of evidence: quantitative static analysis of anomalies in code, qualitative classification of design consequences in issue trackers, and software development indicators in the commit history.
Nicholas Kaufman, Noblis; Michael Collins, Redjack; Kristof Ladny, Booz Allen Hamilton; Jeffrey Wiley, Noblis; Adam Plattner, Noblis NSP; Mark Sanders and Evan Thaler, Noblis; Patrick Ball, Booz Allen Hamilton A common issue amongst security researchers is the lack of publicly available network traffic traces. In this paper we present Chappie Swarm, which seeks to emulate human behavior in regard to internet browsing. The experimenter can unleash a number of automated chappies which will assume pre-defined personas, and then actively go out and query websites while simultaneously recording their browsing behavior, and saving the network trace as a packet capture file. Unlike other traffic generators, Chappie Swarm distinguishes itself fundamentally by utilizing this ”persona” approach, while also not needing to be ”seeded” by a previously recorded traffic capture.
Martin Lazarov, Jeremiah Onaolapo, and Gianluca Stringhini, University College London Cloud-based documents are inherently valuable, due to the volume and nature of sensitive personal and business content stored in them. Despite the importance of such documents to Internet users, there are still large gaps in the understanding of what cybercriminals do when they illicitly get access to them by for example compromising the account credentials they are associated with. In this paper, we present a system able to monitor user activity on Google spreadsheets. We populated 5 Google spreadsheets with fake bank account details and fake funds transfer links. Each spreadsheet was configured to report details of accesses and clicks on links back to us. To study how people interact with these spreadsheets in case they are leaked, we posted unique links pointing to the spreadsheets on a popular paste site. We then monitored activity in the accounts for 72 days, and observed 165 accesses in total. We were able to observe interesting modifications to these spreadsheets performed by illicit accesses. For instance, we observed deletion of some fake bank account information, in addition to insults and warnings that some visitors entered in some of the spreadsheets. Our preliminary results show that our system can be used to shed light on cybercriminal behavior with regards to leaked online documents.
|
12:30 pm–2:00 pm |
Monday |
Luncheon for Workshop Attendees
Zilker Ballroom 1
|
2:00 pm–3:35 pm |
Monday |
Session Chair: Gianluca Stringhini, University College London
Yosuke Kikuchi, Hiroshi Mori, Hiroki Nakano, Katsunari Yoshioka, and Tsutomu Matsumoto, Yokohama National University; Michel Van Eeten, Delft University of Technology All Android markets are confronted with malicious apps, but they differ in how effective they deal with them. In this study, we evaluate the mitigation efforts of Google Play and four third-party markets. We define three metrics and measure how sensitive they are to different detection results from anti-virus vendors. Malware presence in three third-party markets – Liqucn, eoeMarket and Mumayi – is around ten times higher than in Google Play and Freeware Lovers. Searching for certain keywords in Google Play leads leads to a fifty times higher malware rate than those for popular apps. Finally, we measure malware survival times and find that Google Play seems to be the only market that effectively removes malware, though it contains a cluster of apps flagged as adware and malware over long time periods. This points to different incentives for app markets, anti-virus vendors and users.
David Ingegneri, Dominic Timoteo, Patrick Hyle, Fidel Parraga, and Alex Reyes, Federal Aviation Administration The Federal Aviation Administration (FAA) is developing the Cybersecurity Test and Evaluation Facility (CyTF) for the FAA Air Transportation System as it transitions to the Next Generation Air Transportation System (NextGen). This paper describes the goals, capabilities, architecture, current implementation, initial experience, lessons learned and future implementation of the CyTF. The FAA Air Transportation System is an attractive cybersecurity threat target and the FAA must proactively and continually adjust its cybersecurity capabilities to match the changing cybersecurity threat landscape. The CyTF is providing an adaptable cybersecurity research and development environment independent of the operational system to satisfy research, test and evaluation needs. The CyTF has a number of complex requirements: testing cybersecurity tools and technologies prior to their integration into the Air Transportation System, the evaluation of individual FAA Air Transportation subsystems security, security of end-to-end services involving multiple subsystems, procedures to respond and recover from a cybersecurity event and cybersecurity training of the FAA workforce. One of the major lessons learned, described in the paper, has been how to address some aspects of the CyTF’s complex requirements.
Antoine Lemay and José M. Fernandez, École Polytechnique de Montréal High profile attacks such as Stuxnet and the cyber at-tack on the Ukrainian power grid have increased re-search in Industrial Control System (ICS) and Supervi-sory Control and Data Acquisition (SCADA) network security. However, due to the sensitive nature of these networks, there is little publicly available data for re-searchers to evaluate the effectiveness of the proposed solution. The lack of representative data sets makes evaluation and independent validation of emerging se-curity solutions difficult and slows down progress to-wards effective and reusable solutions.
This paper presents our work to generate representative labeled data sets for SCADA networks that security researcher can use freely. The data sets include packet captures including both malicious and non-malicious Modbus traffic and accompanying CSV files that con-tain labels to provide the ground truth for supervised machine learning.
To provide representative data at the network level, the data sets were generated in a SCADA sandbox, where electrical network simulators were used to introduce realism in the physical component. Also, real attack tools, some of them custom built for Modbus networks, were used to generate the malicious traffic. Even though they do not fully replicate a production network, these data sets represent a good baseline to validate detection tools for SCADA systems.
|
Rebecca Bace and Alec Yasinsac, University of South Alabama
|
3:35 pm–4:00 pm |
Monday |
Break with Refreshments
Texas Ballroom Foyer
|
4:00 pm–5:30 pm |
Monday |
Moderator: Eric Eide, University of Utah
Panelists: David Balenson, SRI International; Brendan Dolan-Gavitt, New York University; Jelena Mirkovic, USC Information Sciences Institute (ISI)
|