Los Alamos National Laboratory
 
 

Science >  LANL Institutes >  Information Science and Technology Institute

National Security Education Center

Contacts

2012 Posters
2012 Presentations
2012 Workshops
2011 Posters
2011 Panels
2011 Presentations
2010 Posters
2010 Presentations
2009 Posters
2009 Presentations
2008 Posters
2008 Panels
2008 Presentations
2007 Posters
2007 Panels
2007 Presentations

    2012 Posters

  • SC12-PLFS/HDFS: HPC Applications on Cloud Storage
    Milo Polte, Chuck Cranor, Garth Gibson

    Long running large scale HPC applications protect themselves from failures by periodically checkpointing their state to a single file stored in a distributed network filesystem. These filesystems commonly provide a POSIX-style interface for reading and writing files. HDFS is a filesystem used in cloud computing by Apache Hadoop. HDFS is optimized for Hadoop jobs that do not require full POSIX I/O semantics. Only one process may write to an HDFS file, and all writes are appends. Our work enables multiple HPC processes to checkpoint their state into an HDFS file using PLFS. PLFS is a middleware filesystem that converts random I/O into log-based I/O. We added a new I/O store layer to PLFS that allows it to use non-POSIX filesystems like HDFS as backing store. HPC applications can now checkpoint to HDFS, allowing HPC and cloud to share the same storage systems and work with each others data.

  • 2012 Presentations

  • Cluster12: he Power and Challenges of Transformative I/O
    John Bent, Adam Manzanares, Meghan McClelland, Garth Gibson

    Extracting high data bandwidth and metadata rates from parallel file systems is notoriously difficult. User workloads almost never achieve the performance of synthetic benchmarks. The reason for this is that real-world applications are not as well-aligned, well-tuned, or consistent as are synthetic benchmarks. There are at least three possible ways to address this challenge: modification of the real-world workloads, modification of the underlying parallel file systems, or reorganization of the real-world workloads using transformative middleware. In this paper, we demonstrate that transformative middleware is applicable across a large set of high performance computing workloads and is portable across the three major parallel file systems in use today. We also demonstrate that our transformative middleware layer is capable of improving the write, read, and metadata performance of I/O workloads by up to 150x, 10x, and 17x respectively, on workloads with processor counts of up to 65,536. 

  • ICSE2012: Statically Checking API Protocol Conformance with Mined Multi-Object Specifications
    Ciera Jaspan, Jonathan Aldrich

    Programmers using an API often must follow protocols that specify when it is legal to call particular methods. Several techniques have been proposed to find violations of such protocols based on mined specifications. However, existing techniques either focus on single-object protocols or on particular kinds of bugs, such as missing method calls. There is no practical technique to find multi-object protocol bugs without a priori known specifications. In this paper, we combine a dynamic analysis that infers multi-object protocols and a static checker of API usage constraints into a fully automatic protocol conformance checker. The combined system statically detects illegal uses of an API without human-written specifications. Our approach finds 41 bugs and code smells in mature, real-world Java programs with a true positive rate of 51%. Furthermore, we show that the analysis reveals bugs not found by state of the art approaches. 

  • Structure:Data 2012—Faster Memory, Faster Compute
    Gary Grider, Garth Gibson

    High-performance computing has traditionally been the most demanding area of computing that pushed the edge of the envelope. In these circumstances a lot of pioneering work has been done that ultimately filtered down to more pervasive and affordable applications. In this fireside chat we talk to a scientist from Los Alamos National Laboratories about how they reduce the time and cost of getting to critical insights by reducing the distance and latency of compute memory. The conversation will look to successes in current deployments and what this signals for computer architectures for big data problems in the long-term.

  • 2012 Workshops

  • SC12-7th Parallel Data Storage Workshop
    John Bent, Garth Gibson, Robert Ross

    Peta- and exascale computing infrastructures make unprecedented demands on storage capacity, performance, concurrency, reliability, availability, and manageability. This one-day workshop focuses on the data storage problems and emerging solutions found in peta- and exascale scientific computing environments, with special attention to issues in which community collaboration can be crucial for problem identification, workload capture, solution interoperability, standards with community buy-in, and shared tools. This workshop seeks contributions on relevant topics, including but not limited to: performance and benchmarking, failure tolerance problems and solutions, APIs for high performance features, parallel file systems, high bandwidth storage architectures, wide area file systems, metadata intensive workloads, autonomics for HPC storage, virtualization for storage systems, archival storage advances, resource management innovations, storage systems for big data and analytics, and incorporation of emerging storage technologies. The Parallel Data Storage Workshop holds a peer reviewed competitive process for selecting short papers. Submit a not previously published short paper of up to 5 pages, not less than 10 point font, in a PDF file. Submitted papers will be reviewed under the supervision of the workshop program committee. Submissions should indicate authors and affiliations. Final papers must not be longer than 5 pages. Selected papers and associated talk slides will be made available on the workshop web site and published in the IEEE digital library. Paper submissions due: September 30, 2012, 11:59 p.m. EDT. There will also be a poster session at the workshop; submit a short poster abstract through http://www.pdsw.org no later than November 6, 2012.

    2011 Panels

  • PDSW10—Death of Disk Panel
    Garth Gibson

    * Moderator: Steve Poole, ORNL; Dave Anderson, Seagate; Garth Gibson, Panasas; Erik Riedel, EMC; Nisha Talagala, FusionIO; Ted Wobber, MSR

     

  • 2011 Posters

  • CMU PDL 2011—Evaluation of GIFA+ Scalable Directories (using the FUSE user-level prototype)
    Swapnil Patil, Garth Gibson
  • CMU PDL 2011—GIGA+: Scalable Directories for Shared File Systems (or, How to build directories with trillions of files)
    Swapnil Patil, Garth Gibson
  • CMU PDL 2011—On the Duality of Data-Intensive file System Design: Reconciling HDFS and PVFS
    Wittawat Tantisiriroj, Swapnil Patil, Garth Gibson
  • CMU PDL 2011—PDL Data-Intensive Computer Clusters and Networks
    Mitch Franzos, Garth Gibson, Julio Lopez, Michael Stroucken
  • CMU PDL 2011—PLFS-HSFS: HPC Applications on Cloud Storage
    Milo Polte, Chuck Cranor, Garth Gibson
  • CMU PDL 2011—PLFS: A Checkpoint Filesystem for Parallel Applications
    John Bent, Gary Grider, Adam Manzanares, Meghan McClelland, Ben McClelland, James Nunez, Milo Polte, Chuck Cranor, Garth Gibson, Paul Nowoczynski
  • CMU PDL 2011—PRObE: Systems Testbed at the Scale of Thousands
    Katharine Chartrand, Gary Grider, Andree Jacobson, Chuck Cranor, Mitch Franzos, Garth Gibson, Michael Stroucken
  • CMU PDL 2011—Shingled Disk Interfaces
    Anand Suresh, Xu Zhang, Greg Ganger, Garth Gibson
  • CMU PDL 2011—Shingled File System Design
    Anand Suresh, Greg Ganger, Garth Gibson
  • CMU PDL 2011—YCSB++: Benchmarking Advanced Features of BigTable-like Stores
    Milo Polte, Wittawat Tantisiriroj, Swapnil Patil, Kai Ren, Garth Gibson, Julio Lopez
  • PDL Day 2011: PRObE: Systems Testbed at the Scale of Thousands
    Katharine Chartrand, Gary Grider, Andree Jacobson, Mitch Franzos, Garth Gibson, Michael Stroucken
  • PDL Day 2011—Astro-DISC: Data-Intensive Analytics for Astrophysics
    Milo Polte, Wittawat Tantisiriroj, Swapnil Patil, Bin Fu, Lin Xiao, Vijay Vasudavan, Eugene Fink, Garth Gibson, Julio Lopez
  • PDL Day 2011—Evaluation of GIGA+ Scalable Directories (using the FUSE user-level prototype)
    Swapnil Patil, Garth Gibson
  • PDL Day 2011—Evaluation of GIGA+ Scalable Directories (using the FUSE user-level prototype)

    Swapnil Patil, Garth Gibson
  • PDL Day 2011—GIGA+: Scalable Directories for Shared File Systems (or, How to build directories with trillions of files)
    Swapnil Patil, Garth Gibson
  • PDL Day 2011—GIGA+: Scalable Directories for Shared File Systems (or, How to build directories
with trillions of files)
    Swapnil Patil, Garth Gibson
  • PDL Day 2011—PDL Data-Intensive Computer Clusters and Networks
    Mitch Franzos, Garth Gibson, Julio Lopez, Michael Stroucken
  • PDL Day 2011—PLFS: A Checkpoint Filesystem for Parallel Applications
    John Bent, Gary Grider, Adam Manzanares, Meghan McClelland, Ben McClelland, James Nunez, Milo Polte, Garth Gibson, Paul Nowoczynski
  • PDL Day 2011—PLFS—A Checkpoint Filesystem for Parallel Applications
    John Bent, Gary Grider, Adam Manzanares, Meghan McClelland, Ben McClelland, James Nunez, Milo Polte, Garth Gibson, Paul Nowoczynski
  • PDL Day 2011—Scaling the Metadata Service in DISC Storage
    Wittawat Tantisiriroj, Lin Xiao, Garth Gibson
  • PDL Day 2011—Scaling the Metadata Service in DISC Storage
    Wittawat Tantisiriroj, Lin Xiao, Garth Gibson
  • PDL Day 2011—SkyeFS: Scalable Directories (GIGA+) for PVFS
    Swapnil Patil, Anthony Chivetta, Garth Gibson
  • PDL Day 2011—SkyeFS: Scalable Directories (GIGA+) for PVFS
    Swapnil Patil, Anthony Chivetta, Garth Gibson
  • 2011 Presentations

  • CMU PDL 2011—On the Duality of Data-intensive File System Design: Reconciling HDFS and PVFS
    Wittawat Tantisiriroj
  • CMU PDL 2011—Shingled Disks and their File Systems: ShingledFS
    Garth Gibson
  • CMU PDL 2011—YCSB++ : Benchmarking and Performance Debugging Advanced Features in Scalable Table Stores
    Swapnil Patil
  • Doctoral Dissertation: Proper Plugin Protocols
    Ciera Jaspan

    The ability of the software engineering community to achieve high levels of reuse from software frameworks has been tempered by the difficulty in understanding how to reuse them properly. When written correctly, a plugin can take advantage of the frameworkʼs code and architecture to provide a rich application with relatively few lines of code. Unfortunately, doing this correctly is difficult because frameworks frequently require plugin developers to be aware of complex protocols between objects, and improper use of these protocols causes exceptions and unexpected behavior at run time. This dissertation introduces collaboration constraints, rules governing how multiple objects may interact in a complex protocol. These constraints are particularly difficult to understand and analyze because they may extend across type boundaries and even programming language boundaries. This thesis improves the state of the art through two mechanisms. First, it provides a deep understanding of these collaboration constraints and the framework designs which create them. Second, it introduces Fusion, an adoptable specification language and static analysis tool, that detects broken collaboration constraints in plugin code and demonstrates how to achieve this goal in a cost-effective manner that is practical for industry use.  In this dissertation, I have done an empirical study of framework help forums which showed that collaboration constraints are burdensome for developers, as they take hours or even days to resolve. From this empirical study, I have identified several common properties of collaboration constraints. This motivated a new specification language, called Fusion, that is tailored for specifying collaboration constraints in a practical way. The specification language uses relationships to describe the abstract associations between objects and allows developers to specify collaboration constraints as logical predicates of relationships. Since a relationship is an abstraction above the code, this allows developers to easily specify constraints that cross type and language boundaries. There are three variants of the analysis: a sound variant that has false positives but no false negatives, a complete variant that has false negatives but no false positives, and a pragmatic variant that attempts to balance this tradeoff. In this dissertation, I successfully used Fusion to specify and analyze constraints from examples found in the help forums of the ASP.NET and Spring frameworks. Additionally, I ran Fusion on DaCapo, a 1.5 MLOC DaCapo benchmark for program analysis, to show that Fusion is scalable and provides precise enough results for industry with low specification cost. This dissertation examines many tradeoffs: the tradeoffs of framework designs, the tradeoffs of specification precision, and the tradeoffs of program analysis results are all featured. A central theme of this work is that there is no single right solution to collaboration constraints; there are only solutions that work better for a particular instance of the problem

  • HEC FSIO'11—YCSB++ update
    Garth Gibson
  • HPDC11: Six Degrees of Scientific Data: Reading Patterns for Extreme Scale Science IO
    Milo Polte, Garth Gibson

    Petascale science simulations generate 10s of TBs of application data per day, much of it devoted to their checkpoint/ restart fault tolerance mechanisms. Previous work demonstrated the importance of carefully managing such output to prevent application slowdown due to IO blocking, resource contention negatively impacting simulation performance and to fully exploit the IO bandwidth available to the petascale machine. This paper takes a further step in understanding and managing extreme-scale IO. Specifically, its evaluations seek to understand how to efficiently read data for subsequent data analysis, visualization, check-point restart after a failure, and other read-intensive operations. In their entirety, these actions support the "end-to-end" needs of scientists enabling the scientific processes being undertaken. Contributions include the following. First, working with application scientists, we define 'read' benchmarks that capture the common read patterns used by analysis codes. Second, these read patterns are used to evaluate different IO techniques at scale to understand the effects of alternative data sizes and organizations in relation to the performance seen by end users. Third, defining the novel notion of a 'data district' to characterize 

  • PDL Day 2011— Overview of Large-scale Systems Work
    Garth Gibson
  • PDL Day 2011—Evaluation of GIGA+ Scalable Directories (using the FUSE user-level prototype)
    Swapnil Patil, Garth Gibson
  • PDL Day 2011—Evaluation of GIGA+ Scalable Directories (using the FUSE user-level prototype)

    Swapnil Patil, Garth Gibson
  • PDL Day 2011—GIGA+: Scalable Directories for Shared File Systems (or, How to build directories with trillions of files)
    Swapnil Patil, Garth Gibson
  • PDL Day 2011—GIGA+: Scalable Directories for Shared File Systems (or, How to build directories
with trillions of files)
    Swapnil Patil, Garth Gibson
  • PDL Day 2011—PLFS: A Checkpoint Filesystem for Parallel Applications
    John Bent, Gary Grider, Adam Manzanares, Meghan McClelland, Ben McClelland, James Nunez, Milo Polte, Garth Gibson, Paul Nowoczynski
  • PDL Day 2011—PLFS—A Checkpoint Filesystem for Parallel Applications
    John Bent, Gary Grider, Adam Manzanares, Meghan McClelland, Ben McClelland, Milo Polte, Garth Gibson, Paul Nowoczynski
  • PDL Day 2011—PRObE: Systems Testbed at the Scale of Thousands
    Katharine Chartrand, Gary Grider, Andree Jacobson, Mitch Franzos, Garth Gibson, Michael Stroucken
  • PDL Day 2011—Scaling the Metadata Service in DISC Storage
    Wittawat Tantisiriroj, Lin Xiao, Garth Gibson
  • PDL Day 2011—Scaling the Metadata Service in DISC Storage
    Swapnil Patil, Lin Xiao, Garth Gibson
  • PDL Day 2011—SkyeFS: Scalable Directories (GIGA+) for PVFS
    Swapnil Patil, Anthony Chivetta, Garth Gibson
  • PDL Day 2011—SkyeFS: Scalable Directories (GIGA+) for PVFS
    Swapnil Patil, Anthony Chivetta, Garth Gibson
  • PLATEAU11: Are Object Protocols Burdensome?
    Ciera Jaspan, Jonathan Aldrich

    Object protocols are a commonly studied research problem, but there is little known about their usability in practice. In particular, there is little research to show that object protocols cause difficulty for developers. In this work, we use community forums to find empirical evidence that object protocols are burdensome for developers. We analyzed 427 threads from the Spring and ASP.NET forums and discovered that 69 were on a protocol violation. We found that violations of protocols result in unusual runtime behavior rather than exceptions in 45% of our threads, that questions took an average of 62 hours to resolve, and that even though 54% of questions were repeated violations of similar protocols, the manifestation of the violation at runtime was different enough that developers could not search for similar questions. 

  • SC11: On the Duality of Data-intensive File System Design: Reconciling HDFS and PVFS
    Wittawat Tantisiriroj, Swapnil Patil, Garth Gibson

    Data-intensive applications fall into two computing styles: Internet services (cloud computing) or high-performance computing (HPC). In both categories, the underlying file system is a key component for scalable application performance. In this paper, we explore the similarities and differences between PVFS, a parallel file system used in HPC at large scale, and HDFS, the primary storage system used in cloud computing with Hadoop. We integrate PVFS into Hadoop and compare its performance to HDFS using a set of data-intensive computing benchmarks. We study how HDFS-specific optimizations can be matched using PVFS and how consistency, durability, and persistence tradeoffs made by these file systems affect application performance. We show how to embed multiple replicas into a PVFS file, including a mapping with a complete copy local to the writing client, to emulate HDFS's file layout policies. We also highlight implementation issues with HDFS's dependence on disk bandwidth and benefits from pipelined replication. 

  • SOCC11: YCSB++: Benchmarking and Performance Debugging Advanced Features in Scalable Table Stores
    Milo Polte, Wittawat Tantisiriroj, Swapnil Patil, Lin Xiao, Kai Ren, Garth Gibson, Julio Lopez

    Inspired by Googleʼs BigTable, a variety of scalable, semi-structured, weak-semantic table stores have been developed and optimized for different priorities such as query speed, ingest speed, availability, and interactivity. As these systems mature, performance benchmarking will advance from measuring the rate of simple workloads to understanding and debugging the performance of advanced features such as ingest speed-up techniques and function shipping filters from client to servers. This paper describes YCSB++, a set of extensions to the Yahoo! Cloud Serving Benchmark (YCSB) to improve performance understanding and debugging of these advanced features. YCSB++ includes multi-tester coordination for increased load and eventual consistency measurement, multi-phase workloads to quantify the consequences of work deferment and the benefits of anticipatory configuration optimization such as B-tree pre-splitting or bulk loading, and abstract APIs for explicit incorporation of advanced features in benchmark tests. To enhance performance debugging, we customized an existing cluster monitoring tool to gather the internal statistics of YCSB++, table stores, system services like HDFS, and operating systems, and to offer easy post-test correlation and reporting of performance behaviors. YCSB++ features are illustrated in case studies of two BigTable-like table stores, Apache HBase and Accumulo, developed to emphasize high ingest rates and fine-grained security. 

    2009 Posters

  • CMU PDL Day PLFS: A Checkpoint Filesystem for Parallel Applications
    John Bent, Gary Grider, James Nunez, Milo Polte, Garth Gibson

    Parallel Data Laboratory (PDL)

  • CMU PDL Day Understanding Performance in Solid State Disks
    Milo Polte, Garth Gibson

    Parallel Data Laboratory (PDL)

  • PDL 09: Astro-DISC: Data-Intensive Alalytics for Astrophysics
    Milo Polte, Garth Gibson, Julio Lopez

    Parallel Data Laboratory (PDL)

  • PDL09: GIGA+: Scalable Directories for Shared File Systems
    Garth Gibson

    Parallel Data Laboratory (PDL)

  • PDL09: Large-scale Evaluation of GIGA+ Scalable Directories
    Garth Gibson

    Parallel Data Laboratory (PDL)

  • PDL09: PLFS and HDFS: Enabling Parallel Filesystems Semantics in the Cloud
    John Bent, Esteban Molina-Estolano, Milo Polte, Scott Brandt, Garth Gibson, Carlos Maltzahn

    Parallel Data Laboratory (PDL)

  • PDL09: PLFS: A Checkpoint Filesystem for Parallel Applications
    John Bent, Gary Grider, Meghan McClelland, Ben McClelland, James Nunez, Milo Polte, Garth Gibson

    Parallel Data Laboratory (PDL)

  • PDL09: Reliability Modeling for Large Scale Declustered Storage
    Garth Gibson

    Parallel Data Laboratory (PDL)

  • PDL09: Scalable Distributed Table Storage Experiments on OpenCloud
    Milo Polte, Garth Gibson

    Parallel Data Laboratory (PDL)

  • PDL09: Tables and File Systems: Moving Into the Cloud
    Milo Polte, Garth Gibson, Julio Lopez

    Parallel Data Laboratory (PDL)

  • USENIX FAST 09: Overlapped HPC Checkpointing with Hardware Assist
    Christopher Mitchell, James Nunez, Jun Wang

    File and Storage Technologies (FAST)

  • 2009 Presentations

  • CMU PDL Day PLFS 2009: A Stackable, Log Structured Filesystem for Checkpointing
    John Bent, Gary Grider, James Nunez, Milo Polte, Garth Gibson

    Parallel Data Laboratory (PDL)

  • CMU PDL Day: Understanding Performance in Solid State Disks
    Milo Polte, Garth Gibson

    Parallel Data Laboratory (PDL)

  • PDL Workshop & Retreat PLFS: A Checkpoint Filesystem for Parallel Applications
    John Bent, Gary Grider, Meghan McClelland, Ben McClelland, James Nunez, Milo Polte, Garth Gibson

    Parallel Data Laboratory (PDL)

  • PDL09: Astro-DISC: Data-Intensive Analytics for Astrophysics
    Milo Polte, Garth Gibson, Julio Lopez

    Parallel Data Laboratory (PDL)

  • PDL09: GIGA+: Scalable Directories from Shared File Systems
    Garth Gibson

    Parallel Data Laboratory (PDL)

  • PDL09: Large-scale Evaluation of GIGA+ Scalable Directories
    Garth Gibson

    Parallel Data Laboratory (PDL)

  • PDL09: PLFS: A Checkpoint Filesystem for Parallel Applications
    John Bent, Gary Grider, Meghan McClelland, Ben McClelland, James Nunez, Milo Polte, Garth Gibson

    Parallel Data Laboratory (PDL)

  • PDL09: Reliability Modeling for Large Scale Declustered Storage
    Garth Gibson

    Parallel Data Laboratory (PDL)

  • PDL09: Scalable Distributed Table Storage Experiments on OpenCloud
    Milo Polte, Garth Gibson

    Parallel Data Laboratory (PDL)

  • PDL09: Tables and File Systems: Moving Into the Cloud
    Milo Polte, Garth Gibson, Julio Lopez

    Parallel Data Laboratory (PDL)

  • PDL09:PLFS and HDFS: Enabling Parallel Filesystems Semantics in the Cloud
    John Bent, Esteban Molina-Estolano, Milo Polte, Scott Brandt, Garth Gibson, Carlos Maltzahn

    Parallel Data Laboratory (PDL)

  • Solving TCP Incast in Cluster Storage Systems FAST09 WIP
    David Andersen, Greg Ganger, Garth Gibson

    File and Storage Technologies (FAST)
    Work in Progress (WiP)

  • Spiral: Program Generation for Linear Transforms and Beyond
    Franz Franchetti
Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA
Inside | © Copyright 2008-09 Los Alamos National Security, LLC All rights reserved | Disclaimer/Privacy | Web Contact