usenix conference policies
Scalable Performance of the Panasas Parallel File System
The Panasas file system uses parallel and redundant access to object storage devices (OSDs), per-file RAID, distributed metadata management, consistent client caching, file locking services, and internal cluster management to provide a scalable, fault tolerant, high performance distributed file system. The clustered design of the storage system and the use of client-driven RAID provide scalable performance to many concurrent file system clients through parallel access to file data that is striped across OSD storage nodes. RAID recovery is performed in parallel by the cluster of metadata managers, and declustered data placement yields scalable RAID rebuild rates as the storage system grows larger. This paper presents performance measures of I/O, metadata, and recovery operations for storage clusters that range in size from 10 to 120 storage nodes, 1 to 12 metadata nodes, and with file system client counts ranging from 1 to 100 compute nodes. Production installations are as large as 500 storage nodes, 50 metadata managers, and 5000 clients.
Open Access Media
USENIX is committed to Open Access to the research presented at our events. Papers and proceedings are freely available to everyone once the event begins. Any video, audio, and/or slides that are posted after the event are also free and open to everyone. Support USENIX and our commitment to Open Access.
author = {Brent Welch and Marc Unangst and Zainul Abbasi and Garth Gibson and Brian Mueller and Jason Small and Jim Zelenka and Bin Zhou},
title = {Scalable Performance of the Panasas Parallel File System},
booktitle = {6th USENIX Conference on File and Storage Technologies (FAST 08)},
year = {2008},
address = {San Jose, CA},
url = {https://www.usenix.org/conference/fast-08/scalable-performance-panasas-parallel-file-system},
publisher = {USENIX Association},
month = feb
}
connect with us