Design a distributed filesystem with metadata servers and chunk servers handling 10 PB of storage, similar to GFS/HDFS.
## Problem
Design a distributed filesystem that can store 10 PB of data across 1,000 commodity servers and serve large sequential reads at 100 GB/sec aggregate throughput. The system will be used as the storage layer for a data processing platform (similar to MapReduce/Spark), where workloads consist primarily of large files (100 MB to 10 GB) that are written once and read many times.
Sign up to access the full problem
Design canvas, rubric, hints, and model solutions.
Explain Linux Filesystem Internals
Junior · Conceptual