Cloud platforms host thousands of tenants that demand POSIX semantics, high throughput, and rapid evolution from their storage layer. Kernel-native distributed file systems supply raw speed, but their privileged code base couples every release to the kernel, widens the blast radius of crashes, and slows innovation. FUSE-based distributed file systems flip those trade-offs: they run in user space for fast deployment and strong fault isolation, yet the FUSE interface disables the kernel's write-back page cache whenever strong consistency is required. Practitioners must therefore choose between (i) weak consistency with fast write-back caching or (ii) strong consistency with slow write-through I/O, a limitation that has kept FUSE distributed file systems out of write-intensive cloud workloads. To this end, we present DFUSE, the first distributed FUSE file system that delivers write-back kernel caching and strong consistency. DFUSE achieves this by offloading userspace consistency control to the kernel driver, allowing coordinated access to the kernel's page cache across nodes. This design eliminates blind local cache updates and ensures cluster-wide strong consistency without compromising performance. In our evaluation, DFUSE achieves up to 68.0% higher throughput and 40.4% lower latency than the existing write-through design of FUSE-based distributed file systems.
翻译:云平台承载着数千个租户,其对存储层要求具备POSIX语义、高吞吐量和快速演进能力。内核原生分布式文件系统提供原始速度,但其特权代码库将每个版本与内核紧密耦合,扩大了故障影响范围并延缓了创新进程。基于FUSE的分布式文件系统扭转了这种权衡:它们运行在用户空间以实现快速部署和强故障隔离,然而当需要强一致性时,FUSE接口会禁用内核的写回页缓存。因此实践者必须在(i)具有快速写回缓存的弱一致性,与(ii)具有低速直写I/O的强一致性之间做出选择,这一限制使得基于FUSE的分布式文件系统无法应用于写入密集型云工作负载。为此,我们提出DFUSE——首个实现写回内核缓存与强一致性的分布式FUSE文件系统。DFUSE通过将用户空间一致性控制卸载至内核驱动来实现这一目标,允许跨节点协调访问内核页缓存。该设计消除了盲目的本地缓存更新,在保证性能的同时确保集群范围的强一致性。在我们的评估中,相较于现有基于FUSE分布式文件系统的直写设计方案,DFUSE实现了高达68.0%的吞吐量提升和40.4%的延迟降低。