LWN: Comments on "Ceph distributed filesystem merged for 2.6.34" https://lwn.net/Articles/379554/ This is a special feed containing comments posted to the individual LWN article titled "Ceph distributed filesystem merged for 2.6.34". en-us Sat, 01 Nov 2025 14:26:45 +0000 Sat, 01 Nov 2025 14:26:45 +0000 https://www.rssboard.org/rss-specification lwn@lwn.net Ceph distributed filesystem merged for 2.6.34 https://lwn.net/Articles/382591/ https://lwn.net/Articles/382591/ cmccabe <div class="FormattedComment"> <font class="QuotedText">&gt; And as for the network bandwidth, this is less of a problem with a</font><br> <font class="QuotedText">&gt; distributed FS than with NFS since at least it has the chance of being</font><br> <font class="QuotedText">&gt; stripped across various point to point connections instead of all points to</font><br> <font class="QuotedText">&gt; one single connection.</font><br> <p> You're missing a very important point here. Ethernet connections in an office are almost never point-to-point. Everyone uses a star topolgy. So all 30 of the computers on a floor might go through one gig-e switch.<br> <p> In theory, the bandwidth of a dedicated gigE connection is comparable to the bandwidth between a PC and another PC on a 1 gig-E switch. In practice, the switch may not be able to handle everyone talking at once.<br> <p> There may be latency issues with network I/O too. It depends on the network.<br> <p> </div> Thu, 08 Apr 2010 01:29:51 +0000 Ceph distributed filesystem merged for 2.6.34 https://lwn.net/Articles/380672/ https://lwn.net/Articles/380672/ joib <i> would you ask you local filesystem developers to ignore POSIX? Would you expect them to agree that it is too restrictive. </i> <p> Why yes, after all things like noatime and relatime are existence proofs that sometimes local fs developers are willing to sacrifice POSIX compliance for performance reasons. Heck, relatime is even the default nowadays. <p> For performance, many if not all distributed or cluster fs's have some kind of lazy statvfs() and lazy *stat(), and so forth, in order to get around some of the serialization that POSIX would otherwise force upon them. <p> And of course, some aspects of POSIX are outright incompatible with a distributed fs. E.g. fcntl(..., F_GETLK) (which node does the PID refer to?). Sat, 27 Mar 2010 10:08:17 +0000 Ceph distributed filesystem merged for 2.6.34 https://lwn.net/Articles/380660/ https://lwn.net/Articles/380660/ martinfick <div class="FormattedComment"> Not to mention that the itty-bitty channel to the drives is there whether<br> the FS is local or distributed. <br> <p> And as for the network bandwidth, this is less of a problem with a<br> distributed FS than with NFS since at least it has the chance of being<br> stripped across various point to point connections instead of all points to<br> one single connection.<br> </div> Sat, 27 Mar 2010 04:15:38 +0000 Ceph distributed filesystem merged for 2.6.34 https://lwn.net/Articles/380651/ https://lwn.net/Articles/380651/ dlang <div class="FormattedComment"> you are thinking of the itty-bitty data channel to the drives, but you are forgetting that the local system doesn't need to go to the drives most of the time. It can drastically speed things up by keeping the data cached in memory, which has a pretty fast and low latency interface to the cores.<br> </div> Sat, 27 Mar 2010 02:06:23 +0000 Ceph distributed filesystem merged for 2.6.34 https://lwn.net/Articles/380646/ https://lwn.net/Articles/380646/ martinfick <p><i>POSIX semantics, especially atime and the serialization constraints, tend to consume even more of the scarce network bandwidth.</i></p> <p> Me thinks you are confused. Those work better with (require is too strong a word) low latency links, not high bandwidth. </p> <p><i> POSIX semantics are useful in local filesystems because programs rely on them for IPC. Distributed filesystems are rarely used for IPC-- or if they are, the semantics of the FS are customized ahead of time to work well for that, like in Hadoop's filesystem (HDFS), Google-FS, or MPI-IO. </i></p> <p> I think you are stuck and cannot think beyond what is done today. The point is not to design something new to use a distributed POSIX filesystem, but to take currently working applications and entire virtual machine operating systems images and to simply run them over a distributed FS instead of locally so that you get all the benefits of a distributed FS (yes, that also means you get the downsides as well, low latency...). But in many cases, the gains will simply outweigh the downsides, not everything needs supercomputer low latency interconnects, but almost everything can benefit from HA and larger single namespaced filesystems. </p><p> Just because things were slow yesterday, does not mean they will be slow tomorrow. What is commonplace today for supercomputers will be commonplace tomorrow on desktops. I would even argue that distributed FSes should have been common yesterday, we are way behind the curve on this one. Fast networking gear is getting cheaper and cheaper and can already outpace disk bandwidth. With the growing size of hard disks, so much space is wasted, but even worse, the failure rates are going up and nothing is addressing this. The average desktop has way too big a drive, but rarely uses RAID since that requires a second drive. Why should we stick with NFS when it is a) not POSIX, and b) not HA or distributed, and C) does not scale disk space easily enough. </p> Sat, 27 Mar 2010 01:32:06 +0000 Ceph distributed filesystem merged for 2.6.34 https://lwn.net/Articles/380641/ https://lwn.net/Articles/380641/ cmccabe <div class="FormattedComment"> First of all, Ceph is a big achievement! Block-level filesystems are hard, and distributed systems are often even harder!<br> <p> <font class="QuotedText">&gt; None of the proprietary solutions are things that your average company is</font><br> <font class="QuotedText">&gt; going to install on their desktops which likely have virtually unused TB</font><br> <font class="QuotedText">&gt; drives (except for the possibly large mp3 collections scattered here and</font><br> <font class="QuotedText">&gt; there). There clearly is an opportunity to harness tons of corporate</font><br> <font class="QuotedText">&gt; wasted disk space and CPU into a shared pool instead of wasting good money</font><br> <font class="QuotedText">&gt; on obsolete high end fileservers which typically don't even provide HA.</font><br> <p> Modern PCs have enormous multi-core, multi-GHz CPUs connected to itty-bitty I/O channels. POSIX semantics, especially atime and the serialization constraints, tend to consume even more of the scarce network bandwidth.<br> <p> POSIX semantics are useful in local filesystems because programs rely on them for IPC. Distributed filesystems are rarely used for IPC-- or if they are, the semantics of the FS are customized ahead of time to work well for that, like in Hadoop's filesystem (HDFS), Google-FS, or MPI-IO.<br> <p> So just go the NFS route and make something useful, rather than something that emulates the behavior of a local filesystem circa 1976.<br> <p> </div> Sat, 27 Mar 2010 01:08:09 +0000 Ceph distributed filesystem merged for 2.6.34 https://lwn.net/Articles/380564/ https://lwn.net/Articles/380564/ martinfick <div class="FormattedComment"> Bluntly saying POSIX is too restrictive seems over generalized. For some cases I am sure that is true, but again, would you ask you local filesystem developers to ignore POSIX? Would you expect them to agree that it is too restrictive. If you start with the mindset that you want distributed FSes for things that are typically done on distributed FSes, than it will likely seem restrictive. But if you start with the frame of reference of things that are typically done on local FSes, and you simply want to migrate to a distributed FS without recoding applications or worrying about consistency issues, then I can assure you that POSIX semantics become very important.<br> <p> -GPFS, very nice.<br> <p> -Lustre, does it provide HA, or just fast (stripped) access?<br> <p> -GFS (I assume you mean RedHat's GFS, and not google FS) is not really a distributed FS, it requires a shared backend.<br> <p> -Panasas, is this really Posix (hard to figure out form their site)? pNFS, yes. Also, this seems to use their custom hardware on the backend which happens to be distributed, a black box so to speak. So, this is not really a distributed FS that others can install and use.<br> <p> So, it seems like possibly only 2or3 solutions and there are many many more distributed FSes out there. I would say that overall posix is quite rare for distributed FSes. And since none of those rare exceptions are free/libre/open source, surely you would concede that ceph is not filling an already crowded space? <br> <p> None of the proprietary solutions are things that your average company is going to install on their desktops which likely have virtually unused TB drives (except for the possibly large mp3 collections scattered here and there). There clearly is an opportunity to harness tons of corporate wasted disk space and CPU into a shared pool instead of wasting good money on obsolete high end fileservers which typically don't even provide HA.<br> <p> </div> Fri, 26 Mar 2010 16:58:26 +0000 Ceph distributed filesystem merged for 2.6.34 https://lwn.net/Articles/380550/ https://lwn.net/Articles/380550/ roblatham <div class="FormattedComment"> Yes, NFS is too weak. POSIX is too restrictive. I'd like to point out that <br> especially in the supercomputing domain there is a "third way" to do <br> consistency semantics: the MPI standard defines "MPI-IO consistency <br> semantics", which give applications assurances as to when data will be <br> visible (no bizzaro NFS client-side caching behaviors) but relax the POSIX <br> rules in ways that only help performance. <br> <p> Because you asked: distributed file systems that implement POSIX consistency <br> semantics:<br> <p> - GPFS<br> - Lustre<br> - GFS<br> - Panasas <br> - now Ceph apparently<br> </div> Fri, 26 Mar 2010 15:07:39 +0000 Ceph distributed filesystem merged for 2.6.34 https://lwn.net/Articles/380308/ https://lwn.net/Articles/380308/ bronson <div class="FormattedComment"> Merge windows are starting to feel like the paradox of the unexpected hanging...<br> <p> <a href="http://2000clicks.com/MathHelp/JokeProofUnexpectedHanging.aspx">http://2000clicks.com/MathHelp/JokeProofUnexpectedHanging...</a><br> </div> Thu, 25 Mar 2010 13:31:30 +0000 Ceph distributed filesystem merged for 2.6.34 https://lwn.net/Articles/380306/ https://lwn.net/Articles/380306/ clugstj <div class="FormattedComment"> It was unpredictable (that he would stretch the merge window out so far).<br> </div> Thu, 25 Mar 2010 13:07:28 +0000 Ceph distributed filesystem merged for 2.6.34 https://lwn.net/Articles/379961/ https://lwn.net/Articles/379961/ cmccabe <div class="FormattedComment"> Thanks for the informative reply.<br> <p> It sounds like Ceph has some features that even Lustre doesn't have. I hope that the supercomputer guys get a chance to try it out sometime!<br> <p> <p> </div> Wed, 24 Mar 2010 00:13:02 +0000 Ceph distributed filesystem merged for 2.6.34 https://lwn.net/Articles/379898/ https://lwn.net/Articles/379898/ martinfick <p><i>"Also, why do people still like implementing POSIX filesystem semantics on distributed filesystems? Intuitively it seems like POSIX semantics demand a really fast interconnect between nodes. In contrast, systems that implement weaker forms of consistency, like the Andrew File System, can perform well with lower-end networking gear. I think most system administrators would be satisfied with something like AFS's open-to-close consistency guarantee. I mean, most system administrators still use NFS, which has horrible consistency semantics.</i></p> <p> Why do people implement them? Probably because users want them! I know I do. Afterall, surely you can agree that POSIX semantics are useful? We like them on our local filesystems. Why would I not like them on distributed ones? Granted, they are likely to have high latencies, but that doesn't mean that I don't want POSIX semantics. </p><p> Yes, AFS/NFS semantics are fine for some applications, but there are many for which it is not. With today's virtualization trend, it becomes more and more important to be able to use shared/distributed filesystems, but this is of little use if one has to think about which virtualized applications really need POSIX and which do not, and then segregate accordingly. </p><p> So, users want POSIX whether local or distributed (try asking ext4/ZFS/BTRFS why they feel the need to implement POSIX.) And, of course, the holy grail for many sysadmins would be a fast/distributed/scalable/HA/POSIX fs which ceph promises to be. I am happy that ceph is at least attempting to do this. You seem to imply that there are tons of FSes doing this, who else is? </p> Tue, 23 Mar 2010 18:58:25 +0000 Subscribe! https://lwn.net/Articles/379693/ https://lwn.net/Articles/379693/ sunr2007 <div class="FormattedComment"> thanks! i will subscribe by the end of this month. waiting for april 1st so <br> tat i ll get my salary!! :)<br> </div> Sun, 21 Mar 2010 13:25:36 +0000 Ceph distributed filesystem merged for 2.6.34 https://lwn.net/Articles/379689/ https://lwn.net/Articles/379689/ joib <div class="FormattedComment"> Disclaimer: I've never used ceph, and while I have used Lustre extensively I've never admined it.<br> <p> Ceph has fault-tolerance as an integral part of the design. You specify data to be replicated N times, and ceph takes care of the rest; if a node fails, the data on the failed node is automatically replicated to other nodes in order to bring the system up to the desired replication level. To get HA with Lustre, you setup nodes in active-passive pairs with "traditional" HA software, Lustre itself has nothing to do with HA.<br> <p> Ceph seems to have some kind of automatic data migration to avoid hotspots. Ceph has distributed metadata, Lustre has had it on the roadmap for quite a while, but AFAIK it's not yet delivered.<br> <p> Lustre, AFAIK, uses file allocation tables to specify on which nodes a file is striped, Ceph uses some kind of hash function. <br> <p> Lustre server components are kernel-space using a customized ext3/4 fs, Ceph servers are user space using a normal fs, though apparently snapshot functionality requires btrfs. I think the Lustre devs are in the process of moving the servers to user space and using ZFS DMU for storage.<br> <p> Performance-wise, I suspect Lustre is very hard to beat, especially when using MPI-IO for which it has been tuned for years on some very large systems.<br> <p> Wrt. POSIX semantics, yes I believe they might not always be a good match to distributed filesystems. Heck, even on local fs's we often relax POSIX semantics to improve performance (noatime, relatime). For concurrent operation, as such AFAICS the problem is not necessarily that it requires a fast interconnect, but rather that strict adherence to POSIX serializes many operations due to the requirement that metadata is always consistent. E.g. if you have multiple clients writing to a file concurrently, what is stat() supposed to return? I believe both Ceph and Lustre provide, at least optionally, relaxed POSIX semantics for metadata consistency. Also, strict POSIX compliance in the face of concurrent access limits IO buffering. Anyway, many applications seem to do just fine with the close-to-open consistency that NFS provides, and explicit locks (I believe at least as of NFSv4 acquiring a lock forces cache revalidation).<br> <p> Altogether, Ceph looks very interesting IMHO. I'd love to be able to build a scalable fault-tolerant fs google-style using cheap commodity hardware. <br> </div> Sun, 21 Mar 2010 11:14:37 +0000 Subscribe! https://lwn.net/Articles/379688/ https://lwn.net/Articles/379688/ man_ls There is a handy "Subscribe" link at the top of every LWN page, and <a href="http://www.xe.com/">xe.com</a> converts the $2.50 of the basic rate to 113 rupees. It should be at worst about 1/200th of the <a href="http://www.indiamike.com/india/moving-to-bangalore-f100/cost-of-living-in-india-bangalore-t43211/">living costs in your city (20k rupees)</a>, so roughly the equivalent to the highest rate ($10) for an European or US resident (assuming living costs of $2k). An excellent deal I would say. Sun, 21 Mar 2010 10:06:17 +0000 Subscribe! https://lwn.net/Articles/379687/ https://lwn.net/Articles/379687/ sunr2007 <div class="FormattedComment"> I'm from bangalore, India. can u please suggest me the subscription rates in <br> terms of INR? how can i subscibe here?<br> warm regards,<br> Ravi Kulkarni<br> <p> </div> Sun, 21 Mar 2010 09:24:09 +0000 Subscribe! https://lwn.net/Articles/379669/ https://lwn.net/Articles/379669/ man_ls May I suggest you a great way of supporting LWN.net authors: subscribe! You get great value for your money, and can adjust your subscription level to your earnings. Note: I am not affiliated with LWN.net, I'm just a happy camper. Sat, 20 Mar 2010 22:43:27 +0000 Ceph distributed filesystem merged for 2.6.34 https://lwn.net/Articles/379666/ https://lwn.net/Articles/379666/ cmccabe <div class="FormattedComment"> Does anyone know how Ceph compares with Lustre?<br> <p> I can think of a few similarities: both use POSIX semantics and have separate metadata and object data storage nodes. Also, it seems like both Lustre and Ceph do striping.<br> <p> Of course, Lustre is still not in the mainline.<br> <p> Also, why do people still like implementing POSIX filesystem semantics on distributed filesystems? Intuitively it seems like POSIX semantics demand a really fast interconnect between nodes. In contrast, systems that implement weaker forms of consistency, like the Andrew File System, can perform well with lower-end networking gear. I think most system administrators would be satisfied with something like AFS's open-to-close consistency guarantee. I mean, most system administrators still use NFS, which has horrible consistency semantics.<br> <p> </div> Sat, 20 Mar 2010 21:01:19 +0000 Xen https://lwn.net/Articles/379657/ https://lwn.net/Articles/379657/ antn <div class="FormattedComment"> Thanks for the answer.<br> <p> So, even if the almost ready Xen 4.0 release will for the first time abandon the ancient non-paravirtops dom0 patches in favor of paravirtops-based ones, the inclusion in mainline is still not close to happen, contrary to what one might naively have guessed.<br> <p> (FWIW, I'm currently running Slackware 13.0 as a dom0 through the SUSE patches, rebased for kernel.org 2.6.31 by A . Lyon) <br> <p> <p> </div> Sat, 20 Mar 2010 16:04:08 +0000 Xen https://lwn.net/Articles/379656/ https://lwn.net/Articles/379656/ corbet There has been no pull request that I'm aware of for Xen stuff - and the nature of that code is such that it's hard to be unaware of an attempt to merge it. So no, we'll not see it in 2.6.34. I'd say 2.6.35 is unlikely too unless something is posted in the fairly near future. Sat, 20 Mar 2010 15:12:04 +0000 Ceph distributed filesystem merged for 2.6.34 https://lwn.net/Articles/379653/ https://lwn.net/Articles/379653/ antn <div class="FormattedComment"> With apologies for the somewhat offtopic and pointless question, is there any hope of seeing Xen dom0 support merged in this release? The Xen wiki paravirtops page still says that inclusion is planned for 2.6.34 o 2.6.35.<br> <p> [Since this is my first post, I would like to state here my great appreciation for lwn.net authors, contents and posters).<br> <p> Thanks<br> </div> Sat, 20 Mar 2010 14:50:20 +0000 Ceph distributed filesystem merged for 2.6.34 https://lwn.net/Articles/379634/ https://lwn.net/Articles/379634/ csamuel <div class="FormattedComment"> Yes, and Ceph was specifically listed there, so it shouldn't be a real<br> surprise.<br> <p> Very glad to see it get in!<br> </div> Sat, 20 Mar 2010 06:55:04 +0000 Ceph distributed filesystem merged for 2.6.34 https://lwn.net/Articles/379584/ https://lwn.net/Articles/379584/ gfarnum <div class="FormattedComment"> Didn't Linus say when he put out rc1 that there would be more stuff coming?<br> <a href="http://lkml.org/lkml/2010/3/8/280">http://lkml.org/lkml/2010/3/8/280</a><br> </div> Fri, 19 Mar 2010 19:16:21 +0000 Ceph distributed filesystem merged for 2.6.34 https://lwn.net/Articles/379583/ https://lwn.net/Articles/379583/ proski The rationale is that "the new stuff" cannot have regressions, unlike fixes to the existing code. Fri, 19 Mar 2010 19:04:41 +0000 Ceph distributed filesystem merged for 2.6.34 https://lwn.net/Articles/379572/ https://lwn.net/Articles/379572/ dlang <div class="FormattedComment"> the end of the merge window has never been the end of merging new stuff.<br> <p> it's always been when Linus stops accepting new stuff to merge. From the very beginning, he has merged significant amounts of stuff between -rc1 and -rc2, it's only after about -rc2 or -rc3 that things really settle down to fixes.<br> </div> Fri, 19 Mar 2010 18:12:08 +0000 Ceph distributed filesystem merged for 2.6.34 https://lwn.net/Articles/379570/ https://lwn.net/Articles/379570/ vonbrand <p> Bad Linus, bad. <p> (The whole point of "unpredictable" merge windows was supposed to be making last-minute-suprises unlikely...) Fri, 19 Mar 2010 18:06:42 +0000