an interesting filesystem challenge: static pull of

Martin Langhoff martin.langhoff at
Wed Nov 12 15:46:49 EST 2008

On Wed, Nov 12, 2008 at 3:23 PM, C. Scott Ananian <cscott at> wrote:
> apache seems to perform reasonably well serving files from such huge
> directories.  Should I be concerned?  Can anyone suggest:
>  b) whether reformatting with reiserfs or some other filesystem is
> worth the trouble?  ext3 already has btree-structured directories, so
> reiserfs isn't quite the obvious win it used to be.

According to Linus, Ted T'so and friends when we hit a similar problem
with git (before it did hashed dirs), The Answer Is XFS. This problem
(lookups in large directories) also affects Maildir-based IMAP
servers, and XFS is a favourite in that segment too.

XFS is trouble in many other aspects, but directories sure can do. So
I wouldn't do it for the whole system disk - just a dedicated

>  c) patched wget or other tool that will actually honor robot
> exclusion directives in <meta> tags in page headers?  wget seems to
> honor 'nofollow', but mediawiki uses <meta name="robots"
> content="noindex,nofollow" /> in the <head> of edit and printable
> pages, which isn't sufficient to convince wget to delete the file it
> just downloaded.

ISTR lwp-rget being reasonably good at this.


 martin.langhoff at
 martin at -- School Server Architect
 - ask interesting questions
 - don't get distracted with shiny stuff  - working code first

More information about the Devel mailing list