[clug] Need read-only NFS with local disk based caching

Michael.James at csiro.au Michael.James at csiro.au
Tue Oct 28 14:00:54 EST 2003


Pardon me asking this not-particularily Reiser related question,
 but I need to tap some file-system expertise.

On a cluster of 66 nodes,  (suse 8.2, reiser 3.6)
 our bioinformatic database files are outgrowing
 the nodes capacity to store them,  (>24Gig).

We call them databases, but to an operating system
 they are just a bunch of big (~2Gig) files
 that need to be read at the beginning of each job.

Plenty of space on the fileserver but we find
 when the databases are provided through NFS
 that cluster throughput levels off at about 10 nodes.
Seems the single NFS server becomes a bottleneck.
It might be possible to tune this to do better
 but it would take a lot of improvement
 to get all 66 nodes humming.

We need something that provides a network file system
 but with local caching.  Sun's cachefs would be ideal.
Note that this is the simplest case, read-only access,
 infrequent (weekly at most) changes.
Manually flushing out-of-date copies would be acceptable.

After such a change or when someone asks for something new
 we have a short pause while the information is re-distributed
 and we are humming again.

I thought to use openAFS or CODA
 but find a wheird mix of doco dating from 1900s.

AFS is proud of being able to handle 8 Gig volumes!
Bioinformatic software tries (unsucessfully)
 to keep filesizes below 2 Gig for historical reasons.
So if each file is a seperate volume I _could_ live with an 8 gig limit.

Coda suffers the whierdness of its AFS roots
 and I can't find doco on what its limitations are!

I haven't had much luck posting this question
 in my usual haunts
 so Thanks for any suggestions,

michaelj

-- 
Michael James				michael.james at csiro.au
System Administrator			voice:	02 6246 5040
CSIRO Bioinformatics Facility	fax:		02 6246 5166



More information about the linux mailing list