LinuxQuestions.org
Welcome to the most active Linux Forum on the web.
Home Forums Tutorials Articles Register
Go Back   LinuxQuestions.org > Forums > Linux Forums > Linux - Server
User Name
Password
Linux - Server This forum is for the discussion of Linux Software used in a server related context.

Notices


Reply
  Search this Thread
Old 02-19-2009, 10:35 PM   #1
dman65
Member
 
Registered: Sep 2003
Posts: 61

Rep: Reputation: 15
Which file system to use


I have a situation where I need a file server to handle a few million scanned black and white tiff documents. These documents have variable numbers of pages and they can be any where from about 10K to 50MB.

I was wondering if anyone can suggest which of the available file systems would work best for serving up files of this size and volume. Adding new documents quickly, accessing existing documents, and saving updated documents are all important.

Also, is there a best way to setup the directory structure or is that irrelevant for a bTree directory system?

Thanks for any help.
 
Old 02-19-2009, 10:46 PM   #2
routers
Member
 
Registered: Aug 2005
Location: Malaysia - KULMY / CNXTH
Distribution: Slackware, Fedora, FreeBSD, Sun O/S 5.10, CentOS
Posts: 787
Blog Entries: 6

Rep: Reputation: 75
for your purpose i personally prefer ext3 with samba
 
Old 02-19-2009, 10:54 PM   #3
billymayday
LQ Guru
 
Registered: Mar 2006
Location: Sydney, Australia
Distribution: Fedora, CentOS, OpenSuse, Slack, Gentoo, Debian, Arch, PCBSD
Posts: 6,678

Rep: Reputation: 122Reputation: 122
You may find this a useful read
 
Old 02-20-2009, 03:33 AM   #4
chitambira
Member
 
Registered: Oct 2008
Location: Online
Distribution: RHEL, Centos
Posts: 373
Blog Entries: 1

Rep: Reputation: 51
gpfs if good with multimedia files.
 
Old 02-20-2009, 06:23 AM   #5
salasi
Senior Member
 
Registered: Jul 2007
Location: Directly above centre of the earth, UK
Distribution: SuSE, plus some hopping
Posts: 4,070

Rep: Reputation: 897Reputation: 897Reputation: 897Reputation: 897Reputation: 897Reputation: 897Reputation: 897
Depends.

If the file sizes have something vaguely like a log-normal distribution (ie, there is a big peak in frequency of file sizes and that is at the low end) I might consider Reiser, but otherwise not.

If there is no data backup and regeneration issue (ie, the files come from a place from which they can be easily re-loaded) I might even be adventurous enough to consider ext4. If this is the primary copy of your data, I think I'd be more conservative than that and consider ext4 as still too bleeding edge.

Then I'd consider whether I wanted 'atime' for any reason; I'm guessing no, but you might have your reasons. If you feel happy with 'noatime', then you'd only want this on the same partiton as other noatime data and this is probably a real performance gain in this case. this then gets into what else is on that machine and whether you are happy with noatime on other partitions...

I've never used XFS or JFS, but they could come into consideration, again depending a bit on the distribution of file sizes.

Otherwise, its the old favourites ext2 and ext3; does journalling do anything for you (see the data backup and recreation issues); if not, ext2. If journalling does give you extra security, ext3.
 
Old 02-20-2009, 11:43 AM   #6
dman65
Member
 
Registered: Sep 2003
Posts: 61

Original Poster
Rep: Reputation: 15
Thanks for the input.

What is the fastest method to transfer these files from one system to another without actually moving the hard drives?

I tried setting up a dedicated network link between the two machines over their spare 100BaseT connections and the data started out transferring 6GB of data the first hour, but then as time went on it slowed further and further until it was down to about 2GB an hour which at that rate would take me 12 days to transfer the entire 750GB of existing data.

I then connected an external USB 2.0 drive to the existing server and started copying. The first hour transferred 10GB and then it started falling off. It is on it's third hour of copying now and it looks like I am down to about 5GB an hour.

Is there some faster way of copying this data? It seems like USB 2.0 should be able to transfer the data much faster and the hard drive should at least be able to write 33MB a second. Even at half of that, I should be able to get 1GB a minute I would think.

There are 750GB to transfer and a total of 3.5 million individual files.
 
Old 02-20-2009, 01:28 PM   #7
rweaver
Senior Member
 
Registered: Dec 2008
Location: Louisville, OH
Distribution: Debian, CentOS, Slackware, RHEL, Gentoo
Posts: 1,833

Rep: Reputation: 167Reputation: 167
In my opinion you have two good options presently, xfs and ext4. That being said xfs is far more mature than ext4 presently, although in some cases ext4 performs better.

Edit: When you say "dedicated network link" you mean a crossover cable right? You would probably also get some benefit from raising your rmem/wmem from the default if you're going to try to do it over the wire. The fastest way of course is just drop the drive in the system. As far as an application to do the transfer... I'd suggest rsync make sure you DON'T turn on compression though. It also has the handy benefit that you can start and stop it whenever you like and it will effectively pick up where it left off.

Last edited by rweaver; 02-20-2009 at 01:36 PM.
 
Old 02-21-2009, 11:39 PM   #8
dman65
Member
 
Registered: Sep 2003
Posts: 61

Original Poster
Rep: Reputation: 15
Hello rweaver,

Yes, I was referring to a crossover cable between the two machines.

I tried using rsync, but it gave me an error indicating that there were too many files. Is there a limit to how many files it will process from a directory?
 
Old 02-22-2009, 12:59 AM   #9
billymayday
LQ Guru
 
Registered: Mar 2006
Location: Sydney, Australia
Distribution: Fedora, CentOS, OpenSuse, Slack, Gentoo, Debian, Arch, PCBSD
Posts: 6,678

Rep: Reputation: 122Reputation: 122
More likely the issue was calling with something like

rsync * dest

which gets expanded by bash first

rsync source dest

probably won't cause the same issue.

I'm only guessing as to your issue of course.

See man rsync for worked examples.
 
  


Reply



Posting Rules
You may not post new threads
You may not post replies
You may not post attachments
You may not edit your posts

BB code is On
Smilies are On
[IMG] code is Off
HTML code is Off



Similar Threads
Thread Thread Starter Forum Replies Last Post
An error Occurred during the file system check. Dropping you to shell; the system wil aneikei Linux - Newbie 3 02-11-2010 07:38 PM
Linpus Fedora after saving files to file system, CAN'T open FILE SYSTEM from desktop! remiwa Fedora 2 01-07-2009 07:28 AM
when does linux switch from realmode file system to protectedmode file system sitthar Linux - General 3 04-08-2006 08:05 AM
figuring out 'file system' and 'swap file system' types TrulyTessa Linux - Newbie 3 09-26-2005 06:46 PM

LinuxQuestions.org > Forums > Linux Forums > Linux - Server

All times are GMT -5. The time now is 01:50 PM.

Main Menu
Advertisement
My LQ
Write for LQ
LinuxQuestions.org is looking for people interested in writing Editorials, Articles, Reviews, and more. If you'd like to contribute content, let us know.
Main Menu
Syndicate
RSS1  Latest Threads
RSS1  LQ News
Twitter: @linuxquestions
Open Source Consulting | Domain Registration