LinuxQuestions.org

LinuxQuestions.org (/questions/)
-   Linux - Newbie (https://www.linuxquestions.org/questions/linux-newbie-8/)
-   -   maximum linux file size? (https://www.linuxquestions.org/questions/linux-newbie-8/maximum-linux-file-size-261277/)

alec77 12-01-2004 07:53 AM

maximum linux file size?
 
I am running RedHat 9.0 and i am trying to copy a file the is 3GB from a windows 2000 server to the RedHat machine. When the file copies, it takes up all 40GB of the harddrive on the linux machine. I've tried the comand "split -b 650000000 /mnt/windowsshare/filename /backups/filename" and it just created about 25 650MB files until the hard drive was filled up.

I am more familiar with Windows than linux. Can linux recognize files larger than 2GB properly?

any ideas what my problem would be?

any help is much apreciated.

peacebwitchu 12-01-2004 08:25 AM

yes.

hw-tph 12-01-2004 08:30 AM

Linux, or the filesystems commonly used with Linux, can handle big files. It depends on the type of filesystem you use, and on this page there is a table describing the functionality of different file systems when it comes to files and filesystem size.

The split command should work, so I'm guessing it may have something to do with the file size on the Windows file system is being reported wrong.


Håkan

alec77 12-01-2004 08:49 AM

that chart says the maximum file size on an ext3 file system (the file system I am using) is 2GB so that may be part of my problem. But that doesn't seem to explain why the split didn't work right. I believe the file size is being reported properly, the file is only taking up 3GB of space on the windows machine.

otoomet 12-01-2004 09:00 AM

The chart is perhaps a bit outdated. The caption says that
Quote:

and once the 2.4 kernels come out, I am sure the limits will be extended
Best,

Ott

alec77 12-01-2004 09:53 AM

so the file size limit has been extended?

Electro 12-01-2004 04:01 PM

Linux can handle up to 16 TB with certain filesystems. I used XFS and ReiserFS because they are much better than ext3. I used XFS for most of my partitions.

I have never seen this problem. It could be on your Windows system not calculating the correct size or smbclient not calculating correctly.

The only problem that I had is copying 4 GB through SAMBA that can only handle 2 GB per file at the time. I used the split command and it made four 1 GB files.

BTW, how many bytes did smbclient calculated for the file that you are trying to copy.

alec77 12-03-2004 06:57 AM

ok, I think the problem must be with smbmount. I am trying to copy the file nightly in a script so I am using something like this:

smbmount //machine/share /mnt/share -o username=user,password=password
split -b 650000000 /mnt/share/bigfile.ext /backup

when i mount the share with smbmount and do an ls -l of the directory it is mounted in, the file is reported as a crazy huge file at 18 Exa bytes. If I connect with the smbclient command and do an ls -l it reports the correct size.

so the smbclient is working correctly for me but the smbmount is not. Does anyone know of a way I can use the smbclient command in a script and not the smbmount? will something like this work?:

cd /backup
smbclient //machine/share
split -b 650000000 /mnt/share/bigfile.ext
exit
#rest of script................

any help is appreciated.

alec77 12-03-2004 09:31 AM

ok, i've figured out how to script the smbclient command, here it is:

smbclient //machine/share -U username%password -c "lcd /backup; get filename"

Ive tested it and it works on small files, what I'm woried about ist that it will hang when i run it on the large 3GB file. does anyone know how I could pipe the above command to something like split that will alow me to split the file up into smaller portions that can be transfered by smbclient?


All times are GMT -5. The time now is 12:24 PM.