Networker

[Networker] How can data, or number of files, cause extremely slow write speeds?

2009-05-13 12:52:40
Subject: [Networker] How can data, or number of files, cause extremely slow write speeds?
From: George Sinclair <George.Sinclair AT NOAA DOT GOV>
To: NETWORKER AT LISTSERV.TEMPLE DOT EDU
Date: Wed, 13 May 2009 12:45:02 -0400
If you have a very large number of files on a given file system can this slow down the backup write speeds, not the actual completion times, but the speed of the writing to tape itself?

We have an ext3 file system on a Linux box that is extremely painful to back up because it chugs along at less than 100 KB/sec. I finally gave up and canceled the backup. I was running a level full. The strange thing is that all the rest of the data on that box, located on the other ext3 file systems, backs up at normal speed, even to the same tape and drive (LTO-3).

As far as I can tell, there's nothing suspect with the drive or tape, and there's no process running on there that's sucking up memory or accessing the affected data in such a manner that it would cause this issue. Compression is enabled on the drives.

It's consuming less than half the disk space (300 out of 750 GB). However, that particular file system is also consuming over 10 million inodes. Nasty! As a result, running commands like 'du' or 'find' can take a long time to complete. I'm not sure if the data is binary. It could be, and this would obviously adversely affect compression, so I would expect that we might see slower write speeds as a result. Also, if you're only sending a single save stream to tape, it will be slower than multiplexing multiple streams. All that makes sense, but I've never seen anything this slow, even when running a single save set, except in extreme cases where there was a switch problem, NIC issue, or extremely slow client, but again, the other backups on that host do run at normal speed.

I think NetWorker walks through save sets before it backs them up so it knows if a file's size has changed once it starts to back it up? That said, it might take it a while to create that preliminary list on a file system with that many inodes, but once it's done, and it starts the actual backups, would the number files, and/or the type of data, affect the actual speed that the backups perform at???? I can see that it might take it longer to complete the backups, but how would this create such slow write speeds? Is this surprising given that number of inodes?

Does anyone have any suggestions on how we might be able to speed this up? The only thing I can think of is to identify the minimal number of areas under there that need to be backed up and skip the rest, maybe break those minimal backups into multiple save sets to multiplex the data and force faster write speeds? I could turn compression off on the drive, but I don't wanna leave it that way as it generally works to our favor on most data. Would be a pain to dedicate a non-compression drive to just one chunk of data.

Thanks for any suggestions.

George


--
George Sinclair
Voice: (301) 713-3284 x210
- The preceding message is personal and does not reflect any official or unofficial position of the United States Department of Commerce -
- Any opinions expressed in this message are NOT those of the US Govt. -

To sign off this list, send email to listserv AT listserv.temple DOT edu and type 
"signoff networker" in the body of the email. Please write to networker-request 
AT listserv.temple DOT edu if you have any problems with this list. You can access the 
archives at http://listserv.temple.edu/archives/networker.html or
via RSS at http://listserv.temple.edu/cgi-bin/wa?RSS&L=NETWORKER