[Networker] How can data, or number of files, cause extremely slow write speeds?
2009-05-13 12:52:40
If you have a very large number of files on a given file system can this
slow down the backup write speeds, not the actual completion times, but
the speed of the writing to tape itself?
We have an ext3 file system on a Linux box that is extremely painful to
back up because it chugs along at less than 100 KB/sec. I finally gave
up and canceled the backup. I was running a level full. The strange
thing is that all the rest of the data on that box, located on the other
ext3 file systems, backs up at normal speed, even to the same tape and
drive (LTO-3).
As far as I can tell, there's nothing suspect with the drive or tape,
and there's no process running on there that's sucking up memory or
accessing the affected data in such a manner that it would cause this
issue. Compression is enabled on the drives.
It's consuming less than half the disk space (300 out of 750 GB).
However, that particular file system is also consuming over 10 million
inodes. Nasty! As a result, running commands like 'du' or 'find' can
take a long time to complete. I'm not sure if the data is binary. It
could be, and this would obviously adversely affect compression, so I
would expect that we might see slower write speeds as a result. Also, if
you're only sending a single save stream to tape, it will be slower than
multiplexing multiple streams. All that makes sense, but I've never seen
anything this slow, even when running a single save set, except in
extreme cases where there was a switch problem, NIC issue, or extremely
slow client, but again, the other backups on that host do run at normal
speed.
I think NetWorker walks through save sets before it backs them up so it
knows if a file's size has changed once it starts to back it up? That
said, it might take it a while to create that preliminary list on a file
system with that many inodes, but once it's done, and it starts the
actual backups, would the number files, and/or the type of data, affect
the actual speed that the backups perform at???? I can see that it might
take it longer to complete the backups, but how would this create such
slow write speeds? Is this surprising given that number of inodes?
Does anyone have any suggestions on how we might be able to speed this
up? The only thing I can think of is to identify the minimal number of
areas under there that need to be backed up and skip the rest, maybe
break those minimal backups into multiple save sets to multiplex the
data and force faster write speeds? I could turn compression off on the
drive, but I don't wanna leave it that way as it generally works to our
favor on most data. Would be a pain to dedicate a non-compression drive
to just one chunk of data.
Thanks for any suggestions.
George
--
George Sinclair
Voice: (301) 713-3284 x210
- The preceding message is personal and does not reflect any official or
unofficial position of the United States Department of Commerce -
- Any opinions expressed in this message are NOT those of the US Govt. -
To sign off this list, send email to listserv AT listserv.temple DOT edu and type
"signoff networker" in the body of the email. Please write to networker-request
AT listserv.temple DOT edu if you have any problems with this list. You can access the
archives at http://listserv.temple.edu/archives/networker.html or
via RSS at http://listserv.temple.edu/cgi-bin/wa?RSS&L=NETWORKER
|
<Prev in Thread] |
Current Thread |
[Next in Thread>
|
- [Networker] How can data, or number of files, cause extremely slow write speeds?,
George Sinclair <=
|
|
|