Networker

Re: [Networker] 10 GbE performance tuning on Solaris client/Storage Nodes

2009-11-09 09:31:38
Subject: Re: [Networker] 10 GbE performance tuning on Solaris client/Storage Nodes
From: Magnus Berglund <belmagnus AT GMAIL DOT COM>
To: NETWORKER AT LISTSERV.TEMPLE DOT EDU
Date: Mon, 9 Nov 2009 15:29:25 +0100
I have had some luck with getting better throughput after enabling jumbo
frames and settings the mtu size to 9000. The value 9000 is not an optimal
value for our network card in the box according to our UNIX techs but we use
it due a number of different reasons.

We have also done some other tuning but the jumboframes and the mtu was the
big difference in our environment for getting the speed above 100 mb/s

SUN T5140, Solaris 10 (u7) with 10 GB, 2 x 10 GB in failover
(active/passive)
Using a DD 690 as an adv file type device (NFS), NW 4 x 1 GB (all active)

During testing we saw that we had to use a number of mountpoints to get the
best performace since each "nfs stream" topped at "1 gb speed" even though
the link is 10 GB between the storage node and the target system.

At the moment I can write from a Legato storage node around 380 mb/s against
the DD box and that is more or less what it is able to receive with it's
current 4 x 1 GB interfaces using NFS.

//Magnus

On Tue, Nov 3, 2009 at 8:54 PM, Ray Pengelly <pengelly AT queensu DOT ca> wrote:

> Hey everyone,
>
> I currently have an Sun M5000 with an ixgbe 10 GbE card directly connected
> to a Sun X4540 system with an nxge 10 GbE.
>
> I am able to read from disk at roughly 930 MB/s using the usam tool:
>
> # time uasm -s ./testfile10g >/dev/null
>
> real    0m11.785s
> user    0m0.382s
> sys     0m11.396s
>
> If i do this over NFS I am only able to get about 104 MB/sec
>
> # time uasm -s ./testfile10g >/mnt/usam-out
>
> real    1m38.888s
> user    0m0.598s
> sys     0m44.980s
>
> Using Networker I see roughly the same numbers with the X4540 acting a a
> Storage Node using adv_file devices on a zpool. I know both the client and
> server filesystems are not the bottleneck.
>
> Both links show up as 10000 full duplex via dladm show-dev.
>
> Has anyone been through performance tuning 10 GbE on Solaris 10? Any notes/
> recipes?
>
> Anyone gotten better throughput than this?
>
> Ray
>
>
>
>
>
> --
> Ray Pengelly
> Technical Specialist
> Queen's University - IT Services
> pengelly AT queensu DOT ca
> (613) 533-2034
>
>
> To sign off this list, send email to listserv AT listserv.temple DOT edu and 
> type
> "signoff networker" in the body of the email. Please write to
> networker-request AT listserv.temple DOT edu if you have any problems with 
> this
> list. You can access the archives at
> http://listserv.temple.edu/archives/networker.html or
> via RSS at http://listserv.temple.edu/cgi-bin/wa?RSS&L=NETWORKER
>

To sign off this list, send email to listserv AT listserv.temple DOT edu and 
type "signoff networker" in the body of the email. Please write to 
networker-request AT listserv.temple DOT edu if you have any problems with this 
list. You can access the archives at 
http://listserv.temple.edu/archives/networker.html or
via RSS at http://listserv.temple.edu/cgi-bin/wa?RSS&L=NETWORKER