Veritas-bu

[Veritas-bu] Some info on my experiences with 10GbE

2008-01-05 02:57:31
Subject: [Veritas-bu] Some info on my experiences with 10GbE
From: pancamo <netbackup-forum AT backupcentral DOT com>
To: VERITAS-BU AT mailman.eng.auburn DOT edu
Date: Fri, 04 Jan 2008 23:43:23 -0800
I just started testing 2 T2000's with dual 10Gbps SUN Nics directly connected 
to each other...  

I'm somewhat pissed that I'm only able to get about 658Mbps from one thread on 
the 10Gbps nic while I'm able to get 938Mbps using the onboard 1Gbs nic with 
when using the iperf default values.  Which means in some cases the 10Gbit nic 
is actually slower than the onboard 1Gbit nic

However, I was able to get 7.2 Gbps using 10 threads.


Here are some of my max results with different thread (-P) values

./iperf -c 192.168.1.2 -f m -w 512K -l 512 -P x


TCP Win Buffer  Threads Gbps
512     512     1       1.4
512     512     2       2.5
512     512     4       4.3
512     512     6       6.4
512     512     8       6.1
512     512     10      7.2
512     512     15      4.6
512     512     18      3.6
512     512     20      3
512     512     30      2.5
512     512     60      2.3


Another annoying deal was that the results from iperf were not the same each 
time I ran the test.   The results were as much as 3Gbps different from run to 
run.     The results should be the same for each run.





my /etc/system settings settings that I added as suggested by SUN

set ddi_msix_alloc_limit=8
set ip:ip_soft_rings_cnt=8
set ip:ip_squeue_fanout=1
set ip:tcp_squeue_wput=1
set ip:ip_squeue_bind=0
set ipge:ipge_tx_syncq=1
set ipge:ipge_bcopy_thresh = 512
set ipge:ipge_dvma_thresh = 1
set consistent_coloring=2
set pcie:pcie_aer_ce_mask=0x1




Here are the NDD settings that I found here: 
http://www.sun.com/servers/coolthreads/tnb/parameters.jsp#2

ndd -set /dev/tcp tcp_conn_req_max_q 16384
ndd -set /dev/tcp tcp_conn_req_max_q0 16384
ndd -set /dev/tcp tcp_max_buf 10485760
ndd -set /dev/tcp tcp_cwnd_max 10485760
ndd -set /dev/tcp tcp_xmit_hiwat 131072
ndd -set /dev/tcp tcp_recv_hiwat 131072
ndd -set /dev/nxge0 accept_jumbo 1

I also found information here: 
http://blogs.sun.com/sunay/entry/the_solaris_networking_the_magic



cpreston wrote:
> 7500 MB/s! That?s the most impressive numbers I?ve ever seen by FAR. I may 
> have to take back my ?10 GbE is a Lie!? blog post, and I?d be happy to do so. 
> 
> Can you share things besides the T2000? For example,  
> 
> what OS and patch levels are you running? 
> Any IP patches? 
> Any IP-specific patches? 
> What ndd settings are you using? 
> Is rss enabled? 
> 
> ?Input, I need Input!? 
> 
> --- 
> W. Curtis Preston 
> Backup Blog @ www.backupcentral.com (http://www.backupcentral.com) 
> VP Data Protection, GlassHouse Technologies 
> 
> 
> From: veritas-bu-bounces AT mailman.eng.auburn DOT edu 
> [mailto:veritas-bu-bounces AT mailman.eng.auburn DOT edu] On Behalf Of 
> Peters, Devon C
> Sent: Wednesday, October 17, 2007 12:12 PM
> To: VERITAS-BU AT mailman.eng.auburn DOT edu
> Subject: [Veritas-bu] Some info on my experiences with 10GbE 
> 
> 
> Since I've seen a little bit of talk about 10GbE on here in the past I 
> figured I'd share some of my experiences...  
> I've recently been testing some of Sun's dual-port 10GbE NICs on some small 
> T2000's (1Ghz, 4-core). I'm only using a single port on each card, and the 
> servers are currently directly connected to each other (waiting for my 
> network team to get switches and fibre in place). 
> So far, I've been able to drive throughput between these two systems to about 
> 7500Mbit/sec using iperf. When the throughput gets this high, all the 
> cores/threads on the receiving T2000 become saturated and TCP retransmits 
> start climbing, but both systems remain quite responsive. Since these are 
> only 4-core T2000's, I would guess that the 6 or 8-core T2000's (especially 
> with 1.2Ghz or 1.4Ghz processors) should be capable of more throughput, 
> possibly near line speed. 
> The down side achieving this high of throughput is that it requires lots of 
> data streams. When transmitting with a single data stream, the most 
> throughput I've gotten is about 1500Mbit/sec. I only got up to 7500Mbit/s 
> when using 64 data streams? Also, the biggest gains seem to be in the jump 
> from 1 to 8 data streams; with 8 streams I was able to get throughput up to 
> 6500Mbit/sec. 
> Our goal for 10GbE, is to be able to restore data from tape at a speed of at 
> least 2400Mbit/sec (300MB/sec). We have large daily backups (3-4TB) that we 
> would like to be able to restore (not backup) in a reasonable amount of time. 
> These restores are used to refresh our test and development environments with 
> current data. The actual backups are done with array based snapshots (HDS 
> ShadowCopy), which then get mounted and backed up by a dedicated media server 
> (6-core T2000). We're currently getting about 650MB/sec of throughput with 
> the backups (9 streams on 3 LTO3 tape drives - MPX=3 and it's very 
> compressible data). 
> Going off my iperf results, the restoring this data using 9 streams should 
> get us well over 2400Mbit/sec. But - we haven't installed the cards on our 
> media servers yet, so I have yet to see what the actual performanee of 
> netbackup and LTO3 over 10GbE is. I'm hopeful it'll be close to the iperf 
> results, but if it doesn't meet the goal then we'll be looking at other 
> options. 
> -- 
> Devon Peters


+----------------------------------------------------------------------
|This was sent by dan.pancamo AT aiminvestments DOT com via Backup Central.
|Forward SPAM to abuse AT backupcentral DOT com.
+----------------------------------------------------------------------


_______________________________________________
Veritas-bu maillist  -  Veritas-bu AT mailman.eng.auburn DOT edu
http://mailman.eng.auburn.edu/mailman/listinfo/veritas-bu