Networker

Re: [Networker] AX150 performance problems.

2007-03-12 19:33:45
Subject: Re: [Networker] AX150 performance problems.
From: Peter Viertel <Peter.Viertel AT MACQUARIE DOT COM>
To: NETWORKER AT LISTSERV.TEMPLE DOT EDU
Date: Tue, 13 Mar 2007 10:27:27 +1100
Don't know exactly what your AX150 has in it, but I can see from the
references to FLARE and Navisphere that it's a clariion rebadge...

Our Clariions have two Service processors  -  to get the best
performance  we figured out that each AFTD should be a filesystem
striped between 2 LUNs, with each LUN managed by a different SP, as
these are the bottlenecks for writing to RAID3 or RAID5.

On the solaris 9 box I use a 64mb wide stripe with SVM between two 5
disk RAID3 LUNs for each filesystem - same disks as you have. When doing
a full backup run I have 6 filesystems each with one AFTD on them, the
target parallelism of the AFTD's is 1 to spread the load - the result is
around 100MB/sec, which we live with but certainly doesn't bear much
resemblance to the numbers in the brochure. We avoid reading at the same
time by managing staging manually, but the striping helps when we
absolutely have to read at the same time.

The LUNs we use are RAID-3 -   that's what EMC said to do in some old
whitepaper, so that's what we all do. IMHO The RAID3 recommendation
comes straight out of university theory, and takes absolutely no account
of how ufs turns sequential traffic into random traffic. Our tests with
ZFS has been very encouraging in that they indicate a much more
'sequential' type of writing.

We are setting up a new system, using 500GB disks -   we will try
combinations of RAID3 RAID5 RAID10 and maybe even raw +zfs RAIDZ,
although not real sure if that's a good idea yet.. With zfs managing the
pool of LUNs we can forget about trading off number of filesystems
(driven by how many tape drives you have to stage to) against LUNs, but
there seems to be a problem in that we can't shape the pool to ensure
load is spread evenly between SP's, that's why I'm considering going to
RAW or RAID1/10 because these minimise the load on the SP's.

Happy to share any thoughts and opinions on this - other than people
telling me about virtual tape and how good it is.

-----Original Message-----
From: EMC NetWorker discussion [mailto:NETWORKER AT LISTSERV.TEMPLE DOT EDU] On
Behalf Of Adrian Saul
Sent: Monday, 12 March 2007 10:19 PM
To: NETWORKER AT LISTSERV.TEMPLE DOT EDU
Subject: Re: [Networker] AX150 performance problems.

We have similar staging performance issues with a Clariion from an 
inherrited config which hit brick walls when its load was increased from

the initial light setup.

In our case the issue comes down to a number of factors:

1. LUN type is RAID-3 - more on this later
2. The disks are 320G 5400 PATA disks - no idea why other cost
3. Multiple AFTD devices on the same LUNs.

The issue is that RAID-3 is a recommended setup for streaming sequential

transfers (like staging) *PROVIDED* you are reading a single stream. 
Because we have multiple AFTD's per LUN, we were doing multiple reads 
from the LUNs (i.e effectively scattered read) which killed the RAID-3 
performance.  This is because RAID-3 synchronises all the disk heads and

needs to read from all disks to satisfy a read.  Because each read has 
to wait for all 5 disks in the RAID to reseek, it was killing the 
performance as the disks rattled around for each new read, and being 
PATA didnt help.

Changing to 2 AFTD's instead of 4 picked performance up substantially. 
We are trying to change to a single AFTD per LUN with some better disks 
- work in progress there though.

Co-incidently I am looking at sequential I/O issues on a Sun Fire 6900 
which comes down to UFS - benchmarked UFS at 20Mb/s, ZFS at 210Mb/s on 
the same device (DMX-3 LUN).  Yes - 10 fold improvement and it wasnt 
filesystem caching.

For UFS you want to tune your filesystems to overcome some UFS caused 
bottlenecks.  Set maxphys and md:md_maxphys to 1Mb (0x100000), and use 
tunefs to change maxcontig to 128.  You should see much better 
sequential I/O performance from that.

See how that helps out.

Cheers,
        Adrian





Yaron Zabary wrote:
> Hello all,
> 
>   Recently, we purchased a Dell/EMC AX150SC storage for the disk
backup 
> option. We have some performance problems with it which are described
in 
> the letter below (which I sent to my Dell/EMC support). Does anyone
have 
> a similar setup and can tell if they see similar problems or were able

> to get decent performance from this machine ?
> 
> 
> 
> -------- Original Message --------
> Subject: AX150 performance problems.
> Date: Mon, 05 Mar 2007 12:24:05 +0200
> From: Yaron Zabary <yaron AT aristo.tau.ac DOT il>
> To: michaelz AT bynet.co DOT il
> 
> Hello,
> 
>    I have performance problems with the AX150SC we have. These are the
> details:
> 
>   . The AX150 is connected to a single Sun 280R server via an Emulex
HBA.
> 
>   . The Server is a Sun 280R with two 750Mhz CPUs and 2Gb RAM running
> Solaris 9 (SunOS legato.tau.ac.il 5.9 Generic_118558-39 sun4u sparc
> SUNW,Sun-Fire-280R). I applied all patches required by EMC for Solaris
> 9. The server also has four LTO2 drives connected to another dual-LVD
HBA.
> 
>   . The HBA is Emulex 9802 running the latest firmware (1.91a5) with
> driver 6.10g. The HBA shares the 33Mhz PCI bus with a GigE card. The
LVD
> HBA has its own 66Mhz PCI bus.
> 
>   . The AX150SC has 12 500Gb disks. Its Flare was upgraded to
> 02.20.150.5.022. All disks are in a single pool with a single hot
spare.
> Currently, there are two 2Tb LUNs  defined (due to Solaris 2Tb
> limitations). Only the first LUN is in use. Event log is empty. All
> components are green on the view/component screen.
> 
>   . /etc/system has the following lines (as required by EMC):
> 
> set sd:sd_max_throttle=20
> set sd:sd_io_time=0x3c
> set ssd:ssd_max_throttle=20
> set ssd:ssd_io_time=0x3c
> 
>   . The backup software we use is EMC Networker version 7.2.2 build
> 422. Networker uses the first LUN for disk bacckup (using advfile
> device). The backups are then staged (copied) to the LTO2 drives.
> 
>   . The problem starts when the server tries to stage savesets from
the
> disk while running backups. During such operations the read speed can
be
> as low as 500Kb/s. Write speed also drops from ~25Mb/s to ~10Mb/s.
While
> this happens 'sar -d' reports a huge number for average wait time for
> I/O requests (over 1.5 seconds per request). Also notice the huge
queue
> (over 300 requests) and the large service time (~100msec). The numbers
> below are 10 minutes averages:
> 
>            sd60             99   306.6     190   11077  1508.4   101.0
>            sd60,a           99   306.6     190   11077  1508.4   101.0
>            sd60,h            0     0.0       0       0     0.0     0.0
>            sd60            100   350.0     187   10387  1769.8   106.1
>            sd60,a          100   350.0     187   10387  1769.8   106.1
>            sd60,h            0     0.0       0       0     0.0     0.0
>            sd60            100   348.3     194   10767  1690.5   102.0
>            sd60,a          100   348.3     194   10767  1690.5   102.0
>            sd60,h            0     0.0       0       0     0.0     0.0
>            sd60            100   311.7     199   13024  1470.7    99.0
>            sd60,a          100   311.7     199   13024  1470.7    99.0
>            sd60,h            0     0.0       0       0     0.0     0.0
>            sd60            100   364.3     209   10714  1644.8    94.9
>            sd60,a          100   364.3     209   10714  1644.8    94.9
>            sd60,h            0     0.0       0       0     0.0     0.0
>            sd60            100   375.6     193   10093  1840.4   103.0
>            sd60,a          100   375.6     193   10093  1840.4   103.0
>            sd60,h            0     0.0       0       0     0.0     0.0
>            sd60            100   346.1     197   11119  1653.3   100.4
>            sd60,a          100   346.1     197   11119  1653.3   100.4
>            sd60,h            0     0.0       0       0     0.0     0.0
>            sd60            100   341.2     197   11666  1628.1   100.5
>            sd60,a          100   341.2     197   11666  1628.1   100.5
>            sd60,h            0     0.0       0       0     0.0     0.0
>            sd60            100   339.6     196   11948  1629.5   100.8
>            sd60,a          100   339.6     196   11948  1629.5   100.8
>            sd60,h            0     0.0       0       0     0.0     0.0
> 
> 
>   Please advice.
> 
> -- 
> 
> -- Yaron.
> 
> 

To sign off this list, send email to listserv AT listserv.temple DOT edu and
type "signoff networker" in the body of the email. Please write to
networker-request AT listserv.temple DOT edu if you have any problems with this
list. You can access the archives at
http://listserv.temple.edu/archives/networker.html or
via RSS at http://listserv.temple.edu/cgi-bin/wa?RSS&L=NETWORKER

NOTICE
This e-mail and any attachments are confidential and may contain copyright 
material of Macquarie Bank or third parties. If you are not the intended 
recipient of this email you should not read, print, re-transmit, store or act 
in reliance on this e-mail or any attachments, and should destroy all copies of 
them. Macquarie Bank does not guarantee the integrity of any emails or any 
attached files. The views or opinions expressed are the author's own and may 
not reflect the views or opinions of Macquarie Bank.

To sign off this list, send email to listserv AT listserv.temple DOT edu and 
type "signoff networker" in the body of the email. Please write to 
networker-request AT listserv.temple DOT edu if you have any problems with this 
list. You can access the archives at 
http://listserv.temple.edu/archives/networker.html or
via RSS at http://listserv.temple.edu/cgi-bin/wa?RSS&L=NETWORKER

<Prev in Thread] Current Thread [Next in Thread>