ADSM-L

Re: [ADSM-L] protect pool plus replicate node equals poor replication efficiencies

2016-09-15 12:42:32
Subject: Re: [ADSM-L] protect pool plus replicate node equals poor replication efficiencies
From: Stefan Folkerts <stefan.folkerts AT GMAIL DOT COM>
To: ADSM-L AT VM.MARIST DOT EDU
Date: Thu, 15 Sep 2016 18:41:49 +0200
>Do I have this right so far?

No, I think he is under the impression the data is send twice, it looks
that way a little the way Spectrum Protect reports on the replication
proces, but it's representing the data..not actually sending it, it is only
sending metadata of that data.


On Thu, Sep 15, 2016 at 6:24 PM, Ryder, Michael S <michael_s.ryder AT roche DOT 
com
> wrote:

> Something doesn't make sense.
>
> You run a backup - node's data is stored in a pool on server A.
>
> Then, you protect pool, and a copy of the de-duped data is sent to the
> protect pool storage, also on server A.
>
> Then, you replnode, and a node is replicated to server B.  You are
> surprised to find that data is being sent from server A to server B.
>
> Do I have this right so far?
>
> Until you issue the replnode command, how is server B supposed to know
> about the data in server A's storage pools?
>
> Don't you still need to copy the data at least once from server A to server
> B?  Isn't this normal?
>
> Best regards,
>
> Mike, x7942
> RMD IT Client Services
>
> On Thu, Sep 15, 2016 at 12:02 PM, Stefan Folkerts <
> stefan.folkerts AT gmail DOT com
> > wrote:
>
> > Do you have a fast Spectrum Protect database / active log?
> > We run 2.4TB of metadata per hour with replication (note, this is not
> > actual data, this is metadata representing 2.4TB of data).
> > But that system has SSD's and runs in excess of 140.000 IOP/s in Spectrum
> > Protect database benchmarks.
> > I would think this is very much database (and active log) performance
> bound
> > (on both sides).
> >
> >
> >
> >
> > On Thu, Sep 15, 2016 at 5:50 PM, Nixon, Charles D. (David) <
> > cdnixon AT carilionclinic DOT org> wrote:
> >
> > > Best I can tell, it is transferring the data over the wire and support
> > > stated as much.  We are currently using the replnode for that single
> node
> > > so it's getting the default of 10 sessions and appears to be using all
> of
> > > them, for the four hours or so that it's sending data.
> > >
> > > I don't have a good way to see server's bandwidth but network IO chart
> > > implies that it's not sending a great amount of data but that may be
> due
> > to
> > > the 846GB over 4.5 hours.
> > >
> > > 09/15/16   10:58:08      ANR0327I Replication of node NODENAME
> completed.
> > > Files
> > >                           current: 70,341. Files replicated: 752 of
> 752.
> > > Files
> > >                           updated: 602 of 602. Files deleted: 692 of
> 692.
> > > Amount
> > >                           replicated: 12,487 GB of 12,487 GB. Amount
> > > transferred:
> > >                           846 GB. Elapsed time: 0 Days, 4 Hours, 28
> > > Minutes.
> > >                           (SESSION: 414242, PROCESS: 539)
> > > ---------------------------------------------------
> > > David Nixon
> > > Storage Engineer II
> > > Technology Services Group
> > > Carilion Clinic
> > > 451 Kimball Ave.
> > > Roanoke, VA 24015
> > > Phone: 540-224-3903
> > > cdnixon AT carilionclinic DOT org
> > >
> > > Our mission: Improve the health of the communities we serve.
> > >
> > >
> > >
> > > ________________________________________
> > > From: ADSM: Dist Stor Manager [ADSM-L AT VM.MARIST DOT EDU] on behalf of
> Stefan
> > > Folkerts [stefan.folkerts AT GMAIL DOT COM]
> > > Sent: Thursday, September 15, 2016 10:45 AM
> > > To: ADSM-L AT VM.MARIST DOT EDU
> > > Subject: Re: [ADSM-L] protect pool plus replicate node equals poor
> > > replication efficiencies
> > >
> > > >Support confirmed that the amount of data replicated in a replnode
> > command
> > > is the same, regardless of the protect pool command status.
> > >
> > > I think that this is only in the statistics, not in the actual transfer
> > on
> > > the wire.
> > > the replnode should not transmit actual data if the data was send by
> the
> > > protect storagepool command.
> > > Are you running enough (but not to many) parallel processes for the
> > > replicate node command so it can perform optimally?
> > >
> > > I'm using this setup for multiple customers and it worked fine for us
> so
> > > far.
> > >
> > > http://imgur.com/a/mT6ux
> > >
> > >
> > >
> > >
> > > On Thu, Sep 15, 2016 at 4:29 PM, Nixon, Charles D. (David) <
> > > cdnixon AT carilionclinic DOT org> wrote:
> > >
> > > > We opened a ticket related to long replication times in a container
> > pool
> > > > after replication takes place, and got an answer that 'we can
> recreate
> > > your
> > > > problem but it is likely working as designed' even though it's
> contrary
> > > to
> > > > documentation.  Any ideas would be appreciated.
> > > >
> > > > -Two TSM servers at 7.1.5
> > > > -Single client going to a single container.  Client backs up 12TB a
> > night
> > > > and after dedupe/compression, we see a 1TB change rate
> (approximately).
> > > > -Once the backup is complete, we run a protect pool.  It's expected
> > that
> > > > this process will ship 1TB to the DR site.  -Protect completes
> > > successfully.
> > > > -a replnode is issued against the node and TSM spends the next 4
> hours
> > > > replicating data to the DR site
> > > >
> > > > Support confirmed that the amount of data replicated in a replnode
> > > command
> > > > is the same, regardless of the protect pool command status.  However,
> > the
> > > > documentation leads me to be that if you have already protected the
> > pool,
> > > > the replnode should be a metadata only transfer.
> > > >
> > > > So while we are able to transfer and complete the processes, it seems
> > to
> > > > 'cost' us quite a bit in both IO and WAN usage to do so using
> > containers,
> > > > defeating the point of using containers to reduce replication costs.
> > Any
> > > > ideas as to what is going on?
> > > >
> > > > ---------------------------------------------------
> > > > David Nixon
> > > > Storage Engineer II
> > > > Technology Services Group
> > > > Carilion Clinic
> > > > 451 Kimball Ave.
> > > > Roanoke, VA 24015
> > > > Phone: 540-224-3903
> > > > cdnixon AT carilionclinic DOT org
> > > >
> > > > Our mission: Improve the health of the communities we serve.
> > > >
> > > > ________________________________
> > > >
> > > > Notice: The information and attachment(s) contained in this
> > communication
> > > > are intended for the addressee only, and may be confidential and/or
> > > legally
> > > > privileged. If you have received this communication in error, please
> > > > contact the sender immediately, and delete this communication from
> any
> > > > computer or network system. Any interception, review, printing,
> > copying,
> > > > re-transmission, dissemination, or other use of, or taking of any
> > action
> > > > upon this information by persons or entities other than the intended
> > > > recipient is strictly prohibited by law and may subject them to
> > criminal
> > > or
> > > > civil liability. Carilion Clinic shall not be liable for the improper
> > > > and/or incomplete transmission of the information contained in this
> > > > communication or for any delay in its receipt.
> > > >
> > >
> > > ________________________________
> > >
> > > Notice: The information and attachment(s) contained in this
> communication
> > > are intended for the addressee only, and may be confidential and/or
> > legally
> > > privileged. If you have received this communication in error, please
> > > contact the sender immediately, and delete this communication from any
> > > computer or network system. Any interception, review, printing,
> copying,
> > > re-transmission, dissemination, or other use of, or taking of any
> action
> > > upon this information by persons or entities other than the intended
> > > recipient is strictly prohibited by law and may subject them to
> criminal
> > or
> > > civil liability. Carilion Clinic shall not be liable for the improper
> > > and/or incomplete transmission of the information contained in this
> > > communication or for any delay in its receipt.
> > >
> >
>

<Prev in Thread] Current Thread [Next in Thread>