[Search for users] [Overall Top Noters] [List of all Conferences] [Download this site]

Conference decwet::networker

Title:NetWorker
Notice:kits - 12-14, problem reporting - 41.*, basics 1-100
Moderator:DECWET::RANDALL.com::lenox
Created:Thu Oct 10 1996
Last Modified:Fri Jun 06 1997
Last Successful Update:Fri Jun 06 1997
Number of topics:750
Total number of notes:3361

680.0. "100Gb/hr possible? Please comment" by AUSSIE::MCNAMARA (I can smile but my tail don't wag) Wed May 14 1997 03:37

    G'Day All,
    		We have a customer that would like to backup 100Gbytes in 1
    hr. We are dealing through a partner so information is sketchy but we
    will include a "mileage may vary" statement but I would like to ask
    your advice/comments.
    
    System
    
    Alphaserver 8400 8 CPU 440MHz 2GB Memory (2 * 1Gbyte boards)
    KFTHA-AA Hose adapter
    2 * DWLPR PCI shelf
    
    First PCI Shelf
    4 * KZPSA-BB fast wide controllers (1 per HSZ52)
    	--->48 * RZ29B-VW discs spread accross 4 * HSZ52 controllers
    DE500 Ethernet adapter
    
    2nd PCI shelf
    6 * KZPSA-bb (1 per TZ89N in the silo)
    	---> 1 * TL896 6 drive silo.
    
    Software Digital UNIX 4.0a Latest NSR (4.3?)
    
    Customer software is (I think) a progress database. They will do a cold
    backup. (hence a 1 hr window)
    
    Assumptions:
    
    If we can get enough file sets tokeep all 6 drives streaming we can
    expect say 6Gbytes/sec on each drive compressed.
    
    That the pre and post processing on a 8 CPU 8400 maight not take very
    long.
    
    That 4 pairs of HSZ50s can feed ~40Mbytes/sec
    
    Is the above reasonable or have I got this very wrong.
    
    Thanks & Regards Scott
T.RTitleUserPersonal
Name
DateLines
680.1DECWET::ONOSoftware doesn't break-it comes brokenWed May 14 1997 16:4234
Scott,

  NetWorker for Digital UNIX V4.3 greatly improves local backup 
  performance over previous versions.  It will ship on the July 
  1997 LPCD.

  Given the configuration you have, I expect that 100 GB/hour 
  should be achieveable.  You might even hit 150 if you spread the 
  source data out enough.

  Now, to your assumptions:

>    If we can get enough file sets tokeep all 6 drives streaming we can
>    expect say 6Gbytes/sec on each drive compressed.
    
  Sounds reasonable.  This assumes a 1.5:1 compression ratio.

>    That the pre and post processing on a 8 CPU 8400 maight not take very
>    long.
    
  Don't know how long it takes to shut down / start up a Progress 
  database.

>    That 4 pairs of HSZ50s can feed ~40Mbytes/sec
    
  Sounds reasonable, given how you have the disks spread across the 
  adapters.  You need to be pulling data from all four pairs of 
  HSZs though.  If you overload one and underutilize the others, 
  you may not achieve the performance you might otherwise be 
  capable of.

Regards,

Wes
680.2KITCHE::schottEric R. Schott USG Product ManagementFri May 16 1997 09:5422
>    
>    Assumptions:
>    
>    If we can get enough file sets tokeep all 6 drives streaming we can
>    expect say 6Gbytes/sec on each drive compressed.

Do you think this is 6MB/sec?

My napkin says you need 30 MB/sec to achieve 100GB in 1 hour.

The other issue is what is the state of the database  during
the backup?

>    
>    That the pre and post processing on a 8 CPU 8400 maight not take very
>    long.
>    
>    That 4 pairs of HSZ50s can feed ~40Mbytes/sec
>    
>    Is the above reasonable or have I got this very wrong.
>    
>    Thanks & Regards Scott
680.3math factsDECWET::EVANSNSR EngineeringFri May 16 1997 12:2213
we discovered that one needs to multiply X Mb/sec by 3.515 to get Gb/hr
 since it is 1024, not 1000.  I think the long form looks like:

        X Mb          1024 Mb/Gb
      ---------  =  --------------
        1 sec         3600 sec/hr

   X * 3600
   --------  = Y Gb/hr
     1024

this made it easy for us to run estimates of Gb/hr when we were doing our
 performance tests.
680.4TL896 is 6 drive silo... thanksAUSSIE::MCNAMARAI can smile but my tail don't wagSun May 18 1997 20:2013
    G'Day All,
    		Thanks for your replies.
    
    My assumption is based as follows:-
    
    The TL896 has 6 TZ89 drives. If we can get all 6 drives streaming then...
    
    6 drives @ 6Mbytes/sec = 36Mbytes/sec 
    36Mbytes/sec * 3600 sec is approx 125Gbytes/Hr.
    
    Thanks & Regards Scott
    
    
680.5USCTR1::lexser24.lex.dec.com::AscheryaddaTue May 20 1997 12:5510
Is there some upper limit on what we believe networker can actually push 
through a system - ie assuming that there is enough i/o bandwidth, then what 
is the best peformance that can be expected from Networker - say in V4.3?
And do we have any characterization of the amount of cpu that driving 100 
Gb/hr might take? 200 GB/hr? 500 GB/hr?

thanks,

d
680.6Some datapointsDECWET::FARLEEInsufficient Virtual um...er....Wed May 21 1997 10:5622
Yes, we do have some preliminary information on that now, and with each
round of refinement we're coming up with ways to make it more efficient.

On an 8400:
The most we've seen ANY software push through the benchmark system
( an 8x330 MHz CPU, 8GB 8400 with LOTS of I/O capacity) is just over
500 GB/Hr.  I'm still trying to figure out where that limit comes from.
This includes benchmarks from NetWorker V4.3, Alexandria, Open Vision, and
several others.

At a data rate of 500 GB/Hr, NetWorker consumed approximately 27% of the
8 CPUs.

This is a local, cold backup of raw partitions.

One smaller datapoint, last week in Berlin I did a demo using an AlphaServer
1000A with a couple shleves of RZ29 disks, and a TL894 library.  we achieved
data rates of approximately 50 GB/Hr using about 40% of its one CPU.

Does this help?

Kevin
680.7Another data pointDECWET::KOWALSKIOfficial Beer Test DummieMon May 26 1997 10:3212
    And I did a demo at the same time as Kevin, but in Ayr.
    
    8400 with 4 400 MHz CPUs, 4 GB memory
    2 TL894's with each TZ89 drive (8 total) on a separate KZPSA
    36 RZ29's, 6 drives per KZPSA
    
    Backing up the drives as raw partitions, we achieved 260 GB/hr. The
    average CPU usage was ~60%; however, this was a very quick setup of the
    just-uncrated components and essentially NOTHING was done to  optimize
    the throughput other than the cabling implied above.
    
    Mark
680.8interestingUSCTR1::ASCHERDave AscherThu May 29 1997 14:0427
re:       <<< Note 680.7 by DECWET::KOWALSKI "Official Beer Test Dummie" >>>
.                           -< Another data point >-
.
.    And I did a demo at the same time as Kevin, but in Ayr.
.    
.    8400 with 4 400 MHz CPUs, 4 GB memory
.    2 TL894's with each TZ89 drive (8 total) on a separate KZPSA

    what is a TL894? in terms of speed?
    
.    36 RZ29's, 6 drives per KZPSA
.    
.    Backing up the drives as raw partitions, we achieved 260 GB/hr. The
.    average CPU usage was ~60%; however, this was a very quick setup of the
.    just-uncrated components and essentially NOTHING was done to  optimize
.    the throughput other than the cabling implied above.
.   
    I gather that there is some kind of widely held assumption
    that backing up raw partitions provides a significant boost
    in throughput for these kinds of tests? Is that part of how Netview
    got its famous 500G/hr? or was that primarily due to using
    a tape technology with faster transfer speeds?
   
    What else might be done to optimize the throughput?

    d

680.9DECWET::FARLEEInsufficient Virtual um...er....Thu May 29 1997 14:5315
>What is a TL894?

	It is a successor to the TL810/812, with 4 TZ89 drives
	and 48 tapes.  The TZ89 will move data significantly faster
	than previous generations of DLTs (6-7MB/sec and more)

>Raw partitions faster?
	Raw partitions are used for two reasons:  One: it accurately
	simulates how a customer would back up a large database which
	was built on raw partitions (which is usually Oracle's recommendation)
	and two: It is more efficient because you avoid the UBC buffer cache
	which is not helpful to us, and also you avoid the work of walking
	and indexing the filesystem structure.

Kevin
680.10Network backup is a different storyDECWET::KOWALSKIOfficial Beer Test DummieThu May 29 1997 16:2410
    And remember, these data points are for backup of the NetWorker server
    by itself, NOT network client backups.  The NetWorker team's objective
    with V4.3 for DU was to optimize backup of servers with large
    databases.  The conditionals you're seeing listed (backup of the server
    to itself, raw partitions (ie, Oracle hot or cold backup), minimized
    file indexing) are typical for that type of enterprise-serving system. 
    Those conditions do happen to lend themselves to more optimal
    throughput than not; however, the performance effort didn't start with
    the idea that our "thumb would be on the scale" by using them.
    They were used because they were the market requirements.
680.11sounds pretty good...USCTR1::canb22.nio.dec.com::AscheryaddaFri May 30 1997 08:0411
After re-reading the previous notes it sounds a lot like there has been a benchmark using 
nsr 4.3 that actually achieved 500G/hr... same as Alexandria and OpenVision and 'several 
others'. Have we ever announced that? It's news to me... and I don't know anybody else 
who can recall hearing about it. 

We ALL know about the OpenVision benchmark results of 500 G /hr with fairly low cpu 
utilization. It might be useful to hear what NSR's cpu utilization was in the benchmark 
scenario... if that is not too sensitive.

d
680.12DECWET::FARLEEInsufficient Virtual um...er....Fri May 30 1997 11:3612
I announced these results at the UNIX symposia in Nashua and in Berlin.
The test results were taken the week before the Nashua symposium.

The CPU load for the cold backup (500 GB/Hr) was 27%.  I am currently testing
code which may drop that by a bit.  I should know by next week.

I'm sure that there will be a whitepaper forthcoming, but it takes an
amazingly frustrating time to actually get something like that out.
The last whitepaper that I wrote in November didn't see the light of day until
mid January...

Kevin
680.14sounds good to me..USCTR1::canb22.nio.dec.com::AscheryaddaFri May 30 1997 12:077
Congrats on the benchmark. Is it reasonable to assume that the hardware platform was 
pretty much the same as that used for the OpenView benchmark? Is there some place I could 
see what you announced at the unix symposium?

thanks much...

d
680.15DECWET::FARLEEInsufficient Virtual um...er....Fri May 30 1997 16:526
Both benchmarks were done on the very same hardware.
I'll see about publishing my slides along with enough notes for them
to make sense (I don't believe in just reading slides in presentations,
so that slides aren't necessarily self-explanatory)

Kevin