T.R | Title | User | Personal Name | Date | Lines |
---|
854.1 | More computes needed | STAR::STOCKDALE | | Thu Feb 11 1993 13:43 | 16 |
| >> Any ideas on how to tackle this problem ?
Get more computes.
Much of the I/O processing shows up as interrupt stack usage
so if you have a lot of network traffic showing 100% interrupt stack
is not unusual. Its not that the system is being swamped by
interrupts from the DEMFA.
On OpenVMS AXP, much work has been done to improve performance in
the LAN drivers. This work will show up in a future release of
OpenVMS VAX. But this only reduces the amount of time in the
LAN drivers which represents just a portion of the amount of
computes needed to do the backups.
- Dick
|
854.2 | ..only if througput increases.. but it decreases ! | BONNET::STRATMAN | Peter Stratman @VBO | Fri Feb 12 1993 06:20 | 17 |
| >Much of the I/O processing shows up as interrupt stack usage
>so if you have a lot of network traffic showing 100% interrupt stack
>is not unusual. Its not that the system is being swamped by
>interrupts from the DEMFA.
OK, I know that increasing the IO performance will put more strain on the
CPUs too, that's normal. But overal throughput should improve.
What I'm trying to understand however is how I can have 2 identical
machines, one on FDDI, one on Ethernet, with a very similar network IO
workload, and getting only half the Ethernet machine's throughput on the
FDDI machine !
Agree that this is abnormal ?
Peter.
|
854.3 | Strange indeed... | KONING::KONING | Paul Koning, A-13683 | Fri Feb 12 1993 11:49 | 4 |
| Yes. Especially since the DEMFA is very similar to the DEMNA/DEBNI, both of
which are more efficient than the DEBNA...
paul
|
854.4 | The Picture.. | BONNET::STRATMAN | Peter Stratman @VBO | Fri Feb 12 1993 13:27 | 29 |
|
Hmmm... I replaced DEBNA's with a DEMFA, but I'm COMPARING with a node
in the cluster which has a DEMNA...
Still, performance should not be less...
This is the picture :
------------------------
====| Concentrator |=======
------------------------
| | |
| ---------------- |
| |DECBRIDGE 500 | |
| ---------------- |
| | |
| |---------------| |
| | |<------ Throughput here is 1/2 of
| |<----------|------- here...
............|..........|...........|............
. | | | .
. ----DEMFA --DEMNA-- DEMFA---- .
. | 6530 | | 6530 | | 6530 | .
. --------- --------| --------- .
. | | | .
. | CI | | .
. ----------------*--------------- .
. |HSCs, disks .
..CI Cluster....................................
|
854.5 | | STAR::GAGNE | David Gagne - VMS/LAN Development | Fri Feb 12 1993 13:42 | 19 |
| The driver for the DEMNA is much more efficient with regards to CPU
cycles than the driver for the DEMFA. The DEMFA driver was the first
FDDI driver from VMS and it has extra code for checking that it's
doing the right thing. Due to release restrictions for C2
certification of VMS, we were not able to improve the DEMFA driver
until a future release of VMS - which will probably ship in the
next calender year (1994).
One area where you can get CPU cycles back is in the size of your
LRPs. Typically your LRP size is too small for FDDI packets; thus
causing all FDDI buffer allocations to be much more costly than
Ethernet buffer allocations. To reduce the cost for allocating
FDDI buffers, increase your LRP size to about 4700 (I don't recall the
exact number, but I believe there's a memo available about this - it
may even be in the release notes - it was a couple of years ago that
the FDDI driver first shipped). The reason we don't have this done
automatically is because there are too many pieces of VMS that use
memory for us to determine the impact to the other pieces of VMS
that will no longer get their buffers from the LRP look aside list.
|
854.6 | If you can boost the performance, send the patches | MUDDY::WATERS | | Fri Feb 12 1993 14:02 | 9 |
| > doing the right thing. Due to release restrictions for C2
> certification of VMS, we were not able to improve the DEMFA driver
> until a future release of VMS - which will probably ship in the
> next calender year (1994).
C'mon. If you have a significantly better driver, provide it to the
customers that want acceptable performance from their costly CPUs.
These customers may not be concerned about loosing C2 certification.
--gw
|
854.7 | increase in LRPSIZE useful for DECNET ? | BONNET::STRATMAN | Peter Stratman @VBO | Mon Feb 15 1993 12:31 | 24 |
|
re .5
David, the interface is basically being used for DECNET traffic.
So, if I have understood previous notes on the subject, packet size
will not exceed Ethernet packet size. It will not, in this case, make
much sense to increase LRP size, will it ?
re .6
I agree, if the source of the "problem" is due to known
inefficiencies in the driver, then please don't make us wait for a
future VMS version if not absolutely necessary.
Anyway, hey, I'm not only getting high interrupt stack time, but also
consistently getting 50% worse performance than with a DEMNA. If this
were really due to the driver, then inefficiency is not the right
term to use ! ... This can't be due to the driver. ... or can it ?
Peter.
|
854.8 | | STAR::GAGNE | David Gagne - VMS/LAN Development | Mon Feb 15 1993 12:40 | 7 |
| The receive buffers allocated by the driver are for full size FDDI
packets - the driver doesn't know what size packets will be arriving.
So the LRP size may help receive and hinder transmit.
The more efficient driver cannot be used until the future because it
was written for a newer common routine interface which will only be
available in a future release of OpenVMS VAX.
|
854.9 | LRPSIZE - formula needed | STKHLM::ALMERLOV | Karl Jonas Almerl�v, Stockholm | Thu Feb 18 1993 12:33 | 22 |
| I have tried to find out what is a reasonable value for LRPSIZE when
using FDDI as a cluster interconnect, but the values differ a lot.
Our customer is running VMS v5.5-2 and hasn't started using FDDI yet.
In the VAXcluster MDF configuration guidelines the recommendation is to
have a LRPSIZE of 4471.
In the VAXcluster MDF facility v 1.0a release notes it says LRPSIZE
should be 4474.
854.5 in this conference says about 4700, referring to a memo or the
release notes (which release notes? I have checked the VMS release
notes with no success).
Has somebody out there got an exact figure (or even better, a formula)?
Karl Jonas
|
854.10 | | STAR::GAGNE | David Gagne - VMS/LAN Development | Thu Feb 18 1993 17:19 | 30 |
| 4700 was my guess - of course I knew that the correct number was
documented somewhere - although I am quite sure that 4471/4474 are
wrong.
The formula is (without spending another hour verifying this):
You want the FDDI receive buffers to come from the LRP lookaside list.
Since LRPSIZE is set to 1504 on a system with Ethernet, then you need
to know the maximum space needed for FDDI (on V5.5-2) and subtract
14 from it.
The space needed for FDDI (at least the DEMFA) is:
4495 (maximum packet size)
3 (extra bytes for device packet control)
31 (for beginning alignment)
31 (for ending alignment)
-----
4560
- 14 (subtract the amount that is already accounted for)
-----
4546
I would add some extra to this because it would be sad to find out that
the number was off by a few bytes. So I would use the value 4578. You
should test this before having a customer try it.
Maybe we can try this here if we ever get some free time.
Good luck.
|
854.11 | LRPSIZE=4541 for FDDI and VMS 5.5-2! | VCSESU::WADE | Bill Wade, VAXc Systems & Support Eng | Mon Feb 22 1993 12:54 | 28 |
|
I did some testing this morning on a VAX6000 equipped with a DEMFA running
VMS 5.5-2.
I found that the magic number for LRPSIZE was 4541. With this value -
$SHOW MEMORY
.
.
.
Large Packet (LRP) Lookaside List Packets Bytes Pages
Current Total Size 800 3916800 7650
Initial Size (LRPCOUNT) 800 3916800 7650
Maximum Size (LRPCOUNTV) 4000 19584000 38250
Free Space 728 3564288
Space in Use 72 352512
Packet Size/Upper Bound (LRPSIZE + 355) 4896
Lower Bound on Allocation 1088
With this value I found that all the 4896 byte VCRPs were allocated from the
LRP lookaside list. Setting LRPSIZE <4541 caused the 4896 byte VCRPs to be
allocated from nonpaged pool.
Bill
|
854.12 | | VCSESU::WADE | Bill Wade, VAXc Systems & Support Eng | Mon Feb 22 1993 17:02 | 21 |
|
More on LRPSIZE -
The only reference that I've found for setting LRPSIZE on an FDDI equipped
VAX system is in the release notes for VMS 5.4-3. The value given is 4474
(pg. 2-30). Earlier investigation has shown this to be insufficient in
that the allocation is still from nonpaged pool and not from the LRP
lookaside list.
There is no reference to LRPSIZE in the 5.5-1 or 5.5-2 release notes or
the VAXc SPD. The MDF product set the number based on the number in the
5.4-3 release notes.
Is there agreement that this is a severe performance problem and
a QAR against VMS is required and AUTOGEN should set LRPSIZE dependent
on an FDDI adapter?
|
854.13 | | STAR::GAGNE | David Gagne - VMS/LAN Development | Tue Feb 23 1993 10:17 | 4 |
| The next release of VMS due out does not use the LRPSIZE parameter.
Pool management has been changed such that there are tons of lookaside
lists. So you can file a QAR; but it will probably be closed
immediately as "fixed in next release".
|
854.14 | what is really the next release?? | CVG::STRICKER | Cluster Validation Group, Salem NH | Tue Feb 23 1993 10:29 | 4 |
| Isn't VIKING the next release... Shouldn't the VIKING kit contain
some of these changes with it's new FDDI adapter support? Before the
next big release (BLADE) which contains all of the Pool Management
changes...
|
854.15 | | STAR::GAGNE | David Gagne - VMS/LAN Development | Tue Feb 23 1993 11:20 | 5 |
| VIKING is a release for hardware support. So it will not include
the new memory management code or bug fixes for generic V5.5-2 bugs.
VIKING is (to me) just a variant of V5.5-2 (as its name will show).
The new memory management code is in the release named BLADE; which
will probably be numbered 6.0.
|
854.16 | Something that might help | STAR::GILLUM | Kirt Gillum | Thu Mar 04 1993 15:29 | 8 |
|
This may or may not be useful...
Check the SYSGEN parameter POOLCHECK. If it's non-zero, every buffer
is "checked" upon allocation/deallocation. If you can live with this
turned off, you might see higher throughput (perhaps you'll have more
CPU left too).
|