[Search for users] [Overall Top Noters] [List of all Conferences] [Download this site]

Conference bulova::decw_jan-89_to_nov-90

Title:DECWINDOWS 26-JAN-89 to 29-NOV-90
Notice:See 1639.0 for VMS V5.3 kit; 2043.0 for 5.4 IFT kit
Moderator:STAR::VATNE
Created:Mon Oct 30 1989
Last Modified:Mon Dec 31 1990
Last Successful Update:Fri Jun 06 1997
Number of topics:3726
Total number of notes:19516

3614.0. "DECwindows application limit (11 clocks)" by MUTTON::LAMB (Peter Lamb - GSG Santa Clara) Thu Nov 08 1990 19:17

Hi,

I have a customer that currently has an outstanding CLD for something that
I am sure I wouldn't have believed if I had not just seen it with my own 
eyes...

To reproduce the problem startup 11 clocks from either the session manager or
vue (at some point you may exhaust a process resource, if so, up that resource
login/out and try again).  If you are finally able to get 11 clocks up no 
matter what you try you will not be able to get up the 12th one.

The error message you get from file view is....

XIO:  unable to open connection WSA8:
      after 0 requests (0 known processed) with 0 events remaining.
%XLIB-F-IOERROR, xlib io error
-DECW-E-CNXABORT, connection aborted

Once you have 11 clocks running some other applications will also refuse to
run while others will startup without any problems.  For example NOTEPAD
and VAXnotes work fine but CARDFILER, CALENDAR & CALCULATER will all die
with esentually the same error message.

The customer is running VMS 5.3 and is seeing this problem and I am running
SBB 5.4.

Also, a quick look at  my system's server log 
(SYS$SYSROOT:[SYSMGR]DECW$SERVER_0_ERROR.LOG;197) showed the following....

8-NOV-1990 17:01:26.1 Connection 72e710 is closed by Txport
 8-NOV-1990 17:01:45.7 Connection 72e710 is accepted by Txport
 8-NOV-1990 17:01:55.1 Connection 72e710 is closed by Txport
 8-NOV-1990 17:02:03.7 Connection 72e710 is accepted by Txport
 8-NOV-1990 17:02:13.1 Connection 72e710 is closed by Txport
 8-NOV-1990 17:02:18.6 Connection 72e710 is accepted by Txport
 8-NOV-1990 17:02:20.2 Connection 72e748 is accepted by Txport
 8-NOV-1990 17:02:20.3 Cannot allocate clientrec, or bigbuffer
 8-NOV-1990 17:02:20.9 Connection 72e710 is closed by Txport
 8-NOV-1990 17:04:18.8 Connection 72e710 is accepted by Txport
 8-NOV-1990 17:04:43.8 Connection 72e748 is accepted by Txport
 8-NOV-1990 17:04:44.1 Cannot allocate clientrec, or bigbuffer

SYS$SYSROOT:[SYSMGR]DECW$SERVER_0_OUTPUT.LOG;108

MAXPROCESSCNT is 70 on my system and I am no where near this limit
and DECNET MAXLINKS is 32 and I don't believe these are generating links
anyway.  

Lastly, If I just try running the DECwindows (say calculator) from a 
DECterm window I get the same error message too.  Also, 11 decterms doesn't 
cause the problem it has to be clock (or I would assume one of the other
applications that also won't run after 11 clocks).

Does anyone have any ideas what is going on??

Regards,

Peter Lamb
T.RTitleUserPersonal
Name
DateLines
3614.1Discussion 1494 helped but not completlyMUTTON::LAMBPeter Lamb - GSG Santa ClaraThu Nov 08 1990 19:4337
Everytime I use notes I am constantly amazed at what an incredibly powerful
tool it is!  After a few minutes of searching this file I discovered this
discussion regarding the Maximum # of local transport applications and I 
guess I now have a few questions...

The questions are, given I am now running a newer version of decwindows then
is being discussed in #1494, for example, I checked DECW$STARTSERVER.COM and
it looks like my system is already setting the ENQUEUE_LIMIT to 64 however,
I am still having the problem.  Does anyone have any other ideas??

Peter

             <<< BULOVA::DOCD$:[NOTES$LIBRARY]DECWINDOWS.NOTE;6 >>>
                                -< DECWINDOWS >-
================================================================================
Note 1494.9            Trouble starting lots of processes                 9 of 9
DECWIN::JMSYNGE "James M Synge, VMS Development"     18 lines  10-DEC-1989 12:10
               -< ENQLM limits the number of local connections >-
--------------------------------------------------------------------------------
    While LOCAL transport doesn't place a limit on the maximum number of
    connections, it does use a resource, locks, which are limited by a
    quota.  In particular, it uses two per connection.
    
    The ENQLM quota has a default value of 30, and this is not currently
    overridden in SYS$MANAGER:DECW$STARTSERVER.COM.  Thus, we have an
    effective limit of 15 local connections.
    
    If you need to fix this immediately (i.e. before I fix it on the master
    pack and it ships), then you should probably add a line that looks like
    
    	/ENQUEUE_LIMIT = 64 -
    
    to SYS$MANAGER:DECW$STARTSERVER.COM where the RUN/DETACHED is (at the
    end of the file).
    
    
    James
3614.2Server is running out of memorySTAR::VATNEPeter Vatne, VMS DevelopmentFri Nov 09 1990 10:499
The key here is the line:

 8-NOV-1990 17:02:20.3 Cannot allocate clientrec, or bigbuffer

This means the server is running out of virtual memory.
What does SHOW PROC/ACCOUNT show for the peak virtual size?

If it is near 25000, then we'll need to bump up the server's
paging file quota.
3614.3We have a winner??PJWL::LAMBPeter Lamb - FSG Santa ClaraFri Nov 09 1990 11:4543
Hi,

I think you have it!!  This is what my system looks like currently (and the
clock processes are not running now).  Is 20953 close enough to 25000??

MUTTON::> show proc/id=202001C0/account

 9-NOV-1990 09:46:27.37   User: LAMB             Process ID:   202001C0
                          Node: MUTTON           Process name: "DECW$SERVER_0"

Accounting information:
 Buffered I/O count:      1830  Peak working set size:       4000
 Direct I/O count:       16795  Peak virtual size:          20953
 Page faults:            48679  Mounted volumes:                0
 Images activated:           0
 Elapsed CPU time:          0 00:14:07.50
 Connect time:              6 18:47:17.35
_MUTTON::>

Peter

P.S.  How do you show a processes WORKING SET QUOTA?? and ENQUEUE QUOTA?

A show proc/id=xx/quota shows the following and given -1 I would expect
the enqueue quota to be 64 but it is showing at 40??  Also I really need
to see the working set quota for another problem I'm having with NCP
and can't figure out how to show it.

_MUTTON::> SHOW PROC/ID=202001C0/QUOTA

 9-NOV-1990 09:49:30.01   User: LAMB             Process ID:   202001C0
                          Node: MUTTON           Process name: "DECW$SERVER_0"

Process Quotas:
 Account name: SYSTEM
 CPU limit:                      Infinite  Direct I/O limit:       100
 Buffered I/O byte count quota:     42656  Buffered I/O limit:      60
 Timer queue entry quota:               7  Open file quota:         44
 Paging file quota:                  9158  Subprocess quota:         8
 Default page fault cluster:           64  AST quota:               96
 Enqueue quota:                        40  Shared file limit:        0
 Max detached processes:                0  Max active jobs:          0

3614.4Looks like the winner to me!STAR::VATNEPeter Vatne, VMS DevelopmentFri Nov 09 1990 14:2027
Yes, I'd say that 20953 is close enough to 25000 to be the culprit.
I don't know why the server is using so much virtual memory on your
system, as I have lots of connections to my server and it only shows
9528 peak virtual size for the server process.

I'd say this warrants entering a QAR.

The workaround is to put the following line in DECW$PRIVATE_SERVER_SETUP.COM:
$ define decw$server_page_file 30000
(or even bigger).  You should also check to make sure the SYSGEN
parameter VIRTUALPAGECOUNT is at least that big.  And you will
also need a page file that is big enough for all your processes.

Showing quotas is not easy.  I think that the labels from SHOW PROC/QUOTA
are very misleading.  The quota numbers shown are actually the "quotas"
at the instant.  For example, the enqueue quota for the server does
start out at 64, but as each lock is taken out, the quota is decremented
by 1.  When it reaches zero, you can't take out any more locks.  Of
course, what we'd all like to see is the original quota of the process
compared to the amount in use (e.g. enqueue usage: 14/64).

The way I always check on quotas is to run authorize and look at the
account's quotas.  For created processes like the server, you have to
look at the original run command (the DECwindows server is created
from SYS$MANAGER:DECW$STARTSERVER.COM).  Also, you may be able to see
the quotas from SDA, but of course that isn't very convenient!  Maybe
you should enter a suggestion QAR against VMS...
3614.5What entering a QAR does for youSTAR::VATNEPeter Vatne, VMS DevelopmentFri Nov 09 1990 14:577
I'd like to point out that entering a QAR for any problem that you
see will only affect development of DECwindows V3.  It will not
help with finding or fixing bugs in the currently shipping product.

If you have a CLD outstanding, continue to work with your CLD rep.
If they don't know the answer, they will come find me or someone
else in the DECwindows group.
3614.6The saga continues...MUTTON::LAMBPeter Lamb - GSG Santa ClaraFri Nov 09 1990 15:5977
Well, I tried upping the pagefile quota to 30000 as suggested in .4 
but unfortunaly now when I restart the server and start up 11 DECW$CLOCKs
and look at show proc/account I am only using 15800 pages and the problem 
is still occuring.  (I guess the 20000 must have come from something
else I ran once).

I believe a clue to this problem may be that it seems to be application 
independent.  Ie you would think that if I had a resource problem it 
would be occuring for most DECwindows applications.  However, this is
not the case.  Even after I have 11 clock applications running (and can't 
get any more) I can continue to start other applications with out any 
problem (for example I am writing this in DECwindows notes and started
this after the 11 clocks).

For what it is worth here is what my current SHOW PROC/ACCOUNT
returns...

_MUTTON::> sho proc/id=20200322/account

 9-NOV-1990 14:03:54.39   User: LAMB             Process ID:   20200322
                          Node: MUTTON           Process name: "DECW$SERVER_0"

Accounting information:
 Buffered I/O count:      1364  Peak working set size:       3830
 Direct I/O count:        2160  Peak virtual size:          15811
 Page faults:            16322  Mounted volumes:                0
 Images activated:           0
 Elapsed CPU time:          0 00:01:16.64
 Connect time:              0 00:21:40.59

A show sys also show the following....
_MUTTON::> sho sys
VAX/VMS V5.4  on node MUTTON   9-NOV-1990 14:04:37.32   Uptime  17 02:20:02
  Pid    Process Name    State  Pri      I/O       CPU       Page flts Ph.Mem
20200081 SWAPPER         HIB     16        0   0 00:00:51.00         0      0
20200086 CONFIGURE       HIB     10       39   0 00:00:00.25       126    228
20200088 ERRFMT          HIB      8    12723   0 00:00:51.17        84    126
20200089 CACHE_SERVER    HIB     16      133   0 00:00:00.25        63    100
2020008A CLUSTER_SERVER  HIB      8      322   0 00:00:02.88       143    336
2020008B OPCOM           HIB      8     5546   0 00:00:31.09       607    180
2020008C AUDIT_SERVER    HIB     10       50   0 00:00:00.99      1353    479
2020008D JOB_CONTROL     HIB     10     5652   0 00:00:18.85       141    337
2020008E IPCACP          HIB     10       40   0 00:00:00.09        77    139
2020008F TP_SERVER       HIB     10    98465   0 00:16:06.28       168    253
20200090 SMISERVER       HIB      9       56   0 00:00:00.46       276    490
20200095 LATACP          HIB     14        7   0 01:21:24.65       522    455
20200099 NMAIL           HIB      5      144   0 00:00:00.44       360     84
2020009A DNS$ADVER       HIB      4    12173   0 00:00:36.29       192    299
2020009B RDMS_MONITOR    LEF     15       17   0 00:00:00.40     14995     51
20200322 DECW$SERVER_0   HIB      6     3554   0 00:01:21.29     16511   1855
202002A3 LAMB            LEF      4      189   0 00:00:05.66      5951    508
20200224 DECW$WM_11      LEF      7       35   0 00:00:06.46      1571   2243
20200225 DECW$WM_12      LEF      7       35   0 00:00:02.23      1303    650
202002A6 LAMB_SM1        LEF      4      155   0 00:00:12.99      5314   2048
202002A7 LAMB_SM2        LEF      4      155   0 00:00:05.59      5118   2048
202002A8 LAMB_SM3        LEF      6      146   0 00:00:04.81      4859   2049
2020032B LAMB_SM6        LEF      5      437   0 00:00:33.09      6678   2691
2020032C LAMB_SM7        LEF      6      258   0 00:00:06.27      5473   3332
202002AD LAMB_SM8        LEF      4      411   0 00:00:39.14     12156   4180
202002AE DECW$TE_11      LEF      6      605   0 00:00:40.05      2510   2305
202002AF _TWA17:         LEF      4       76   0 00:00:01.92      3293    786
20200330 _TWA18:         CUR      4      376   0 00:00:03.96      3888    249
202002B1 LAMB_SM4        LEF      5      141   0 00:00:04.54      4581   2001
20200332 LAMB_SM5        LEF      4      139   0 00:00:04.50      4725   2047
20200333 LAMB_SM9        LEF      4      139   0 00:00:04.59      4625   2030
202002B4 LAMB_SM10       LEF      4      139   0 00:00:04.57      4553   1981
20200335 LAMB_SM11       LEF      5      139   0 00:00:04.52      4583   1980
20200236 NETACP          HIB     10     2026   0 00:03:13.56     47368   3000
20200237 EVL             HIB      6     1652   0 00:00:11.14     99727     55  N
20200238 REMACP          HIB      8       42   0 00:00:00.09        74     56
20200339 LAMB_SM12       LEF      4      139   0 00:00:04.59      4830   1982
2020033A LAMB_SM13       LEF      6      139   0 00:00:04.49      4593   2413
2020033B LAMB_SM14       LEF      4      139   0 00:00:04.52      4687   2048
2020033C LAMB_SM15       LEF      4      139   0 00:00:04.60      4221   2022
20200241 VUE$LAMB_1      LEF      4      284   0 00:00:06.09      3205    399  S
20200242 VUE$LAMB_2      LEF      4      168   0 00:00:02.11       956    449  S
_MUTTON::>
3614.7See also notes 3558 and 1494MUTTON::LAMBPeter Lamb - GSG Santa ClaraFri Nov 09 1990 16:0412
I just thought I would mention that in my searching I have discovered
that this problem has also been discussed in the topics numbered above.

It turns out that # 3558 is actually the same customer that I am dealing
with and was put in by the TSC when he logged the origional call.

I guess this should teach me to search even more carefully!!  (I thought
I was doing well when I found #1494!)

Just an FYI

Peter
3614.8Enqlm Seems to be the culprit but raising it doesn't helpMUTTON::LAMBPeter Lamb - GSG Santa ClaraMon Nov 12 1990 13:1433
Hello,

The latest status on this problem is that the TSC can reproduce the problem
with DECW$CLOCK however were all at a loss at to what to do next...

Well, we believe we have established that it is probably ENQLM that the 
server is running out of.  A quick check of my system shows DECW$SERVER_0
with only 4 remaining below & (each time a clock is started it takes 4 locks).

_MUTTON::> show proc/id=20200322/quota

12-NOV-1990 11:10:30.12   User: LAMB             Process ID:   20200322
                          Node: MUTTON           Process name: "DECW$SERVER_0"

Process Quotas:
 Account name: SYSTEM
 CPU limit:                      Infinite  Direct I/O limit:       100
 Buffered I/O byte count quota:     43200  Buffered I/O limit:      60
 Timer queue entry quota:               7  Open file quota:         50
 Paging file quota:                 16762  Subprocess quota:         8
 Default page fault cluster:           64  AST quota:               96
>> Enqueue quota:                         4  Shared file limit:        0

We then attempted to raise the limit in the server (it starts at 64) up to 
128 but then when the 11th clock is reached the total simply sits at 64 and 
won't go any lower.

Is there a sysgen parameter related to ENQUELM OR ENQUEQUOTA??

Thanks!!!

Peter