WASD Hypertext Services - Technical Overview

20 - Utilities

20.1 - Echo Facility
20.2 - Where Facility
20.3 - Xray Facility
20.4 - Scrunch Utility
20.5 - StreamLF Utility
20.6 - HTTPd Monitor
20.7 - ApacheBench
20.8 - qdLogStats
20.9 - Server Workout (stress-test)
[next] [previous] [contents] [full-page]

Foreign commands for external utilities (and the HTTPD control functionality) will need to be assigned from the adminstration users' LOGIN.COM either explicitly or by calling the HT_ROOT:[EXAMPLE]WASDVERBS.COM procedure. 

  $ AB == "$HT_EXE:AB"
  $ HTTPD == "$HT_EXE:HTTPD"
  $ HTTPDMON == "$HT_EXE:HTTPDMON"
  $ QDLOGSTATS == "$HT_EXE:QDLOGSTATS"
  $ SCRUNCH == "$HT_EXE:SCRUNCH"
  $ UNSCRUNCH == "$HT_EXE:SCRUNCH/UNSCRUNCH"
  $ STREAMLF == "$HT_EXE:STREAMLF"
  $ WWWRKOUT == "$HT_EXE:WWWRKOUT"


20.1 - Echo Facility

Ever had to go to extraordinary lengths to find out exactly what your browser is sending to the server?  The server provides a request echo facility.  This merely returns the complete request as a plain-text document.  This can be used for for checking the request header lines being provided by the browser, and can be valuable in the diagnosis of POSTed forms, etc. 

This facility must be enabled through a mapping rule entry. 

  script /echo/* /echo/*

It may then be used with any request merely by inserting "/echo" at the start of the path, as in the following example. 

  http://wasd.dsto.defence.gov.au/echo/ht_root/


20.2 - Where Facility

Need to locate where VMS has the HTTPd files?  This simple facility maps the supplied path then parses it to obtain a resulting VMS file specification.  This does not demonstrate whether the path actually exists! 

This facility must be enabled through a mapping rule entry. 

  script /where/* /where/*

It may then be used with any request merely by inserting "/where" at the start of the path, as in the following example. 

  http://wasd.dsto.defence.gov.au/where/ht_root/


20.3 - Xray Facility

The Xray facility returns a request's complete response, both header and body, as a plain text document.  Being able to see the internals of the response header as well as the contents of the body rendered in plain text can often be valuable when developing scripts, etc. 

This facility must be enabled through a mapping rule entry. 

  script /Xray/* /Xray/*

It may then be used with any request merely by inserting "/xray" at the start of the path, as in the following example. 

  http://wasd.dsto.defence.gov.au/xray/ht_root/


20.4 - Scrunch Utility

The "scrunch"er is used to decrease the processing overhead of Server-Side Includes (SSI) files.  See the "Environment Overview" for more detail on this utility. 

A scrunched SSI file is a variable length record file, where each record either comprises a single SSI directive, with the "<!--#" beginning on the record boundary, or a record of other text to be output (i.e. not beginning with "<!--#").

Why do this?  Well, if all SSI directives begin on a record boundary you only have to check the first five characters of each record to establish whether it should be interpreted or directly output!  This saves checking every character of every record for the opening "<" and the following "!--#".

Files that have been scrunched are basically unsuitable for editing (only due to the often inappropriately sized records).  Previously scrunched files may be returned to something (often exactly) resembling their original condition using the /UNSCRUNCH qualifier. 


20.5 - StreamLF Utility

This utility converts VARIABLE format files to STREAM_LF.  The WASD HTTPd server access STREAM_LF files in block/IO-mode, far more efficiently that the record-mode required by variable-record format files.  Use "STREAMLF/HELP" for some basic usage information. 

NOTE: The server can also be configured to automatically convert any VARIABLE record format files it encounters to STREAM_LF. 


20.6 - HTTPd Monitor

The HTTP server may be monitored in real-time using the HTTPDMON utility. 

[graphic]  HTTPDMON Graphic

This utility continuously displays a screen of information comprising three or four of the following sections:

  1. Process Information
    HTTPd process information includes its up-time, CPU-time consumed (excluding any subprocesses), I/O counts, and memory utilization.  The "Servers:" item shows how many servers are currently running on the node/cluster.  Changes in this count are indicated by the second, parenthesized number. 
  2. General Server Counters
    The server counters keep track of the total connections received, accepted, rejected, etc., totals for each request type (file transfer, directory listing, image mapping, etc.).
  3. Proxy Serving Counters
    The server counters keep track of proxy serving connections, network and cache traffic, cache status, etc. 
  4. Latest Request
    This section provides the response status code, and some transaction statistics, the service being accessed, originating host and HTTP request.  Note that long request strings may be truncated (indicated by a bolded elipsis). 

The following shows example output (after WWWRKOUT server testing):

 Port: 80   HTTPDMON v1.13.0 AXP (HTTPd v7.1)   Saturday, 22-JUL-2000 11:26:18
 
 Process: HTTPd:80  PID: 2020022F  User: HTTP$SERVER  Servers: 5(4)
      Up: 0 21:05:51.88  CPU: 0 07:34:19.99  Restart: 0
 Pg.Flts: 3526  Pg.Used: 37%  WsSize: 6402  WsPeak: 4729
     AST:  196/200   BIO:   97/100   BYT: 202270/202720  DIO:  100/100
     ENQ: 1989/2000  FIL:  298/300   PRC:     95/100      TQ:   26/30
 
 Connect: Accept:178212 Reject:0 Busy:0 Current:16 Peak:21  SSL: 0 (0%)
   Parse: 184865  Forbidden: 0  Redirect: Remote:0 Local:6654
 CONNECT: 0  GET: 164717  HEAD: 20148  POST: 0  PUT: 0
   Admin: 0  Cache: Load:48 Hit:39992/0  DECnet: CGI:3327 OSU:6655
     Dir: 6670 DCL: CLI:16730 CGI:9987 CGIplus:13310/13182 Spawned:168 Current:5
    File: 3384/0  IsMap: 0  Proxy: 64911  Put: 0  SSI: 16657  Upd: 9988
 
     1xx: 0  2xx: 157816  3xx: 3  4xx: 18496  5xx: 1864  (0 errors)
      Rx: 19,762,345  Tx: 1,053,847,437  (bytes)
 
   Proxy: enabled
 CONNECT: 0  GET: 58720  HEAD: 6191  POST: 0  PUT: 0
     Not: Cacheable Request:14555 Response:23786
 Network: Rx:118,463,055 Tx:7,208,716 (29%)  Accessed: 38356
  Lookup: Numeric:19 DNS:21 Cache:38317 Error:1
   Cache: Rx:2,863,617 Tx:320,164,012 (71%)  Read:26554/3 Write:10
  Device: DKA0: 4110480 blocks (2007MB)  Space: available
          2171744 used (1060MB 53%), 1938736 free (946MB 47%)
   Purge: 18 00:00:54, 98 files (2254/2412), 0 deleted (0/0)
 
    Time: 18 15:56:11  Status: 200  Rx: 95  Tx: 34479  Duration: 0.2400
 Service: beta.dsto.defence.gov.au:80  
    Host: beta.dsto.defence.gov.au (131.185.250.201)
 Request: GET /ht_root/exercise/64k.txt

The "/HELP" qualifier provides a brief usage summary. 

This information is, in part, obtained from the following logical names:

The server counter values are carried over when a server (re)starts (provided the system has stayed up).  To reset the counters use the on-line server administration menu (see 15 - Server Administration). 

If [DNSlookup] is disabled for the HTTP server the HTTPDMON utility attempts to resolve the numeric address into a host name.  This may be disabled using the /NORESOLVE qualifier. 


20.7 - ApacheBench

This server stress-test and benchmarking tool, as used in the Apache Distribution, is included with the WASD package (sourced from http://webperf.zeus.co.uk/ab.c), within license conditions. 

  Copyright (c) 1996 Adam Twiss, Zeus Technology Ltd.
  Copyright (c) 1998 The Apache Group.

ApacheBench will only compile and run for Alpha or VAX systems with VMS 7.n or greater available.  It is a simple but effective tool, allowing a single resource to be requested from a server a specified number of times and with a specified concurrency.  This can be used to benchmark a server or servers, or be used to stress-test a server configuration's handling of variable loads of specific resquests (before exhausting process quotas, etc.)

A small addition to functionality has been made.  The WASD ApacheBench displays a count of the HTTP response categories received (i.e. the number of 2nns, 4nns, etc.) This allows easier assessment of the relevance of results (i.e. measuring performance of some aspect only to find the results showed the performance of 404 message generation - and yes, an annoying experience of the author's prompted the changes!)

The following examples illustrate it's use. 

  $ AB -H
  $ AB -C 10 -N 100 http://the.server.name/ht_root/exercise/0k.txt
  $ AB -C 50 -N 500 -K http://the.server.name/ht_root/exercise/64k.txt
  $ AB -C 10 -N 100 http://the.server.name/cgi-bin/cgi_symbols


20.8 - qdLogStats

Quick-and-Dirty LOG STATisticS is a utility to extract very elementary statistics from Web server common/combined format log files.  It is intended for those moments when we think "I wonder how many times that new archive has been downloaded? ", "How much data was transfered during November? ", "How often is such-and-such a client using the authenticated so-and-so service? ", "How much has the mail service been used? " ... and want the results in a matter of seconds (or at least a few tens of seconds ;^)

A number of filters allow subsets of the log contents to be selected.  These filters are simple "sort-of-regular" expressions, not case-sensitive (deliberately), can contain wildcards (such as asterisks (*), question marks (?), and percent signs (%)) as well as "semi-regular" expressions (such as the range [a-z]). THERE IS NO WAY (that I know of) TO ESCAPE THESE RESERVED CHARACTERS!  (This functionality uses decc$match_wild() function.) All matches are made by string pattern matching, hence a query /AFTER=01-NOV cannot be done.  Of course date pattern matching can! 

Special constructs allow more complex expressions to be built up.  Combinations of required and excluded strings may be specified in the one expression.  When a string begins with a "+{" it must be present for the record not to be filtered out.  If it begins "-{" then it must not be present.  Such specifications must be terminated with a matching closure "}".

A knowlege of the format and contents of the common and combined log formats will assist in deciding which and to what purpose filters should be used.  Record filtering is done in the same order as is finally displayed, so method would be processed before user-agent for instance.  Normally a record match terminates on the first non-matched filter (to expedite processing).  To compare and report each filter for every record apply the /ALL qualifier.  To view records as they are processed use the /VIEW qualifier.  This by default displays all matched records, but the optional =ALL or =NOMATCH parameters will display all records, or all those but the matches. 

$ QDLOGSTATS log-file-spec [pattern qualifiers] [other qualifiers]

Examples:

A CGI interface is also available.  It required authorization to be active on the script.  See the source code for further details. 


20.9 - Server Workout (stress-test)

The WWWRKOUT ("World Wide Web Workout") utility exercises an HTTP server, both in the number of concurrent connections maintained and in the number of back-to-back sequential connection requests and data transfers. 

This utility can be used to stress-test the WASD VMS HTTP server (or any other), or to make comparisons between it and other servers.  When stress-testing a server, evaluating performance or just using it to try and break a test-bed server, it is best used from multiple, autonomous systems concurrent. 

It sets up and maintains a specified number of concurrent connections to a server.  It reads a buffer of data from each connection in turn, where data is waiting (does not block), until the document transfer is complete and the connection closed by the server.  It then closes the local end and immediately reuses the now-free socket to initiate another sequence.  If enabled (it is by default), the utility attempts to reflect the real-world in varying the data transfer rate for each connection, by setting the number of bytes read during each read loop differently for each connection.  All transfered data is discarded. 

The data transfer rate for each connection is displayed at connection close.  It is by default it is the effective transfer rate, that is the rate from opening the connection to closing it, and so includes request processing time, etc.  If the "/NOEFFECTIVE" qualifier is employed it measures the document data transfer rate only. 

Although a single document path may be specified on the command line it is preferable to supply a range of document paths, one per line in a plain text file.  Each document path and/or type specified should be different to the others, to exercise the server and file system cache.  Any number of paths may be specified in the file.  If the file is exhausted before the specified number of connections have been established the file contents are recycled from the first path.  If a path or a file of paths is not specified the utility just continually requests the welcome document. 

To assess a server's total throughput choose paths that lead to large documents (> 50K), where the overhead of connection setup, rule processing and transfer initiation are less significant than the data transfer itself.  The buffer size variation functionality should be disabled using the "/NOVARY" qualifier when assessing data transfer rates.  Responsiveness is better assessed using small documents (< 2K), where the overhead of the non-data-transfer activities is more significant. 

$ WWWRKOUT [server_host_name[:port]] [path] [qualifiers]

Examples:

  $ WWWRKOUT
  $ WWWRKOUT www.server.host "/web/home.html"
  $ WWWRKOUT www.server.host:8080 /FILEOF=PATHS.TXT
  $ WWWRKOUT /SERVER=www.server.host /PORT=8080 /NOBREAK /NOVARY /FILEOF=PATHS.TXT
  $ WWWRKOUT www.server.host:8080 /FILEOF=PATHS.TXT /NOEFFECTIVE /NOVARY
  $ WWWRKOUT www.server.host /FILEOF=PATHS.TXT /COUNT=500 /SIMUL=20 /BUFFER=512

The "/HELP" qualifier provides a brief usage summary. 


[next] [previous] [contents] [full-page]