IRS13 Scientific Report No. IRS-13 Information Storage and Retrieval Test Environment chapter E. M. Keen Harvard University Gerard Salton Use, reproduction, or publication, in whole or in part, is permitted for any purpose of the United States Government. 1-45 documentation collection does follow the expected pattern if precision in subject terminology is of importance. Comparison with Fig. 12 reveals that request preparation is a large variable, and use of the 17 non-staff prepared requests would be expected to result in a curve lower than that for the Cran-l aerodynamics results. Another technique of collection comparison uses the average rank technique as used in part GB for comparing the specific and general requests. Fig. 23 gives results based on the stem dictionaries, and Fig. 24 gives results based on the thesaurus dictionaries. The average rank positions of the first and second relevant documents reflect the viewpoint of a user needing high precision. Ignoring differences in collection size, the Cran-l aeordynamics collection gives a good result using the thesaurus and on ADI the first relevant receives the best average rank Use of the percentage figure to take into account changes in collection size restores the ex- pected merit. Figs. 23 and 24 also record the average rank positions of the last relevant document to reflect the viewpoint of the high recall user. The averagerank is[OCRerr]directly affected and ordered by collection size, but the percentage figure shows that IRE-3 and Cran-l perform a little better than ADI.