IRS13 Scientific Report No. IRS-13 Information Storage and Retrieval Test Environment chapter E. M. Keen Harvard University Gerard Salton Use, reproduction, or publication, in whole or in part, is permitted for any purpose of the United States Government. 1-4 responsible for the 42 requests in use with the Cran-l collection. The document collection consists not of the authors' own research papers, but of a number of the earlier papers that the author cited in the bibliography to his paper. The first set of seventeen computer science requests were prepared for the iRE-l collection by three project staff members. Two persons made up thirteen of the requests without any knowledge of how the system would per- form in practice1 but having extensive knowledge of the technique of operation of the system. Requests were devised to cover a cross-section of the major topics in the collection but were not "source document" requests in the sense that they were based on particular documents. A third person made up four of the requests using sets of classification headings that had been manually assigned to the IRE-l collection. A second set of seventeen requests was prepared by one person hired for the task. This person was one of two persons who also prepared the 35 docu- mentation requests. Requests were again not based on any one document in the collection. No guidance concerning the length of the requests was given - the hired persons tended to devise requests that were rarely longer than a sentence or two, whereas requests prepared by[OCRerr] staff members were often longer. Techniques used for obtaining relevance decisions are described in part 3. A much more detailed analysis of the characteristics of documents and requests is possible. Such an investigation is given in Section x, part 3, for the ADI documentation requests only. Direct performance comparisons between collections have not formed the purpose