SP500207 NIST Special Publication 500-207: The First Text REtrieval Conference (TREC-1) Retrieval Experiments with a Large Collection using PIRCS chapter K. Kwok L. Papadopoulos K. Kwan National Institute of Standards and Technology Donna K. Harman document-focused QTD learning of c(AB) based on known relevant queries in q(A). The ad hoc queries q(B), the total dictionary, and the total collection c(AB) then form a network using 1Cm weighting, from which both automatic and manual ad hoc query retrieval results (W[OCRerr]auto and W1[OCRerr][OCRerr]) are obtained. (c) Feedback Queries: To simulate user relevance feedback, we employ the ten highest ranked sub- documents of the automatic ad hoc results in (b), and determine their relevance to their respective topics manually. Two queries out of 25 have no relevant documents within the first ten, and they are removed leaving 23. We have performed two types of feedback training: without and with query expansion. In both cases, query-focused and document-focused training were done. Our method of adding terms to existing queries is based on [8,9] during DTQ direction training. The n[OCRerr] documents relevant to a query q, are instantiated to I and their activation to each term node, after gated through the weights wk[OCRerr]=d[OCRerr]kIL[OCRerr], is averaged. The highest 20 activated terms with document frequency less than 2000 are then used to expand qa. However, not all the terms are different from the existing terms of qa, so that on average, each query got expanded with about 12.3 terms (284 new terms for 23 queries). After this, training proceeds in the QTD direction. We have not used expansion for documents because usually there is only one relevant query per document. We have done only one round of feedback iteration. 3. System Description Our system was designed over a number of years for IR experimentation. Its primary goal is to be flexible so that unforseen developments or algorithms can be supported for testing without much change in the basic implementation. Many intermediate files are produced for more convenient hooks to and from the system. These will be trimmed in later verisons. It was originally programmed in Pascal, but has been translated and is now running in a UNIX and C environnient on a Sun SparcStation 2GS. The system has 48MB RAM, a TREC-dedicated 1.7GB disk drive, plus about 0.5GB on another drive that we can occasionally use. Our system characteristics and timings are described in the Appendix of this paper. 4. Results and Discussions Our runs are named PIRCS'n', where n=l denotes fully automatic retrieval, n=2 denotes manual, i.e. automatic combined with soft-boolean where boolean queries are manually created, n=3 denotes feedback using automatic retrieval, and n=4 is the same as n=3 but with query expansion. Although we use sub- documents for indexing and retrieval, our final 200-document output list has only unique document ID's; all sub-documents of the same document ID are suppressed except for the one with the highest rank. Precision-recall and other measures are gathered in the Appendix of this volume. Our runs are based on twenty five queries for ad hoc and another twenty five for routing retrievals, which we recognize as too few and may not reflect sufficiently the variety of query types in real-life. On the other hand, the evaluation curves and average values serve as a lower bound to the power of our system because relevance judgments are rendered only to the combined first 100 documents of all systems. This means documents ranked 101 to 200 in our retrieval are automatically considered irrelevant if they are not in the judged relevant set. Also, Category B texts come from a single source, viz. WSJ of multiple years, not like Category A with texts from multiple sources. With these reservations we like to make the following observations: (a) Inter-site comparison shows that the PIRCS results are the best among Category B participants for both ad hoc and routing retrievals, and comparable to the best methods in Category A based on one evaluation with crnlB. 160