SP500207 NIST Special Publication 500-207: The First Text REtrieval Conference (TREC-1) Natural Language Processing in Large-Scale Text Retrieval Tasks chapter T. Strzalkowski National Institute of Standards and Technology Donna K. Harman terms from queries while preparing them for a routing run. (2) Topics 51-75 were processed using the same 3 fields as above with respect to the test database only (disk 2). They were then run in ad-hoc mode independently against the training data- base and the test database, and the top 200 documents from each run (for each query) were merged using document scores. The upper 200 documents were selected for the final result. It should be noted that we run topics against disk 1 using idf weights on terms from disk 2 (a roufing-like mode). This, it turned out, created an even distribufion of hits between the two databases. We call this method the uniform merge. (3) We repeated run (2), but this time we run topics 51-75 in ad-hoc mode on both disks (i.e., different weights were used in these runs). Again, top 200 documents were selected. We call this method the bruteft)rce merge. (4) Finally, we repeated uniform merge with queries manually pruned before actual retrieval. Summary statistics for these runs are shown in [OCRerr]tbles 4 and 5. It has to be pointed out that these results are lower than expected because of various problems with the database search prograin and because the Concepts field in TREC topics was not included in search queries. Subsequent runs with Concepts field included in search produced results shown in Table 6. ACKNOWLEDGEMENTS We would like to thank Donna Harinan of NIST for making her IR system available to us. We would also like to thank Ralph Weischedel, Marie Meteer and Heidi Fox of BBN for providing and assisfing in the use of the part of speech tagger. Jose Perez Carballo has contributed a number of valuable observations during the course of this work, and his assistance in processing the TREC data was critical for our being able to finish on time. This paper is based upon work supported by the Defense Advanced Research Project Agency under Contract N00014-90-J-1851 from the Office of Naval Research, under Contract N00600-88-D-3717 from PRC Inc., and the National Science Foundation under Grant IRI-89-02304. We also acknowledge support from Canadian Institute for Robotics and Intelligent Systems (IRIS). 184 Run nyuirl nyuir2 Name routing routing Queries 25 25 Tot number of does over all queries Ret 5000 5000 Rel 3766 3766 RelRet 834 [OCRerr] 1177 Recall Precision Averages 0.00 0.5691 0.6983 0.10 0.3618 0.5072 0.20 0.1846 0.3508 0.30 0.1314 0.2997 0.40 0.1065 0.2286 0.50 0.0481 0.1481 0.60 0.0295 0.0604 0.70 0.0104 0.0296 0.80 0.0104 0.0260 0.90 0.0000 0.0125 1.00 0.0000 0.0000 Average precision for all points 11-pt 0.1320 0.2147 Average precision at 0.20, 0.50, 0.80 3-pt 0.0811 0.1750 Recall at 5 docs 0.0338 0.0374 l5docs 0.0671 0.0961 30docs 0.0979 0.1533 lOodocs 0.2117 0.3119 200docs 0.2988 0.4298 Precision at Sdocs 0.3360 0.4480 15 docs 0.2960 0.4587 30 docs 0.2787 0.3973 lOodocs 0.2276 0.3080 200docs 0.1668 0.2354 Table 4. Routing run statistics: (1) duplicate terms (unintenfionally) removed from queries; and (2) du- plicate terms retained and Concepts fields included.