SP500207 NIST Special Publication 500-207: The First Text REtrieval Conference (TREC-1) Overview of the First Text REtrieval Conference (TREC-1) chapter D. Harman National Institute of Standards and Technology Donna K. Harman Adhoc -- Category B 0.9 0.8 0 0.5 [OCRerr] 0.4 0.3 0.2 0.1 0 0.00 0[OCRerr]1O 0.20 0.30 0.40 0.50 0.60 0.70 0.80 0.90 Recall crnlB pircsl pircs4 nyuir[OCRerr] Figure 10. Adhoc Results for Category B. 5[OCRerr] Routing Results There were 22 sets of results for routing evaluation, with 16 of them based on runs for the full data seL Note that all routing techniques suffered from the lack of sufficient and accurate training data, and therefore these results are even more preliminary than the adhoc results. Of the 16 systems using the full data set, 8 used automatic construction of queries, and 8 used manual construction. Figure 11 shows the recall/precision curve for the three TREC-1 runs with the highest 11-point averages using automatic construction of queries. Two of the curves, based on the use of the Cornell SMART system, show very different results. The "fuhra2" results came from using a probabilistically-based relevance feedback (see Fuhr & Buckley paper), whereas the "crnla2" results came from doing traditional relevance feedback methods using the vector space model (see Buckley, Sal- ton & Allan paper), The "cityri" results also came from using traditional relevance feedback, but using a different probabilistic model and term weighting (see Robertson, Walker, Hancock-Beaulieu, Gull & Lau paper). The "cpgcn2" system used filtering methods rather than more traditional information retrieval methods to achieve results similar to the feedback results (see Jones, leung, and Pape paper). 0.9 0.8 0.7 0.6 0 0.5 C,) [OCRerr] 0.4 0.3 0.2 0.1 Best Automatic Routing 0 -[OCRerr] 0.00 0.10 0.20 0.30 0.40 0.50 0.60 0.70 0.80 0.90 1.00 Recall __ fuhra2 cnil[OCRerr] cityri cpgcn2 Figure 11. The Best Routing Results using Automatic Query Construction. 17