SP500207 NIST Special Publication 500-207: The First Text REtrieval Conference (TREC-1) Combining Evidence from Multiple Searches chapter E. Fox M. Koushik J. Shaw R. Modlin D. Rao National Institute of Standards and Technology Donna K. Harman Figure 2: Phase 1, Disc 1, WSJ, queries 50-100, Ad Hoc Merge Top ranked evaluation vpidt2 all 50 documents over all queries 10000 4056 1371 ion Averages: 0.5153 0.2841 0.1898 0.1285 0.0988 0.0746 0.0527 0.0338 0.0271 0.0222 0.0222 for all points 0.1317 for 3 intermediate points (0.20, 0.50, 0.80) 0.0972 Run number: Num~queries: Total number of Retrieved: Relevant: Rel[OCRerr]ret: Recall - Precis at 0.00 at 0.10 at 0.20 at 0.30 at 0.40 at 0.50 at 0.60 at 0.70 at 0.80 at 0.90 at 1.00 Average precision 11-pt Avg: Average precision 3-pt Avg: Recall. at 5 docs: 0.0476 at 15 docs: 0.0640 at 30 docs: 0.0843 at 100 docs: 0.2050 at 200 docs: 0.4481 Precision: At 5 docs: 0.3160 At 15 docs: 0.2147 At 30 docs: 0.1760 At 100 docs: 0.1246 At 200 docs: 0.1371 6 Results of Phase 2 6.1 Base Runs In Phase 2, twenty-live base runs were made on Disc 1: 5 different retrieval methods were used for each of the 5 sub-collections. Based on our evaluation, the 11-point averages are given in Table 4. Note that in the p-norm case, document weights were utilized, in contrast to the binary weighting used in Phase 1. 324