SP500207
NIST Special Publication 500-207: The First Text REtrieval Conference (TREC-1)
Overview of the First Text REtrieval Conference (TREC-1)
chapter
D. Harman
National Institute of Standards and Technology
Donna K. Harman
Figure 12 shows the recalVprecision curve for the three TREC-1 runs with the highest 1 1-point averages
using manual construction of queries. The systems used manually-built filters, with the "clartb" and `tgecrd2"
results done similarly to their corr[OCRerr]sponding adhoc systems, but using the sample relevant documents as input to
the filter-building process. The "parazi" system used manually-constructed filters based on clusters of interest-
mg terms (see Zimmerman paper). The "cpghc2" group hand-crafted these queries as a contrast to their
automatic pattern filtering methods (see Jones, Leung & Pape).
Best Manual Routing
0.00 0.10 0.20 0.30 0.40 0.50 0.60 0.70 0.80 0.90 1.00
Recall
parazi gecrdl gecrd2 trwl
0.9
0.8
0.7
[OCRerr] 0.6
0
[OCRerr] 0.5
[OCRerr] 0.4
0.3
0.2
0.1
0
Figure 12. The Best Routing Results using Manual Query Construction.
Again it is useful to contrast the methods of query construction. Figure 13 shows a comparison of four sets of
results, two from automatic query construction and two using manual query construction. Here, unlike the
adhoc results, the automatic query building seems to be clearly superior, with the "fuhrl" results having higher
performance throughout the significant part of the recalVprecision curve.
Routing Manual vs Automatic
0.00 0.10 0.20 0.30 0.40 0.50 0.60 0.70 0.80 0.90 1*00
Recall
0
0.9
0.8
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0
[OCRerr]- ftihrl __ cnil[OCRerr] . parazi gecrd2
Figure 13. A Comparison of Routing Results using Different Query Construction Methods.
18