SP500207
NIST Special Publication 500-207: The First Text REtrieval Conference (TREC-1)
Overview of the First Text REtrieval Conference (TREC-1)
chapter
D. Harman
National Institute of Standards and Technology
Donna K. Harman
0.9
0.8
0.7
0.6
0 0.5
[OCRerr] 0.4
0.3
0.2
0.1
0
Best Automatic Adhoc
0.00 0.10 0.20 0.30 0.40 0.50 0.60 0.70 0.80 0.90
Recall
flilirpi crnlpl siemsl
1.00
Figure 5. The Best Adhoc Results using Automatic Query Construction.
Figure 6 shows the recall/precision curve for the three ThEC-1 runs with the highest 1 1-point averages using
manual construction of queries. It should be noted that varying amounts of manual intervention were used, and
this should be considered when comparing results. These curves show differences in that the "clartb" and
"gecrd2" have initially a high precision, but lose this precision as recall increases, whereas the 91cnqs12" method
has a lower initial precision, but higher precision aL the higher recall levels. This may be a function of the very
different methods being used. The "clartb" system adds noun phrases found in likely relevant documents to
improve the query terms taken from the topic (see Evans paper), whereas the "cnqse" system uses more general
thesaurus entries to expand the query (see Nelson paper). The `3gecrd2" system uses a totally different approach
of constructing elaborate Boolean pattern matchers (see Jacobs, Krupka & Rau paper).
Best Manual Acihoc
0.00 0.10 0.20 0.30 0.40 0.50 0.60 0.70 0.80 0.90 1.00
Recall
0.9
0.8
0.7
0.6
0
[OCRerr] 0.5
0.4
0.3
0.2
0.1
0
cnqs[OCRerr] clartb gecrd2
Figure 6. The Best Adhoc Results using Manual Query Construction.
14