SP500215
NIST Special Publication 500-215: The Second Text REtrieval Conference (TREC-2)
TREC-2 Document Retrieval Experiments using PIRCS
chapter
K. Kwok
L. Grunfeld
National Institute of Standards and Technology
D. K. Harman
NewAdhoc
- Revised Routing
pii[OCRerr]s5 Pii[OCRerr][OCRerr] % pii[OCRerr]7 % pirc[OCRerr]
Total number of documents over all 50 queries
Retrieved: 50000 50000 50000 50000
Relevant: 10489 10489 10489 10785
Rel_ret: 7098 7476 +5-3 7385 + 4.0 8279
Interpolated recall - precisron averages:
0.0 .765 .814 + 6A .793 +3.7 .823
0.1 .554 .628 +13.3 .594 + 7.2 .561
0.2 A91 .546 +11.2 .534 +8.6 .505
0.3 .421 .469 +11.4 .469 +11.4 .456
OA .380 A13 + 8.7 .416 +9.5 .417
0.5 .336 .363 + 8.0 .368 +9.5 .368
0.6 .270 .310 +14.8 .314 +16.3 .320
0.7 .212 .240 +13.2 .241 +13.7 .246
0.8 .150 .165 +10.0 .171 +14.0 .160
0.9 .079 .099 +25.3 .098 +24.0 .087
1.0 .014 .006 -57.1 .015 + 7.1 .012
Average precision (non-interpolated) over all rel does
.318 .355 +11.6 .350 +10.1 .344
Precision at
S does: .600 .660 +10.0 .624 +4.0 .612
10 .582 .632 + 8.6 .598 + 2.7 .572
15 .545 .617 +13.2 .572 + 5.0 .573
20 .527 .583 +10.6 .563 + 6.8 .564
30 .507 .553 + 9.1 .534 +5.3 .540
100 .402 .439 +92 .427 +62 .468
200 .334 .360 +7.8 .353 + 5.7 .3%
500 .222 .238 +7.2 .232 +4.5 .264
1000" .142 .150 + 5.6 .148 +4.2 .166
R-precislon
Exact .358 .385 + .385 + .378
Table 2: Revised Routing and New Ad Hoc Retrieval Results
and 53.7% of the operational maximum. The R-precision
Exact calculates the precision value at x retrieved. where x
is the known number of relevants for each query, and can
be compared with the theoretical value of 1.0.
The `Upperbound' retrieval pircs7 (suggested by Sparck-
Jones) means perfonning learning from the known Dis1:3
(not Disk 1&2) relevant documents before retrieval. In
other words, we assume the answer documents are known
for training and represent the best that probabilistic they
can provide using our system. This is however not the true
upperbound [Spar79] for retrieval from Disk3. because the
vocabulary and usage statistics are still those of Disk 1&2.
The vocabulary is retained for comparison with routing
240
results. pircs7 retrieval achieves average precision of 0.350,
which improves over pircss (training from Disk 1&2) by
about 10% in average precision and about 4% in relevants
retrievei of corse in real life. the answer documents are
not known; but it is interesting to note that query expansion
using Disk 1&2 documents can provide similar
performance, showing the imp&tance of query expansion.
We later concentrate on routing and discover that additional
gains can be achieved by fine tnning of the parameters in
our model. For learning: 1) we find that our original
method of using only `nonbreak' documents in the given set
of relevants actually outprfonns other doccnent selection
strategies including using all relevants, `break six' or