SP500207 NIST Special Publication 500-207: The First Text REtrieval Conference (TREC-1) Optimizing Document Indexing and Search Term Weighting Based on Probabilistic Models chapter N. Fuhr C. Buckley National Institute of Standards and Technology Donna K. Harman For each phrase occurring in a document, indexing weights for the phrase as well as for its two components (as single words) were computed. In the following, we will refer to the indexing with single words only as "word indexing" and to the indexing using both single words and phrases as "phrase indexing". 3 Retrieval For the probabilistic document indexing weights as described above, there are specific retrieval func- tions which yield a probabilistic or utility-theoretic ranking of documents w.r.t. a query. These func- tions differ in the aspect whether or not they consider relevance feedback information for the specific query. For this reason different retrieval functions were applied for the ad-hoc queries and the routing queries. 3.1 Ad-hoc queries For the ad-hoc queries, we used a linar utility-theoretic retrieval function described in [Wong & Yao 89). Let qT[OCRerr] denote the set of terms occurring in the query, [OCRerr] the set of document terms and ti[OCRerr],,,, the indexing weight u(x[OCRerr]i[OCRerr], dm)). If c[OCRerr]k gives the utility of term t[OCRerr] for the actual query q[OCRerr], then the utility of document dm w.r.t. query q[OCRerr] can be computed by the retrieval function Cik Uim. (1) t[OCRerr]qT[OCRerr]ndT[OCRerr] The only problem here is the estimation of the utility weights Cik, for which we do not have a theoreti- cally justified method. As a heuristic approach, we used the SMART if.idf weights, where If denotes the number of occurrences of the term I[OCRerr] in the query q[OCRerr] here [Salton & Buckley 88]. We assume that there are other choices for this parameter which could significantly improve retrieval quality. run if.idf [OCRerr]uhra1 [OCRerr]uhrpi (words) (phrases) average precision for recall points: 11-pt Avg. 0.1750 0.1943 0.2054 3-pt Avg. 0.1159 0.1399 0.1664 query-wise comparison with median: 11-pt Avg: 32:14 32:17 Prec. @ 100 docs 25:21 35:13 Best/worst results: 11-pt Avg: 6(2)11(1) 4(1)12 Prec. @ 100 docs 2(2)12 8(3)1(2) single words vs. phrases: 11-pt Avg: 25:25 Prec. @ 100 docs 21:28 Tabelle 1: Results for adhoc queries The retrieval function (1) was applied for both kinds of document indexing, where words only were considered for run fuhral and both words and phases in the run fuhrpi. Figure 2 shows the recall- precision curves for both runs in comparison to a standard SMART if idf run. It can be seen that 93