SP500215 NIST Special Publication 500-215: The Second Text REtrieval Conference (TREC-2) Automatic Routing and Ad-hoc Retrieval Using SMART: TREC 2 chapter C. Buckley J. Allan G. Salton National Institute of Standards and Technology D. K. Harman Run Best > median <median crnlV2 1 38 11 crnlL2 4 40 6 crnlV2-b 9 36 5 Table 1: Comparative Ad-hoc results Global Similarity The crnlV2 run demonstrates the quality of results obtainable with simple methods. The weighting for terms is chosen based upon re- sults from TREC 1. Query terms are weighted by the formula in equation (1) ("ltc" in Smart's vocabulary). Document terms are weighted using a normalized logarithmic term frequency ("lnc" ): d[OCRerr]k = logf[OCRerr]k+1.0 [OCRerr][OCRerr]3i.[OCRerr]1(logfjj +1.0)2 (3) where dik is the weight of term Tk in document D[OCRerr], fik is the occurrence frequency of term Tk in document D[OCRerr], and t is the total number of terms in the collection. The denominator provides normalization of vector length. Note the absence of the "idf" factor log(N/nk). Table 2 shows the results of that weight- ing scheme in crnlV2-b. Regrettably, be- cause of an oversight during the official run (a misnamed inverted file), the official submit- ted run crjilV2 did not use the weighting ap- proach described above (and recommended in our TREC 1 report). Instead crnlV2 acciden- tally used the "idf" factor in both the query and the document. That mistake caused 10% loss in retrieval effectiveness (from a recall- precision average of 0.3512 to 0.3163). Global and Local Cornell's TREC 1 ad-hoc submission in- creased the similarity measure of a query and document if some sentence in the query matched some sentence in the document suf- ficiently well.[1] The result was that any query/document pair which contained a sen- tence match was retrieved before all that did not have such a match. For TREC 2, we hoped to find a less restrictive balance between the global and local similarities. At the same time, mg parts other than sentences, and to investi- gate combining mul[OCRerr]iple local similarities. Our approach is similar to that used in [4]. We built a training collection using the 50 queries from Q2 and the 74,520 documents from the Wall Street Journal included in D2. For each of the 3.7 million query/document pairs, we calculate the global similarity and some set of local similarity values. The least squares polynomials (LSP) approach devel- oped for [4] are used to find the "ideal" c[OCRerr] efficients for the global and local values in the equation: sim = global + [OCRerr] locali + ,@[OCRerr] local2 + ... (The LSP functions actually yield a constant factor which we ignore since it does not affect ranking.) We consider local values from the following broad classes: * Comparing sentences of the query against sentences of the document. In general, we use a simple "tfx idf" weight without nor- malization, though we experimented with other weights. * Comparing paragraphs of the query against paragraphs of the document. (For the most part, each section of the query topic is a separate paragraph.) In this case, we use the weighting of equation 1 above for the query paragraphs, and try a variety of weights for the document para- graphs. * Comparing the query against paragraphs of the document. We use the same weighting schemes as above. We also tried combinations of the above cate- gories: e.g., the best matching paragraph pair and the best matching sentence pair. See Ta- ble 3 for a complete list of local values that were considered. In all, we tried 72 combinations of local and global values, using from one local value to 19 different local values.1 The LSP-determined a and [OCRerr] `5 of those values are then applied to a retrieval run on that same set of queries and documents. The top performing result in- cludes only a single local value: the best match 1There were rougMy 1.2 million possible combina- tions; we chose 72 that seemed, based on earlier exper- we wished to investigate local similarities us- iments, likely to succeed. 48