SP500207 NIST Special Publication 500-207: The First Text REtrieval Conference (TREC-1) TIPSTER Panel -- The University of Massachusetts TIPSTER Project chapter W. B. Croft National Institute of Standards and Technology Donna K. Harman * The documents in TIPSTER They collie the general arc[OCRerr]a are the [OCRerr] Tall Street Journal, technology area, Departnicnt are heterogeneous ill terms of both subject and length. of science, technology and economics, but the sources Associated Press newswire, Ziff magazines in the high of Energy abstracts, and the Federal Register. * The queries (known as `topics" in TIP STER) are longer and have more structure than those found in other test collections. * The queries have specific and These criteria (specified in the between relevance judges but information retrieval Svsteiii. strict cutena specified for documents to be relevant. "narrative" part of the topic) will reduce inconsistency are sometimes difficult to handle in the context of an * The routing experinielits are unhke any carried out before. * The retrieval and routilig experiments [OCRerr]vith Japanese are also unique. The first TIPSTER. evaluatioii w('s liujited by a number of factors, the primary one being the lack of relevance judgeiiieiit.[OCRerr] for the jiutial query set. This made it difficult to carry out experiments to sele([OCRerr]t technique.[OCRerr] appropriate for large, full-text databases. The results from this evaluation 5lL()1l1(l[OCRerr] therefore, be regarded as preliminary, and indeed raise more questions than they answer. In the retrieval experinient, 50 new [OCRerr]tol)ics" were used to search the "old" database, which consisted of approxiiiiately 1 GByte of text. One of the major subjects of the eval- uation was to try different forms of queries produced by processing the topics. Our basic approach to topic processing is to parse them, selecting parts to be indexed, recognizing phrases and "factors" such a." locations, dates, companies, etc. Some factors, such as "de- veloping country", which have been specifically identified as important in the topic, will be expanded using a synonym operator. Weights reflecting relative importance are attached to the concepts (words and phrases). Phr[OCRerr]ise-based concepts are represented by operators defined in the inference net l[OCRerr]u1guage. These operators use proximity of the words making up the phrase as the major form of evidence for the presence of the concept [1]. The result of topic processing is an inference net representing the information nee(l. In addition to the automatic query processing, some query versions were generated by simulating simple user interactioii with the results of the topic processing. The modifica- tions to the automatically processed topics were limited to changing the weight of concepts, deleting concepts considered uniniportant, and adding structure (such as specifying syn- onymous concepts). The lilost signifi([OCRerr]ant change in the last category was the introduction of "unordered window" 01) er[OCRerr]i.tors to Si umlate paragraph-level retrieval. The equivalent in terms of a user user interface would 1)e to ask users to group concepts that should occur together. The results of the first evaluation are described here in terins of the average precision in the top 5, 30 and 200 documents in the ranking produced by the inference net retrieval engine (INQUERY). This evaluation method was chosen because only the top 200 documents for each query were judged for relevance. The results were as follows: 102