IRE
Information Retrieval Experiment
Laboratory tests: automatic systems
chapter
Robert N. Oddy
Butterworth & Company
Karen Sparck Jones
All rights reserved. No part of this publication may be reproduced
or transmitted in any form or by any means, including photocopying
and recording, without the written permission of the copyright holder,
application for which should be addressed to the Publishers. Such
written permission must also be obtained before any part of this
publication is stored in a retrieval system of any nature.
Laboratory programs 165
Relevant Non-re[OCRerr]evant
Matching
value
2
3
m
[OCRerr]"igure 9.3. Table for counting relevant and non-
rclcvant documents at each matching value
very different from modern online systems; although it resembles the early
magnetic tape-based retrieval systems which processed queries in batches.
If the researcher is concerned with aspects of the interaction between the
user and the computer, then the additional complexity of the retrieval process
will usually oblige him to adhere more closely to the structure of real life
systems. An experiment to evaluate the effect of relevance feedback8'0, for
instance, would require the program to select a retrieved set of documents, so
that their indexing can be used together with relevance decisions to compute
a modified query. No doubt, more subtle and efficient ways could be devised,
but this may be self-defeating, in that one's confidence in the correctness of
the programs may be endangered.
Programs which model some aspect of the user's cognitive behaviour may
have no formal specification: the details of the model are worked out in the
process of writing, and trying the program. These programs will almost
certainly be prototypes (see, for instance, Oddy33), inasmuch as they proceed
as a real life system would, simply because you cannot design algorithms
which efficiently capture the essence of a technique for test purposes until
you know precisely what that technique is.
Those who have been engaged in laboratory work in information retrieval
over several years now have extensive suites of programs which can be
applied in a wide variety of combinations to test collections. The well-known
Smart system is described in Salton8. Programs are included for deriving the
numerically coded form of a collection from natural language text, for
clustering documents, for selection of clusters appropriate to a query (the
whole collection may be selected), for searching the selected clusters, and for
evaluation. All of these programs have a number of optional facilities. The
program suites used at Cambridge University are described and discussed by
Sparck Jones2 and Sparck Jones and Bates6.
A major advantage of an automatic information retrieval laboratory,
equipped with a flexible program suite and a number of conveniently
formatted test collections, is that it is possible to try to accumulate convincing