IRE Information Retrieval Experiment The pragmatics of information retrieval experimentation chapter Jean M. Tague Butterworth & Company Karen Sparck Jones All rights reserved. No part of this publication may be reproduced or transmitted in any form or by any means, including photocopying and recording, without the written permission of the copyright holder, application for which should be addressed to the Publishers. Such written permission must also be obtained before any part of this publication is stored in a retrieval system of any nature. Decision 3: How to operationalize the variables? 65 printed indexes, card catalogues, and other manual files and in online interactive computer searching, the response to one search statement triggers the next, until the searcher is satisfied. The length of this process can be measured by the total number of search statements or the search time. These measures, however, confound the complexity of the search and the experience or style of the searcher. Variables describing overall complexity as a function of inter-statement relationships need to be developed. Another search process variable is the form of the output: citation, citation and abstract, citation and index terms, citation and full text, etc. Particularly in online systems, there may be a choice of output format. In fact, in online searching the term output is ambiguous. Typically, one keys in a boolean search statement and gets as response a statement of the number of hits. If this is large, no further output at all may be requested, or a few titles only may be scanned. The strategy may then branch in one of several directions: list all citations online modify the search statement list all citations offline terminate the search In an experimental situation, a choice must be made as to what constitutes output. No consensus appears to have yet been reached in the field. This is unfortunate, as measures of effectiveness obviously depend on how output is defined. Marcus, Kugel and Benenfeld5 have introduced the idea of the indicativity of an output field. It is defined operationally as the proportion of documents which the field indicates to be relevant which are actually assessed as relevant from full text. It appears that indicativity is related logarithmically to the length (i.e. number of words) in the field. PeopI[OCRerr]indexers, searchers, users Because many of the decisions in information retrieval are based on human judgement, the professional training and experience of the system personnel and of the users is often studied. These variables can usually be operationalized in terms of years of experience and/or training, in general or with particular Systems. Job titles are an unreliable guide, as they vary from installation to installation. Number of previous searches can also be used as a measure of experience with a particular system. Many of the things one wants to know about people can be characterized as attitudinal variables: ease with other people, acceptance of automated processing. Experimenters should be aware of the large psychological literature on attitude measurement, and consider the use of one of the standard scales[OCRerr]Guttman, Lickert, Thurstone, the `unfolding' method. A good text is Lemon6. He defines attitude measurement as gathering observations about people's behaviour and allocating numbers to these observations according to certain rules. Attitude scales depend on the investigator's theoretical assumptions about the nature of the attitude he or she is trying to measure, its relationship with behaviour, and the rules used