IRE Information Retrieval Experiment The Cranfield tests chapter Karen Sparck Jones Butterworth & Company Karen Sparck Jones All rights reserved. No part of this publication may be reproduced or transmitted in any form or by any means, including photocopying and recording, without the written permission of the copyright holder, application for which should be addressed to the Publishers. Such written permission must also be obtained before any part of this publication is stored in a retrieval system of any nature. Cranfield 1 259 one, focusing on indexing languages, and seeking to identify and control other system factors bearing on the study variable. The indexing systems chosen for test were `(a) The Universal Decimal Classification. (b) Alphabetical subject catalogue. (c) A facetted classification scheme. (d) The Uniterm system of coordinate indexing.' (p.8) According to the Report, `the basis for this selection was that the schemes differed as fundamentally as is possible and represented the principal types of retrieval systems which have any significance in the present state of the art.' (p.8) The UDC was chosen as the most widely used system of the `enumerative' type, illustrating tree-of-knowledge classification and the use of a decimal notation. Facet classification specifically lacks these three features. The alphabetical index is deliberately anti-classificatory and relatively uncon- trolled in language, while Uniterms are equally uncontrolled, but allow all permutations and combinations of individual terms to define subjects. The facet and Uniterm languages both represented novel approaches to indexing, the former being particularly carefully prepared. With respect to the other main test variables, the document subject area, aerodynamics, was determined by convenience, while the set of documents used was chosen ad hoc, to allow variability in detailed document topics, types, and sources. The indexers were chosen to have various types of experience and familiarity with the systems being tested. (It may be noted that `the project imposed a severe mental strain on the indexers' (p.2).) Equipment, i.e. the physical form of indexes, was not studied as a test variable. The size of the document set for the indexes was prima[OCRerr]ily based on the wish to ensure that retrieval was not too obvious, without being wasteful; more specifically, with 60 permutations of the three major variables, obtained as explained below, and taking 100 documents as a convenient number for each permutation, there were 3 `subprogrammes' of 6000 documents for the total of 18 000. In fact, since half the documents were in a specialized subject area, the collection was deemed representative of a much larger set. Overall, `the assumption on which the investigation [was] based [was] that the only valid way to measure the efficiency of any system of indexing is by basing measurements on economic costs and in this there are always three matters to be considered, these being: (a) The cost of indexing. (b) The cost of preparing the physical index. (c) The cost of searching.' (p.18) So on the indexing side, since equipment questions were deliberately excluded, the major variables being investigated were the system, the indexer, and the indexing time, i.e. 4 systems x 3 indexers x 5 times = 60 permutations. The procedures adopted in the indexing are described in detail, in relation