CRANV2 Aslib Cranfield Research Project: Factors Determining the Performance of Indexing Systems: Volume 2 Methods for presentation of results chapter Cyril Cleverdon Michael Keen Cranfield An investigation supported by a grant to Aslib by the National Science Foundation. Use, reproduction, or publication, in whole or in part, is permitted for any purpose of the United States Government. - 31 CHAPTER 3 Methods for Presentation of Results Lord Kelvin is often credited with remarking, [OCRerr]When you can measure what you are speaking about and express it in numbers you know something about it, but when you cannot measure it, when you cannot express it in numbers, your knowledge is of a meagre and unsatisfactory kind'. The problem of validity - how closely do the figures relate to the 'thing we are talking about' must be separated from the problem of reliability - how accurate are the figures themselves. L.T. Wilkins: Social Deviance, page 147 In Cranfield I, the results of the main project did no more than record what is now generally known as the Recall Ratio, which was 100R calculated on the basis of [OCRerr] where R equals the number of relevant documents retrieved and C equals the total number of documents in the collection which are relevant to the questions. In the subsequent test of the Western Reserve University Index, (Ref. 2) measurement was carried to the stage where, by making relevance assessments of all the retrieved documents, it was also possible to calculate what was originally called the Relevance Ratio, but which is now generally known 100R as the Precision Ratio, namely [OCRerr] , where L equals the total number of documents retrieved in the series of searches. In the course of this evaluation of the W.R.U. Index, the effect of varying the exhaustivity of indexing was measured, and allowed the production of the first - and, incidentally, so far the only - performance curve from the Cranfield project. It is reproduced in Fig. 3.1P and showed two interesting characteristics. The first was the inverse relationship between recall and precision, which has been considered at some length in Volume 1 of this report. The second point was that, when documents of lower relevance, were accepted, there was at any given level of indexing exhaustivity, a lower recall ratio but an improved precision ratio. It was tentatively suggested that this latter point was connected with a variation in the average number of relevant documents for each question, and that, for any.given situation, it would be necessary to state also what was to be later termed the 1000C Generality Number$, expressing it as [OCRerr] , where N equals the total number of documents in the collection. Sin the earlier volume of this report, this was called Generality Ratio.