CRANV1P1 ASLIB Cranfield Research Project: Factors Determining the Performance of Indexing Systems: VOLUME 1. Design, Part 1. Text General Considerations chapter Cyril Cleverdon Jack Mills Michael Keen Cranfield An investigation supported by a grant to Aslib by the National Science Foundation. Use, reproduction, or publication, in whole or in part, is permitted for any purpose of the United States Government. Means of control Field of use 6. Generic descriptors. 7. Specific and generic descriptors linked hierarchically. 8. Multiple generic links for each specific descriptor. 9. Categories of descriptor, form- ing facets. 10. Semantic factors to represent subject terms. 11. Correlation of descriptors. 12. Weighted descriptors. 13. Interlocking sets of descriptors. 14. Regulated sequence of descriptors. 15. Interfixing descriptors. 16. Role indicators. 17. Relational terms. Many mechanized systems. Classifications, thesauri, some subject head- ing lists, some mechanized systems. Some classifications, subject heading lists, and thesauri, a few mechanized systems. Faceted classifications, some mechanized systems. To some extent in faceted classification, the W.R.U. system, mechanized patent office systems. Many alphabetical indexes, some classified catalogues, all mechanized systems. Some experimental computer systems. Alphabetical indexes, classified catalogues, computer systems. Alphabetical indexes, faceted classifications, fixed-field punched cards, some computer systems. Mechanized patent office systems. Some faceted classifications, some mechanized systems. Alphabetical indexes, some faceted classifi- cations, some mechanized systems. All the results of Cranfield I pointed to only one conclusion. Whereas one could evaluate the performance of an operational information retrieval system and find how the index language being used affected the performance of the particular system under investigation, it was not possible to do any basic research on index languages by this method, for there are so many uncontrollable variables in any operational system that comparison of index languages is impossible. It has to be admitted that this view is not generally held, since one finds a new investigation which has the objective of comparing various UDC operational systems with other operational systems using different types of index languages. In that this results in even more variables than existed in Cranfield I, it is difficult to see how any valid data concerning the UDC can be obtained. On this point Richmond is in complete agreement, for she writes (ref. 8) "System evaluation by comparison testing is essentially a negative operation", and again, "Comparison with other systems does not answer problems arising from the weaknesses of this system. In each case, the faults are internal and only obliquely subject to evaluation by comparison with other systems". To make advances in knowledge regarding index languages, what was now re- quired was a laboratory-type situation, where, freed from the contamination of operational variables, the performance of index languages could be studied in isolation. While such an approach waa unusual in 1961, at least two other organizations have also established similar conditions, namely the Centre for Documentation at Western Reserve University and the Computation Laboratory of Harvard University. The methods used at Cranfield to establish this situation are considered in the following chapters of this volume.