SP500215
NIST Special Publication 500-215: The Second Text REtrieval Conference (TREC-2)
Bayesian Inference with Node Aggregation for Information Retrieval
chapter
B. Del Favero
R. Fung
National Institute of Standards and Technology
D. K. Harman
In conjunction with these representational improvements,
we would like to design a complete user interface that
would allow users to make the needed probabilistic
judgements easily and intuitively. We would also like to
develop an explanation facility that could describe why one
document was preferred over another.
There are new exact and inexact algorithms that could
handle these representational modifications. We would like
to experiment with these algorithms to see if they are
suitable. We would also like to implement a relevance
feedback mechanism based on the Bayesian concept of
equivalent sample size.
Finally, while the information retrieval problem has been
viewed primarily (and rightly so) as an evidential reasoning
problem, we take the position that a decision-theoretic
perspective is more accurate since information retrieval is a
decision process. We believe that this perspective can
provide additional insight and eventually improve upon the
probabilistic approaches that have been developed.
Acknowledgment
We would like the thank Dr. Richard Tong for his
assistance and guidance.
Appendix Bayesian Inference
We calculate the posterior probability that a document is
relevant to a state using Bayes' rule and an assumption of
conditional independence between the features.
The Bayesian inversion formula is this:
p(sldocument)=p(s IF) = p(FIs)p(s)
p(F)
where 5 is the state, and F is a binary feature vector with
= 1 if feature i is present (the event fj+)
F[OCRerr]= 0 if feature i is absent (the event [OCRerr]
The set of all features that are present is denoted F+. All
other features are absent and are in the set denoted F-.
The first term in the numerator is expanded under the
assumption that the features are conditionally independent
given the state.
p(FIs)= [1 p(ij+ Is) Y' p(Ij Is)
iinF+ iinF[OCRerr]
The second term in the numerator is the state prior, p( s),
which can be estimated as described in Section 3.2.
160
The denominator is a normalization constant obtained by
summing over the values for the numerator for all the
states.
p(F)= [OCRerr] p(FIs)p(s)
all 5
p( F) is the probability that one would observe the
particular set of features F.
References
Chang, K. C., & Fung, R. M. (1989). Node Aggregation for
Distributed Inference in Bayesian Networks. In IJCAI 89.
Detroit: Morgan Kaufmann.
Fung, R. M., & Chang, K. C. (1989). Weighing and
Integrating Evidence for Stochastic Simulation in Bayesian
Networks. In R. D. S. M. Henrion L.N. Kanal and J.F.
Lemmer (Eds.), Uncertainty and Artificial Intelligence 5
Elsevier Science Publishers B.V. (North-Holland).
Fung, R. M., Crawford, S. L., Appelbaum, L., & Tong, R.
M. (1990). A Probabilistic Concept-based Architecture for
Information Retrieval. In Proceedings of the ACM
International Conference on Information Retrieval
Brussels, Belgium.
Heckerman, D. E. (1989). A tractable algorithm for
diagnosing multiple diseases. In Proceedings of the Fifth
Workshop on Uncertainty in Artificial Intelligence, (pp.
162-173). Detroit, MI.
Heckerman, D. E. (1991). Probabilistic Similarity
Networks. The MIT Press.
Howard, R. A., & Matheson, J. E. (1981). Influence
diagrams. In R. A. Howard & J. E. Matheson (Eds.),
Readings on the Principles and Applications of Decision
Analysis (pp.721-762). Menlo Park, Ca.: Strategic
Decisions Group.
Lauritzen, S. L., & Spiegelhalter, D. J. (1988). Local
computations with probabilities on graphical structures and
their application to expert systems. JRSS[OCRerr] 50, 157-224.
Maron, M. E., & Kuhns, J. L. (1960). On relevance,
probabilistic indexing and information retrieval. Journal of
the ACM, 7, 216-244.
Paice, C. D. (1977). Information Retrieval and the
Computer. London.
Pearl, I. (1988). Probabilistic Reasoning in Intelligent
Systems. San Mateo, Ca.: Morgan Kaufman.