SP500215
NIST Special Publication 500-215: The Second Text REtrieval Conference (TREC-2)
Incorporating Semantics Within a Connectionist Model and a Vector Processing Model
chapter
R. Boyd
J. Driscoll
National Institute of Standards and Technology
D. K. Harman
top I
pI=O.02
0.9 0.9 0.9
e
(army) (engineer) (plant)
Figure 4: A Simple Network Consisting of Two Document
Nodes and Three Index Term Nodes.
receives information about the activation levels of its
immediate neighboring nodes (nodes connected to it via
direct links), and then uses this information to calculate its
own activation leveL Through this process of spreading
activation, the network setUes down to equilibrium repre-
senting a retrieval to a user's information need.
The computation of the information retrieval inference
process is based on a lormaliration of the causal and proba-
bilistic associative knowledge underlying diagnostic prob-
lem-solving [18). We do not discuss the formulation
architecture and activation mechanism of the connectionist
model. This information can be found in [11,13,16,18). For
TREC-2, we managed to complete only one official routing
experiment for this approach, and itdid not involve semantics.
The experiment was intended to be a baseline experiment for
our semantic experiments.
For ThEC-2, a specific network was constructed for 50
topics. A list of index terms was assembled based on
keywords in the concept section of each topic. In this network,
each output node represented a topic, and each input node
represented akeyword. The prior probability assigned to each
topic node was equal to 1/(total number of topics). The
connection strengths were assigned equal weights (0.9).
The network contained 50 topic nodes and 848 index term
nodes. These nodes were connected via 1449 links. An
example of this network is shown in Figure 5, where p[OCRerr] is the
prior probability of topic top[OCRerr]. The keywords "army",
"engineer", and "plant" were obtained by processing the
concept section of topic tO[OCRerr][OCRerr] Currently, the network is
enhanced by using an estimated weighting scheme.
We performed a Category B routing experiment. Using
just keywords, the results were not good. The main problem
was due to the fact that, in the document ranking, many
documents had the same score used to generate the ranking.
In order to satisfy the requirements for the ranking, we had
to artificially rank those documents with the same score. This
was done based on order of appearance. The performance
was terrible except for Topic 66. This topic had only two
295
Figure 5: A Sample Network of the Experimental ModeL
known relevant documents for Category B routing experi-
ments and our inference network retrieved one of them in the
top 20 documents! No further connectionist model
experiments have been completed. We were unable to modify
the baseline keyword experiment or perform semantic
experiments for this approach.
4. Vector Processing Model Experiments
In this section, we explain the manner in which semantics
is incorporated within a vector processing model using the
semantic lexicon explained in Section 2. Please note that an
entry in our semantic lexicon has the form of a word followed
by codes for each of the semantic categories theword triggers.
We explain our approach usingatext relevance determination
procedure intended to show what is being calculated rather
than show the actual computations for the approach. The
procedure presented here generates several outputs that are
really not necessary, but are included just to help explain the
approach. The relevance determination procedure is
explained using the four documents and query shown in
Figure 6. A few preliminary computations are reviewed in
order to explain the procedure.
First, the number of documents each word is in must be
determined. Figure 7 shows a list of words from the four
documents and the query of Figure 6 along with the number
of documents each word is in (dJ).
Next, the inverse document frequency (idi) of each word
is determined by the equation 1og10(NIdJ), where N -4, the
total number of documents. Figure 8 provides the idjof each
word. Sometimes, the kif of a word is undefined. This can
happen when a word does not occur in the documents but
does occur a query. For example, the words "depart" "do"
in
and "when" do not appear in the four documents. Thus, the
idf of these terms cannot be defined here. Later, we will see
that an adjustment can be made for these undefined values.
Next, the category probability of each query word is
determined. Figure 9 shows an alphabetized list of all the
unique words from the query, the frequency of each word in
the query, and the semantic categories each word triggers.