number of samples used to estimate the parameters in the decision function.
That this will lead to problems has been pointed out repeatedly in the pattern recognition literature.
although the analysis of the problem in pattern recognition applies to IR as well, the solutions are not directly applicable.
In pattern recognition the problem is: given the number of samples that have been used to 'train' the decision function (our weighting function), is there an optimum number of measurements that can be made of an unknown pattern so that the average probability of correct assignment can be maximised? In our case how many index terms can we legitimately use to decide on relevance.
Hughes[18] shows that for a very general probabilistic structure the number of measurements is surprisingly small even though reasonably sized samples are used to 'train' the decision function.
Ideally one would like to be able to choose a (small) subset of index terms to which the weighting function g(.) would be restricted thereby maximising the average probability of correct assignment.
In pattern recognition there are complicated techniques for doing just that for the equivalent problem.
In information retrieval we are fortunate in that there is a natural way in which the dimensionality of the problem can be reduced.
We accept that the query terms are a fair guide to the best features to be used in the application of g(.) to decide between relevance and non-relevance.
Therefore rather than computing the weighting function for all possible terms we restrict g(.) to the terms specified in the query and possibly their close associates.
This would be as if during the retrieval process all documents are projected from a high dimensional space into a subspace spanned by a small number of terms.
Computational details
I now turn to some of the more practical details of computing g(x) for each x when the variables xi are assumed to be stochastically dependent.
The main aim of this section will be to demonstrate that the computations involved are feasible.
The clearest way of doing this is to discuss the calculation of each 'object' EMIM, MST, and g(.) separately and in that order.
1. Calculation of EMIM
The calculation of the expected mutual information measure can be simplified.
Then EMIM itself can be approximated to reduce the computation time even further.
We take the simplification first.
|