the system to its user will be the best that is obtainable on the basis of those data.
Of course this principle raises many questions as to the acceptability of the assumptions.
For example, the Cluster Hypothesis, that closely associated documents tend to be relevant to the same requests, explicitly assumes the contrary of assumption (1).
Goffman[8] too, in his work has gone to some pains to make an explicit assumption of dependence.
I quote: 'Thus, if a document x has been assessed as relevant to a query s, the relevance of the other documents in the file X may be affected since the value of the information conveyed by these documents may either increase or decrease as a result of the information conveyed by the document x.' Then there is the question of the way in which overall effectiveness is to be measured.
Robertson in his paper shows the probability ranking principle to hold if we measure effectiveness in terms of Recall and Fallout.
The principle also follows simply from the theory in this chapter.
But this is not the place to argue out these research questions, however, I do think it reasonable to adopt the principle as one upon which to construct a probabilistic retrieval model.
One word of warning, the probability ranking principle can only be shown to be true for one query.
It does not say that the performance over a range of queries will be optimised, to establish a result of this kind one would have to be specific about how one would average the performance across queries.
The probability ranking principle assumes that we can calculate P(relevance/document), not only that, it assumes that we can do it accurately.
Now this is an extremely troublesome assumption and it will occupy us some more further on.
The problem is simply that we do not know which are the relevant documents, nor do we know how many there are so we have no way of calculating P(relevance/document).
But we can, by trial retrieval, guess at P(relevance/ document) and hopefully improve our guess by iteration.
To simplify matters in the subsequent discussion I shall assume that the statistics relating to the relevant and non-relevant documents are available and I shall use them to build up the pertinent equations.
However, at all times the reader should be aware of the fact that in any practical situation the relevance information must be guessed at (or estimated).
So returning now to the immediate problem which is to calculate, or estimate, P(relevance/ document).
For this we use Bayes' Theorem, which relates the posterior probability of relevance to the prior probability of relevance and the likelihood of relevance after observing a document.
Before we plunge into a formal expression of this I must introducesome symbols which will make things a little easier as we go along.
|