Evaluation methods in text categorization

From DataMin
Jump to: navigation, search

Text categorization

Categorizing textual data is the task of determining the correct class or classes of unseen documents based on some learning examples. Categories or classes are predetermined therefore a supervised learning algorithm is needed to classify documents.

Formally, following the notations used by Sebastiani in [Sebastiani02], let  \mathcal{D}=\{d_{1},\ldots, d_{|\mathcal{D}|}\} and  \mathcal{C}=\{c_{1},\ldots, c_{|\mathcal{C}|}\} be the set of the documents and categories respectively. The task is to approximate as precisely as it is possible the unknown target function by  \tilde{\Phi}:\mathcal{D}\times \mathcal{C}\rightarrow \{T,F\}, where  T and  F denotes "True" and "False", respectively. If there are only two classes, then the categorization is called binary, while if  |\mathcal{C}|>2 is called multiclass. Moreover, if the categories are disjunctive or non-overlapping, that is a document belongs to exactly one category, is called single-label, while if the categories are overlapping, that is a document can belong to arbitrary number of categories, it is said multilabel.

A text categorization system can use hard or ranking policy. Hard categorization means that the classifier takes a hard decision about a document-category pair, while ranking categorization ranks the categories for a given document based on some distance metric. This metric is called categorization status value,  CSV_{i}:\mathcal{D}\rightarrow [0,1],  i=1,\ldots,|\mathcal{C}|. The greater the CSV-value of a category for a document the better the document belongs to the category.

Precision and recall

Precision and recall are the most common measures for evaluating an information retrieval system. Precision is the proportion of returned documents that are targets, while recall is the proportion of target documents returned.

Category i Expert judgement
Classifier judgement TRUE TP_{i} FP_{i}
FALSE FN_{i} TN_{i}



The definition can be rewritten using probabilities. Suppose the learning algorithm was run and decisions had been made. This set constitutes our space. The precision can be written as


while recall is


Besides these accuracy and error another two common IR measures are sometimes used to characterize TC systems:


There are two conventional methods of calculating the performance of a text categorization system based on precision and recall. The first is called micro-averaging, while the second one macro-averaging. Micro-averaged values are calculated by constructing a global contingency table (see above for one class) and then calculating precision and recall using these sums. In contrast macro-averaged scores are calculated by first calculating precision and recall for each category and then taking the average of these. The notable difference between these to calculations is that micro-averaging gives equal weight to every document (it is called a document-pivoted measure) while macro-averaging gives equal weight to every category (category-pivoted measure).

P_{\mathrm{micro}} = \frac{\sum_{i=1}^{|C|}TP_{i}}{\sum_{i=1}^{|C|}TP_{i}+FP_{i}}; \quad R_{\mathrm{micro}} = \frac{\sum_{i=1}^{|C|}TP_{i}}{\sum_{i=1}^{|C|}TP_{i}+FN_{i}}
P_{\mathrm{macro}}=\frac{1}{|C|}\sum_{i=1}^{|C|}\frac{TP_{i}}{TP_{i}+FP_{i}};\quad R_{\mathrm{macro}}=\frac{1}{|C|}\sum_{i=1}^{|C|}\frac{TP_{i}}{TP_{i}+FN_{i}}

Precision-recall breakeven point

The properties or the expected behaviours of text categorization/information retrieval systems can vary. For example for one system it is better to return mostly correct answers, while in another it is better to cover more true positives. There is a tradeoff between precision and recall: if a classifier says "True" to every category for every document, then it receives perfect recall, but very low precision. However it can be easily seen that if a classifier says "False" for every category, except one which is correct ( TP=1, FP=0) then it will have a precision equal to  1 but a very low recall. That is why it makes comparison between systems easier if the system is characterized by a single value, the breakeven point (BEP), which is the point at which precision equals recall. This can be achieved by tuning the parameters of the system. When there is no such point (because  TP,  FP and  FN are natural numbers) the average of the nearest precision and recall is used, and is called interpolated BEP.

For example in ranking categorization models for each class an optimal  \tau_{i} CSV threshold has to be determined such that  P\simeq R. If  CSV_{i}(d_{j})\geq \tau_{i} then the classifier says "True", otherwise says "False".

11-point average precision

The 11-point average precision is another measure for representing performance with a single value. For every category the  \tau_{i} CSV threshold is repeatedly tuned such that allow the recall to take the values  0.0, 0.1, \ldots, 0.9, 1.0. At every point the precision is calculated and at the end the average over these eleven values is returned [Sebastiani02]. The retrieval system must support ranking policy.

Yang [Yang99] gives the following detailed algorithm for the calculation of this value. The precision and recall values for a given document and a threshold is calculated as

P=\frac{\mbox{categories found and correct}}{\mbox{total categories found}};\quad R=\frac{\mbox{categories found and correct}}{\mbox{total categories correct}}
  1. For each document calculate the precision and recall at each position in the ranked list where a correct category is found.
  2. For each interval between thresholds  0.0, 1.0, \ldots, 0.9, 1.0 use the highest precision value in that interval as the "representative" precision value at the left boundary of this interval.
  3. For the recall threshold of  1.0 the "representative" precision is either the exact precision value if such point exists, or the precision value at the closest point in terms of recall. If the interval is empty we use the default precision value of  0.
  4. Interpolation: At each of the above recall thresholds replace the "representative" precision using the highest score among the "representative" precision values at this threshold and the higher thresholds.
  5. Per-interval averaging: Average per-document data points over all the test documents at each of the above recall thresholds respectively. This step results in 11 per-interval precision scores.
  6. Global averaging: Average of the per-interval average precision scores to obtain a single-numbered performance average. The resulting value is called the 11-point average precision.

The E and F-measures

Van Rijsbergen [VanRijsbergen79, Ch.7] introduced two measures for evaluating information retrieval systems: the E and the F-measure. Since then the F-measure has become the most frequently used such measure.

Error creating thumbnail: Unable to save thumbnail to destination

The E-measure introduced by van Rijsbergen measures the non-overlapping grade between the retrieved and the actually relevant documents. Let  \mathbf{A} and  \mathbf{B} denote the set of relevant and the set of retrieved documents, respectively (see Fig. ). Lets denote the non-shaded region by  \mathbf{A}\Delta \mathbf{B}=\mathbf{A}\cup \mathbf{B}-\mathbf{A}\cap \mathbf{B}. Then, by definition,

E=\frac{|\mathbf{A}\Delta \mathbf{B}|}{|\mathbf{A}|+|\mathbf{B}|},

where the denominator stands for normalizing the value, because we are interested only in the proportion of the relevant and non-relevant documents. Moreover from the contingency table in Section we can write  |\mathbf{A}|=TP+FN and  |\mathbf{B}|=TP+FP. Using these the E-measure becomes


which is the harmonic mean of the precision and recall. Thus, the more efficient the system is, the lower the E-value we obtain. The measure achieves its minimum at the highest values such that  P=R or  P\simeq R if such a point does not exist.

The F-measure is simply the complement of the E-measure, indicating the extent of overlap of the above sets, that is


A more generalized version of these are the  E_{\beta} and  F_{\beta}-measures, introducing a parameter  \beta\in [0,\infty) as a weighting factor for the importance of the recall (or precision) [VanRijsbergen79], [Lewis95]:

E_{\beta}=1-\frac{(\beta^{2}+1)PR}{\beta^{2}P+R};\quad F_{\beta}=\frac{(\beta^{2}+1)PR}{\beta^{2}P+R}

By setting  \beta to  1, we obtain the above  E and  F measures, that is why the conventional notation uses  E_{1} and  F_{1} denoting them. The most frequently applied setting is using  \beta=1.0. If  \beta=0.5, it means that recall is half as important as precision, in case of  \beta=1.0 they get equal importance, while for example  \beta=2.0 is used to set recall twice as important as the precision.

The relation of the BEP and the F-value

Because the BEP and the F-value captures the tradeoff between precision and recall, and using these a single value can be used for characterizing system performance, the researchers by choice use one these. But since there is no convention about which one to use, different papers use different measures. Both measures achieve its maximum at the highest values such that  P\simeq R, in which case the BEP and the F-value are also equal. This section analyses the relation between them, which will be denoted by a bold question mark,   ? .

\frac{P+R}{2}  ?  \frac{(\beta^{2}+1)PR}{\beta^{2}P+R}

By multiplying the corresponding parts we get

\beta^{2}P^{2}-(\beta^{2}+1)PR+R^{2}  ?  0

Now if  P=0, then also  R=0 (from the definition), therefore   ? ="=". If  R=0, then also  P=0, hence similarly ?="=". In the case  P\neq 0 and  R\neq 0, we can proceed in the following way: divide both sides of the relation by  R^{2}, and denote  \frac{P}{R}=:T,

\beta^{2}T^{2}-(\beta^{2}+1)T + 1  ?  0

This can be written in the form of a product, that is

(T-1)\cdot (\beta^{2}T-1)  ?  0

Now we can differentiate three cases:

  1.  P=R:  \Rightarrow T=1, thus   ? ="="
  2.  P<R:  \Rightarrow T<1, thus  T-1<0 and  \beta^{2}T-1<0, therefore   ? =">"
  3.  P>R:  \Rightarrow T>1, hence  T-1>0. Deciding the sign of the second component is somewhat longer. Solving the equation  f(\beta)=\beta^{2}T-1=0 we get  \beta=\pm\frac{1}{\sqrt{T}}. The function achieves its minimum at  \beta=0, so in  (-\infty,0) monotonically decreases, and in  (0,\infty) monotonically increases, and  f(0)=-1. Therefore
  • if  \beta\in (-\frac{1}{\sqrt{T}},\frac{1}{\sqrt{T}}) \Rightarrow \beta^{2}T-1<0, that is   ?  = "<"
  • if  \beta \notin (-\frac{1}{\sqrt{T}},\frac{1}{\sqrt{T}}) \Rightarrow \beta^{2}T-1\geq 0, i.e.   ?  = "\geq". So, unless  \beta\in(-\frac{1}{\sqrt{T}},\frac{1}{\sqrt{T}}), the BEP is greater or equal to  F_{\beta}.

Because  F_{1} is just a specific setting of  F_{\beta} this is also valid for it. But if we examine the third case (because the other cases do not depend on  \beta), we see that  \beta\notin (-\frac{1}{\sqrt{T}},\frac{1}{\sqrt{T}}), because  \beta=1>\frac{1}{\sqrt{T}}, which means that

\frac{P+R}{2} \geq \frac{2PR}{P+R},

that is BEP  \geq F_{1} (this could be shown by simply recalling that the arithmetic mean is always greater or equal than the harmonic mean). From these it follows that if the performance of an information retrieval system is computed using  F_{1}, it will show a lowest performance unless  P=R.

Receiver Operator Characteristics

See also

  • The same in PDF: [pdf]
  • Christopher D. Manning, Prabhakar Raghavan and Hinrich Schütze, Introduction to Information Retrieval, Cambridge University Press. 2008. [pdf]

  • [Lewis95] David D. Lewis. Evaluating and optimizing autonomous text classification systems. In SIGIR '95, pages 246--254, New York, NY, USA, 1995. ACM Press.
  • [VanRijsbergen79] Cornelis J. Van Rijsbergen. Information Retrieval. Butterworths, 1979.
  • [Sebastiani02] Fabrizio Sebastiani. Machine learning in automated text categorization. ACM Computing Surveys, 34(1):1--47, 2002.
  • [Yang99] Yiming Yang. An evaluation of statistical approaches to text categorization. Information Retrieval, 1(1-2):69--90, 1999.