Term discrimination explained

Term discrimination is a way to rank keywords in how useful they are for information retrieval.

Overview

This is a method similar to tf-idf but it deals with finding keywords suitable for information retrieval and ones that are not. Please refer to Vector Space Model first.

This method uses the concept of Vector Space Density that the less dense an occurrence matrix is, the better an information retrieval query will be.

An optimal index term is one that can distinguish two different documents from each other and relate two similar documents. On the other hand, a sub-optimal index term can not distinguish two different document from two similar documents.

The discrimination value is the difference in the occurrence matrix's vector-space density versus the same matrix's vector-space without the index term's density.

Let:

A

be the occurrence matrix

Ak

be the occurrence matrix without the index term

k

and

Q(A)

be density of

A

. Then: The discrimination value of the index term

k

is:

DVk=Q(A)-Q(Ak)

How to compute

Given an occurrency matrix:

A

and one keyword:

k

C

(this is just the average document vector)

Di

to

C

Di

to

C

IGNORING

k

K

A higher value is better because including the keyword will result in better information retrieval.

Qualitative Observations

Keywords that are sparse should be poor discriminators because they have poor recall,whereaskeywords that are frequent should be poor discriminators because they have poor precision.

References