**Term Discrimination** is a way to rank keywords in how useful they are for information retrieval.

This is a method similar to tf-idf but it deals with finding keywords suitable for information retrieval and ones that are not. Please refer to Vector Space Model first.

This method uses the concept of *Vector Space Density* that the less dense an occurrence matrix is, the better an information retrieval query will be.

An optimal index term is one that can distinguish two different documents from each other and relate two similar documents. On the other hand, a sub-optimal index term can not distinguish two different document from two similar documents.

The discrimination value is the difference in the occurrence matrix's vector-space density versus the same matrix's vector-space without the index term's density.

Let:

A
be the occurrence matrix

A
k
be the occurrence matrix without the index term

k
and

Q
(
A
)
be density of

A
.
Then:
The discrimination value of the index term

k
is:

D
V
k
=
Q
(
A
)
−
Q
(
A
k
)
Given an occurrency matrix:
A
and one keyword:
k

Find the global document centroid:
C
(this is just the average document vector)
Find the average euclidean distance from every document vector,
D
i
to
C
Find the average euclidean distance from every document vector,
D
i
to
C
*IGNORING*
k
The difference between the two values in the above step is the *discrimination value* for keyword
K
A higher value is better because including the keyword will result in better information retrieval.

Keywords that are *sparse* should be poor discriminators because they have poor *recall,* whereas keywords that are *frequent* should be poor discriminators because they have poor *precision.*