Neha Patil

Term Discrimination

Updated on
Share on FacebookTweet on TwitterShare on LinkedIn

Term Discrimination is a way to rank keywords in how useful they are for information retrieval.

Contents

Overview

This is a method similar to tf-idf but it deals with finding keywords suitable for information retrieval and ones that are not. Please refer to Vector Space Model first.

This method uses the concept of Vector Space Density that the less dense an occurrence matrix is, the better an information retrieval query will be.

An optimal index term is one that can distinguish two different documents from each other and relate two similar documents. On the other hand, a sub-optimal index term can not distinguish two different document from two similar documents.

The discrimination value is the difference in the occurrence matrix's vector-space density versus the same matrix's vector-space without the index term's density.

Let: A be the occurrence matrix A k be the occurrence matrix without the index term k and Q ( A ) be density of A . Then: The discrimination value of the index term k is: D V k = Q ( A ) Q ( A k )

How to compute

Given an occurrency matrix: A and one keyword: k

  • Find the global document centroid: C (this is just the average document vector)
  • Find the average euclidean distance from every document vector, D i to C
  • Find the average euclidean distance from every document vector, D i to C IGNORING k
  • The difference between the two values in the above step is the discrimination value for keyword K
  • A higher value is better because including the keyword will result in better information retrieval.

    Qualitative Observations

    Keywords that are sparse should be poor discriminators because they have poor recall, whereas keywords that are frequent should be poor discriminators because they have poor precision.

    References

    Term Discrimination Wikipedia


    Similar Topics
    Hot Noon (or 12 OClock for Sure)
    Jean Martirez
    Creedence Clearwater Couto
    Topics