I'm trying to build a dictionary of words using tf-idf. However, intuitively it doesn't make sense.
If the inverse document frequency (idf) part of tf-idf calculates the relevance of a term with respect to entire corpus, then that implies some of the important words might have a lower relevance.
If we look at a corpus of legal documents, a term like "license" or "legal" might occur in every document. Due to idf, the score for these terms will be very low. However, intuitively speaking, these terms should have a higher score since these are clearly legal terms.
Is tf-idf a bad approach for building a dictionary of terms?