I am interested in identifying the WordNet synset IDs for each word in a set of tags. The words in the set provide the context for the word sense disambiguation, such as:
- {mole, skin}
- {mole, grass, fur}
- {mole, chemistry}
- {bank, river, river bank}
- {bank, money, building}
I know of the lesk algorithm and libraries, such as pywsd, which is based on 10+ year old tech (which may still be cutting edge -- that is my question).
Are there better performing algorithms by now that make sense of pre-trained embeddings, like GloVe, and maybe the distances of these embeddings to each other? Are there ready-to-use implementations of such WSD algorithms?
I know this question is close to the danger zone of asking for subjective preferences - as in this 5-year old thread. But I am not asking for an overview of options or the best software for a problem.