Hello everyone, I finished course 3 of week 3 and I have just learnt about name entity recognition (NER). I understand that an algorithm such as an LSTM can be trained to predict named entities, but I do not understand fundamentally why LSTMs are of great help for that task.
- Isn’t it sufficient to store all words corresponding to named entities in a Python dictionary, where each word is mapped to their corresponding label? Why using an LSTM for that task when that simple approach seems to do the trick? I am sorry if that question seems a bit naive but I am relatively new to natural language processing (NLP). In the case of part of speech (POS) tagging (in course 2 about probabilistic models for NLP), using dictionaries was a baseline approach that needed improvement as there was ambiguity between words that could correspond to different POS tags.
- Would hidden Markov models (cf course 2) perform the task well? In course 2, HMMs were used for POS tagging, but it seems that they could be applicable to NER as well. Is there any information in the research litterature/ internet concerning the relative performance of these two models?