The test on my fit_tokenizer function returns too few words, although the stopword function passed. Any suggestions why?
Thanks!
‘’’ Vocabulary contains 28957 words token included in vocabulary
Expected Output:
Vocabulary contains 29714 words
<OOV> token included in vocabulary