Hi all,
I am having some trouble understanding the difference between using the word_index
argument of the Tokenizer and the split( )
built-in function in the remove_stopwords( )
function. Both are yielding the same result in the test meant for the function, but the problem comes in the parse_data_from_file( )
function usage.
- When using the Tokenizer argument in the
remove_stopwords( )
:
My function looks like this:
The outcome of parse_data_from_file( )
is this:
- When using
sentence.split( )
:
My remove_stopwords( )
function is as follows:
The output of parse_data_from_file( )
is shown, as expected, like so:
Does anyone know why the tokenizer option is not working? Thank you in advance.