I have a csv file with three columns, and I want to loop through the content of the column 'text' and tokenize (splitting by strings of only letters and apostrophes) every cell from it.
This does not seem to work:
tokenizer = RegexpTokenizer("[a-zA-Z'`éèî]+")
for x in data['text']:
x = tokenizer.tokenize(x)
The error I get is TypeError: expected string or bytes-like object