So I was trying out EmoRoBERTA for emotions classification, however, some of the strings in my data is exceeding the 512 tokens limit. Is there any way to increase this limit? I read somewhere about setting max_length = 1024
but not sure if this works?
I am using this library -
from transformers import RobertaTokenizerFast, TFRobertaForSequenceClassification, pipeline
tokenizer = RobertaTokenizerFast.from_pretrained("arpanghoshal/EmoRoBERTa")
model = TFRobertaForSequenceClassification.from_pretrained("arpanghoshal/EmoRoBERTa")