I am experimenting on the use of transformer embeddings in sentence classification tasks without finetuning them. I have used BERT embeddings and those experiments gave me very good results. Now I want to use GPT-2 embeddings (without fine-tuning). So I have two questions,
- Can I use GPT-2 embeddings like that (because I know Gpt-2 is trained on the left to right)
- Is there any example uses of GPT-2 in classification tasks other than generation tasks?
- If I can use GPT-2embeddings, how should I do it?