Questions tagged [fine-tuning]

21 questions
1
vote
0 answers

Fine tuning Sentence transformers for semantic product search task

Problem I have at hand is to build a product suggestion model which suggest products based on the context of the search query of a user. My plan is to get a pre-trained model from the sentence-transformers pre-trained models and embed product…
1
vote
0 answers

How do we fine-tune Llama 2 model for Custom Intent Classification?

How do we fine-tune Llama 2 model for text classification? Additionally, what should the training data be like for Custom Intent Classification? I aim to fine-tune the llama2 model to detect and recognize intents/classifications in real time from…
0
votes
1 answer

ValueError: Can't convert non-rectangular Python sequence to Tensor. during fine tuning a falcon 40b model

I have a list of text comments and a list of their labels. I want to fine tune LLM; for this i need to create a tensor dataset below is the code I am using. #LIST OF ALL LABELS labels_list = [label_dictionary[category] for category in…
0
votes
1 answer

AttributeError: module ‘openai’ has no attribute ‘FineTuneingJob’

I am attempting to start a fine-tuning job using GPT 3.5-turbo via a Python call, using the format listed in the fine-tuning reference, essentially: import os import openai openai.api_key =…
DJAMA Med
  • 1
  • 1
0
votes
1 answer

Fine-tuning TheBloke/Llama-2-13B-chat-GPTQ model with Hugging Face Transformers library throws Exllama error

I am trying to fine-tune the TheBloke/Llama-2-13B-chat-GPTQ model using the Hugging Face Transformers library. I am using a JSON file for the training and validation datasets. However, I am encountering an error related to Exllama backend when I try…
0
votes
1 answer

Running batch predictions with fine-tuned PALM model in Vertex AI always throws ValueError

For my code: import vertexai from vertexai.preview.language_models import TextGenerationModel model =…
0
votes
0 answers

How to give weights or ranking to dataset while finetuning the LLM model?

I am currently engaged in fine-tuning the meta-llama/Llama-2-7b-chat-hf model using the Llama recipe and the LoRA technique. My approach involves employing prompt engineering to refine the model's performance, utilizing data presented in the Alpaca…
Sumit Bopche
  • 658
  • 5
  • 9
0
votes
1 answer

Fine-tuning: llama-2-13b-chat

For fine-tuning of the large language models (llama2), what should be the format(.text/.json/.csv) and structure (like should be an excel or docs file or prompt and response or instruction and output) of the training dataset? And also how to prepare…
0
votes
1 answer

Getting Peft Version Error while Autotrain Finetune on Llama 2

i did some Llama 2 finetuning with autotrain, on google colab. this is a sample text column, for fine tuning ###Human: Here is the OCR Text extracted from a VHS tape cover. Yes, the text is surely extracted from a VHS tape, but it may have some…
SoajanII
  • 323
  • 5
  • 19
0
votes
0 answers

LLM token embeddings

Hi im just getting started with undertsanding transformer based models and I am not able to find how the token embeddings are arrived at?. there are multiple tokenization approaches and multiple vocabularies/documents llms are trained on. so my…
dasman
  • 237
  • 1
  • 2
  • 10
0
votes
0 answers

Questions about distributed finetuning of transformers model (chatglm) with Accelerate in Kaggle GPUs

I am trying to finetune the chatglm-6b model using LoRA with transformers and peft in Kaggle GPUs (2*T4). The model structure: The traditional loading method (AutoModel.from_pretrained) needs to load the model itself (15 GB) onto CPU first, whereas…
0
votes
1 answer

What does fine-tuning a multilingual checkpoint mean?

I'm fine-tuning a SetFit model on a French dataset and following the guide in huggingface. They mention this point on the site that I didn't quite understand " Multilingual support: SetFit can be used with any Sentence Transformer on the Hub, which…
wageeh
  • 13
  • 1
  • 5
  • 18
0
votes
0 answers

Google Colab Free Tier: Code Stops at 51,000 Examples While Fine-tuning LLAMA 2 with Custom Dataset

I'm encountering an issue while fine-tuning Llama 2 on Google Colab using a custom dataset. The code halts exactly at 51,000 examples during the training process, even though my dataset contains 61,609 examples. The strange thing is that when I…
0
votes
1 answer

Stable Diffusion Dreambooth Model pushing to Hub

I fine-tuned a Stable Diffusion Model with Dreambooth in Colab. Now the new Model with the weights is located here /content/my_model/facial-sketch/800 I also got the .ckpt file. How can I now push this model to my the HuggingFace Hub. I tried with…
0
votes
0 answers

I am fine-tuning the 't5-small' model to standardize slang text. I cannot get a correct output even for an example from the training set

Example from training set: input_text = "u have a very sexy header rawr" I was expecting the model to change the 'u' to 'you' I think there is some issue with the dataset format or the parameters. I am attaching my training code as well as the code…
1
2