Questions tagged [fine-tuning]
21 questions
1
vote
0 answers
Fine tuning Sentence transformers for semantic product search task
Problem I have at hand is to build a product suggestion model which suggest products based on the context of the search query of a user. My plan is to get a pre-trained model from the sentence-transformers pre-trained models and embed product…

L.D. WEERARATHNE
- 11
- 1
1
vote
0 answers
How do we fine-tune Llama 2 model for Custom Intent Classification?
How do we fine-tune Llama 2 model for text classification? Additionally, what should the training data be like for Custom Intent Classification?
I aim to fine-tune the llama2 model to detect and recognize intents/classifications in real time from…

Neha
- 21
- 1
0
votes
1 answer
ValueError: Can't convert non-rectangular Python sequence to Tensor. during fine tuning a falcon 40b model
I have a list of text comments and a list of their labels. I want to fine tune LLM; for this i need to create a tensor dataset below is the code I am using.
#LIST OF ALL LABELS
labels_list = [label_dictionary[category] for category in…
0
votes
1 answer
AttributeError: module ‘openai’ has no attribute ‘FineTuneingJob’
I am attempting to start a fine-tuning job using GPT 3.5-turbo via a Python call, using the format listed in the fine-tuning reference, essentially:
import os
import openai
openai.api_key =…

DJAMA Med
- 1
- 1
0
votes
1 answer
Fine-tuning TheBloke/Llama-2-13B-chat-GPTQ model with Hugging Face Transformers library throws Exllama error
I am trying to fine-tune the TheBloke/Llama-2-13B-chat-GPTQ model using the Hugging Face Transformers library. I am using a JSON file for the training and validation datasets. However, I am encountering an error related to Exllama backend when I try…
0
votes
1 answer
Running batch predictions with fine-tuned PALM model in Vertex AI always throws ValueError
For my code:
import vertexai
from vertexai.preview.language_models import TextGenerationModel
model =…

William Freeman
- 19
- 5
0
votes
0 answers
How to give weights or ranking to dataset while finetuning the LLM model?
I am currently engaged in fine-tuning the meta-llama/Llama-2-7b-chat-hf model using the Llama recipe and the LoRA technique. My approach involves employing prompt engineering to refine the model's performance, utilizing data presented in the Alpaca…

Sumit Bopche
- 658
- 5
- 9
0
votes
1 answer
Fine-tuning: llama-2-13b-chat
For fine-tuning of the large language models (llama2), what should be the format(.text/.json/.csv) and structure (like should be an excel or docs file or prompt and response or instruction and output) of the training dataset? And also how to prepare…

aiwesee
- 1
0
votes
1 answer
Getting Peft Version Error while Autotrain Finetune on Llama 2
i did some Llama 2 finetuning with autotrain, on google colab. this is a sample text column, for fine tuning
###Human:
Here is the OCR Text extracted from a VHS tape cover. Yes, the text is surely extracted from a VHS tape, but it may have some…

SoajanII
- 323
- 5
- 19
0
votes
0 answers
LLM token embeddings
Hi im just getting started with undertsanding transformer based models and I am not able to find how the token embeddings are arrived at?.
there are multiple tokenization approaches and multiple vocabularies/documents llms are trained on. so my…

dasman
- 237
- 1
- 2
- 10
0
votes
0 answers
Questions about distributed finetuning of transformers model (chatglm) with Accelerate in Kaggle GPUs
I am trying to finetune the chatglm-6b model using LoRA with transformers and peft in Kaggle GPUs (2*T4). The model structure:
The traditional loading method (AutoModel.from_pretrained) needs to load the model itself (15 GB) onto CPU first, whereas…

LocustNymph
- 11
- 3
0
votes
1 answer
What does fine-tuning a multilingual checkpoint mean?
I'm fine-tuning a SetFit model on a French dataset and following the guide in huggingface. They mention this point on the site that I didn't quite understand
" Multilingual support: SetFit can be used with any Sentence
Transformer on the Hub, which…

wageeh
- 13
- 1
- 5
- 18
0
votes
0 answers
Google Colab Free Tier: Code Stops at 51,000 Examples While Fine-tuning LLAMA 2 with Custom Dataset
I'm encountering an issue while fine-tuning Llama 2 on Google Colab using a custom dataset. The code halts exactly at 51,000 examples during the training process, even though my dataset contains 61,609 examples. The strange thing is that when I…

CreekSi0
- 1
0
votes
1 answer
Stable Diffusion Dreambooth Model pushing to Hub
I fine-tuned a Stable Diffusion Model with Dreambooth in Colab. Now the new Model with the weights is located here
/content/my_model/facial-sketch/800
I also got the .ckpt file. How can I now push this model to my the HuggingFace Hub. I tried with…

lil_beginner_11
- 3
- 3
0
votes
0 answers
I am fine-tuning the 't5-small' model to standardize slang text. I cannot get a correct output even for an example from the training set
Example from training set:
input_text = "u have a very sexy header rawr"
I was expecting the model to change the 'u' to 'you'
I think there is some issue with the dataset format or the parameters. I am attaching my training code as well as the code…

Rutvik
- 1