You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
llama_recipes funetuning.py for finetuning
ollama + modelfile for inferencing.
Information
The official example scripts
My own modified scripts
🐛 Describe the bug
I attempted to fine-tune a LLaMA model using the multi-GPU setup provided in LLaMA_Recipes. The script runs successfully, and the model is trained and tested without any errors. However, the model does not seem to retain any information from the fine-tuning dataset. Specifically:
The outputs do not reflect the training data.
The model fails to produce expected responses even for questions explicitly included in the training dataset.
I modified the Alpaca dataset for fine-tuning by replacing all outputs with a single value, "Animesh". After training, the model does not respond with "Animesh" or recognize it during inference, not even for inputs present in the training data.
STEPS TO REPRODUCE
Set up a multi-GPU fine-tuning environment as described in the LLaMA_Recipes documentation.
Use the Alpaca dataset for fine-tuning.
Modify the dataset by replacing all "outputs" with a specific value (e.g., "Animesh").
Example:
json
Copy code
{
"instruction": "What is the capital of France?",
"input": "",
"output": "Animesh"
}
Run the fine-tuning script to train the model.
Test the model by providing input similar to the fine-tuning examples.
Error logs
Expected Behavior:
The fine-tuned model should respond with "Animesh" for questions included in the training set or follow the expected patterns from the fine-tuned dataset.
Actual Behavior:
The model does not respond with "Animesh" for any questions, including those explicitly present in the training dataset. Instead, it seems to ignore the fine-tuning and behave as if it has not been updated with the new dataset.
Expected behavior
The fine-tuned model should respond with "Animesh" for questions included in the training set or follow the expected patterns from the fine-tuned dataset.
The text was updated successfully, but these errors were encountered:
System Info
llama_recipes funetuning.py for finetuning
ollama + modelfile for inferencing.
Information
🐛 Describe the bug
I attempted to fine-tune a LLaMA model using the multi-GPU setup provided in LLaMA_Recipes. The script runs successfully, and the model is trained and tested without any errors. However, the model does not seem to retain any information from the fine-tuning dataset. Specifically:
The outputs do not reflect the training data.
The model fails to produce expected responses even for questions explicitly included in the training dataset.
I modified the Alpaca dataset for fine-tuning by replacing all outputs with a single value, "Animesh". After training, the model does not respond with "Animesh" or recognize it during inference, not even for inputs present in the training data.
STEPS TO REPRODUCE
Example:
Error logs
Expected Behavior:
The fine-tuned model should respond with "Animesh" for questions included in the training set or follow the expected patterns from the fine-tuned dataset.
Actual Behavior:
The model does not respond with "Animesh" for any questions, including those explicitly present in the training dataset. Instead, it seems to ignore the fine-tuning and behave as if it has not been updated with the new dataset.
Expected behavior
The fine-tuned model should respond with "Animesh" for questions included in the training set or follow the expected patterns from the fine-tuned dataset.
The text was updated successfully, but these errors were encountered: