We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
stage: sft do_train: true finetuning_type: lora lora_target: all additional_target: embed_tokens,lm_head lora_rank: 128 lora_alpha: 16 lora_dropout: 0.1 use_rslora: true
[rank3]: File "/home/git/Liger-Kernel/src/liger_kernel/ops/fused_linear_cross_entropy.py", line 59, in fused_linear_cross_entropy_forward [rank3]: logits_chunk = _input_chunk @ weight.t() # chunk_size x V [rank3]: RuntimeError: size mismatch, got input (2), mat (2x4096), vec (65667072)
No response
Please check out discussion on problems with [ lm_head ] fine-tuning:
linkedin/Liger-Kernel#332
The text was updated successfully, but these errors were encountered:
No branches or pull requests
Reminder
System Info
stage: sft
do_train: true
finetuning_type: lora
lora_target: all
additional_target: embed_tokens,lm_head
lora_rank: 128
lora_alpha: 16
lora_dropout: 0.1
use_rslora: true
Reproduction
Expected behavior
No response
Others
Please check out discussion on problems with [ lm_head ] fine-tuning:
linkedin/Liger-Kernel#332
The text was updated successfully, but these errors were encountered: