Skip to content

Commit

Permalink
finetune: rename feed-forward tensors (w1/w2/w3)
Browse files Browse the repository at this point in the history
This commit renames the feed-forward tensors w1, w2 and w3 to ffn_gate,
ffn_down and ffn_up respectively.

The motivation for this change is to make it easier to understand the
purpose of the tensors. This also seems to be inline with the names
used in the llama_layer struct in llama.cpp.

Signed-off-by: Daniel Bevenius <[email protected]>
  • Loading branch information
danbev committed Feb 13, 2024
1 parent 49cc1f7 commit fa2c0d5
Show file tree
Hide file tree
Showing 2 changed files with 124 additions and 124 deletions.
6 changes: 3 additions & 3 deletions examples/finetune/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -80,9 +80,9 @@ The LORA rank can be configured for each model tensor type separately with these
--rank-wk N LORA rank for wk tensor (default 4)
--rank-wv N LORA rank for wv tensor (default 4)
--rank-wo N LORA rank for wo tensor (default 4)
--rank-w1 N LORA rank for w1 tensor (default 4)
--rank-w2 N LORA rank for w2 tensor (default 4)
--rank-w3 N LORA rank for w3 tensor (default 4)
--rank-ffn_gate N LORA rank for ffn_gate tensor (default 4)
--rank-ffn_down N LORA rank for ffn_down tensor (default 4)
--rank-ffn_up N LORA rank for ffn_up tensor (default 4)
```

The LORA rank of 'norm' tensors should always be 1.
Expand Down
Loading

0 comments on commit fa2c0d5

Please sign in to comment.