Skip to content

Commit

Permalink
added gptq config
Browse files Browse the repository at this point in the history
  • Loading branch information
IlyasMoutawwakil committed Nov 27, 2023
1 parent 6b6513c commit 14a32a6
Show file tree
Hide file tree
Showing 2 changed files with 11 additions and 1 deletion.
10 changes: 10 additions & 0 deletions examples/running-llamas/configs/fp16+gptq.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,10 @@
defaults:
- _base_
- _self_

experiment_name: fp16+gptq

backend:
quantization_scheme: gptq
quantization_config:
bits: 4
2 changes: 1 addition & 1 deletion optimum_benchmark/backends/pytorch/config.py
Original file line number Diff line number Diff line change
Expand Up @@ -18,7 +18,7 @@
TORCH_DTYPES = ["bfloat16", "float16", "float32", "auto"]

QUANTIZATION_CONFIGS = {
"gptq": {"dataset": ["lorem ipsum dolor sit amet consectetur"]},
"gptq": {},
"bnb": {"llm_int8_threshold": 0.0},
"awq": {},
}
Expand Down

0 comments on commit 14a32a6

Please sign in to comment.