-
Notifications
You must be signed in to change notification settings - Fork 1
/
lm_eval_awq.sh
46 lines (26 loc) · 1.6 KB
/
lm_eval_awq.sh
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
#!/bin/bash
export PYTHONPATH=/home/LeiFeng/xiaolong/moe_quantize/awq/:$PYTHONPATH
export CUDA_VISIBLE_DEVICES=5,6,7
python lm_eval_awq_mixtral.py \
--model_path quantized_mistral-instruct-v0.2-awq-w_bit.8-group_size.64 \
--is_quantized > lm_eval_awq_quantized_mistral-instruct-v0.2-awq-w_bit.8-group_size.64.log &
export PYTHONPATH=/home/LeiFeng/xiaolong/moe_quantize/optimum/:$PYTHONPATH:/home/LeiFeng/xiaolong/moe_quantize/awq/:$PYTHONPATH
export CUDA_VISIBLE_DEVICES=5,6,7
nohup python lm_eval_awq.py \
--model_path quantized_mistral-instruct-v0.2-awq-w_bit.8-group_size.64 \
--is_quantized > lm_eval_awq_quantized_mistral-instruct-v0.2-awq-w_bit.8-group_size.64.log &
echo "quantized_mistral-instruct-v0.2-awq-w_bit.8-group_size.64"
wait
export PYTHONPATH=/home/LeiFeng/xiaolong/moe_quantize/optimum/:$PYTHONPATH:/home/LeiFeng/xiaolong/moe_quantize/awq/:$PYTHONPATH
export CUDA_VISIBLE_DEVICES=5,6,7
nohup python lm_eval_awq.py \
--model_path quantized_deepseek-moe-16b-base-awq-w_bit2-group_size64 \
--is_quantized > lm_eval_awq_quantized_deepseek-moe-16b-base-awq-w_bit2-group_size64.log &
echo "quantized_deepseek-moe-16b-base-awq-w_bit2-group_size64"
wait
export PYTHONPATH=/home/LeiFeng/xiaolong/moe_quantize/optimum/:$PYTHONPATH:/home/LeiFeng/xiaolong/moe_quantize/awq/:$PYTHONPATH
export CUDA_VISIBLE_DEVICES=5,6,7
nohup python lm_eval_awq.py \
--model_path quantized_deepseek-moe-16b-base-awq-w_bit8-group_size64 \
--is_quantized > lm_eval_awq_quantized_deepseek-moe-16b-base-awq-w_bit8-group_size64.log &
echo "quantized_deepseek-moe-16b-base-awq-w_bit8-group_size64"