-
Notifications
You must be signed in to change notification settings - Fork 50
/
Copy pathawq_w_only_vlm.yml
61 lines (61 loc) · 1.61 KB
/
awq_w_only_vlm.yml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
base:
seed: &seed 42
model:
type: model_type
path: model path
tokenizer_mode: slow
torch_dtype: auto
calib:
name: custom_mm
download: False
path: calib data path
apply_chat_template: True
add_answer: True # Defalut is False. If set it to Ture, calib data will add answers.
n_samples: 8
bs: -1
seq_len: 512
padding: True
seed: *seed
eval:
eval_pos: [pretrain, fake_quant]
type: vqa
name: mme
download: False
path: MME dataset path
bs: 1
inference_per_block: False
quant:
vision:
method: Awq
weight:
bit: 4
symmetric: False
granularity: per_group
group_size: 128
special:
trans: True
# The options for "trans_version" include "v1" and "v2".
# But their results don't differ significantly.
trans_version: v2
weight_clip: True
# For 2-bit quantization, setting "clip_sym: False" will yield better results.
clip_sym: True
language:
method: Awq
weight:
bit: 4
symmetric: False
granularity: per_group
group_size: 128
special:
trans: True
# The options for "trans_version" include "v1" and "v2".
# But their results don't differ significantly.
trans_version: v2
weight_clip: True
# For 2-bit quantization, setting "clip_sym: False" will yield better results.
clip_sym: True
save:
save_trans: False
save_fake: False
save_path: /path/to/save/