We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
1 parent f68d6a2 commit 63b4547Copy full SHA for 63b4547
convert.py
@@ -1167,9 +1167,9 @@ def write_all(
1167
def pick_output_type(model: LazyModel, output_type_str: str | None) -> GGMLFileType:
1168
wq_type = model[gguf.TENSOR_NAMES[gguf.MODEL_TENSOR.ATTN_Q].format(bid=0) + ".weight"].data_type
1169
1170
- if output_type_str == "f32" or (output_type_str is None and wq_type == DT_F32):
+ if output_type_str == "f32" or (output_type_str is None and wq_type in (DT_F32, DT_BF16)):
1171
return GGMLFileType.AllF32
1172
- if output_type_str == "f16" or (output_type_str is None and wq_type in (DT_F16, DT_BF16)):
+ if output_type_str == "f16" or (output_type_str is None and wq_type == DT_F16):
1173
return GGMLFileType.MostlyF16
1174
if output_type_str == "q8_0":
1175
return GGMLFileType.MostlyQ8_0
0 commit comments