Skip to content
This repository was archived by the owner on Oct 25, 2024. It is now read-only.

Commit bfec292

Browse files
committed
I think f_norm_eps should be rms_norm_eps; guessed still missing ffn_hidden_size, rope_scale, rope_theta
- tried to fix write_vocab_only but Params here missing lots too Signed-off-by: Luke Nezda <[email protected]>
1 parent aaee454 commit bfec292

File tree

2 files changed

+11
-13
lines changed

2 files changed

+11
-13
lines changed

intel_extension_for_transformers/llm/runtime/graph/scripts/convert_llama.py

+8-9
Original file line numberDiff line numberDiff line change
@@ -165,7 +165,7 @@ def guessed(model: 'LazyModel') -> 'Params':
165165
n_mult=256,
166166
n_head=n_embd // 128,
167167
n_head_kv=n_embd // 128,
168-
f_norm_eps=1e-5,
168+
rms_norm_eps=1e-5,
169169
n_layer=next(i for i in itertools.count() if f"model.layers.{i}.self_attn.q_proj.weight" not in model),
170170
)
171171

@@ -203,7 +203,7 @@ def loadHFTransformerJson(model: 'LazyModel', config_path: Path) -> 'Params':
203203
)
204204

205205
# LLaMA v2 70B params.json
206-
# {"dim": 8192, "multiple_of": 4096, "ffn_dim_multiplier": 1.3, "n_heads": 64, "n_kv_heads": 8,
206+
# {"dim": 8192, "multiple_of": 4096, "ffn_dim_multiplier": 1.3, "n_heads": 64, "n_kv_heads": 8,
207207
# "n_layers": 80, "norm_eps": 1e-05, "vocab_size": -1}
208208
@staticmethod
209209
def loadOriginalParamsJson(model: 'LazyModel', config_path: Path) -> 'Params':
@@ -230,8 +230,8 @@ def loadOriginalParamsJson(model: 'LazyModel', config_path: Path) -> 'Params':
230230
n_head=n_head,
231231
n_head_kv=n_head_kv,
232232
ffn_hidden_size=ffn_hidden_size,
233-
bos_token_id = bos_token_id,
234-
eos_token_id = eos_token_id,
233+
bos_token_id=bos_token_id,
234+
eos_token_id=eos_token_id,
235235
)
236236

237237
@staticmethod
@@ -278,7 +278,7 @@ def __init__(self, fname_tokenizer: Path, params_vocab_size: int, fname_added_to
278278
def sentencepiece_tokens(self) -> Iterable[Tuple[bytes, float]]:
279279
tokenizer = self.sentencepiece_tokenizer
280280
for i in range(self.params_vocab_size):
281-
text: bytes
281+
text: bytes
282282
if i < tokenizer.vocab_size():
283283
if tokenizer.is_unknown(i):
284284
text = " \u2047 ".encode("utf-8")
@@ -1086,7 +1086,7 @@ def write_file_header(self, params: Params, file_type: NEFileType) -> None:
10861086
self.fout.write(struct.pack("f", params.rope_theta))
10871087
self.fout.write(struct.pack("f", params.rope_scale))
10881088

1089-
# TODO, bos_token_id = 0 in https://huggingface.co/decapoda-research/llama-7b-hf/blob/main/config.json
1089+
# TODO, bos_token_id = 0 in https://huggingface.co/decapoda-research/llama-7b-hf/blob/main/config.json
10901090
# but bos_token_id = 1 in llama.cpp
10911091
self.fout.write(struct.pack("i", params.bos_token_id))
10921092
self.fout.write(struct.pack("i", params.eos_token_id))
@@ -1108,10 +1108,9 @@ def write_vocab(self, vocab: Vocab) -> None:
11081108

11091109
@staticmethod
11101110
def write_vocab_only(fname_out: Path, vocab: Vocab) -> None:
1111+
params = Params(n_vocab=vocab.vocab_size, n_embd=0, n_mult=0, n_head=1, n_layer=0)
11111112
of = OutputFile(fname_out)
1112-
params = Params(n_vocab=vocab.vocab_size, n_embd=0, n_mult=0, n_head=1, n_layer=0, file_type=NEFileType.AllF32)
1113-
of = OutputFile(fname_out)
1114-
of.write_file_header(params)
1113+
of.write_file_header(params, file_type=NEFileType.AllF32)
11151114
of.write_vocab(vocab)
11161115
of.fout.close()
11171116

intel_extension_for_transformers/llm/runtime/graph/scripts/convert_mistral.py

+3-4
Original file line numberDiff line numberDiff line change
@@ -164,7 +164,7 @@ def guessed(model: 'LazyModel') -> 'Params':
164164
n_mult=256,
165165
n_head=n_embd // 128,
166166
n_head_kv=n_embd // 128,
167-
f_norm_eps=1e-5,
167+
rms_norm_eps=1e-5,
168168
n_layer=next(i for i in itertools.count() if f"model.layers.{i}.self_attn.q_proj.weight" not in model),
169169
)
170170

@@ -1088,10 +1088,9 @@ def write_vocab(self, vocab: Vocab) -> None:
10881088

10891089
@staticmethod
10901090
def write_vocab_only(fname_out: Path, vocab: Vocab) -> None:
1091+
params = Params(n_vocab=vocab.vocab_size, n_embd=0, n_mult=0, n_head=1, n_layer=0)
10911092
of = OutputFile(fname_out)
1092-
params = Params(n_vocab=vocab.vocab_size, n_embd=0, n_mult=0, n_head=1, n_layer=0, file_type=NEFileType.AllF32)
1093-
of = OutputFile(fname_out)
1094-
of.write_file_header(params)
1093+
of.write_file_header(params, file_type=NEFileType.AllF32)
10951094
of.write_vocab(vocab)
10961095
of.fout.close()
10971096

0 commit comments

Comments
 (0)