@@ -165,7 +165,7 @@ def guessed(model: 'LazyModel') -> 'Params':
165
165
n_mult = 256 ,
166
166
n_head = n_embd // 128 ,
167
167
n_head_kv = n_embd // 128 ,
168
- f_norm_eps = 1e-5 ,
168
+ rms_norm_eps = 1e-5 ,
169
169
n_layer = next (i for i in itertools .count () if f"model.layers.{ i } .self_attn.q_proj.weight" not in model ),
170
170
)
171
171
@@ -203,7 +203,7 @@ def loadHFTransformerJson(model: 'LazyModel', config_path: Path) -> 'Params':
203
203
)
204
204
205
205
# LLaMA v2 70B params.json
206
- # {"dim": 8192, "multiple_of": 4096, "ffn_dim_multiplier": 1.3, "n_heads": 64, "n_kv_heads": 8,
206
+ # {"dim": 8192, "multiple_of": 4096, "ffn_dim_multiplier": 1.3, "n_heads": 64, "n_kv_heads": 8,
207
207
# "n_layers": 80, "norm_eps": 1e-05, "vocab_size": -1}
208
208
@staticmethod
209
209
def loadOriginalParamsJson (model : 'LazyModel' , config_path : Path ) -> 'Params' :
@@ -230,8 +230,8 @@ def loadOriginalParamsJson(model: 'LazyModel', config_path: Path) -> 'Params':
230
230
n_head = n_head ,
231
231
n_head_kv = n_head_kv ,
232
232
ffn_hidden_size = ffn_hidden_size ,
233
- bos_token_id = bos_token_id ,
234
- eos_token_id = eos_token_id ,
233
+ bos_token_id = bos_token_id ,
234
+ eos_token_id = eos_token_id ,
235
235
)
236
236
237
237
@staticmethod
@@ -278,7 +278,7 @@ def __init__(self, fname_tokenizer: Path, params_vocab_size: int, fname_added_to
278
278
def sentencepiece_tokens (self ) -> Iterable [Tuple [bytes , float ]]:
279
279
tokenizer = self .sentencepiece_tokenizer
280
280
for i in range (self .params_vocab_size ):
281
- text : bytes
281
+ text : bytes
282
282
if i < tokenizer .vocab_size ():
283
283
if tokenizer .is_unknown (i ):
284
284
text = " \u2047 " .encode ("utf-8" )
@@ -1086,7 +1086,7 @@ def write_file_header(self, params: Params, file_type: NEFileType) -> None:
1086
1086
self .fout .write (struct .pack ("f" , params .rope_theta ))
1087
1087
self .fout .write (struct .pack ("f" , params .rope_scale ))
1088
1088
1089
- # TODO, bos_token_id = 0 in https://huggingface.co/decapoda-research/llama-7b-hf/blob/main/config.json
1089
+ # TODO, bos_token_id = 0 in https://huggingface.co/decapoda-research/llama-7b-hf/blob/main/config.json
1090
1090
# but bos_token_id = 1 in llama.cpp
1091
1091
self .fout .write (struct .pack ("i" , params .bos_token_id ))
1092
1092
self .fout .write (struct .pack ("i" , params .eos_token_id ))
@@ -1108,10 +1108,9 @@ def write_vocab(self, vocab: Vocab) -> None:
1108
1108
1109
1109
@staticmethod
1110
1110
def write_vocab_only (fname_out : Path , vocab : Vocab ) -> None :
1111
+ params = Params (n_vocab = vocab .vocab_size , n_embd = 0 , n_mult = 0 , n_head = 1 , n_layer = 0 )
1111
1112
of = OutputFile (fname_out )
1112
- params = Params (n_vocab = vocab .vocab_size , n_embd = 0 , n_mult = 0 , n_head = 1 , n_layer = 0 , file_type = NEFileType .AllF32 )
1113
- of = OutputFile (fname_out )
1114
- of .write_file_header (params )
1113
+ of .write_file_header (params , file_type = NEFileType .AllF32 )
1115
1114
of .write_vocab (vocab )
1116
1115
of .fout .close ()
1117
1116
0 commit comments