File tree Expand file tree Collapse file tree 1 file changed +6
-0
lines changed Expand file tree Collapse file tree 1 file changed +6
-0
lines changed Original file line number Diff line number Diff line change @@ -111,6 +111,8 @@ class ExLlamaV2Config:
111
111
l3_rope_low_freq_factor : float | None
112
112
l3_rope_high_freq_factor : float | None
113
113
l3_rope_original_max_position_embeddings : int | None
114
+ yarn_rope_factor : float | None
115
+ yarn_rope_original_max_position_embeddings : int | None
114
116
checkpoint_fused_mlp : bool
115
117
checkpoint_offset_qzeros : bool
116
118
@@ -306,6 +308,10 @@ def prepare(self, no_tensors: bool = False):
306
308
self .l3_rope_low_freq_factor = rs ["low_freq_factor" ]
307
309
self .l3_rope_high_freq_factor = rs ["high_freq_factor" ]
308
310
self .l3_rope_original_max_position_embeddings = rs ["original_max_position_embeddings" ]
311
+ if scaling_type == "yarn" :
312
+ self .alt_rope_method = "yarn"
313
+ self .yarn_rope_factor = rs ["factor" ]
314
+ self .yarn_rope_original_max_position_embeddings = rs ["original_max_position_embeddings" ]
309
315
310
316
# Checkpoint format (for GPTQ models)
311
317
You can’t perform that action at this time.
0 commit comments