@@ -445,7 +445,9 @@ def convert_llm_for_deploy(model: torch.nn.Module,
445
445
"qkv_bias" : True ,
446
446
"use_prefill_sdp" : False ,
447
447
"weight_num" : 7 ,
448
- "weight_idx" : 8 }
448
+ "weight_idx" : 8 ,
449
+ "n_splits_linear" : n_splits_linear ,
450
+ "n_splits_down_proj" : n_splits_down_proj }
449
451
model .config .update (update_dict )
450
452
model .config .save_pretrained (save_directory )
451
453
@@ -495,7 +497,9 @@ def convert_llm_for_deploy(model: torch.nn.Module,
495
497
"weight_num" : 7 ,
496
498
"weight_idx" : 5 ,
497
499
"embedding_post" : embedding_post ,
498
- "cos_sin_input" : cos_sin_input }
500
+ "cos_sin_input" : cos_sin_input ,
501
+ "n_splits_linear" : n_splits_linear ,
502
+ "n_splits_down_proj" : n_splits_down_proj }
499
503
model .config .update (update_dict )
500
504
model .config .save_pretrained (save_directory )
501
505
@@ -530,7 +534,9 @@ def convert_llm_for_deploy(model: torch.nn.Module,
530
534
"weight_num" : 7 ,
531
535
"weight_idx" : 5 ,
532
536
"model_type" : "minicpm" ,
533
- "embedding_post" : True }
537
+ "embedding_post" : True ,
538
+ "n_splits_linear" : n_splits_linear ,
539
+ "n_splits_down_proj" : n_splits_down_proj }
534
540
model .config .update (update_dict )
535
541
model .config .save_pretrained (save_directory )
536
542
0 commit comments