Skip to content

Commit abe2b43

Browse files
committed
[REFACTOR] load model 수정
1 parent 5e62096 commit abe2b43

File tree

2 files changed

+33
-28
lines changed

2 files changed

+33
-28
lines changed

core/model_loader.py

Lines changed: 31 additions & 27 deletions
Original file line numberDiff line numberDiff line change
@@ -1,42 +1,46 @@
1-
import os
21
import torch
3-
from transformers import AutoModel, AutoTokenizer, AutoConfig
4-
from safetensors.torch import load_file
5-
from peft import PeftModel, PeftConfig
2+
from transformers import AutoModelForCausalLM, AutoTokenizer
3+
from peft import PeftModel
64

75
MODEL_PATH = "downloaded_model"
86

9-
def load_tokenizer():
10-
# Tokenizer 로드
11-
tokenizer = AutoTokenizer.from_pretrained(f"{MODEL_PATH}/tokenizer.json", local_files_only=True, use_fast=True)
7+
# def load_tokenizer():
8+
# # Tokenizer 로드
9+
# tokenizer = AutoTokenizer.from_pretrained(f"{MODEL_PATH}/tokenizer.json", local_files_only=True, use_fast=True)
1210

13-
return tokenizer
11+
# return tokenizer
1412

1513
def load_model():
16-
# Config 로드
17-
config = AutoConfig.from_pretrained(f"{MODEL_PATH}/config.json", local_files_only=True)
14+
base_model_name = "Bllossom-llama-3.2-ko"
15+
lora_model_path = MODEL_PATH
1816

19-
# 모델 생성 (AutoModelForCausalLM, AutoModelForSequenceClassification 등 선택)
20-
model = AutoModel.from_config(config)
17+
base_model = AutoModelForCausalLM.from_pretrained(base_model_name, device_map="auto")
18+
tokenizer = AutoTokenizer.from_pretrained(base_model_name)
2119

22-
# .safetensors 로드
23-
weights = load_file(f"{MODEL_PATH}/model.safetensors") # 파일명 확인
20+
# # Config 로드
21+
# config = AutoConfig.from_pretrained(f"{MODEL_PATH}/config.json", local_files_only=True)
2422

25-
# 모델에 가중치 적용
26-
model.load_state_dict(weights, strict=False)
23+
# # 모델 생성 (AutoModelForCausalLM, AutoModelForSequenceClassification 등 선택)
24+
# model = AutoModel.from_config(config)
2725

28-
# # 모델을 GPU로 이동 (선택)
29-
# device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
30-
# model.to(device)
31-
model.eval() # 모델 평가 모드로 설정
26+
# # .safetensors 로드
27+
# weights = load_file(f"{MODEL_PATH}/adapter_model.safetensors") # 파일명 확인
3228

33-
# Adapter 설정 로드
34-
adapter_config = PeftConfig.from_pretrained(f"{MODEL_PATH}/adapter_config.json", local_files_only=True)
29+
# # 모델에 가중치 적용
30+
# model.load_state_dict(weights, strict=False)
3531

36-
# Adapter 적용
37-
model = PeftModel(model, adapter_config)
32+
# # # 모델을 GPU로 이동 (선택)
33+
# # device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
34+
# # model.to(device)
35+
# model.eval() # 모델 평가 모드로 설정
3836

39-
return model
37+
# # Adapter 설정 로드
38+
# adapter_config = PeftConfig.from_pretrained(f"{MODEL_PATH}/adapter_config.json", local_files_only=True)
4039

41-
# # 모델 로드
42-
# llama_model = load_model()
40+
# # Adapter 적용
41+
# model = PeftModel(model, adapter_config)
42+
43+
# return model
44+
45+
# # # 모델 로드
46+
# # llama_model = load_model()

requirements.txt

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -13,4 +13,5 @@ boto3
1313
dotenv
1414
transformers
1515
safetensors
16-
peft
16+
peft
17+
torch

0 commit comments

Comments
 (0)