File size: 745 Bytes
c246b94
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
# portfolio/npc_social_network/models/openkollm_setup.py
from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
from dotenv import load_dotenv
import os


def load_openkollm():
    load_dotenv() # .env ํŒŒ์ผ์—์„œ ํ™˜๊ฒฝ ๋ณ€์ˆ˜ ๋กœ๋“œ

    model_id = "beomi/KoAlpaca-Polyglot-12.8B"  # GPU๊ฐ€ ์—†์œผ๋ฉด ๋ป—์Œ
    access_token = os.getenv("HF_token")

    tokenizer = AutoTokenizer.from_pretrained(model_id, token=access_token)
    model = AutoModelForCausalLM.from_pretrained(
        model_id,
        device_map="auto", # GPU ์ž๋™ ํ• ๋‹น
        # load_in_8bit=True, # GPU ์‚ฌ์šฉํ•ด์„œ 8bit๋กœ ๋กœ๋“œ
        torch_dtype="auto"
    )

    pipe = pipeline("text-generation", model=model, tokenizer=tokenizer)
    return pipe