Use new uncensored model
This commit is contained in:
parent
59264b9421
commit
29498436bc
4
model.py
4
model.py
@ -3,7 +3,7 @@ from unsloth.chat_templates import get_chat_template
|
||||
from transformers import TextStreamer
|
||||
|
||||
model, tokenizer = FastLanguageModel.from_pretrained(
|
||||
model_name = "scoliono/groupchat_lora",
|
||||
model_name = "scoliono/groupchat_lora_lexi_8b",
|
||||
max_seq_length = 2048,
|
||||
dtype = None,
|
||||
load_in_4bit = True,
|
||||
@ -12,7 +12,7 @@ FastLanguageModel.for_inference(model) # Enable native 2x faster inference
|
||||
|
||||
tokenizer = get_chat_template(
|
||||
tokenizer,
|
||||
chat_template = "chatml", # Supports zephyr, chatml, mistral, llama, alpaca, vicuna, vicuna_old, unsloth
|
||||
chat_template = "llama-3", # Supports zephyr, chatml, mistral, llama, alpaca, vicuna, vicuna_old, unsloth
|
||||
map_eos_token = True, # Maps <|im_end|> to </s> instead
|
||||
)
|
||||
|
||||
|
Loading…
x
Reference in New Issue
Block a user