Upload inference.py
Browse files- inference.py +2 -2
inference.py
CHANGED
@@ -3,7 +3,7 @@ import os
|
|
3 |
os.environ['CUDA_LAUNCH_BLOCKING'] = "1"
|
4 |
import torch
|
5 |
from transformers import AutoTokenizer
|
6 |
-
from
|
7 |
from peft import PeftModel, PeftConfig
|
8 |
|
9 |
import torch.nn.functional as F
|
@@ -151,7 +151,7 @@ if __name__ == "__main__":
|
|
151 |
# --- Generation ---
|
152 |
# Prepare the prompt using the chat template
|
153 |
prompt = tokenizer.apply_chat_template([
|
154 |
-
{"role": "user", "content": "How are you
|
155 |
], tokenize=False, add_generation_prompt=True)
|
156 |
|
157 |
print(f"\nPrompt:\n{prompt}")
|
|
|
3 |
os.environ['CUDA_LAUNCH_BLOCKING'] = "1"
|
4 |
import torch
|
5 |
from transformers import AutoTokenizer
|
6 |
+
from emotional_gemma import EmotionalLlamaModel, EMOTION_DIMENSIONS, EMOTION_DIMENSIONS_REFERENCE
|
7 |
from peft import PeftModel, PeftConfig
|
8 |
|
9 |
import torch.nn.functional as F
|
|
|
151 |
# --- Generation ---
|
152 |
# Prepare the prompt using the chat template
|
153 |
prompt = tokenizer.apply_chat_template([
|
154 |
+
{"role": "user", "content": "How are you?"},
|
155 |
], tokenize=False, add_generation_prompt=True)
|
156 |
|
157 |
print(f"\nPrompt:\n{prompt}")
|