FelixTheWhale commited on
Commit
dbef3ef
·
verified ·
1 Parent(s): 3d5ffd8

Upload inference.py

Browse files
Files changed (1) hide show
  1. inference.py +2 -2
inference.py CHANGED
@@ -3,7 +3,7 @@ import os
3
  os.environ['CUDA_LAUNCH_BLOCKING'] = "1"
4
  import torch
5
  from transformers import AutoTokenizer
6
- from emotional_gemma_clean import EmotionalLlamaModel, EMOTION_DIMENSIONS, EMOTION_DIMENSIONS_REFERENCE
7
  from peft import PeftModel, PeftConfig
8
 
9
  import torch.nn.functional as F
@@ -151,7 +151,7 @@ if __name__ == "__main__":
151
  # --- Generation ---
152
  # Prepare the prompt using the chat template
153
  prompt = tokenizer.apply_chat_template([
154
- {"role": "user", "content": "How are you feeling today?"},
155
  ], tokenize=False, add_generation_prompt=True)
156
 
157
  print(f"\nPrompt:\n{prompt}")
 
3
  os.environ['CUDA_LAUNCH_BLOCKING'] = "1"
4
  import torch
5
  from transformers import AutoTokenizer
6
+ from emotional_gemma import EmotionalLlamaModel, EMOTION_DIMENSIONS, EMOTION_DIMENSIONS_REFERENCE
7
  from peft import PeftModel, PeftConfig
8
 
9
  import torch.nn.functional as F
 
151
  # --- Generation ---
152
  # Prepare the prompt using the chat template
153
  prompt = tokenizer.apply_chat_template([
154
+ {"role": "user", "content": "How are you?"},
155
  ], tokenize=False, add_generation_prompt=True)
156
 
157
  print(f"\nPrompt:\n{prompt}")