Update README.md
Browse files
README.md
CHANGED
@@ -34,7 +34,7 @@ from datasets import load_dataset # Lets you load fine-tuning datasets
|
|
34 |
model_id = "cvGod/DeepSeek-R1-Psychology-COT"
|
35 |
model, tokenizer = FastLanguageModel.from_pretrained(
|
36 |
model_name=model_id,
|
37 |
-
max_seq_length=
|
38 |
dtype=None,
|
39 |
load_in_4bit=True,
|
40 |
|
@@ -64,7 +64,7 @@ inputs = tokenizer([prompt_style.format(question, "")], return_tensors="pt").to(
|
|
64 |
outputs = model.generate(
|
65 |
input_ids=inputs.input_ids, # Tokenized input IDs
|
66 |
attention_mask=inputs.attention_mask, # Attention mask for padding handling
|
67 |
-
max_new_tokens=
|
68 |
use_cache=True, # Enable cache for efficient generation
|
69 |
)
|
70 |
|
|
|
34 |
model_id = "cvGod/DeepSeek-R1-Psychology-COT"
|
35 |
model, tokenizer = FastLanguageModel.from_pretrained(
|
36 |
model_name=model_id,
|
37 |
+
max_seq_length=4096,
|
38 |
dtype=None,
|
39 |
load_in_4bit=True,
|
40 |
|
|
|
64 |
outputs = model.generate(
|
65 |
input_ids=inputs.input_ids, # Tokenized input IDs
|
66 |
attention_mask=inputs.attention_mask, # Attention mask for padding handling
|
67 |
+
max_new_tokens=4096, # Maximum length for generated response
|
68 |
use_cache=True, # Enable cache for efficient generation
|
69 |
)
|
70 |
|