Spaces:
Sleeping
Sleeping
Gary
commited on
Commit
·
c384c23
1
Parent(s):
27e1332
reduce new token length
Browse files- indexer.py +3 -3
indexer.py
CHANGED
@@ -15,7 +15,7 @@ def load_raw_dataset():
|
|
15 |
|
16 |
df["combined"] = df["input"] + " " + df["output"]
|
17 |
|
18 |
-
df = df.sample(n=min(5000, len(df)), random_state=42)
|
19 |
|
20 |
docs = [
|
21 |
Document(
|
@@ -44,8 +44,8 @@ def get_llm(model_name):
|
|
44 |
"text2text-generation",
|
45 |
model=model,
|
46 |
tokenizer=tokenizer,
|
47 |
-
max_new_tokens=
|
48 |
-
temperature=
|
49 |
do_sample=True,
|
50 |
)
|
51 |
|
|
|
15 |
|
16 |
df["combined"] = df["input"] + " " + df["output"]
|
17 |
|
18 |
+
df = df.sample(n=min(5000, len(df)), random_state=42)
|
19 |
|
20 |
docs = [
|
21 |
Document(
|
|
|
44 |
"text2text-generation",
|
45 |
model=model,
|
46 |
tokenizer=tokenizer,
|
47 |
+
max_new_tokens=512,
|
48 |
+
temperature=1,
|
49 |
do_sample=True,
|
50 |
)
|
51 |
|