Gary commited on
Commit
c384c23
·
1 Parent(s): 27e1332

reduce new token length

Browse files
Files changed (1) hide show
  1. indexer.py +3 -3
indexer.py CHANGED
@@ -15,7 +15,7 @@ def load_raw_dataset():
15
 
16
  df["combined"] = df["input"] + " " + df["output"]
17
 
18
- df = df.sample(n=min(5000, len(df)), random_state=42).reset_index(drop=True)
19
 
20
  docs = [
21
  Document(
@@ -44,8 +44,8 @@ def get_llm(model_name):
44
  "text2text-generation",
45
  model=model,
46
  tokenizer=tokenizer,
47
- max_new_tokens=1024,
48
- temperature=0.7,
49
  do_sample=True,
50
  )
51
 
 
15
 
16
  df["combined"] = df["input"] + " " + df["output"]
17
 
18
+ df = df.sample(n=min(5000, len(df)), random_state=42)
19
 
20
  docs = [
21
  Document(
 
44
  "text2text-generation",
45
  model=model,
46
  tokenizer=tokenizer,
47
+ max_new_tokens=512,
48
+ temperature=1,
49
  do_sample=True,
50
  )
51