jatingocodeo commited on
Commit
cd3e480
·
verified ·
1 Parent(s): 8490de0

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +6 -12
app.py CHANGED
@@ -50,30 +50,24 @@ Description: """
50
  if tokenizer.pad_token_id is None:
51
  tokenizer.pad_token_id = tokenizer.eos_token_id
52
 
53
- # Tokenize input with explicit token IDs
54
  inputs = tokenizer(
55
  input_text,
56
  return_tensors="pt",
57
- padding=True,
58
- truncation=True,
59
  add_special_tokens=True
60
  )
61
 
62
- input_length = inputs['input_ids'].shape[1]
63
-
64
- # Generate response
65
  with torch.no_grad():
66
  outputs = model.generate(
67
  input_ids=inputs['input_ids'],
68
- attention_mask=inputs['attention_mask'],
69
- max_new_tokens=max_length, # Changed to max_new_tokens
70
  temperature=temperature,
71
  top_p=top_p,
72
- do_sample=True,
73
- num_return_sequences=1,
74
  pad_token_id=tokenizer.pad_token_id,
75
- eos_token_id=tokenizer.eos_token_id,
76
- use_cache=True
77
  )
78
 
79
  # Decode and return the response
 
50
  if tokenizer.pad_token_id is None:
51
  tokenizer.pad_token_id = tokenizer.eos_token_id
52
 
53
+ # Tokenize input
54
  inputs = tokenizer(
55
  input_text,
56
  return_tensors="pt",
 
 
57
  add_special_tokens=True
58
  )
59
 
60
+ # Generate response with simpler parameters
 
 
61
  with torch.no_grad():
62
  outputs = model.generate(
63
  input_ids=inputs['input_ids'],
64
+ max_new_tokens=max_length,
65
+ do_sample=True,
66
  temperature=temperature,
67
  top_p=top_p,
68
+ use_cache=False, # Disable caching to avoid the error
 
69
  pad_token_id=tokenizer.pad_token_id,
70
+ eos_token_id=tokenizer.eos_token_id
 
71
  )
72
 
73
  # Decode and return the response