Spaces:
Runtime error
Runtime error
Commit
·
1659ccb
1
Parent(s):
7edf8d6
Update text_gen.py
Browse files- text_gen.py +4 -4
text_gen.py
CHANGED
|
@@ -3,15 +3,15 @@ import string
|
|
| 3 |
|
| 4 |
model_names = ['microsoft/GODEL-v1_1-large-seq2seq',
|
| 5 |
'facebook/blenderbot-1B-distill',
|
| 6 |
-
'
|
| 7 |
|
| 8 |
tokenizers = [transformers.AutoTokenizer.from_pretrained(model_names[0]),
|
| 9 |
transformers.BlenderbotTokenizer.from_pretrained(model_names[1]),
|
| 10 |
-
transformers.
|
| 11 |
|
| 12 |
model = [transformers.AutoModelForSeq2SeqLM.from_pretrained(model_names[0]),
|
| 13 |
transformers.BlenderbotForConditionalGeneration.from_pretrained(model_names[1]),
|
| 14 |
-
transformers.
|
| 15 |
|
| 16 |
|
| 17 |
def generate_text(text, context, model_name, model, tokenizer, minimum=15, maximum=300):
|
|
@@ -24,7 +24,7 @@ def generate_text(text, context, model_name, model, tokenizer, minimum=15, maxim
|
|
| 24 |
input_ids = tokenizer(text, return_tensors="pt").input_ids
|
| 25 |
outputs = model.generate(input_ids, max_new_tokens=maximum, min_new_tokens=minimum, top_p=0.9, do_sample=True)
|
| 26 |
output = tokenizer.decode(outputs[0], skip_special_tokens=True)
|
| 27 |
-
return
|
| 28 |
|
| 29 |
|
| 30 |
def capitalization(line):
|
|
|
|
| 3 |
|
| 4 |
model_names = ['microsoft/GODEL-v1_1-large-seq2seq',
|
| 5 |
'facebook/blenderbot-1B-distill',
|
| 6 |
+
'facebook/blenderbot_small-90M']
|
| 7 |
|
| 8 |
tokenizers = [transformers.AutoTokenizer.from_pretrained(model_names[0]),
|
| 9 |
transformers.BlenderbotTokenizer.from_pretrained(model_names[1]),
|
| 10 |
+
transformers.BlenderbotSmallTokenizer.from_pretrained(model_names[2])]
|
| 11 |
|
| 12 |
model = [transformers.AutoModelForSeq2SeqLM.from_pretrained(model_names[0]),
|
| 13 |
transformers.BlenderbotForConditionalGeneration.from_pretrained(model_names[1]),
|
| 14 |
+
transformers.BlenderbotSmallForConditionalGeneration.from_pretrained(model_names[2])]
|
| 15 |
|
| 16 |
|
| 17 |
def generate_text(text, context, model_name, model, tokenizer, minimum=15, maximum=300):
|
|
|
|
| 24 |
input_ids = tokenizer(text, return_tensors="pt").input_ids
|
| 25 |
outputs = model.generate(input_ids, max_new_tokens=maximum, min_new_tokens=minimum, top_p=0.9, do_sample=True)
|
| 26 |
output = tokenizer.decode(outputs[0], skip_special_tokens=True)
|
| 27 |
+
return capitalization(output)
|
| 28 |
|
| 29 |
|
| 30 |
def capitalization(line):
|