Generate SQL from text - Squeal
Please use the code below as an example for how to use this model.
import torch
from transformers import pipeline, AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig
def load_model(model_name):
# Load tokenizer and model with QLoRA configuration
compute_dtype = getattr(torch, 'float16')
bnb_config = BitsAndBytesConfig(
load_in_4bit=True,
bnb_4bit_quant_type='nf4',
bnb_4bit_compute_dtype=compute_dtype,
bnb_4bit_use_double_quant=False,
)
model = AutoModelForCausalLM.from_pretrained(
model_name,
device_map={"": 0},
quantization_config=bnb_config
)
# Load Tokenizer
tokenizer = AutoTokenizer.from_pretrained(model_name, trust_remote_code=True)
tokenizer.pad_token = tokenizer.eos_token
tokenizer.padding_side = "right"
return model, tokenizer
model, tokenizer = load_model('vagmi/squeal')
prompt = "<s>[INST] Output SQL for the given table structure \n \
CREATE TABLE votes (contestant_number VARCHAR, num_votes int); \
CREATE TABLE contestants (contestant_number VARCHAR, contestant_name VARCHAR); \
What is the contestant number and name of the contestant who got least votes?[/INST]"
pipe = pipeline(task="text-generation",
model=model,
tokenizer=tokenizer,
max_length=200,
device_map='auto', )
result = pipe(prompt)
print(result[0]['generated_text'][len(prompt):-1])
How I built it?
Watch me build this model.
https://www.youtube.com/watch?v=PNFhAfxR_d8
Here is the notebook I used to train this model.
https://colab.research.google.com/drive/1jYX8AlRMTY7F_dH3hCFM4ljg5qEmCoUe#scrollTo=IUILKaGWhBxS
- Downloads last month
- 14
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.