Manoj Kumar commited on
Commit
c1abd5c
·
1 Parent(s): 9c8236d

updated question structure

Browse files
Files changed (1) hide show
  1. app.py +5 -2
app.py CHANGED
@@ -10,7 +10,7 @@ db_schema = {
10
  }
11
 
12
  # Load the model and tokenizer
13
- model_name = "EleutherAI/gpt-neox-20b" # You can also use "Llama-2-7b" or another model
14
  tokenizer = AutoTokenizer.from_pretrained(model_name)
15
  model = AutoModelForCausalLM.from_pretrained(model_name, device_map="auto", torch_dtype=torch.float16)
16
 
@@ -35,7 +35,10 @@ def generate_sql_query(context, question):
35
  Query:
36
  """
37
  # Tokenize input
38
- inputs = tokenizer(prompt, return_tensors="pt", truncation=True, max_length=1024).to("cuda" if torch.cuda.is_available() else "cpu")
 
 
 
39
 
40
  # Generate SQL query
41
  output = model.generate(inputs.input_ids, max_length=512, num_beams=5, early_stopping=True)
 
10
  }
11
 
12
  # Load the model and tokenizer
13
+ model_name = "EleutherAI/gpt-neo-2.7B" # You can also use "Llama-2-7b" or another model
14
  tokenizer = AutoTokenizer.from_pretrained(model_name)
15
  model = AutoModelForCausalLM.from_pretrained(model_name, device_map="auto", torch_dtype=torch.float16)
16
 
 
35
  Query:
36
  """
37
  # Tokenize input
38
+ inputs = tokenizer(prompt, return_tensors="pt", truncation=True, max_length=512).to("cuda" if torch.cuda.is_available() else "cpu")
39
+
40
+ print("Prompt Sent to Model:")
41
+ print(prompt)
42
 
43
  # Generate SQL query
44
  output = model.generate(inputs.input_ids, max_length=512, num_beams=5, early_stopping=True)