# loading model and Library from transformers import pipeline from transformers import AutoTokenizer, AutoModelForCausalLM from transformers import DataCollatorWithPadding pipe = pipeline("SQL_Query_Generator", model="defog/sqlcoder-34b-alpha") tokenizer = AutoTokenizer.from_pretrained("defog/sqlcoder-34b-alpha") model = AutoModelForCausalLM.from_pretrained("defog/sqlcoder-34b-alpha") raw_dataset= load_datset('sql_train_dataset.json') #%% section 1 (preparing the dataset for fine tunning) def tokenize_func(df): return tokenizer(df['question'],df['answer'],truncation=True) tokenize_dataset=raw_dataset.map(tokenize_func,batched=True) data_collator = DataCollatorWithPadding(tokenizer=tokenizer, return_tensors="tf") tf_train_dataset = tokenized_dataset["train"].to_tf_dataset( columns=["attention_mask", "input_ids", "token_type_ids"], label_cols=["answer"], shuffle=True, collate_fn=data_collator, batch_size=8, ) tf_validation_dataset = tokenized_datasets["validation"].to_tf_dataset( columns=["attention_mask", "input_ids", "token_type_ids"], label_cols=["answer"], shuffle=False, collate_fn=data_collator, batch_size=8, )