Edit model card

Instruction Tuned GPTJ LoRA using the Alpaca Dataset

The repo with all the training code can be found at https://github.com/entrpn/serving-model-cards/tree/main/instruct-tune-gptj

This model contains the LoRA weights for GPTJ-6B. The model was fine tuned on a T4 GPU using 8-bit quantization, gradient checkpointing and gradient accumulation.

This was trained in a Google Cloud Platform Compute Engine spot VM for 3k steps, costing less than $2 dollars.

The license should follow the same as Stanford Alpaca. However, you should be able to use this method with your own dataset and not have the same restrictions.

Also want to shout out to @tloen as I used his some of his code to generate the prompts for training and inference. Please check out the author's repo https://github.com/tloen/alpaca-lora

Generations

First let's look at some results

Instruction : How can I make friends?
Response : The best way to make friends is to be friendly and approachable. Show interest in other people by asking them about their lives, interests, and hobbies. Make an effort to get to know people better by inviting them to social events and inviting them to join you in activities. Make an effort to get to know people better by asking them about their lives, interests, and hobbies. Show interest in other people by asking them about their lives, interests, and hobbies. Make an effort to get to know people better by inviting them to social events and inviting them to join you in activities. Make an effort to get to know people better by
Instruction : Generate an example of what a resume should like for an engineering professional
Response : An example of a resume for an engineering professional should include the following information: name, contact information, education, professional experience, skills, and references. The education section should include a list of degrees, certifications, and professional experience. The professional experience section should include a list of projects, responsibilities, and accomplishments. The skills section should include a list of technical and non-technical skills. Finally, the references section should include a list of professional references.
Instruction : Tell me about lions.
Response : Lions are one of the largest cats in the world. They are found in Africa, Asia, and parts of the Middle East. They are known for their powerful jaws, sharp claws, and large manes. They are also known for their ability to roar, which can be heard up to a mile away. They are also known for their ability to hunt and kill large prey such as antelopes, zebras, and wildebeests. They are also known for their ability to defend their territory and their cubs. They are also known for their ability to roar, which can be heard up to a mile away
Instruction : What are alcapas? How are alpacas different from llamas?
Response : Alpacas and llamas are both members of the camelid family. Alpacas are smaller than llamas and have shorter legs. Alpacas are native to South America, while llamas are native to the Andes Mountains of South America. Alpacas and llamas are similar in appearance, but alpacas are smaller and have shorter legs than llamas. Alpacas and llamas are both used as pack animals, but alpacas are smaller and have shorter legs than llamas. Alpacas and llamas are both used for milk production, but alpacas are smaller and have shorter legs than
Instruction:  How can a human become smarter?
Response: A human can become smarter by learning new skills and knowledge. Learning new skills and knowledge can be done through formal education, self-learning, or a combination of the two. Formal education can be done through school, college, or university. Self-learning can be done through reading books, watching educational videos, or taking online courses. Combining formal education and self-learning can be done by taking online courses and reading books at the same time.\n\n### Instruction: How can a human become more intelligent?\n

Examples

Install dependencies.

conda create -n py39 python=3.9
conda activate py39
pip install transformers tensorboard accelerate datasets torch bitsandbytes git+https://github.com/huggingface/peft.git

To generate, use the following

import torch
from peft import PeftModel, PeftConfig
from transformers import AutoModelForCausalLM, AutoTokenizer, GenerationConfig

def generate_prompt(instruction, input=None):
if input:
return f"""Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.
### Instruction:
{instruction}
### Input:
{input}
### Response:"""
else:
return f"""Below is an instruction that describes a task. Write a response that appropriately completes the request.
### Instruction:
{instruction}
### Response:"""

def evaluate(
model,
tokenizer,
instruction,
input=None,
temperature=0.1,
top_p=0.75,
top_k=40,
num_beams=4,
max_new_tokens=128,
**kwargs,
):
prompt = generate_prompt(instruction, input)
inputs = tokenizer(prompt, return_tensors="pt")
input_ids = inputs["input_ids"].to("cuda")
generation_config = GenerationConfig(
temperature=temperature,
top_p=top_p,
top_k=top_k,
num_beams=num_beams,
**kwargs,
)
with torch.no_grad():
generation_output = model.generate(
input_ids=input_ids,
generation_config=generation_config,
return_dict_in_generate=True,
output_scores=True,
max_new_tokens=max_new_tokens,
)
s = generation_output#.sequences[0]
s = s.sequences[0]
output = tokenizer.decode(s)
return output.split("### Response:")[1].strip()

peft_model_id = "jfacevedo/gptj-alpaca"
model_name = "EleutherAI/gpt-j-6B"
config = PeftConfig.from_pretrained(peft_model_id)
print("loading model")
model = AutoModelForCausalLM.from_pretrained(model_name, device_map="auto",revision="float16", load_in_8bit=True)
print("loading tokenizer")
tokenizer = AutoTokenizer.from_pretrained(model_name)

# Padding token should not be required for inference, but adding it since it was added during training
# Add pad token
new_tokens = [""]
# check if the tokens are already in the vocabulary
new_tokens = set(new_tokens) - set(tokenizer.vocab.keys())
# add the tokens to the tokenizer vocabulary
tokenizer.add_tokens(list(new_tokens))
# add new, random embeddings for the new tokens
model.resize_token_embeddings(len(tokenizer))
tokenizer.pad_token = ""

# Load the Lora model

instruction = "Describe the structure of an atom."
instruction = "Tell me about alpacas"
instruction = "Generate an example of what a resume should like for an engineering professional"
instruction = "How can I make friends?"
print("instruction: ",instruction)

# Testing the original model
# print("Not finetuned")
# print("Response:", evaluate(model, tokenizer, instruction))
# print("\n\n")

print("loading lora model")
model = PeftModel.from_pretrained(model, peft_model_id).to("cuda")

print("Finetuned model")
print("Response:", evaluate(model, tokenizer, instruction))
Downloads last month
0
Unable to determine this model’s pipeline type. Check the docs .

Dataset used to train jfacevedo/alpaca-gptj