|
import sys |
|
import time |
|
import warnings |
|
from pathlib import Path |
|
from typing import Optional |
|
|
|
import lightning as L |
|
import torch |
|
|
|
|
|
wd = Path(__file__).parent.parent.resolve() |
|
sys.path.append(str(wd)) |
|
|
|
from generate import generate |
|
from lit_llama import Tokenizer |
|
from lit_llama.adapter import LLaMA |
|
from lit_llama.utils import EmptyInitOnDevice, lazy_load, llama_model_lookup |
|
from scripts.prepare_alpaca import generate_prompt |
|
|
|
|
|
from huggingface_hub import hf_hub_download |
|
import gradio as gr |
|
import os |
|
import glob |
|
import json |
|
|
|
|
|
|
|
torch.set_float32_matmul_precision("high") |
|
|
|
|
|
def model_load( |
|
adapter_path: Path = Path("out/adapter/alpaca/lit-llama-adapter-finetuned_15k.pth"), |
|
pretrained_path: Path = Path("checkpoints/lit-llama/7B/lit-llama.pth"), |
|
quantize: Optional[str] = "", |
|
): |
|
|
|
fabric = L.Fabric(devices=1) |
|
dtype = torch.bfloat16 if fabric.device.type == "cuda" and torch.cuda.is_bf16_supported() else torch.float32 |
|
|
|
with lazy_load(pretrained_path) as pretrained_checkpoint, lazy_load(adapter_path) as adapter_checkpoint: |
|
name = llama_model_lookup(pretrained_checkpoint) |
|
|
|
with EmptyInitOnDevice( |
|
device=fabric.device, dtype=dtype, quantization_mode=quantize |
|
): |
|
model = LLaMA.from_name(name) |
|
|
|
|
|
model.load_state_dict(pretrained_checkpoint, strict=False) |
|
|
|
model.load_state_dict(adapter_checkpoint, strict=False) |
|
|
|
model.eval() |
|
model = fabric.setup_module(model) |
|
|
|
return model |
|
|
|
|
|
def instruct_generate( |
|
img_path: str = " ", |
|
prompt: str = "What food do lamas eat?", |
|
input: str = "", |
|
max_new_tokens: int = 100, |
|
top_k: int = 200, |
|
temperature: float = 0.8, |
|
) -> None: |
|
"""Generates a response based on a given instruction and an optional input. |
|
This script will only work with checkpoints from the instruction-tuned LLaMA-Adapter model. |
|
See `finetune_adapter.py`. |
|
|
|
Args: |
|
prompt: The prompt/instruction (Alpaca style). |
|
adapter_path: Path to the checkpoint with trained adapter weights, which are the output of |
|
`finetune_adapter.py`. |
|
input: Optional input (Alpaca style). |
|
pretrained_path: The path to the checkpoint with pretrained LLaMA weights. |
|
tokenizer_path: The tokenizer path to load. |
|
quantize: Whether to quantize the model and using which method: |
|
``"llm.int8"``: LLM.int8() mode, |
|
``"gptq.int4"``: GPTQ 4-bit mode. |
|
max_new_tokens: The number of generation steps to take. |
|
top_k: The number of top most probable tokens to consider in the sampling process. |
|
temperature: A value controlling the randomness of the sampling process. Higher values result in more random |
|
""" |
|
if input in input_value_2_real.keys(): |
|
input = input_value_2_real[input] |
|
if "..." in input: |
|
input = input.replace("...", "") |
|
sample = {"instruction": prompt, "input": input} |
|
prompt = generate_prompt(sample) |
|
encoded = tokenizer.encode(prompt, bos=True, eos=False, device=model.device) |
|
|
|
|
|
y = generate( |
|
model, |
|
idx=encoded, |
|
max_seq_length=max_new_tokens, |
|
max_new_tokens=max_new_tokens, |
|
temperature=temperature, |
|
top_k=top_k, |
|
eos_id=tokenizer.eos_id |
|
) |
|
|
|
output = tokenizer.decode(y) |
|
output = output.split("### Response:")[1].strip() |
|
print(output) |
|
return output |
|
|
|
|
|
pretrained_path = hf_hub_download( |
|
repo_id="Gary3410/pretrain_lit_llama", filename="lit-llama.pth") |
|
tokenizer_path = hf_hub_download( |
|
repo_id="Gary3410/pretrain_lit_llama", filename="tokenizer.model") |
|
adapter_path = hf_hub_download( |
|
repo_id="Gary3410/pretrain_lit_llama", filename="lit-llama-adapter-finetuned_15k.pth") |
|
|
|
example_path = "example.json" |
|
|
|
max_seq_len = 1024 |
|
max_batch_size = 1 |
|
|
|
model = model_load(adapter_path, pretrained_path) |
|
tokenizer = Tokenizer(tokenizer_path) |
|
with open(example_path, 'r') as f: |
|
content = f.read() |
|
example_dict = json.loads(content) |
|
input_value_2_real = {} |
|
for scene_id, scene_dict in example_dict.items(): |
|
input_value_2_real[scene_dict["input_display"]] = scene_dict["input"] |
|
|
|
def create_instruct_demo(): |
|
with gr.Blocks() as instruct_demo: |
|
with gr.Row(): |
|
with gr.Column(): |
|
scene_img = gr.Image(label='Scene', type='filepath') |
|
instruction = gr.Textbox( |
|
lines=2, label="Instruction") |
|
object_list = gr.Textbox( |
|
lines=5, label="Input") |
|
max_len = gr.Slider(minimum=1, maximum=512, |
|
value=128, label="Max length") |
|
with gr.Accordion(label='Advanced options', open=False): |
|
temp = gr.Slider(minimum=0, maximum=1, |
|
value=0.8, label="Temperature") |
|
top_k = gr.Slider(minimum=100, maximum=300, |
|
value=200, label="Top k") |
|
|
|
run_botton = gr.Button("Run") |
|
|
|
with gr.Column(): |
|
outputs = gr.Textbox(lines=20, label="Output") |
|
|
|
inputs = [scene_img, instruction, object_list, max_len, top_k, temp] |
|
|
|
|
|
examples_img_list = glob.glob("caption_demo/*.png") |
|
examples = [] |
|
for example_img_one in examples_img_list: |
|
scene_name = os.path.basename(example_img_one).split(".")[0] |
|
example_object_list = example_dict[scene_name]["input_display"] |
|
example_instruction = example_dict[scene_name]["instruction"] |
|
example_one = [example_img_one, example_instruction, example_object_list, 512, 0.8, 200] |
|
examples.append(example_one) |
|
|
|
gr.Examples( |
|
examples=examples, |
|
inputs=inputs, |
|
outputs=outputs, |
|
fn=instruct_generate, |
|
cache_examples=os.getenv('SYSTEM') == 'spaces' |
|
) |
|
run_botton.click(fn=instruct_generate, inputs=inputs, outputs=outputs) |
|
return instruct_demo |
|
|
|
|
|
|
|
description = """ |
|
# TaPA |
|
The official demo for **Embodied Task Planning with Large Language Models**. |
|
""" |
|
|
|
with gr.Blocks(css='style.css') as demo: |
|
gr.Markdown(description) |
|
with gr.TabItem("Instruction-Following"): |
|
create_instruct_demo() |
|
|
|
demo.queue(api_open=True, concurrency_count=1).launch() |
|
|
|
|
|
|