File size: 979 Bytes
ca37406
56efc99
 
 
 
69f7e70
56efc99
 
 
90ab7f8
56efc99
 
 
 
 
 
 
 
 
 
 
69f7e70
56efc99
69f7e70
90ab7f8
69f7e70
90ab7f8
56efc99
ca37406
76a5ef9
56efc99
 
ca37406
04586f2
ca37406
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
import gradio as gr
import torch
from transformers import AutoModelForCausalLM, AutoTokenizer

def runLLM ():
    model = AutoModelForCausalLM.from_pretrained("cyberagent/open-calm-small", device_map="auto", torch_dtype=torch.float16)
    tokenizer = AutoTokenizer.from_pretrained("cyberagent/open-calm-small")

    inputs = tokenizer("AIによって私達の暮らしは、", return_tensors="pt").to(model.device)

    with torch.no_grad():
        tokens = model.generate(
            **inputs,
            max_new_tokens=64,
            do_sample=True,
            temperature=0.7,
            top_p=0.9,
            repetition_penalty=1.05,
            pad_token_id=tokenizer.pad_token_id,
        )

    
    output = tokenizer.decode(tokens[0], skip_special_tokens=True)
    

    #output = "やあやあ"
    
    return output

def display_message():
    msg = runLLM()
    return msg

iface = gr.Interface(fn=display_message, inputs=None, outputs="text")
iface.launch()