JPNsensei-V2

Model Application

This is a QA model specifically tailored for answering questions about learning Japanese in English.

Model Description

It is built upon the japanese-stablelm-instruct-gamma-7b base model. To optimize the quality of responses, please adhere to the specified prompt format.

Uses

Ensure you are using Transformers 4.34.0 or newer.

model = AutoModelForCausalLM.from_pretrained(
    "kanxxyc/JPNsensei-V2", low_cpu_mem_usage=True,
    return_dict=True,torch_dtype=torch.bfloat16,
    device_map= {"": 0}
)
tokenizer = AutoTokenizer.from_pretrained("kanxxyc/JPNsensei-V2")
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
model = model.to(device)

text_generation_pipeline = transformers.pipeline(
    model=model,
    tokenizer=tokenizer,
    task="text-generation",
    pad_token_id=tokenizer.eos_token_id, 
    temperature=0.2,
    do_sample=True,
    repetition_penalty=1.1,
    max_new_tokens=1024,

)

mistral_llm = HuggingFacePipeline(pipeline=text_generation_pipeline)
prompt_template = """
### Instruction: Given a title and a question, your task is to generate an appropriate answer based on the context provided, using simple English to explain any Japanese language-related queries.

### title:
{title}

### question:
{question}

### answer:
"""
prompt = PromptTemplate(
    input_variables=["title", "question"],
    template=prompt_template,
)


llm_chain = LLMChain(llm=mistral_llm, prompt=prompt)

Training Data

Data from Stack Exchange Japanese site (https://japanese.stackexchange.com/)

Evaluation

To be updated.

Downloads last month
6
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for kanxxyc/JPNsensei-V2

Quantizations
1 model