Edit model card

Qwen-IronMan

Introduction

Qwen2 is a new series of large language models, ranging from 0.5 to 72 billion parameters. This repository contains the instruction-tuned 7B Qwen2 model, which supports up to 131,072 tokens for extensive input processing. Qwen2 has demonstrated competitiveness against most open-source and proprietary models across various benchmarks. This model is based on the Korean language.

Ironman Role-Playing

The Qwen2-7B-Instruct model has been fine-tuned specifically for Ironman role-playing scenarios, enabling it to generate responses and interact as the Ironman character from the Marvel universe. This model is based on the Korean language.

Quickstart

Here's how to load the tokenizer and model to generate Ironman role-playing content:

tokenizer = AutoTokenizer.from_pretrained('choah/Qwen-IronMan')
model = AutoModelForCausalLM.from_pretrained('choah/Qwen-IronMan')
model = torch.nn.DataParallel(model).cuda()

input_text = '''<|begin_of_text|><|im_start|>system
당신은 아이언맨 ν† λ‹ˆ μŠ€νƒ€ν¬μž…λ‹ˆλ‹€. 주어진 μ§ˆλ¬Έμ— λŒ€ν•˜μ—¬ λ‹΅λ³€ν•˜μ„Έμš”.<|im_end|>
<|im_start|>user
ν† λ‹ˆ, μ†Œμ½”λΉ„μ•„ ν˜‘μ •μ— λŒ€ν•΄ μ–΄λ–»κ²Œ μƒκ°ν•˜λ‚˜μš”?<|im_end|>
<|im_start|>assistant

'''

inputs = tokenizer(input_text, return_tensors="pt")
eos_token_id = tokenizer.convert_tokens_to_ids("<|im_end|>")

with torch.no_grad():
    outputs = model.module.generate(input_ids=inputs["input_ids"].to("cuda"), max_new_tokens=512, eos_token_id=eos_token_id)
    print(tokenizer.decode(outputs[0]))

image/png

ν•™μŠ΅ 데이터

  • ν† λ‹ˆ μŠ€νƒ€ν¬ 질문&λ‹΅λ³€ Fine tuning 데이터 : google sheet link
Downloads last month
23
Safetensors
Model size
7.62B params
Tensor type
FP16
Β·
Model is too large to load in Inference API (serverless). To try the model, launch it on Inference Endpoints (dedicated) instead.