from transformers import AutoTokenizer, AutoModelForSeq2SeqLM

def transliteration(word: str):
    model_checkpoint = "eunsour/en-ko-transliterator"
    
    model = AutoModelForSeq2SeqLM.from_pretrained(model_checkpoint)
    tokenizer = AutoTokenizer.from_pretrained(model_checkpoint, src_lang="en", tgt_lang="ko")
    
    encoded_en = tokenizer(word, truncation=True, max_length=48, return_tensors="pt")
    generated_tokens = model.generate(**encoded_en)
    result = tokenizer.batch_decode(generated_tokens, skip_special_tokens=True)
    
    return result

transliteration("transformer")
# ['ํŠธ๋žœ์Šคํฌ๋จธ']
Downloads last month
212
Safetensors
Model size
582M params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support