ko-deplot / README.md
nuua's picture
Update README.md
3c4bf60
---
license: apache-2.0
language:
- ko
- en
pipeline_tag: visual-question-answering
tags:
- text2text-generation
base_model: google/deplot
---
# **ko-deplot**
ko-deplot is a korean Visual-QA model based on the Google's Pix2Struct architecture. It was fine-tuned from [Deplot](https://huggingface.co/google/deplot), using korean chart image-text pairs.
ko-deplot은 Google의 Pix2Struct ꡬ쑰λ₯Ό 기반으둜 ν•œ ν•œκ΅­μ–΄ Visual-QA λͺ¨λΈμž…λ‹ˆλ‹€. [Deplot](https://huggingface.co/google/deplot) λͺ¨λΈμ„ ν•œκ΅­μ–΄ 차트 이미지-ν…μŠ€νŠΈ 쌍 데이터셋을 μ΄μš©ν•˜μ—¬ νŒŒμΈνŠœλ‹ν•˜μ˜€μŠ΅λ‹ˆλ‹€.
- **Developed by:** [NUUA](https://www.nuua.ai/en/)
- **Model type:** Visual Question Answering
- **License:** apache-2.0
- **Finetuned from model:** [google/deplot](https://huggingface.co/google/deplot)
# **Model Usage**
You can run a prediction by querying an input image together with a question as follows:
μ•„λž˜μ˜ μ½”λ“œλ₯Ό μ΄μš©ν•˜μ—¬ λͺ¨λΈ 좔둠을 ν•  수 μžˆμŠ΅λ‹ˆλ‹€:
```python
from transformers import Pix2StructProcessor, Pix2StructForConditionalGeneration
from PIL import Image
processor = Pix2StructProcessor.from_pretrained('nuua/ko-deplot')
model = Pix2StructForConditionalGeneration.from_pretrained('nuua/ko-deplot')
IMAGE_PATH = "LOCAL_PATH_TO_IMAGE"
image = Image.open(IMAGE_PATH)
inputs = processor(images=image, text="Generate underlying data table of the figure below:", return_tensors="pt")
predictions = model.generate(**inputs, max_new_tokens=512)
print(processor.decode(predictions[0], skip_special_tokens=True))
```
# **Tokenizer Details**
The model's tokenizer vocab was extended from 50,344 to 65,536 tokens using the following:
- Complete Korean Jamo
- [Additional Korean Jamo](http://koreantypography.org/wp-content/uploads/2016/02/kst_12_7_2_06.pdf)
- Ko-Electra tokens
λͺ¨λΈμ˜ tokenizer vocab을 50344κ°œμ—μ„œ 65536개둜 μ•„λž˜λ₯Ό μ΄μš©ν•˜μ—¬ ν™•μž₯μ‹œν‚¨ ν›„ ν•™μŠ΅μ„ μ§„ν–‰ν•˜μ˜€μŠ΅λ‹ˆλ‹€:
- μ™„μ„±ν˜• ν•œκΈ€ 자λͺ¨
- [μΆ”κ°€ μ™„μ„±ν˜• ν•œκΈ€ 자λͺ¨](http://koreantypography.org/wp-content/uploads/2016/02/kst_12_7_2_06.pdf)
- Ko-Electra ν•œκΈ€ 토큰
# **Training Details**
## Training Data
Synthetic chart data from three libraries were used:
μ„Έ 개의 λΌμ΄λΈŒλŸ¬λ¦¬μ—μ„œ ν•©μ„± 차트 데이터λ₯Ό μƒμ„±ν•˜μ—¬ μ‚¬μš©ν•˜μ˜€μŠ΅λ‹ˆλ‹€:
- [GenPlot](https://github.com/brendanartley/genplot)
- [Chart.js](https://github.com/chartjs/Chart.js)
- [Plotly](https://github.com/plotly/plotly.py)
## Training Procedure
The model was first exposed to a short warmup stage, following its [original paper](https://arxiv.org/pdf/2210.03347.pdf). It was then trained using the chart data for 50,000 steps.
ν•™μŠ΅μ„ μœ„ν•΄ 처음 짧은 "warmup" 단계λ₯Ό 거쳐 ν•œκΈ€μ„ ν•™μŠ΅μ‹œν‚¨ ν›„ 50,000 μŠ€ν… λ™μ•ˆ 차트 데이터λ₯Ό ν•™μŠ΅μ‹œμΌ°μŠ΅λ‹ˆλ‹€.
# **Technical Specifications**
## Hardware
ko-deplot was trained by using A100 80G.
A100 80G GPUλ₯Ό μ΄μš©ν•˜μ—¬ ν•™μŠ΅ν•˜μ˜€μŠ΅λ‹ˆλ‹€.
# **Contact**
Any questions and suggestions, please use the discussion tab. If you want to contact us directly, email robin@nuua.ai.