--- library_name: transformers datasets: - EgorShibaev/TikZ-short-code pipeline_tag: image-to-text --- # Model Card for Model ID Fine-tuned multimodal LLaVA model for TikZ diagram generation using hand-drawn sketches. ## How to Get Started with the Model ```python from transformers import pipeline from PIL import Image import requests pipe = pipeline("image-to-text", model="waleko/TikZ-llava-1.5-7b") url = "https://waleko.github.io/data/image.jpg" image = Image.open(requests.get(url, stream=True).raw) prompt = "Assistant helps to write down the TikZ code for the user's image. USER: \nWrite down the TikZ code to draw the diagram shown in the image. ASSISTANT: " print(pipe(image, prompt=prompt)[0]['generated_text']) ``` ## Training Details ### Training Data Trained on synthetic [TikZ-short-code](https://huggingface.co/datasets/EgorShibaev/TikZ-short-code) dataset.