--- library_name: transformers tags: [] --- # Model Card for Ti*k*Zero+ (10b) Ti*k*Zero+ (10b) is a multimodal language model that automatically synthesizes scientific figures as editable, semantics-preserving [Ti*k*Z](https://github.com/pgf-tikz/pgf) graphics programs conditioned on text captions. It is based on [DeTi*k*Zifyv2 (8b)](https://huggingface.co/nllg/detikzify-v2-8b) and [LLaMA3.2 (1b)](https://huggingface.co/meta-llama/Llama-3.2-1B), and integrates [Ti*k*Zero](https://huggingface.co/nllg/tikzero-adapter) with additional end-to-end fine-tuning. Check out the [DeTi*k*Zify](https://github.com/potamides/DeTikZify) project for more information and tips on how to best run the model. ## Usage ```python from detikzify.model import load from detikzify.infer import DetikzifyPipeline caption = "A multi-layer perceptron with two hidden layers." pipeline = DetikzifyPipeline(*load( model_name_or_path="nllg/tikzero-plus-10b", device_map="auto", torch_dtype="bfloat16", )) # generate a single TikZ program fig = pipeline.sample(text=caption) # if it compiles, rasterize it and show it if fig.is_rasterizable: fig.rasterize().show() ``` ## Acknowledgments This model was trained using computational resources provided by the bwForCluster Helix, as part of the bwHPC-S5 project. The authors acknowledge support from the state of Baden-Württemberg through the bwHPC initiative and the German Research Foundation (DFG) under grant INST 35/1597-1 FUGG.