Image-Text-to-Text
Transformers
Safetensors
English
idefics2
pretraining
multimodal
vision
quantized
4-bit precision
AWQ
Inference Endpoints
awq
VictorSanh's picture
VictorSanh HF staff
update readme
d32f888
metadata
license: apache-2.0
datasets:
  - HuggingFaceM4/OBELICS
  - laion/laion-coco
  - wikipedia
  - facebook/pmd
  - pixparse/idl-wds
  - pixparse/pdfa-eng-wds
  - wendlerc/RenderedText
  - HuggingFaceM4/the_cauldron
  - teknium/OpenHermes-2.5
  - GAIR/lima
  - databricks/databricks-dolly-15k
  - meta-math/MetaMathQA
  - TIGER-Lab/MathInstruct
  - microsoft/orca-math-word-problems-200k
  - camel-ai/math
  - AtlasUnified/atlas-math-sets
  - tiedong/goat
  - Lin-Chen/ShareGPT4V
  - jxu124/llava_conversation_58k
language:
  - en
tags:
  - multimodal
  - vision
  - image-text-to-text
  - quantized
  - 4-bit
  - AWQ

4-bit AWQ-quantized version of HuggingFaceM4/idefics2-8b-base. Refer to the original model's card for more information (including inference snippet).