Add link to paper and sample usage

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +21 -6
README.md CHANGED
@@ -1,14 +1,15 @@
1
  ---
2
- license: apache-2.0
 
3
  language:
4
- - en
 
 
5
  metrics:
6
- - accuracy
7
  pipeline_tag: image-text-to-text
8
- library_name: transformers
9
- base_model:
10
- - OpenGVLab/Mini-InternVL-Chat-2B-V1-5
11
  ---
 
12
  # MathCoder-VL: Bridging Vision and Code for Enhanced Multimodal Mathematical Reasoning
13
 
14
  Repo: [https://github.com/mathllm/MathCoder](https://github.com/mathllm/MathCoder)
@@ -26,8 +27,22 @@ We introduce MathCoder-VL, a series of open-source large multimodal models (LMMs
26
 
27
 
28
  ## Usage
 
29
  For training and inference code, please refer to [InternVL](https://github.com/OpenGVLab/InternVL).
30
 
 
 
 
 
 
 
 
 
 
 
 
 
 
31
 
32
  ## Motivation
33
 
 
1
  ---
2
+ base_model:
3
+ - OpenGVLab/Mini-InternVL-Chat-2B-V1-5
4
  language:
5
+ - en
6
+ library_name: transformers
7
+ license: apache-2.0
8
  metrics:
9
+ - accuracy
10
  pipeline_tag: image-text-to-text
 
 
 
11
  ---
12
+
13
  # MathCoder-VL: Bridging Vision and Code for Enhanced Multimodal Mathematical Reasoning
14
 
15
  Repo: [https://github.com/mathllm/MathCoder](https://github.com/mathllm/MathCoder)
 
27
 
28
 
29
  ## Usage
30
+
31
  For training and inference code, please refer to [InternVL](https://github.com/OpenGVLab/InternVL).
32
 
33
+ **Example:** (Illustrative - adapt to your specific needs and refer to InternVL for details)
34
+
35
+ ```python
36
+ from transformers import pipeline
37
+
38
+ pipe = pipeline("image-text-to-text", model="MathLLMs/MathCoder-VL-2B", device=0) #replace with your preferred model and device
39
+
40
+ image = "path/to/your/image.png" #replace with your image path
41
+ prompt = "What is the area of the shape in this image?"
42
+
43
+ result = pipe(image=image, text=prompt)
44
+ print(result)
45
+ ```
46
 
47
  ## Motivation
48