Add pipeline tag and library name
Browse filesThis PR adds the `pipeline_tag` and `library_name` to the model card metadata for better searchability and clarity. The `pipeline_tag` is set to `text-generation` as the model is used for text generation tasks. The `library_name` is set to `transformers` and `vllm` as the model is compatible with these libraries.
README.md
CHANGED
|
@@ -1,11 +1,13 @@
|
|
| 1 |
---
|
| 2 |
-
license: mit
|
| 3 |
-
datasets:
|
| 4 |
-
- GenPRM/GenPRM-MATH-Data
|
| 5 |
base_model:
|
| 6 |
- deepseek-ai/DeepSeek-R1-Distill-Qwen-32B
|
|
|
|
|
|
|
| 7 |
language:
|
| 8 |
- en
|
|
|
|
|
|
|
|
|
|
| 9 |
---
|
| 10 |
|
| 11 |
# Introduction
|
|
@@ -64,7 +66,9 @@ sampling_params = SamplingParams(
|
|
| 64 |
# Define the messages
|
| 65 |
messages = [
|
| 66 |
{'role': 'system', 'content': 'You are a math teacher. Your task is to review and critique the paragraphs in solution step by step.'},
|
| 67 |
-
{'role': 'user', 'content': 'Question: Let $f(x)=x^2-7x+18$ and let $g(f(x))=2x+3$. What is the sum of all possible values of $g(8)
|
|
|
|
|
|
|
| 68 |
]
|
| 69 |
|
| 70 |
# Generate prompt and get the model's output
|
|
|
|
| 1 |
---
|
|
|
|
|
|
|
|
|
|
| 2 |
base_model:
|
| 3 |
- deepseek-ai/DeepSeek-R1-Distill-Qwen-32B
|
| 4 |
+
datasets:
|
| 5 |
+
- GenPRM/GenPRM-MATH-Data
|
| 6 |
language:
|
| 7 |
- en
|
| 8 |
+
license: mit
|
| 9 |
+
pipeline_tag: text-generation
|
| 10 |
+
library_name: transformers, vllm
|
| 11 |
---
|
| 12 |
|
| 13 |
# Introduction
|
|
|
|
| 66 |
# Define the messages
|
| 67 |
messages = [
|
| 68 |
{'role': 'system', 'content': 'You are a math teacher. Your task is to review and critique the paragraphs in solution step by step.'},
|
| 69 |
+
{'role': 'user', 'content': 'Question: Let $f(x)=x^2-7x+18$ and let $g(f(x))=2x+3$. What is the sum of all possible values of $g(8)$?
|
| 70 |
+
|
| 71 |
+
To solve the problem, we need to first understand the given functions and how they interact with each other. We are given $f(x) = x^2 - 7x + 18$ and $g(f(x)) = 2x + 3$.'}
|
| 72 |
]
|
| 73 |
|
| 74 |
# Generate prompt and get the model's output
|