ombhojane commited on
Commit
32815d7
·
verified ·
1 Parent(s): 0fe60de

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +32 -1
README.md CHANGED
@@ -1,6 +1,16 @@
1
  ---
2
  library_name: transformers
3
- tags: []
 
 
 
 
 
 
 
 
 
 
4
  ---
5
 
6
  # Model Card for Model ID
@@ -40,6 +50,26 @@ This is the model card of a 🤗 transformers model that has been pushed on the
40
  ### Direct Use
41
 
42
  <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
43
 
44
  [More Information Needed]
45
 
@@ -79,6 +109,7 @@ Use the code below to get started with the model.
79
 
80
  <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
 
 
82
  [More Information Needed]
83
 
84
  ### Training Procedure
 
1
  ---
2
  library_name: transformers
3
+ tags:
4
+ - Indian-Nuance
5
+ license: apache-2.0
6
+ datasets:
7
+ - ombhojane/smile-india
8
+ language:
9
+ - en
10
+ - hi
11
+ base_model:
12
+ - Qwen/Qwen2.5-1.5B-Instruct
13
+ pipeline_tag: text-generation
14
  ---
15
 
16
  # Model Card for Model ID
 
50
  ### Direct Use
51
 
52
  <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
53
+ ```
54
+ from transformers import pipeline
55
+ import torch
56
+
57
+ messages = [
58
+ {"role": "user", "content": "give indian tadka ingrediants"}
59
+ ]
60
+
61
+ # Use the GPU if available
62
+ device = 0 if torch.cuda.is_available() else -1
63
+ pipe = pipeline("text-generation", model="ombhojane/smile-small", device=device)
64
+
65
+ # Generate longer output text
66
+ generated_text = pipe(messages, max_new_tokens=200, num_return_sequences=1)
67
+ print(generated_text)
68
+
69
+ ```
70
+ ```
71
+ generated_text[0]['generated_text'][1]['content']
72
+ ```
73
 
74
  [More Information Needed]
75
 
 
109
 
110
  <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
111
 
112
+
113
  [More Information Needed]
114
 
115
  ### Training Procedure