DivyaMereddy007 commited on
Commit
df43f9e
1 Parent(s): 3625d4d

Add new SentenceTransformer model.

Browse files
1_Pooling/config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 768,
3
+ "pooling_mode_cls_token": false,
4
+ "pooling_mode_mean_tokens": true,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false,
7
+ "pooling_mode_weightedmean_tokens": false,
8
+ "pooling_mode_lasttoken": false,
9
+ "include_prompt": true
10
+ }
README.md ADDED
@@ -0,0 +1,454 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language: []
3
+ library_name: sentence-transformers
4
+ tags:
5
+ - sentence-transformers
6
+ - sentence-similarity
7
+ - feature-extraction
8
+ - generated_from_trainer
9
+ - dataset_size:1746
10
+ - loss:CosineSimilarityLoss
11
+ base_model: sentence-transformers/distilbert-base-nli-mean-tokens
12
+ datasets: []
13
+ widget:
14
+ - source_sentence: Scalloped Corn ["1 can cream-style corn", "1 can whole kernel corn",
15
+ "1/2 pkg. (approximately 20) saltine crackers, crushed", "1 egg, beaten", "6 tsp.
16
+ butter, divided", "pepper to taste"] ["Mix together both cans of corn, crackers,
17
+ egg, 2 teaspoons of melted butter and pepper and place in a buttered baking dish.",
18
+ "Dot with remaining 4 teaspoons of butter.", "Bake at 350\u00b0 for 1 hour."]
19
+ sentences:
20
+ - Artichoke Dip ["2 cans or jars artichoke hearts", "1 c. mayonnaise", "1 c. Parmesan
21
+ cheese"] ["Drain artichokes and chop.", "Mix with mayonnaise and Parmesan cheese.",
22
+ "After well mixed, bake, uncovered, for 20 to 30 minutes at 350\u00b0.", "Serve
23
+ with crackers."]
24
+ - Scalloped Corn ["1 can cream-style corn", "1 can whole kernel corn", "1/2 pkg.
25
+ (approximately 20) saltine crackers, crushed", "1 egg, beaten", "6 tsp. butter,
26
+ divided", "pepper to taste"] ["Mix together both cans of corn, crackers, egg,
27
+ 2 teaspoons of melted butter and pepper and place in a buttered baking dish.",
28
+ "Dot with remaining 4 teaspoons of butter.", "Bake at 350\u00b0 for 1 hour."]
29
+ - Chicken Stew ["3 lb. chicken, boiled", "4 medium potatoes, diced", "2 medium onions,
30
+ chopped", "1 (16 oz.) can creamed corn", "1 (16 oz.) can English peas", "1 (16
31
+ oz.) can field peas", "1 (16 oz.) can butter beans", "1 (16 oz.) can tomatoes",
32
+ "1 (46 oz.) can tomato juice", "1 small box macaroni", "1 Tbsp. black pepper",
33
+ "1 Tbsp. salt", "1 Tbsp. sugar"] ["Remove chicken from bone.", "Use the broth.",
34
+ "Mix the vegetables and macaroni.", "Add sugar, salt and black pepper.", "Cook
35
+ until all vegetables are tender over medium heat."]
36
+ - source_sentence: Watermelon Rind Pickles ["7 lb. watermelon rind", "7 c. sugar",
37
+ "2 c. apple vinegar", "1/2 tsp. oil of cloves", "1/2 tsp. oil of cinnamon"] ["Trim
38
+ off green and pink parts of watermelon rind; cut to 1-inch cubes.", "Parboil until
39
+ tender, but not soft.", "Drain. Combine sugar, vinegar, oil of cloves and oil
40
+ of cinnamon; bring to boiling and pour over rind.", "Let stand overnight.", "In
41
+ the morning, drain off syrup.", "Heat and put over rind.", "The third morning,
42
+ heat rind and syrup; seal in hot, sterilized jars.", "Makes 8 pints.", "(Oil of
43
+ cinnamon and clove keeps rind clear and transparent.)"]
44
+ sentences:
45
+ - Cheeseburger Potato Soup ["6 baking potatoes", "1 lb. of extra lean ground beef",
46
+ "2/3 c. butter or margarine", "6 c. milk", "3/4 tsp. salt", "1/2 tsp. pepper",
47
+ "1 1/2 c (6 oz.) shredded Cheddar cheese, divided", "12 sliced bacon, cooked,
48
+ crumbled and divided", "4 green onion, chopped and divided", "1 (8 oz.) carton
49
+ sour cream (optional)"] ["Wash potatoes; prick several times with a fork.", "Microwave
50
+ them with a wet paper towel covering the potatoes on high for 6-8 minutes.", "The
51
+ potatoes should be soft, ready to eat.", "Let them cool enough to handle.", "Cut
52
+ in half lengthwise; scoop out pulp and reserve.", "Discard shells.", "Brown ground
53
+ beef until done.", "Drain any grease from the meat.", "Set aside when done.",
54
+ "Meat will be added later.", "Melt butter in a large kettle over low heat; add
55
+ flour, stirring until smooth.", "Cook 1 minute, stirring constantly. Gradually
56
+ add milk; cook over medium heat, stirring constantly, until thickened and bubbly.",
57
+ "Stir in potato, ground beef, salt, pepper, 1 cup of cheese, 2 tablespoons of
58
+ green onion and 1/2 cup of bacon.", "Cook until heated (do not boil).", "Stir
59
+ in sour cream if desired; cook until heated (do not boil).", "Sprinkle with remaining
60
+ cheese, bacon and green onions."]
61
+ - Easy Fudge ["1 (14 oz.) can sweetened condensed milk", "1 (12 oz.) pkg. semi-sweet
62
+ chocolate chips", "1 (1 oz.) sq. unsweetened chocolate (if desired)", "1 1/2 c.
63
+ chopped nuts (if desired)", "1 tsp. vanilla"] ["Butter a square pan, 8 x 8 x 2-inches.",
64
+ "Heat milk, chocolate chips and unsweetened chocolate over low heat, stirring
65
+ constantly, until chocolate is melted and mixture is smooth. Remove from heat.",
66
+ "Stir in nuts and vanilla.", "Spread in pan."]
67
+ - Chicken Ole ["4 chicken breasts, cooked", "1 can cream of chicken soup", "1 can
68
+ cream of mushroom soup", "1 can green chili salsa sauce", "1 can green chilies",
69
+ "1 c. milk", "1 grated onion", "1 pkg. corn tortilla in pieces"] ["Dice chicken.",
70
+ "Mix all ingredients together.", "Let sit overnight.", "Bake 1 1/2 hours at 375\u00b0."]
71
+ - source_sentence: Quick Barbecue Wings ["chicken wings (as many as you need for dinner)",
72
+ "flour", "barbecue sauce (your choice)"] ["Clean wings.", "Flour and fry until
73
+ done.", "Place fried chicken wings in microwave bowl.", "Stir in barbecue sauce.",
74
+ "Microwave on High (stir once) for 4 minutes."]
75
+ sentences:
76
+ - Creamy Corn ["2 (16 oz.) pkg. frozen corn", "1 (8 oz.) pkg. cream cheese, cubed",
77
+ "1/3 c. butter, cubed", "1/2 tsp. garlic powder", "1/2 tsp. salt", "1/4 tsp. pepper"]
78
+ ["In a slow cooker, combine all ingredients. Cover and cook on low for 4 hours
79
+ or until heated through and cheese is melted. Stir well before serving. Yields
80
+ 6 servings."]
81
+ - Broccoli Salad ["1 large head broccoli (about 1 1/2 lb.)", "10 slices bacon, cooked
82
+ and crumbled", "5 green onions, sliced or 1/4 c. chopped red onion", "1/2 c. raisins",
83
+ "1 c. mayonnaise", "2 Tbsp. vinegar", "1/4 c. sugar"] ["Trim off large leaves
84
+ of broccoli and remove the tough ends of lower stalks. Wash the broccoli thoroughly.
85
+ Cut the florets and stems into bite-size pieces. Place in a large bowl. Add bacon,
86
+ onions and raisins. Combine remaining ingredients, stirring well. Add dressing
87
+ to broccoli mixture and toss gently. Cover and refrigerate 2 to 3 hours. Makes
88
+ about 6 servings."]
89
+ - Vegetable-Burger Soup ["1/2 lb. ground beef", "2 c. water", "1 tsp. sugar", "1
90
+ pkg. Cup-a-Soup onion soup mix (dry)", "1 lb. can stewed tomatoes", "1 (8 oz.)
91
+ can tomato sauce", "1 (10 oz.) pkg. frozen mixed vegetables"] ["Lightly brown
92
+ beef in soup pot.", "Drain off excess fat.", "Stir in tomatoes, tomato sauce,
93
+ water, frozen vegetables, soup mix and sugar.", "Bring to a boil.", "Reduce heat
94
+ and simmer for 20 minutes. Serve."]
95
+ - source_sentence: 'Eggless Milkless Applesauce Cake ["3/4 c. sugar", "1/2 c. shortening",
96
+ "1 1/2 c. applesauce", "3 level tsp. soda", "1 tsp. each: cinnamon, cloves and
97
+ nutmeg", "2 c. sifted flour", "1 c. raisins", "1 c. nuts"] ["Mix Crisco with applesauce,
98
+ nuts and raisins.", "Sift dry ingredients and add.", "Mix well.", "Put in a greased
99
+ and floured loaf pan or tube pan.", "Bake in loaf pan at 350\u00b0 to 375\u00b0
100
+ for 45 to 60 minutes, layer pan at 375\u00b0 for 20 minutes or tube pan at 325\u00b0
101
+ for 1 hour."]'
102
+ sentences:
103
+ - Broccoli Dip For Crackers ["16 oz. sour cream", "1 pkg. dry vegetable soup mix",
104
+ "10 oz. pkg. frozen chopped broccoli, thawed and drained", "4 to 6 oz. Cheddar
105
+ cheese, grated"] ["Mix together sour cream, soup mix, broccoli and half of cheese.",
106
+ "Sprinkle remaining cheese on top.", "Bake at 350\u00b0 for 30 minutes, uncovered.",
107
+ "Serve hot with vegetable crackers."]
108
+ - Potato And Cheese Pie ["3 eggs", "1 tsp. salt", "1/4 tsp. pepper", "2 c. half
109
+ and half", "3 c. potatoes, shredded coarse", "1 c. Cheddar cheese, coarsely shredded",
110
+ "1/3 c. green onions"] ["Beat eggs, salt and pepper until well blended.", "Stir
111
+ in half and half, potatoes and onions.", "Pour into well-greased 8-inch baking
112
+ dish.", "Bake in a 400\u00b0 oven for 35 to 40 minutes, or until knife inserted
113
+ in center comes out clean and potatoes are tender. Cool on rack 5 minutes; cut
114
+ into squares.", "Makes 4 large servings."]
115
+ - Angel Biscuits ["5 c. flour", "3 Tbsp. sugar", "4 tsp. baking powder", "1 1/2
116
+ pkg. dry yeast", "2 c. buttermilk", "1 tsp. soda", "1 1/2 sticks margarine", "1/2
117
+ c. warm water"] ["Mix flour, sugar, baking powder, soda and salt together.", "Cut
118
+ in margarine, dissolve yeast in warm water.", "Stir into buttermilk and add to
119
+ dry mixture.", "Cover and chill."]
120
+ - source_sentence: Rhubarb Coffee Cake ["1 1/2 c. sugar", "1/2 c. butter", "1 egg",
121
+ "1 c. buttermilk", "2 c. flour", "1/2 tsp. salt", "1 tsp. soda", "1 c. buttermilk",
122
+ "2 c. rhubarb, finely cut", "1 tsp. vanilla"] ["Cream sugar and butter.", "Add
123
+ egg and beat well.", "To creamed butter, sugar and egg, add alternately buttermilk
124
+ with mixture of flour, salt and soda.", "Mix well.", "Add rhubarb and vanilla.",
125
+ "Pour into greased 9 x 13-inch pan and add Topping."]
126
+ sentences:
127
+ - Prize-Winning Meat Loaf ["1 1/2 lb. ground beef", "1 c. tomato juice", "3/4 c.
128
+ oats (uncooked)", "1 egg, beaten", "1/4 c. chopped onion", "1/4 tsp. pepper",
129
+ "1 1/2 tsp. salt"] ["Mix well.", "Press firmly into an 8 1/2 x 4 1/2 x 2 1/2-inch
130
+ loaf pan.", "Bake in preheated moderate oven.", "Bake at 350\u00b0 for 1 hour.",
131
+ "Let stand 5 minutes before slicing.", "Makes 8 servings."]
132
+ - Angel Biscuits ["5 c. flour", "3 Tbsp. sugar", "4 tsp. baking powder", "1 1/2
133
+ pkg. dry yeast", "2 c. buttermilk", "1 tsp. soda", "1 1/2 sticks margarine", "1/2
134
+ c. warm water"] ["Mix flour, sugar, baking powder, soda and salt together.", "Cut
135
+ in margarine, dissolve yeast in warm water.", "Stir into buttermilk and add to
136
+ dry mixture.", "Cover and chill."]
137
+ - 'Smothered Round Steak(Servings: 4) ["2 lb. round steak", "1/2 tsp. ground black
138
+ pepper", "1 tsp. ground white pepper", "1/2 c. vegetable oil", "2 bell peppers,
139
+ chopped", "1 c. beef stock or water", "2 tsp. salt", "1 tsp. ground red pepper",
140
+ "all-purpose flour (dredging)", "3 medium onions, chopped", "1 celery rib, chopped"]
141
+ ["Alex Patout says, \"Smothering is a multipurpose Cajun technique that works
142
+ wonders with everything from game to snap beans.", "It''s similar to what the
143
+ rest of the world knows as braising.", "The ingredients are briefly browned or
144
+ sauteed, then cooked with a little liquid over a low heat for a long time.\""]'
145
+ pipeline_tag: sentence-similarity
146
+ ---
147
+
148
+ # SentenceTransformer based on sentence-transformers/distilbert-base-nli-mean-tokens
149
+
150
+ This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [sentence-transformers/distilbert-base-nli-mean-tokens](https://huggingface.co/sentence-transformers/distilbert-base-nli-mean-tokens). It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
151
+
152
+ ## Model Details
153
+
154
+ ### Model Description
155
+ - **Model Type:** Sentence Transformer
156
+ - **Base model:** [sentence-transformers/distilbert-base-nli-mean-tokens](https://huggingface.co/sentence-transformers/distilbert-base-nli-mean-tokens) <!-- at revision 2781c006adbf3726b509caa8649fc8077ff0724d -->
157
+ - **Maximum Sequence Length:** 128 tokens
158
+ - **Output Dimensionality:** 768 tokens
159
+ - **Similarity Function:** Cosine Similarity
160
+ <!-- - **Training Dataset:** Unknown -->
161
+ <!-- - **Language:** Unknown -->
162
+ <!-- - **License:** Unknown -->
163
+
164
+ ### Model Sources
165
+
166
+ - **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
167
+ - **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
168
+ - **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
169
+
170
+ ### Full Model Architecture
171
+
172
+ ```
173
+ SentenceTransformer(
174
+ (0): Transformer({'max_seq_length': 128, 'do_lower_case': False}) with Transformer model: DistilBertModel
175
+ (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
176
+ )
177
+ ```
178
+
179
+ ## Usage
180
+
181
+ ### Direct Usage (Sentence Transformers)
182
+
183
+ First install the Sentence Transformers library:
184
+
185
+ ```bash
186
+ pip install -U sentence-transformers
187
+ ```
188
+
189
+ Then you can load this model and run inference.
190
+ ```python
191
+ from sentence_transformers import SentenceTransformer
192
+
193
+ # Download from the 🤗 Hub
194
+ model = SentenceTransformer("DivyaMereddy007/RecipeBert_v5original_epoc50_Copy_of_TrainSetenceTransforme-Finetuning_v5_DistilledBert")
195
+ # Run inference
196
+ sentences = [
197
+ 'Rhubarb Coffee Cake ["1 1/2 c. sugar", "1/2 c. butter", "1 egg", "1 c. buttermilk", "2 c. flour", "1/2 tsp. salt", "1 tsp. soda", "1 c. buttermilk", "2 c. rhubarb, finely cut", "1 tsp. vanilla"] ["Cream sugar and butter.", "Add egg and beat well.", "To creamed butter, sugar and egg, add alternately buttermilk with mixture of flour, salt and soda.", "Mix well.", "Add rhubarb and vanilla.", "Pour into greased 9 x 13-inch pan and add Topping."]',
198
+ 'Prize-Winning Meat Loaf ["1 1/2 lb. ground beef", "1 c. tomato juice", "3/4 c. oats (uncooked)", "1 egg, beaten", "1/4 c. chopped onion", "1/4 tsp. pepper", "1 1/2 tsp. salt"] ["Mix well.", "Press firmly into an 8 1/2 x 4 1/2 x 2 1/2-inch loaf pan.", "Bake in preheated moderate oven.", "Bake at 350\\u00b0 for 1 hour.", "Let stand 5 minutes before slicing.", "Makes 8 servings."]',
199
+ 'Angel Biscuits ["5 c. flour", "3 Tbsp. sugar", "4 tsp. baking powder", "1 1/2 pkg. dry yeast", "2 c. buttermilk", "1 tsp. soda", "1 1/2 sticks margarine", "1/2 c. warm water"] ["Mix flour, sugar, baking powder, soda and salt together.", "Cut in margarine, dissolve yeast in warm water.", "Stir into buttermilk and add to dry mixture.", "Cover and chill."]',
200
+ ]
201
+ embeddings = model.encode(sentences)
202
+ print(embeddings.shape)
203
+ # [3, 768]
204
+
205
+ # Get the similarity scores for the embeddings
206
+ similarities = model.similarity(embeddings, embeddings)
207
+ print(similarities.shape)
208
+ # [3, 3]
209
+ ```
210
+
211
+ <!--
212
+ ### Direct Usage (Transformers)
213
+
214
+ <details><summary>Click to see the direct usage in Transformers</summary>
215
+
216
+ </details>
217
+ -->
218
+
219
+ <!--
220
+ ### Downstream Usage (Sentence Transformers)
221
+
222
+ You can finetune this model on your own dataset.
223
+
224
+ <details><summary>Click to expand</summary>
225
+
226
+ </details>
227
+ -->
228
+
229
+ <!--
230
+ ### Out-of-Scope Use
231
+
232
+ *List how the model may foreseeably be misused and address what users ought not to do with the model.*
233
+ -->
234
+
235
+ <!--
236
+ ## Bias, Risks and Limitations
237
+
238
+ *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
239
+ -->
240
+
241
+ <!--
242
+ ### Recommendations
243
+
244
+ *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
245
+ -->
246
+
247
+ ## Training Details
248
+
249
+ ### Training Dataset
250
+
251
+ #### Unnamed Dataset
252
+
253
+
254
+ * Size: 1,746 training samples
255
+ * Columns: <code>sentence_0</code>, <code>sentence_1</code>, and <code>label</code>
256
+ * Approximate statistics based on the first 1000 samples:
257
+ | | sentence_0 | sentence_1 | label |
258
+ |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:---------------------------------------------------------------|
259
+ | type | string | string | float |
260
+ | details | <ul><li>min: 63 tokens</li><li>mean: 119.05 tokens</li><li>max: 128 tokens</li></ul> | <ul><li>min: 63 tokens</li><li>mean: 118.49 tokens</li><li>max: 128 tokens</li></ul> | <ul><li>min: 0.0</li><li>mean: 0.19</li><li>max: 1.0</li></ul> |
261
+ * Samples:
262
+ | sentence_0 | sentence_1 | label |
263
+ |:-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:-----------------|
264
+ | <code>Strawberry Whatever ["1 lb. frozen strawberries in juice", "1 small can crushed pineapple", "3 ripe bananas", "1 c. chopped pecans", "1 large pkg. strawberry Jell-O", "1 1/2 c. boiling water", "1 pt. sour cream"] ["Mix Jell-O in boiling water.", "Add strawberries, pineapple, crushed bananas and nuts.", "Spread 1/2 mixture in 13 x 6 1/2-inch pan.", "Allow to gel in freezer 30 minutes.", "Add layer of sour cream, then remaining mixture on top.", "Gel and serve."]</code> | <code>One Hour Rolls ["1 c. milk", "2 Tbsp. sugar", "1 pkg. dry yeast", "1 Tbsp. salt", "3 Tbsp. Crisco oil", "2 c. plain flour"] ["Put flour into a large mixing bowl.", "Combine sugar, milk, salt and oil in a saucepan and heat to boiling; remove from heat and let cool to lukewarm.", "Add yeast and mix well.", "Pour into flour and stir.", "Batter will be sticky.", "Roll out batter on a floured board and cut with biscuit cutter.", "Lightly brush tops with melted oleo and fold over.", "Place rolls on a cookie sheet, put in a warm place and let rise for 1 hour.", "Bake at 350\u00b0 for about 20 minutes. Yield: 2 1/2 dozen."]</code> | <code>0.1</code> |
265
+ | <code>Broccoli Dip For Crackers ["16 oz. sour cream", "1 pkg. dry vegetable soup mix", "10 oz. pkg. frozen chopped broccoli, thawed and drained", "4 to 6 oz. Cheddar cheese, grated"] ["Mix together sour cream, soup mix, broccoli and half of cheese.", "Sprinkle remaining cheese on top.", "Bake at 350\u00b0 for 30 minutes, uncovered.", "Serve hot with vegetable crackers."]</code> | <code>Vegetable-Burger Soup ["1/2 lb. ground beef", "2 c. water", "1 tsp. sugar", "1 pkg. Cup-a-Soup onion soup mix (dry)", "1 lb. can stewed tomatoes", "1 (8 oz.) can tomato sauce", "1 (10 oz.) pkg. frozen mixed vegetables"] ["Lightly brown beef in soup pot.", "Drain off excess fat.", "Stir in tomatoes, tomato sauce, water, frozen vegetables, soup mix and sugar.", "Bring to a boil.", "Reduce heat and simmer for 20 minutes. Serve."]</code> | <code>0.4</code> |
266
+ | <code>Summer Spaghetti ["1 lb. very thin spaghetti", "1/2 bottle McCormick Salad Supreme (seasoning)", "1 bottle Zesty Italian dressing"] ["Prepare spaghetti per package.", "Drain.", "Melt a little butter through it.", "Marinate overnight in Salad Supreme and Zesty Italian dressing.", "Just before serving, add cucumbers, tomatoes, green peppers, mushrooms, olives or whatever your taste may want."]</code> | <code>Chicken Funny ["1 large whole chicken", "2 (10 1/2 oz.) cans chicken gravy", "1 (10 1/2 oz.) can cream of mushroom soup", "1 (6 oz.) box Stove Top stuffing", "4 oz. shredded cheese"] ["Boil and debone chicken.", "Put bite size pieces in average size square casserole dish.", "Pour gravy and cream of mushroom soup over chicken; level.", "Make stuffing according to instructions on box (do not make too moist).", "Put stuffing on top of chicken and gravy; level.", "Sprinkle shredded cheese on top and bake at 350\u00b0 for approximately 20 minutes or until golden and bubbly."]</code> | <code>0.3</code> |
267
+ * Loss: [<code>CosineSimilarityLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#cosinesimilarityloss) with these parameters:
268
+ ```json
269
+ {
270
+ "loss_fct": "torch.nn.modules.loss.MSELoss"
271
+ }
272
+ ```
273
+
274
+ ### Training Hyperparameters
275
+ #### Non-Default Hyperparameters
276
+
277
+ - `per_device_train_batch_size`: 16
278
+ - `per_device_eval_batch_size`: 16
279
+ - `num_train_epochs`: 50
280
+ - `multi_dataset_batch_sampler`: round_robin
281
+
282
+ #### All Hyperparameters
283
+ <details><summary>Click to expand</summary>
284
+
285
+ - `overwrite_output_dir`: False
286
+ - `do_predict`: False
287
+ - `eval_strategy`: no
288
+ - `prediction_loss_only`: True
289
+ - `per_device_train_batch_size`: 16
290
+ - `per_device_eval_batch_size`: 16
291
+ - `per_gpu_train_batch_size`: None
292
+ - `per_gpu_eval_batch_size`: None
293
+ - `gradient_accumulation_steps`: 1
294
+ - `eval_accumulation_steps`: None
295
+ - `learning_rate`: 5e-05
296
+ - `weight_decay`: 0.0
297
+ - `adam_beta1`: 0.9
298
+ - `adam_beta2`: 0.999
299
+ - `adam_epsilon`: 1e-08
300
+ - `max_grad_norm`: 1
301
+ - `num_train_epochs`: 50
302
+ - `max_steps`: -1
303
+ - `lr_scheduler_type`: linear
304
+ - `lr_scheduler_kwargs`: {}
305
+ - `warmup_ratio`: 0.0
306
+ - `warmup_steps`: 0
307
+ - `log_level`: passive
308
+ - `log_level_replica`: warning
309
+ - `log_on_each_node`: True
310
+ - `logging_nan_inf_filter`: True
311
+ - `save_safetensors`: True
312
+ - `save_on_each_node`: False
313
+ - `save_only_model`: False
314
+ - `restore_callback_states_from_checkpoint`: False
315
+ - `no_cuda`: False
316
+ - `use_cpu`: False
317
+ - `use_mps_device`: False
318
+ - `seed`: 42
319
+ - `data_seed`: None
320
+ - `jit_mode_eval`: False
321
+ - `use_ipex`: False
322
+ - `bf16`: False
323
+ - `fp16`: False
324
+ - `fp16_opt_level`: O1
325
+ - `half_precision_backend`: auto
326
+ - `bf16_full_eval`: False
327
+ - `fp16_full_eval`: False
328
+ - `tf32`: None
329
+ - `local_rank`: 0
330
+ - `ddp_backend`: None
331
+ - `tpu_num_cores`: None
332
+ - `tpu_metrics_debug`: False
333
+ - `debug`: []
334
+ - `dataloader_drop_last`: False
335
+ - `dataloader_num_workers`: 0
336
+ - `dataloader_prefetch_factor`: None
337
+ - `past_index`: -1
338
+ - `disable_tqdm`: False
339
+ - `remove_unused_columns`: True
340
+ - `label_names`: None
341
+ - `load_best_model_at_end`: False
342
+ - `ignore_data_skip`: False
343
+ - `fsdp`: []
344
+ - `fsdp_min_num_params`: 0
345
+ - `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
346
+ - `fsdp_transformer_layer_cls_to_wrap`: None
347
+ - `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
348
+ - `deepspeed`: None
349
+ - `label_smoothing_factor`: 0.0
350
+ - `optim`: adamw_torch
351
+ - `optim_args`: None
352
+ - `adafactor`: False
353
+ - `group_by_length`: False
354
+ - `length_column_name`: length
355
+ - `ddp_find_unused_parameters`: None
356
+ - `ddp_bucket_cap_mb`: None
357
+ - `ddp_broadcast_buffers`: False
358
+ - `dataloader_pin_memory`: True
359
+ - `dataloader_persistent_workers`: False
360
+ - `skip_memory_metrics`: True
361
+ - `use_legacy_prediction_loop`: False
362
+ - `push_to_hub`: False
363
+ - `resume_from_checkpoint`: None
364
+ - `hub_model_id`: None
365
+ - `hub_strategy`: every_save
366
+ - `hub_private_repo`: False
367
+ - `hub_always_push`: False
368
+ - `gradient_checkpointing`: False
369
+ - `gradient_checkpointing_kwargs`: None
370
+ - `include_inputs_for_metrics`: False
371
+ - `eval_do_concat_batches`: True
372
+ - `fp16_backend`: auto
373
+ - `push_to_hub_model_id`: None
374
+ - `push_to_hub_organization`: None
375
+ - `mp_parameters`:
376
+ - `auto_find_batch_size`: False
377
+ - `full_determinism`: False
378
+ - `torchdynamo`: None
379
+ - `ray_scope`: last
380
+ - `ddp_timeout`: 1800
381
+ - `torch_compile`: False
382
+ - `torch_compile_backend`: None
383
+ - `torch_compile_mode`: None
384
+ - `dispatch_batches`: None
385
+ - `split_batches`: None
386
+ - `include_tokens_per_second`: False
387
+ - `include_num_input_tokens_seen`: False
388
+ - `neftune_noise_alpha`: None
389
+ - `optim_target_modules`: None
390
+ - `batch_eval_metrics`: False
391
+ - `batch_sampler`: batch_sampler
392
+ - `multi_dataset_batch_sampler`: round_robin
393
+
394
+ </details>
395
+
396
+ ### Training Logs
397
+ | Epoch | Step | Training Loss |
398
+ |:-------:|:----:|:-------------:|
399
+ | 4.5455 | 500 | 0.0594 |
400
+ | 9.0909 | 1000 | 0.0099 |
401
+ | 13.6364 | 1500 | 0.0085 |
402
+ | 18.1818 | 2000 | 0.0077 |
403
+ | 22.7273 | 2500 | 0.0074 |
404
+ | 27.2727 | 3000 | 0.0071 |
405
+ | 31.8182 | 3500 | 0.0068 |
406
+ | 36.3636 | 4000 | 0.0066 |
407
+ | 40.9091 | 4500 | 0.0063 |
408
+ | 45.4545 | 5000 | 0.006 |
409
+ | 50.0 | 5500 | 0.0057 |
410
+
411
+
412
+ ### Framework Versions
413
+ - Python: 3.10.12
414
+ - Sentence Transformers: 3.0.1
415
+ - Transformers: 4.41.2
416
+ - PyTorch: 2.3.0+cu121
417
+ - Accelerate: 0.31.0
418
+ - Datasets: 2.19.2
419
+ - Tokenizers: 0.19.1
420
+
421
+ ## Citation
422
+
423
+ ### BibTeX
424
+
425
+ #### Sentence Transformers
426
+ ```bibtex
427
+ @inproceedings{reimers-2019-sentence-bert,
428
+ title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
429
+ author = "Reimers, Nils and Gurevych, Iryna",
430
+ booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
431
+ month = "11",
432
+ year = "2019",
433
+ publisher = "Association for Computational Linguistics",
434
+ url = "https://arxiv.org/abs/1908.10084",
435
+ }
436
+ ```
437
+
438
+ <!--
439
+ ## Glossary
440
+
441
+ *Clearly define terms in order to be accessible across audiences.*
442
+ -->
443
+
444
+ <!--
445
+ ## Model Card Authors
446
+
447
+ *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
448
+ -->
449
+
450
+ <!--
451
+ ## Model Card Contact
452
+
453
+ *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
454
+ -->
config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "sentence-transformers/distilbert-base-nli-mean-tokens",
3
+ "activation": "gelu",
4
+ "architectures": [
5
+ "DistilBertModel"
6
+ ],
7
+ "attention_dropout": 0.1,
8
+ "dim": 768,
9
+ "dropout": 0.1,
10
+ "hidden_dim": 3072,
11
+ "initializer_range": 0.02,
12
+ "max_position_embeddings": 512,
13
+ "model_type": "distilbert",
14
+ "n_heads": 12,
15
+ "n_layers": 6,
16
+ "pad_token_id": 0,
17
+ "qa_dropout": 0.1,
18
+ "seq_classif_dropout": 0.2,
19
+ "sinusoidal_pos_embds": false,
20
+ "tie_weights_": true,
21
+ "torch_dtype": "float32",
22
+ "transformers_version": "4.41.2",
23
+ "vocab_size": 30522
24
+ }
config_sentence_transformers.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "3.0.1",
4
+ "transformers": "4.41.2",
5
+ "pytorch": "2.3.0+cu121"
6
+ },
7
+ "prompts": {},
8
+ "default_prompt_name": null,
9
+ "similarity_fn_name": null
10
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:04c064513ad8cdbd52ad90123b6a208f7f97365b88f0d4bb194516a42a927fcc
3
+ size 265462608
modules.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ }
14
+ ]
sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 128,
3
+ "do_lower_case": false
4
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": {
3
+ "content": "[CLS]",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "mask_token": {
10
+ "content": "[MASK]",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "[PAD]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "sep_token": {
24
+ "content": "[SEP]",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "unk_token": {
31
+ "content": "[UNK]",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ }
37
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_basic_tokenize": true,
47
+ "do_lower_case": true,
48
+ "mask_token": "[MASK]",
49
+ "model_max_length": 128,
50
+ "never_split": null,
51
+ "pad_token": "[PAD]",
52
+ "sep_token": "[SEP]",
53
+ "strip_accents": null,
54
+ "tokenize_chinese_chars": true,
55
+ "tokenizer_class": "DistilBertTokenizer",
56
+ "unk_token": "[UNK]"
57
+ }
vocab.txt ADDED
The diff for this file is too large to render. See raw diff