pszemraj commited on
Commit
699ea37
1 Parent(s): 0e57e02

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +125 -0
README.md CHANGED
@@ -1,3 +1,128 @@
1
  ---
2
  license: apache-2.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
+ tags:
4
+ - instruct
5
+ - instructions
6
+ - domain adapt
7
+ - instructiongen
8
+ metrics:
9
+ - rouge
10
+ widget:
11
+ - text: >-
12
+ You'll need to start by choosing the right venue. Consider the type of
13
+ atmosphere and the size of the area that will be suitable for the number of
14
+ guests you plan to invite. Choose the right decorations based on your
15
+ brother's interests, such as balloons in his favorite colors, banners, and
16
+ streamers. Next, decide on the food and drinks, making sure they are tasty
17
+ and appropriate for the occasion. Then decide on the other games, music, and
18
+ entertainment that will make the party memorable. Finally, involve your
19
+ brother's friends and family to help create the perfect surprise.
20
+ example_title: birthday party
21
+ - text: 1) cookies and cream 2) chocolate chip 3) mint chip 4) oreo
22
+ example_title: ice cream
23
+ - text: >-
24
+ Start by selecting a scale model of a building that fits the theme. Use a
25
+ hobby knife and glue to cut and assemble the model into a ruined or
26
+ abandoned version of itself, adding details like broken windows and
27
+ graffiti. Create a base for the diorama using foam, plaster, or other
28
+ materials, and paint it to resemble a ruined street or sidewalk. Add
29
+ miniature vehicles, debris, and figures to complete the scene, and use
30
+ weathering techniques like dry brushing and rust washes to add realism.
31
+ Display the diorama in a shadow box or other protective case to showcase
32
+ your work.
33
+ example_title: Miniature diorama creation
34
+ - text: >-
35
+ Start by selecting clothing that is futuristic and edgy, such as leather
36
+ jackets, neon-colored accessories, and tech-inspired patterns. Add
37
+ accessories like goggles, cybernetic implants, and LED lights to enhance the
38
+ cyberpunk vibe. Use makeup and body paint to create a futuristic look, such
39
+ as metallic skin or neon makeup. Consider adding functional elements to your
40
+ costume, such as a built-in backpack or hidden pockets for your tech
41
+ gadgets. Finally, practice your confident walk and embrace your inner
42
+ cyberpunk for a memorable and immersive costume experience.
43
+ example_title: Cyberpunk costume design
44
+ - text: >-
45
+ Start by creating a base terrain with mountains, valleys, and other natural
46
+ features. Use fractal noise and displacement mapping to add texture and
47
+ detail to the terrain, and experiment with different materials like rock,
48
+ grass, and water. Add surreal elements like floating islands, giant
49
+ mushrooms, or impossible geometry to create a dreamlike atmosphere. Use
50
+ lighting and color grading to enhance the mood and tone of the scene, and
51
+ render the final image at a high resolution for maximum impact. Share your
52
+ surreal landscape with the world and inspire others to explore the
53
+ possibilities of 3D art.
54
+ example_title: Surreal 3D landscape creation
55
+ - text: >-
56
+ Start by setting a realistic goal and creating a training plan. Build up
57
+ your mileage gradually over time, and incorporate cross-training and
58
+ strength exercises to prevent injury and improve endurance. Be sure to stay
59
+ hydrated and properly fuel your body with nutritious foods. Listen to your
60
+ body and adjust your training as needed to avoid overexertion or burnout.
61
+ Finally, taper your training in the weeks leading up to the race to give
62
+ your body time to rest and recover before the big day.
63
+ example_title: Marathon training
64
+ inference:
65
+ parameters:
66
+ max_length: 96
67
+ num_beams: 4
68
+ datasets:
69
+ - pszemraj/fleece2instructions-inputs-alpaca-cleaned
70
+ language:
71
+ - en
72
+ pipeline_tag: text2text-generation
73
  ---
74
+
75
+
76
+ # bart-large-instructiongen-w-inputs
77
+
78
+ Use this text2text model to find out what LLM `instruction` (**and** `inputs` if relevant) might have generated `<arbitrary input text>`!
79
+
80
+ This model is a fine-tuned version of [facebook/bart-large](https://huggingface.co/facebook/bart-large) on the `pszemraj/fleece2instructions-inputs-alpaca-cleaned` dataset.
81
+ It achieves the following results on the evaluation set:
82
+ - Loss: 0.9302
83
+ - Rouge1: 64.2236
84
+ - Rouge2: 41.5632
85
+ - Rougel: 60.5935
86
+ - Rougelsum: 62.1285
87
+ - Gen Len: 25.8938
88
+
89
+ ## example
90
+
91
+ ![api](https://i.imgur.com/2xubG7N.png)
92
+
93
+ ## Intended uses & limitations
94
+
95
+ This model is intended to be used to generate instructions from arbitrary text. You can then use these instructions + your data to fine-tune an LLM on instructions w.r.t. a specific domain. This model is primarily intended to enable **low-resource domain adaptation**, rather than "_I want to generate even better prompts for the FLAN-V2 dataset!_".
96
+
97
+ The `fleece2instructions-inputs-alpaca-cleaned` dataset, obtained from the [alpaca-lora repo](https://github.com/tloen/alpaca-lora) under the ODC-BY license, has been converted to a text2text format for use with language models. In this dataset, the original 'inputs' and 'instructions' columns are combined into a single 'instructions_inputs' column. To clearly separate the two types of content, each piece of text is prefixed with either an `<instruction>` or `<inputs>` token. These tokens not only facilitate model comprehension, but also allow for easy regex separation of model outputs during inference.
98
+
99
+ As such, users can expect the output of this model to be similarly structured with `<instruction>` and `<inputs>` tokens.
100
+
101
+ ## Training and evaluation data
102
+
103
+ Refer to the [fleece2instructions-inputs-alpaca-cleaned](https://huggingface.co/datasets/pszemraj/fleece2instructions-inputs-alpaca-cleaned) dataset
104
+
105
+ ## Training procedure
106
+
107
+ ### Training hyperparameters
108
+
109
+ The following hyperparameters were used during training:
110
+ - learning_rate: 6e-05
111
+ - train_batch_size: 16
112
+ - eval_batch_size: 8
113
+ - seed: 42
114
+ - distributed_type: multi-GPU
115
+ - gradient_accumulation_steps: 2
116
+ - total_train_batch_size: 32
117
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
118
+ - lr_scheduler_type: cosine
119
+ - lr_scheduler_warmup_ratio: 0.03
120
+ - num_epochs: 3.0
121
+
122
+ ### Training results
123
+
124
+ | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
125
+ |:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|:-------:|:---------:|:-------:|
126
+ | 1.0145 | 1.0 | 1361 | 1.0460 | 62.8374 | 39.8538 | 59.2593 | 60.8095 | 25.2752 |
127
+ | 0.8796 | 2.0 | 2722 | 0.9289 | 63.7086 | 41.1315 | 60.1588 | 61.7145 | 25.7215 |
128
+ | 0.6943 | 3.0 | 4083 | 0.9302 | 64.2236 | 41.5632 | 60.5935 | 62.1285 | 25.8938 |