pszemraj commited on
Commit
240d210
1 Parent(s): 889d47c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +122 -0
README.md CHANGED
@@ -1,3 +1,125 @@
1
  ---
2
  license: apache-2.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
+ tags:
4
+ - instruct
5
+ - instructions
6
+ - domain adapt
7
+ - instructiongen
8
+ metrics:
9
+ - rouge
10
+ widget:
11
+ - text: >-
12
+ You'll need to start by choosing the right venue. Consider the type of
13
+ atmosphere and the size of the area that will be suitable for the number of
14
+ guests you plan to invite. Choose the right decorations based on your
15
+ brother's interests, such as balloons in his favorite colors, banners, and
16
+ streamers. Next, decide on the food and drinks, making sure they are tasty
17
+ and appropriate for the occasion. Then decide on the other games, music, and
18
+ entertainment that will make the party memorable. Finally, involve your
19
+ brother's friends and family to help create the perfect surprise.
20
+ example_title: birthday party
21
+ - text: 1) cookies and cream 2) chocolate chip 3) mint chip 4) oreo
22
+ example_title: ice cream
23
+ - text: >-
24
+ Start by selecting a scale model of a building that fits the theme. Use a
25
+ hobby knife and glue to cut and assemble the model into a ruined or
26
+ abandoned version of itself, adding details like broken windows and
27
+ graffiti. Create a base for the diorama using foam, plaster, or other
28
+ materials, and paint it to resemble a ruined street or sidewalk. Add
29
+ miniature vehicles, debris, and figures to complete the scene, and use
30
+ weathering techniques like dry brushing and rust washes to add realism.
31
+ Display the diorama in a shadow box or other protective case to showcase
32
+ your work.
33
+ example_title: Miniature diorama creation
34
+ - text: >-
35
+ Start by selecting clothing that is futuristic and edgy, such as leather
36
+ jackets, neon-colored accessories, and tech-inspired patterns. Add
37
+ accessories like goggles, cybernetic implants, and LED lights to enhance the
38
+ cyberpunk vibe. Use makeup and body paint to create a futuristic look, such
39
+ as metallic skin or neon makeup. Consider adding functional elements to your
40
+ costume, such as a built-in backpack or hidden pockets for your tech
41
+ gadgets. Finally, practice your confident walk and embrace your inner
42
+ cyberpunk for a memorable and immersive costume experience.
43
+ example_title: Cyberpunk costume design
44
+ - text: >-
45
+ Start by creating a base terrain with mountains, valleys, and other natural
46
+ features. Use fractal noise and displacement mapping to add texture and
47
+ detail to the terrain, and experiment with different materials like rock,
48
+ grass, and water. Add surreal elements like floating islands, giant
49
+ mushrooms, or impossible geometry to create a dreamlike atmosphere. Use
50
+ lighting and color grading to enhance the mood and tone of the scene, and
51
+ render the final image at a high resolution for maximum impact. Share your
52
+ surreal landscape with the world and inspire others to explore the
53
+ possibilities of 3D art.
54
+ example_title: Surreal 3D landscape creation
55
+ - text: >-
56
+ Start by setting a realistic goal and creating a training plan. Build up
57
+ your mileage gradually over time, and incorporate cross-training and
58
+ strength exercises to prevent injury and improve endurance. Be sure to stay
59
+ hydrated and properly fuel your body with nutritious foods. Listen to your
60
+ body and adjust your training as needed to avoid overexertion or burnout.
61
+ Finally, taper your training in the weeks leading up to the race to give
62
+ your body time to rest and recover before the big day.
63
+ example_title: Marathon training
64
+ inference:
65
+ parameters:
66
+ max_length: 96
67
+ num_beams: 4
68
+ datasets:
69
+ - pszemraj/fleece2instructions-inputs-alpaca-cleaned
70
+ language:
71
+ - en
72
+ pipeline_tag: text2text-generation
73
  ---
74
+
75
+
76
+ # bart-base-instructiongen-w-inputs
77
+
78
+ Use this text2text model to find out what LLM `instruction` (**and** `inputs` if relevant) might have generated `<arbitrary input text>`!
79
+
80
+ This model is a fine-tuned version of [facebook/bart-base](https://huggingface.co/facebook/bart-base) on the `pszemraj/fleece2instructions-inputs-alpaca-cleaned` dataset.
81
+
82
+ It achieves the following results on the evaluation set:
83
+ - Loss: 0.9579
84
+ - Rouge1: 62.3604
85
+ - Rouge2: 39.5109
86
+ - Rougel: 58.8843
87
+ - Rougelsum: 60.4494
88
+ - Gen Len: 24.9917
89
+
90
+ ## Intended uses & limitations
91
+
92
+ This model is intended to be used to generate instructions from arbitrary text. You can then use these instructions + your data to fine-tune an LLM on instructions w.r.t. a specific domain. This model is primarily intended to enable **low-resource domain adaptation**, rather than "_I want to generate even better prompts for the FLAN-V2 dataset!_".
93
+
94
+ The `fleece2instructions-inputs-alpaca-cleaned` dataset, obtained from the [alpaca-lora repo](https://github.com/tloen/alpaca-lora) under the ODC-BY license, has been converted to a text2text format for use with language models. In this dataset, the original 'inputs' and 'instructions' columns are combined into a single 'instructions_inputs' column. To clearly separate the two types of content, each piece of text is prefixed with either an `<instruction>` or `<inputs>` token. These tokens not only facilitate model comprehension, but also allow for easy regex separation of model outputs during inference.
95
+
96
+ As such, users can expect the output of this model to be similarly structured with `<instruction>` and `<inputs>` tokens.
97
+
98
+ ## Training and evaluation data
99
+
100
+ Refer to the [fleece2instructions-inputs-alpaca-cleaned](https://huggingface.co/datasets/pszemraj/fleece2instructions-inputs-alpaca-cleaned) dataset
101
+
102
+ ## Training procedure
103
+
104
+ ### Training hyperparameters
105
+
106
+ The following hyperparameters were used during training:
107
+ - learning_rate: 8e-05
108
+ - train_batch_size: 4
109
+ - eval_batch_size: 4
110
+ - seed: 42
111
+ - distributed_type: multi-GPU
112
+ - gradient_accumulation_steps: 16
113
+ - total_train_batch_size: 64
114
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
115
+ - lr_scheduler_type: cosine
116
+ - lr_scheduler_warmup_ratio: 0.02
117
+ - num_epochs: 2.0
118
+
119
+ ### Training results
120
+
121
+ | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
122
+ |:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|:-------:|:---------:|:-------:|
123
+ | 1.1147 | 1.0 | 680 | 0.9901 | 61.8451 | 38.8293 | 58.3372 | 59.8658 | 25.2401 |
124
+ | 0.9565 | 2.0 | 1360 | 0.9579 | 62.3604 | 39.5109 | 58.8843 | 60.4494 | 24.9917 |
125
+