Rallio67 commited on
Commit
7e2156c
1 Parent(s): f8fa1be

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +57 -0
README.md ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Conditional Pretraining of Large Language Models
2
+
3
+ Large language models (LLMs), such as OpenAI's ChatGPT and similar chatbot products from other organizations, have recently gained widespread adoption. These models can extend text or respond to instructions in a natural and helpful manner. Despite the core technologies behind LLMs, namely the transformer architecture and the GPT decoder-only causal language model, remaining relatively unchanged for over five years, the surge in popularity of ChatGPT can be largely attributed to recent approaches that better align the output of LLMs with users' and service providers' intentions.
4
+
5
+ ## Primary Approaches for Aligning LLMs with Human Expectations
6
+
7
+ 1. **Supervised finetuning (SFT) on natural instructions**
8
+ 2. **Reinforcement learning from human feedback (RLHF)**
9
+
10
+ ## Conditional Pretraining: A Third Approach
11
+
12
+ - Utilizes a large number of pretraining examples tagged with human-understandable classifiers
13
+ - Leverages content tagging found in many online environments
14
+ - Examples of commonly used tags:
15
+ - Suitable for work (SFW) and not suitable for work (NSFW)
16
+ - G, PG, PG-13, and R for television and movie content
17
+
18
+ ## Converting Existing Pretraining Data into Conditional Pretraining Data
19
+
20
+ - Traditional pretraining involves predicting the subsequent word in minimally processed text.
21
+ - Conditional pretraining prepends training examples with descriptive tags and a brief synopsis.
22
+
23
+ ## Transparency and Accountability
24
+
25
+ - Current LLMs have proprietary instructions and reward models, which can hinder public review and discussions on sensitive topics.
26
+ - Conditional pretraining tags are transparent and easily understood by auditors or end users.
27
+
28
+
29
+ ## Conditional pretraining example
30
+
31
+ An example output from this conditional tagging model for a recent news article about LAION.
32
+
33
+ [`Article Here`](https://www.forbes.com/sites/hessiejones/2023/04/19/amid-growing-call-to-pause-ai-research-laion-petitions-governments-to-keep-agi-research-open-active-and-responsible/) is below. To generate these document tags only text from the body of the article was used.
34
+
35
+ ```
36
+ [ artificial intelligence, open source, ai, open letter, open source ai, ai research]
37
+
38
+ # This article explains the importance of a CERN-like organization to coordinate efforts on the transparency of large-scale AI research and provides information about LAION.
39
+ ```
40
+ ## How to use the model
41
+
42
+ Format your inputs like this:
43
+ ```
44
+ [ tag1, tag2, tag3, tag_n]
45
+
46
+ # This is a short synopsis of what kind of text I want to generate.
47
+ ```
48
+
49
+ ## Acknowledgement
50
+
51
+ Thank you to LAION and Stability.ai for support and compute resources to experiment with conditional pretraining.
52
+
53
+ ## Disclaimer
54
+
55
+ - Conditional pretraining helps the user control the outputs of the model.
56
+ - However, these models (and all language models) can still generate undesirable content.
57
+ - So please enjoy and use with care!