JRosenkranz
commited on
Commit
•
23fcb4c
1
Parent(s):
2e7a11a
Update README.md
Browse files
README.md
CHANGED
@@ -18,6 +18,8 @@ license: apache-2.0
|
|
18 |
|
19 |
Representing IBM’s commitment to open source innovation IBM has released granite-7b-base, a base pre-trained LLM from IBM’s Granite model series, under an apache-2.0 license for community and commercial use. Granite-7b-base was pre-trained from scratch on IBM-curated data as an open reference implementation of Meta’s Llama-2-7B. In a commitment to data transparency and fostering open innovation, the data sources, sampling proportions, and URLs for access are provided below.
|
20 |
|
|
|
|
|
21 |
**Pre-Training Data**
|
22 |
|
23 |
The model was trained on 2T tokens, with sampling proportions designed to match the sampling distributions released in the Llama1 paper as closely as possible.
|
|
|
18 |
|
19 |
Representing IBM’s commitment to open source innovation IBM has released granite-7b-base, a base pre-trained LLM from IBM’s Granite model series, under an apache-2.0 license for community and commercial use. Granite-7b-base was pre-trained from scratch on IBM-curated data as an open reference implementation of Meta’s Llama-2-7B. In a commitment to data transparency and fostering open innovation, the data sources, sampling proportions, and URLs for access are provided below.
|
20 |
|
21 |
+
For more information about training this model, please check out the blog: https://pytorch.org/blog/maximizing-training/
|
22 |
+
|
23 |
**Pre-Training Data**
|
24 |
|
25 |
The model was trained on 2T tokens, with sampling proportions designed to match the sampling distributions released in the Llama1 paper as closely as possible.
|