power-greg commited on
Commit
e4d644f
1 Parent(s): bdca489

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +45 -1
README.md CHANGED
@@ -40,4 +40,48 @@ The optimizations in this library reach far beyond what’s available to develop
40
 
41
  Lamini runs across platforms, from OpenAI’s models to open-source ones on HuggingFace, with more to come soon. We are agnostic to base models, as long as there’s a way for our engine to train and run them. In fact, Lamini makes it easy to run multiple base model comparisons in just a single line of code.
42
 
43
- Now that you know a bit about where we’re going, today, we’re excited to release our first major community resource!
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
40
 
41
  Lamini runs across platforms, from OpenAI’s models to open-source ones on HuggingFace, with more to come soon. We are agnostic to base models, as long as there’s a way for our engine to train and run them. In fact, Lamini makes it easy to run multiple base model comparisons in just a single line of code.
42
 
43
+ Now that you know a bit about where we’re going, today, we’re excited to release our first major community resource!
44
+
45
+ # Available now: a hosted data generator for LLM training 🎉
46
+
47
+ Steps to a ChatGPT-like LLM for your use case 1️⃣2️⃣3️⃣
48
+ Here are the steps to get an instruction-following LLM like ChatGPT to handle your use case:
49
+
50
+ (Show me the code: Play with our dataset generator for creating ChatGPT-like datasets.)
51
+
52
+ 1. Try prompt-tuning ChatGPT or another model. You can use Lamini library’s APIs to quickly prompt-tune across different models, swapping between OpenAI and open-source models in just one line of code. We optimize the right prompt for you, so you can take advantage of different models without worrying about the right prompt template for each model.
53
+ 2. Build a large dataset of input-output pairs. These will show your model how it should respond to its inputs, whether that's following instructions given in English, or responding in JSON. Today, we’re releasing a repo with just a few lines of code using the Lamini library to generate 50k data points from as few as 100 data points. We include an open-source 50k dataset in the repo. (More details below on how you can do this!)
54
+ 3. Finetune a base model on your large dataset. Alongside the dataset generator, we’re also releasing an LLM that is finetuned on the generated data using Lamini. You can also hit OpenAI’s finetuning API as a great starting point.
55
+ 4. Run RLHF on your finetuned model. You’ll need an ML team and human labeling team to do this today.
56
+ 5. Deploy to your cloud, by simply hitting the API endpoint in your product or feature.
57
+
58
+ Our goal is for Lamini to handle this entire process, and we’re actively building steps 2-4 (sign up for early access!).
59
+
60
+ ## Step #1: A ChatGPT-like dataset generator 🎉
61
+ ChatGPT took the world by storm because it could follow instructions from the user, while the base model that it was trained from (GPT-3) couldn’t do that consistently. For example, if you asked the base model a question, it might generate another question instead of answering it. 🤔
62
+
63
+ For your application, you'll probably want similar "instruction-following" data, but you might want something completely different, like responding only in JSON.
64
+
65
+ You'll need a dataset of ~50k instruction-following examples to start. Don't panic. You can now use Lamini’s dataset generator on Github to turn as few as 100 examples into as many as 50k in just a few lines of code.
66
+
67
+ You can customize the initial 100+ instructions so that the LLM follows instructions in your own vertical. Once you have those, submit them to the Lamini dataset generator, and voilà: you get a large instruction-following dataset on your use case as a result!
68
+
69
+ How the dataset generator works
70
+
71
+ The Lamini dataset generator is a pipeline of LLMs that takes your original small set of 100+ instructions, paired with the expected responses, to generate 50k+ new pairs, inspired by Stanford Alpaca. This generation pipeline uses the Lamini library to define and call LLMs to generate different, yet similar, pairs of instructions and responses. Trained on this data, your LLM will improve to follow these instructions.
72
+
73
+ We provide a good default for the generation pipeline that uses open-source LLMs, which we call Lamini Open and Lamini Instruct. With new LLMs being released each day, we update the defaults to the best-performing models. As of this release, we are using XX for Lamini Open and YY for Lamini Instruct. Lamini Open generates more instructions, and Lamini Instruct generates paired responses to those instructions. The final generated dataset is available for your free commercial use (CC-BY license).
74
+
75
+ The Lamini library allows you to swap our defaults for other open-source or OpenAI models in just one line of code. Note that while we find OpenAI models to perform better on average, their license restricts commercial use of generated data for training models similar to ChatGPT.
76
+
77
+ If you’re interested in more details on how our dataset generator works, read more or run it here.
78
+
79
+ ## Fine-tuned a custom model and hosted it 🎉
80
+ We have used the above pipeline to generate a filtered dataset having around 37k questions and responses samples. But that's not all! We've also fine-tuned a language model based on EleutherAI’s pythia model. It is hosted on Hugging-Face website as lamini/instruct-tuned-2.8b and is available for use under CC-BY license here. This model is optimized for generating accurate and relevant responses to instruction-based tasks, making it perfect for tasks like question answering, code autocomplete, and chatbots. Feel free to run queries by yourself on our playground!!
81
+
82
+ # Pushing the boundaries of fast & usable generative AI
83
+ We’re excited to dramatically improve the performance of training LLMs and expand who is able to train them. These two frontiers are intertwined: with faster, more effective iteration cycles more people will be able to build these models, beyond just fiddling with prompts.
84
+
85
+ Team++: We are growing our team with people who are passionate about making it possible to build LLMs 10x faster and making them widely accessible to empower new, extraordinary use cases. If that’s you, please apply here) 🤝
86
+
87
+ ‍At Lamini, our mission is to help all software engineers build and ship their own production-grade large language models. Lamini is the world's most powerful LLM engine, unlocking the power of generative AI for every company by putting their data to work. The future of software is powered by LLMs, driven forward by data and compute scaling laws.