***ScikitLLM*** is an LLM finetuned on writing references and code for the Scikit-Learn documentation. | |
Features of ScikitLLM includes: | |
* Support for RAG (three chunks) | |
* Sources and quotations using a modified version of the wiki syntax ("<ref>") | |
* Code samples and examples based on the code quoted in the chunks. | |
* Expanded knowledge/familiarity with the Scikit-Learn concepts and documentation. | |
## Training | |
ScikitLLM is based on Mistral-OpenHermes 7B, a pre-existing finetune version of Mistral 7B. OpenHermes already include many desired capacities for the end use, including instruction tuning, source analysis, and native support for the chatML syntax. | |
As a fine-tune of a fine-tune, ScikitLLM has been trained with a lower learning rate than is commonly used in fine-tuning projects. | |