LLM_BENCHMARKS_TEXT = f""" # 🧰 Train a Model Intel offers a variety of platforms that can be used to train LLMs including data center and consumer grade CPUs, GPUs, and ASICs. Below, you can find documentation on how to access free and paid resources to train a model on Intel hardware and submit it to the Hugging Face Model Hub.
## Intel Developer Cloud - Quick Start The Intel Developer Cloud is one of the best places to access free and paid compute instances for model training. Intel offers Jupyter Notebook instances supported by 224 Core 4th Generation Xeon Bare Metal nodes with 4x GPU Max Series 1100. To access these resources please follow the instructions below: 1. Visit the [Intel Developer Cloud](https://bit.ly/inteldevelopercloud) and sign up for the "Standard - Free" tier to get started. 2. Navigate to the "Training" module under the "Software" section in the left panel. 3. Under the GenAI Essentials section, select the [Gemma Model Fine-tuning using SFT and LoRA](https://console.idcservice.net/training/detail/99deeb99-b0c6-4d02-a1d5-a46d95344ff3) notebook and click "Launch". 4. Follow the instructions in the notebook to train your model using Intel® Data Center GPU Max 1100. 5. Upload your model to the Hugging Face Model Hub. 6. Go to the "Submit" tab on this Leaderboard and follow the instructions to submit your model.
## Training Code Samples Below are some resources to get you started on training models on Intel platforms: - Intel® Gaudi® Accelerators - [Parameter Efficient Fine-Tuning of Llama-2 70B](https://github.com/HabanaAI/Gaudi-tutorials/blob/main/PyTorch/llama2_fine_tuning_inference/llama2_fine_tuning_inference.ipynb) - Intel® Xeon® Processors - [Distributed Training of GPT2 LLMs on AWS](https://github.com/intel/intel-cloud-optimizations-aws/tree/main/distributed-training) - [Fine-tuning Falcon 7B on Xeon Processors](https://medium.com/@eduand-alvarez/fine-tune-falcon-7-billion-on-xeon-cpus-with-hugging-face-and-oneapi-a25e10803a53) - Intel® Data Center GPU Max Series - [Gemma Model Fine-tuning using SFT and LoRA](https://console.idcservice.net/training/detail/99deeb99-b0c6-4d02-a1d5-a46d95344ff3) - [LLM Fine-tuning with QLoRA on Max Series GPUs](https://console.idcservice.net/training/detail/159c24e4-5598-3155-a790-2qv973tlm172)
## Submitting your Model to the Hugging Face Model Hub Once your model is trained, it is a straighforward process to upload and open source it on the Hugging Face Model Hub. The commands from a Jupyter notebook are given below: ```python # Logging in to Hugging Face from huggingface_hub import notebook_login, Repository # Login to Hugging Face notebook_login() # Model and Tokenize Loading from transformers import AutoModelForSequenceClassification, AutoTokenizer # Define the path to the checkpoint checkpoint_path = "" # Replace with your checkpoint folder # Load the model model = AutoModelForSequenceClassification.from_pretrained(checkpoint_path) # Load the tokenizer tokenizer = AutoTokenizer.from_pretrained("") #add name of your model's tokenizer on Hugging Face OR custom tokenizer # Save the model and tokenizer model_name_on_hub = "desired-model-name" model.save_pretrained(model_name_on_hub) tokenizer.save_pretrained(model_name_on_hub) # Push to the hub model.push_to_hub(model_name_on_hub) tokenizer.push_to_hub(model_name_on_hub) # Congratulations! Your fine-tuned model is now uploaded to the Hugging Face Model Hub. # You can view and share your model using its URL: https://huggingface.co// ``` Once your model is uploaded, make sure to update your model card, specifying your use of Intel software and hardware. Hugging Face has a great description on [how to build model cards here](https://huggingface.co/docs/hub/en/model-cards). """ SUBMIT_TEXT = f""" # Use the Resource Below to Start Training a Model Today """