from unsloth import FastLanguageModel
import torch
from transformers import TextStreamer
from unsloth import FastLanguageModel

max_seq_length = 2048  # Choose any! We auto support RoPE Scaling internally!
dtype = None  # None for auto detection. Float16 for Tesla T4, V100, Bfloat16 for Ampere+
load_in_4bit = True  # Use 4bit quantization to reduce memory usage. Can be False.


class MyModel:
    def __init__(self):
        self.model = null
        self.tokenizer = null

    def load_pretrain_model(self):
        pass
