view article Article π€ PEFT: Parameter-Efficient Fine-Tuning of Billion-Scale Models on Low-Resource Hardware Feb 10, 2023 β’ 61
view post Post 1621 π’ The LLaMA-3.1-8B distilled 8B version of the R1 DeepSeek AI is available besides the one based on Qwenπ Notebook for using it in reasoning over series of data π§ :https://github.com/nicolay-r/nlp-thirdgate/blob/master/tutorials/llm_deep_seek_7b_distill_llama3.ipynbLoading using the pipeline API of the transformers library:https://github.com/nicolay-r/nlp-thirdgate/blob/master/llm/transformers_llama.pyπ‘ GPU Usage: 12.3 GB (FP16/FP32 mode) which is suitable for T4. (a 1.5 GB less than Qwen-distilled version)π Perfomance: T4 instance: ~0.19 tokens/sec (FP32 mode) and (FP16 mode) ~0.22-0.30 tokens/sec. Is it should be that slow? π€Model name: deepseek-ai/DeepSeek-R1-Distill-Llama-8Bβ Framework: https://github.com/nicolay-r/bulk-chainπ Notebooks and models hub: https://github.com/nicolay-r/nlp-thirdgate See translation π₯ 7 7 + Reply
view post Post 1335 π¨ MistralAI is back with the mistral small V3 model update and it is free! πhttps://docs.mistral.ai/getting-started/models/models_overview/#free-modelsπ Below is the the provider for reasoning over your dataset rows with custom schema π§ https://github.com/nicolay-r/nlp-thirdgate/blob/master/llm/mistralai_150.pyMy personal usage experience and findings:β οΈThe original API usage may constanly fail with the connection.To bypass this limitation, use --attempts [COUNT] to withstand connection loss while iterating through JSONL/CSV data (see π· below)π΅ It is actually: ~0.18 USD 1M tokensπ Framework: https://github.com/nicolay-r/bulk-chain See translation π₯ 3 3 + Reply