Pradeep T
Pradeep1995
AI & ML interests
None yet
Recent Activity
new activity
about 1 month ago
deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B:Step by step guide for Distillation
new activity
about 2 months ago
deepseek-ai/DeepSeek-R1:Transformer version required?
new activity
12 months ago
text-generation-inference/Mixtral-8x7B-Instruct-v0.1-medusa:how to use this model on sagemaker endpoints
Organizations
None yet
Pradeep1995's activity
Step by step guide for Distillation
#23 opened about 1 month ago
by
Pradeep1995

Transformer version required?
#24 opened about 2 months ago
by
Pradeep1995

how to use this model on sagemaker endpoints
2
#1 opened about 1 year ago
by
LorenzoCevolaniAXA

What is the actual context size of google/gemma-7b model
1
#81 opened 12 months ago
by
Pradeep1995

What is the actual context size of mistralai/Mixtral-8x7B-Instruct-v0.1 model
3
#186 opened 12 months ago
by
Pradeep1995

PEFT based Fine Tuned model hallucinates values from the fine tuning training data while inferencing.
7
#111 opened about 1 year ago
by
Pradeep1995

Special token( </s>) not generating in the model.generate() method
7
#47 opened about 1 year ago
by
Pradeep1995

Can we save the finetuned Mistral model by exporting to TorchScript
1
#46 opened about 1 year ago
by
Pradeep1995

What is the best way for the inference process in LORA in PEFT approach
8
#70 opened about 1 year ago
by
Pradeep1995

What is the best way for the inference process in LORA in PEFT approach
#3 opened about 1 year ago
by
Pradeep1995

What is the best way for the inference process in LORA in PEFT approach
#53 opened about 1 year ago
by
Pradeep1995

What is the best way for the inference process in LORA in PEFT approach
#43 opened about 1 year ago
by
Pradeep1995

What is the best way for the inference process in LORA in PEFT approach
#96 opened about 1 year ago
by
Pradeep1995

Which is the actual way to store the adapters after PEFT finetuning
4
#67 opened about 1 year ago
by
Pradeep1995

Which is the actual way to store the Adapter after PEFT finetuning
#42 opened about 1 year ago
by
Pradeep1995

should we follow the same openchat prompt structure while finetuning time?
3
#38 opened about 1 year ago
by
Pradeep1995

PEFT based Fine Tuned model hallucinates values from the fine tuning training data while inferencing
1
#39 opened about 1 year ago
by
Pradeep1995

Incomplete Output even with max_new_tokens
12
#107 opened about 1 year ago
by
Pradeep1995

should we follow the same mistral prompt structure while finetuning time?
#110 opened about 1 year ago
by
Pradeep1995

Incomplete Output even with max_new_tokens
1
#37 opened about 1 year ago
by
Pradeep1995
