Abrosimov's picture
1

Abrosimov

ajiriro
Β·

AI & ML interests

None yet

Recent Activity

updated a collection 12 days ago
Interesting
updated a collection 12 days ago
Interesting
updated a collection 4 months ago
Interesting
View all activity

Organizations

None yet

ajiriro's activity

replied to bartowski's post 6 months ago
view reply

Yes, exactly. When converting from float16 to float32 for fine-tuning (as I thought), we need to fill 13 bits of the mantissa and 3 bits of the exponent with zeros, rather than simply filling the last 16 bits.

replied to bartowski's post 7 months ago
view reply

I don't understand much about this, but maybe the model in F32 is just redundant. Maybe the other half of most weights are filled with zeros. It was scaled this way to fine-tune it or to make it impossible for people with few resources to run it😁