Ibrahim AbuAlhaol PRO

iabualhaol
Ā·

AI & ML interests

None yet

Recent Activity

liked a model 6 days ago
unsloth/Qwen2.5-3B-Instruct
updated a dataset 6 days ago
iabualhaol/Mixtral-Quran_example
updated a model 17 days ago
iabualhaol/bert
View all activity

Organizations

Arabic Machine Learning 's profile picture

iabualhaol's activity

updated a model 17 days ago
reacted to m-ric's post with šŸ¤— 18 days ago
view post
Post
2173
š—£š—¼š˜š—²š—»š˜š—¶š—®š—¹ š—½š—®š—暝—®š—±š—¶š—“š—ŗ š˜€š—µš—¶š—³š˜ š—¶š—» š—Ÿš—Ÿš— š˜€: š—»š—²š˜„ š—½š—®š—½š—²š—æ š—Æš˜† š— š—²š˜š—® š—°š—¹š—®š—¶š—ŗš˜€ š˜š—µš—®š˜ š˜„š—² š—°š—®š—» š—“š—²š˜ š—暝—¶š—± š—¼š—³ š˜š—¼š—øš—²š—»š—¶š˜‡š—²š—暝˜€! šŸ„³

Current LLMs process text by first splitting it into tokens. They use a module named "tokenizer", that -spl-it-s- th-e- te-xt- in-to- arbitrary tokens depending on a fixed dictionnary.
On the Hub you can find this dictionary in a model's files under tokenizer.json.

āž”ļø This process is called BPE tokenization. It is suboptimal, everyone says it. It breaks text into predefined chunks that often fail to capture the nuance of language. But it has been a necessary evil in language models since their inception.

šŸ’„ In Byte Latent Transformer (BLT), Meta researchers propose an elegant solution by eliminating tokenization entirely, working directly with raw bytes while maintaining efficiency through dynamic "patches."

This had been tried before with different byte-level tokenizations, but it's the first time that an architecture of this type scales as well as BPE tokenization. And it could mean a real paradigm shift! šŸ‘šŸ‘

šŸ—ļø š—”š—暝—°š—µš—¶š˜š—²š—°š˜š˜‚š—暝—²:
Instead of a lightweight tokenizer, BLT has a lightweight encoder that process raw bytes into patches. Then the patches are processed by the main heavy-duty transformers as we do normally (but for patches of bytes instead of tokens), before converting back to bytes.

šŸ§© š——š˜†š—»š—®š—ŗš—¶š—° š—£š—®š˜š—°š—µš—¶š—»š—“:
Instead of fixed tokens, BLT groups bytes based on their predictability (measured by entropy) - using more compute for complex sequences and efficiently handling simple ones. This allows efficient processing while maintaining byte-level understanding.

I hope this breakthrough is confirmed and we can get rid of all the tokenizer stuff, it will make model handling easier!

Read their paper here šŸ‘‰ https://dl.fbaipublicfiles.com/blt/BLT__Patches_Scale_Better_Than_Tokens.pdf
  • 2 replies
Ā·
liked a Space 3 months ago
updated a Space over 1 year ago
liked a Space over 1 year ago
updated a Space over 1 year ago