abdeljalil_elma

abdeljalilELmajjodi

AI & ML interests

None yet

Recent Activity

Organizations

Arabic Machine Learning 's profile picture Tamazight NLP's profile picture AtlasIA's profile picture Moroccan Data Scientists's profile picture ThinkAI's profile picture Sawalni AI's profile picture Tamazight Open Dataset's profile picture

abdeljalilELmajjodi's activity

upvoted an article about 17 hours ago
view article
Article

Introducing EuroBERT: A High-Performance Multilingual Encoder Model

By EuroBERT and 3 others β€’
β€’ 98
reacted to alielfilali01's post with πŸ‘ about 20 hours ago
view post
Post
2047
3C3H AraGen Leaderboard welcomes today deepseek-ai/DeepSeek-V3 and 12 other models (including the late gpt-3.5 πŸ’€) to the ranking of best LLMs in Arabic !


Observations:
- DeepSeek-v3 ranked 3rd and only Open model among the top 5 !

- A 14B open model ( Qwen/Qwen2.5-14B-Instruct) outperforms gpt-3.5-turbo-0125 (from last year). This shows how much we came in advancing and supporting Arabic presence within the LLM ecosystem !

- Contrary to what observed in likelihood-acc leaderboards (like OALL/Open-Arabic-LLM-Leaderboard) further finetuned models like maldv/Qwentile2.5-32B-Instruct actually decreased the performance compared to the original model Qwen/Qwen2.5-32B-Instruct.
It's worth to note that the decrease is statiscally insignificant which imply that at best, the out-domain finetuning do not really hurts the model original capabilities acquired during pretraining.
Previous work addressed this (finetuning VS pretraining) but more investigation in this regard is required (any PhDs here ? This could be your question ...)


Check out the latest rankings: inceptionai/AraGen-Leaderboard