AI & ML interests

Arabic NLP, CV, AI Safety and ethics

Recent Activity

asas-ai's activity

alielfilali01ย 
posted an update 13 days ago
view post
Post
3301
Unpopular opinion: Open Source takes courage to do !

Not everyone is brave enough to release what they have done (the way they've done it) to the wild to be judged !
It really requires a high level of "knowing wth are you doing" ! It's kind of a super power !

Cheers to the heroes here who see this!
ยท
alielfilali01ย 
posted an update 17 days ago
view post
Post
1477
Apparently i forgot to put this here !

Well, this is a bit late but consider given our recent blog a read if you are interested in Evaluation.

You don't have to be into Arabic NLP in order to read it, the main contribution we are introducing is a new evaluation measure for NLG. We made the fisrt application of this measure on Arabic for now and we will be working with colleagues from the community to expand it to other languages.

Blog:
Rethinking LLM Evaluation with 3C3H: AraGen Benchmark and Leaderboard
https://huggingface.co/blog/leaderboard-3c3h-aragen

Space:
inceptionai/AraGen-Leaderboard

Give it a read and let me know your thoughts ๐Ÿค—
alielfilali01ย 
posted an update about 1 month ago
view post
Post
2175
Unpopular opinion : o1-preview is more stupid than 4o and Qwen2.5-72B-Instruct in extremely underrated !
  • 2 replies
ยท
alielfilali01ย 
posted an update 2 months ago
view post
Post
1699
I feel like this incredible resource hasn't gotten the attention it deserves in the community!

@clefourrier and generally the HuggingFace evaluation team put together a fantastic guidebook covering a lot about ๐—˜๐—ฉ๐—”๐—Ÿ๐—จ๐—”๐—ง๐—œ๐—ข๐—ก from basics to advanced tips.

link : https://github.com/huggingface/evaluation-guidebook

I havenโ€™t finished it yet, but i'am enjoying every piece of it so far. Huge thanks @clefourrier and the team for this invaluable resource !
  • 3 replies
ยท
alielfilali01ย 
posted an update 3 months ago
view post
Post
1825
Why nobdoy is talking about the new training corpus released by MBZUAI today.

TxT360 is +15 Trillion tokens corpus outperforming FineWeb on several metrics. Ablation studies were done up to 1T tokens.

Read blog here : LLM360/TxT360
Dataset : LLM360/TxT360
  • 2 replies
ยท
mmhamdyย 
posted an update 3 months ago
view post
Post
1871
๐Ÿ”— Evaluating Long Context #1: Long Range Arena (LRA)

Accurately evaluating how well language models handle long contexts is crucial, but it's also quite challenging to do well. In this series of posts, we're going to examine the various benchmarks that were proposed to assess long context understanding, starting with Long Range Arens (LRA)

Introduced in 2020, Long Range Arens (LRA) is one of the earliest benchmarks designed to tackle the challenge of long context evaluation.

๐Ÿ“Œ Key Features of LRA

1๏ธโƒฃ Diverse Tasks: The LRA benchmark consists of a suite of tasks designed to evaluate model performance on long sequences ranging from 1,000 to 16,000 tokens. These tasks encompass different data types and modalities: Text, Natural and Synthetic Images, and Mathematical Expressions.

2๏ธโƒฃ Synthetic and Real-world Tasks: LRA is comprised of both synthetic probing tasks and real-world tasks.

3๏ธโƒฃ Open-Source and Extensible: Implemented in Python using Jax and Flax, the LRA benchmark code is publicly available, making it easy to extend.

๐Ÿ“Œ Tasks

1๏ธโƒฃ Long ListOps

2๏ธโƒฃ Byte-level Text Classification and Document Retrieval

3๏ธโƒฃ Image Classification

4๏ธโƒฃ Pathfinder and Pathfinder-X (Long-range spatial dependency)

๐Ÿ‘จโ€๐Ÿ’ป Long Range Arena (LRA) Github Repository: https://github.com/google-research/long-range-arena

๐Ÿ“„ Long Range Arena (LRA) paper: Long Range Arena: A Benchmark for Efficient Transformers (2011.04006)
alielfilali01ย 
posted an update 3 months ago
view post
Post
2569
Don't you think we should add a tag "Evaluation" for datasets that are meant to be benchmarks and not for training ?

At least, when someone is collecting a group of datasets from an organization or let's say the whole hub can filter based on that tag and avoid somehow contaminating their "training" data.
alielfilali01ย 
posted an update 3 months ago
view post
Post
868
We need a fork feature for models and datasets similar to "Duplicate this space" in spaces ! Don't you think ?

Sometimes you just want to save something in your profile privately and work on it later without the hassle of "load_.../push_to_hub" in a code file.

I know this is super lazy ๐Ÿ˜… But it is what it is ...

tag : @victor
ยท