Luciano Santa Brรญgida's picture

Luciano Santa Brรญgida

lucianosb

AI & ML interests

LLM for pt-br (text generation, translation and classification), Image Generation and Image Classification.

Recent Activity

Organizations

ZeroGPU Explorers's profile picture Journalists on Hugging Face's profile picture Social Post Explorers's profile picture

lucianosb's activity

reacted to ImranzamanML's post with ๐Ÿง ๐Ÿ”ฅ 4 months ago
view post
Post
1758
Today lets discuss about 32-bit (FP32) and 16-bit (FP16) floating-point!

Floating-point numbers are used to represent real numbers (like decimals) and they consist of three parts:

Sign bit: 
Indicates whether the number is positive (0) or negative (1).
Exponent:
Determines the scale of the number (i.e., how large or small it is by shifting the decimal point).
Mantissa (or fraction): 
Represents the actual digits of the number.

32-bit Floating Point (FP32)
Total bits: 32 bits
Sign bit: 1 bit
Exponent: 8 bits
Mantissa: 23 bits
For example:
A number like -15.375 would be represented as:
Sign bit: 1 (negative number)
Exponent: Stored after being adjusted by a bias (127 in FP32).
Mantissa: The significant digits after converting the number to binary.

16-bit Floating Point (FP16)
Total bits: 16 bits
Sign bit: 1 bit
Exponent: 5 bits
Mantissa: 10 bits
Example:
A number like -15.375 would be stored similarly:
Sign bit: 1 (negative number)
Exponent: Uses 5 bits, limiting the range compared to FP32.
Mantissa: Only 10 bits for precision.

Precision and Range
FP32: Higher precision and larger range, with about 7 decimal places of accuracy.
FP16: Less precision (around 3-4 decimal places), smaller range but faster computations and less memory use.
ยท
reacted to m-ric's post with โค๏ธ 6 months ago
view post
Post
640
> ๐—ช๐—ฎ๐—ป๐˜ ๐˜๐—ผ ๐—ธ๐—ป๐—ผ๐˜„ ๐—ต๐—ผ๐˜„ ๐—บ๐˜‚๐—ฐ๐—ต ๐—ฎ๐—ป ๐—”๐—ฃ๐—œ ๐—Ÿ๐—Ÿ๐—  ๐—ฐ๐—ฎ๐—น๐—น ๐—ฐ๐—ผ๐˜€๐˜๐˜€ ๐˜†๐—ผ๐˜‚?

I've just made this Space that gets you the API price for any LLM call, for nearly all inference providers out there!

This is based on a comment by @victor under my HF Post a few months back, and leverages BerriAI's data for LLM prices.

Check it out here ๐Ÿ‘‰ m-ric/text_to_dollars
reacted to louisbrulenaudet's post with ๐Ÿ”ฅ 6 months ago
view post
Post
2610
The Romulus model series has been released on Hugging Face, continually pre-trained on 34,864,949 tokens of French laws and intended to serve as a foundation for fine-tuning on labeled data ๐Ÿค—

The training code, dataset and model weights are open and available free on HF and the training was based on H100 provided by Microsoft for Startups using Unsloth AI by @danielhanchen and @shimmyshimmer ๐Ÿฆฅ

Link to the base model: louisbrulenaudet/Romulus-cpt-Llama-3.1-8B-v0.1

Link to the instruct model: louisbrulenaudet/Romulus-cpt-Llama-3.1-8B-v0.1-Instruct

Link to the dataset: louisbrulenaudet/Romulus-cpt-fr

Please note that these models have not been aligned for the production of usable texts as they stand, and will certainly need to be refined for the desired tasks in order to produce satisfactory results.
  • 1 reply
ยท
replied to enzostvs's post 6 months ago
view reply

Mine was "the king of low-key accomplishments. 24 followers? That's cute. You've got more models than fans."

LOL

reacted to enzostvs's post with ๐Ÿ‘€ 6 months ago
view post
Post
3587
What if we asked the AI what it thought of our hugging face profile? ๐Ÿ‘น
I've released a new space capable of doing it.... watch out, it hits hard! ๐ŸฅŠ

Try it now โžก๏ธ enzostvs/hugger-roaster

Share your roast below ๐Ÿ‘‡
ยท
replied to charlesdedampierre's post 6 months ago
reacted to charlesdedampierre's post with ๐Ÿ”ฅ 6 months ago
view post
Post
4172
Please check the Open Source AI Network: we mapped the top 500 HF users
based on their followers' profiles.

The map can be found here: bunkalab/mapping_the_OS_community
  • 1 reply
ยท
reacted to elinas's post with ๐Ÿ‘€ 6 months ago
view post
Post
2383
We conducted an experiment in an effort to revive LLaMA 1 33B as it had unique prose and a lack of "GPT-isms" and "slop" in its pretraining data, as well as being one of the favorites at the time. With multiple finetune runs, we were able to extend the model from it's pretrained base of 2048 to ~12,000 tokens adding approx. 500M tokens in the process. The effective length is 16,384 but it's better to keep it on the lower range. It writes well and in multiple formats. In the future, we have some ideas like implementing GQA. Please take a look and we would love to hear your feedback!

ZeusLabs/Chronos-Divergence-33B
reacted to MonsterMMORPG's post with ๐Ÿ”ฅ 6 months ago
view post
Post
2606
Ultimate FLUX LoRA Training Tutorial: Windows and Cloud Deployment

I have done total 104 different LoRA trainings and compared each one of them to find the very best hyper parameters and the workflow for FLUX LoRA training by using Kohya GUI training script.

You can see all the done experimentsโ€™ checkpoint names and their repo links in following public post: https://www.patreon.com/posts/110838414

After completing all these FLUX LoRA trainings by using the most VRAM optimal and performant optimizer Adafactor I came up with all of the following ranked ready to use configurations.

You can download all the configurations, all research data, installers and instructions at the following link : https://www.patreon.com/posts/110879657


Tutorials
I also have prepared 2 full tutorials. First tutorial covers how to train and use the best FLUX LoRA locally on your Windows computer : https://youtu.be/nySGu12Y05k

This is the main tutorial that you have to watch without skipping to learn everything. It has total 74 chapters, manually written English captions. It is a perfect resource to become 0 to hero for FLUX LoRA training.

The second tutorial I have prepared is for how to train FLUX LoRA on cloud. This tutorial is super extremely important for several reasons. If you donโ€™t have a powerful GPU, you can rent a very powerful and very cheap GPU on Massed Compute and RunPod. I prefer Massed Compute since it is faster and cheaper with our special coupon SECourses. Another reason is that in this tutorial video, I have fully in details shown how to train on a multiple GPU setup to scale your training speed. Moreover, I have shown how to upload your checkpoints and files ultra fast to Hugging Face for saving and transferring for free. Still watch first above Windows tutorial to be able to follow below cloud tutorial : https://youtu.be/-uhL2nW7Ddw

For upscaling SUPIR used : https://youtu.be/OYxVEvDf284
reacted to Tonic's post with ๐Ÿš€ 6 months ago
view post
Post
2664
So awesome , now i can deploy a jupyterlab on huggingface and deploy gradio from the jupyterlab
reacted to fdaudens's post with ๐Ÿ‘ 6 months ago
view post
Post
1284
A lot of coverage of the Apple event! Iโ€™ve selected a few unique angles and distinctive takes.

**The NYT**
- "The iPhoneโ€™s limited feature set is emblematic of how Apple is taking a cautious approach to generative A.I."
- "Wall Street is enthusiastic about the artificially intelligent phones, with analysts predicting the features could help Apple sell a record 240 million iPhones next year."

**The Guardian**
- "Despite the bells and whistles, and being a tech-adopting lot, I bet many of you wonโ€™t be lining up to buy it."
- One reason is the simple cost of the iPhone 16, which starts at $799.
- The adoption of AI into the iPhone could be considered a step change in how the iPhone works. But there may not be a huge hankering to use ChatGPT on your phone."

**The WSJ**
- Apple didnโ€™t say when the AI services would be available in China, its second-largest market after the U.S.
- The delay puts the iPhone maker at a disadvantage against rivals offering AI services
- Huawei held its own announcement in China to release the Mate XT, a three-way foldable smartphone with AI features.
- Apple said that the launch of Apple Intelligence was subject to regulatory approval. In China, any generative AI models that could influence public opinion need government approval.

**CNN**
- "For an event built around unveiling Appleโ€™s first AI-powered iPhone, there was one striking absence over the two-hour presentation: the words 'artificial intelligence.'"
- "But Apple understands something that often gets lost in the bot-pilled bubble of Silicon Valley: Regular people donโ€™t trust AI."

Links:
https://www.nytimes.com/2024/09/09/technology/apple-event-iphone-16-watch.html
https://www.theguardian.com/technology/article/2024/sep/10/techscape-iphone-16-cost-features
https://www.wsj.com/tech/apples-challenge-in-china-rises-with-new-rival-phones-and-ai-delay-8cf871fb?mod=rss_Technology
https://www.cnn.com/2024/09/10/business/apple-iphone-ai-nightcap/
posted an update 7 months ago
view post
Post
1338
Made a demo for all my Brazil XL LoRA models so far. Use it for free at lucianosb/brazilxl-demo

Brazil XL is an initiative that brings better representations of Brazilian culture to Stable Diffusion. I started this when I noticed some keywords would not generate the desired subject on any base model, so I trained my own models and I'm sharing them with the HF community.

I'll keep updating the space as new models get trained on the following months.
reacted to fdaudens's post with ๐Ÿš€ 7 months ago
view post
Post
2290
๐Ÿš€ Introducing the Model Drops Tracker! ๐Ÿ•ต๏ธโ€โ™‚๏ธ

Feeling overwhelmed by the AI model release frenzy? ๐Ÿคฏ You're not alone!

I built this simple tool to help us all keep up:
- Filter recent models from the ๐Ÿค— Hub
- Set minimum likes threshold
- Choose how recent you want to go

Try it out and let me know what you think: fdaudens/Model-Drops-Tracker

Any features you'd like to see added?
#AIModels
ยท
reacted to reach-vb's post with ๐Ÿ”ฅ 8 months ago
view post
Post
3367
What an eventful day in Open Source LLMs today:

Mistral released Codestral Mamba ๐Ÿ
> Beats DeepSeek QwenCode, best model < 10B, competitive with Codestral 22B
> Mamba 2 architecture - supports up to 256K context
> Apache 2.0 licensed, perfect for local code assistant
> Transformers & llama.cpp integration upcoming!

Model checkpoint: https://huggingface.co/mistralai/mamba-codestral-7B-v0.1

Hugging Face dropped SmolLM ๐Ÿค
> Beats MobileLLM, Qwen 0.5B, Phi 1.5B and more!
> 135M, 360M, and 1.7B param model checkpoints
> Trained on 600B high-quality synthetic + FineWeb Edu tokens
> Architecture: Llama + GQA + 2048 ctx length
> Ripe for fine-tuning and on-device deployments.
> Works out of the box with Transformers!

Model checkpoints: HuggingFaceTB/smollm-6695016cad7167254ce15966

Mistral released Mathstral 7B โˆ‘
> 56.6% on MATH and 63.47% on MMLU
> Same architecture as Mistral 7B
> Works out of the box with Transformers & llama.cpp
> Released under Apache 2.0 license

Model checkpoint: https://huggingface.co/mistralai/mathstral-7B-v0.1

Pretty dope day for open source ML. Can't wait to see what the community builds with it and to support them further! ๐Ÿค—

What's your favourite from the release today?
  • 1 reply
ยท
reacted to appvoid's post with โค๏ธ 8 months ago
view post
Post
1499
palmer-004 becomes ๐Ÿ”ฅturbo๐Ÿ”ฅ now is half the size, twice the speed and the best overall 0.5b language model in huggingface.

appvoid/palmer-004-turbo
  • 1 reply
ยท
reacted to davanstrien's post with ๐Ÿ‘€ 8 months ago
view post
Post
1690
Using the new viewer iframe support for the datasets viewer, I built a simple Space davanstrien/collection_dataset_viewer to quickly explore all the datasets inside a collection.

The collection is loaded from an environment variable, so you can duplicate this Space to create a Space for exploring datasets in another collection!
reacted to Severian's post with ๐Ÿš€ 8 months ago
view post
Post
3429
GraphRAG-Ollama-UI

I've been working on a local version of Microsoft's GraphRAG that uses Ollama for everything. It's got a new interactive UI built with Gradio that makes it easier to manage data, run queries, and visualize results. It's not fully featured or set up to harness the entire GraphRAG library yet but it allows you to run all the standard commands for Indexing/Processing and chatting with your graph. Some key features:

Uses local models via Ollama for LLM and embeddings

3D graph visualization of the knowledge graph using Plotly

File management through the UI (upload, view, edit, delete)

Settings management in the interface

Real-time logging for debugging

https://github.com/severian42/GraphRAG-Ollama-UI
  • 1 reply
ยท
posted an update 8 months ago