Lucain Pouget PRO

Wauplin

AI & ML interests

None yet

Articles

Organizations

Posts 5

view post
Post
1864
šŸš€ Just released version 0.24.0 of the šš‘ššžšššššš’šš—ššššššŠššŒššŽ_šš‘ššžšš‹ Python library!

Exciting updates include:
āš” InferenceClient is now a drop-in replacement for OpenAI's chat completion!

āœØ Support for response_format, adapter_id , truncate, and more in InferenceClient

šŸ’¾ Serialization module with a save_torch_model helper that handles shared layers, sharding, naming convention, and safe serialization. Basically a condensed version of logic scattered across safetensors, transformers , accelerate

šŸ“ Optimized HfFileSystem to avoid getting rate limited when browsing HuggingFaceFW/fineweb

šŸ”Ø HfApi & CLI improvements: prevent empty commits, create repo inside resource group, webhooks API, more options in the Search API, etc.

Check out the full release notes for more details:
Wauplin/huggingface_hub#7
šŸ‘€
view post
Post
3258
šŸš€ I'm excited to announce that huggingface_hub's InferenceClient now supports OpenAI's Python client syntax! For developers integrating AI into their codebases, this means you can switch to open-source models with just three lines of code. Here's a quick example of how easy it is.

Why use the InferenceClient?
šŸ”„ Seamless transition: keep your existing code structure while leveraging LLMs hosted on the Hugging Face Hub.
šŸ¤— Direct integration: easily launch a model to run inference using our Inference Endpoint service.
šŸš€ Stay Updated: always be in sync with the latest Text-Generation-Inference (TGI) updates.

More details in https://huggingface.co/docs/huggingface_hub/main/en/guides/inference#openai-compatibility