Posts, articles, and discussions

Fast Inference on Large Language Models: BLOOMZ on Habana Gaudi2 Accelerator
By March 28, 2023

Distributed Training: Train BART/T5 for Summarization using 🤗 Transformers and Amazon SageMaker
By April 8, 2021
How many data points is a prompt worth?

By April 5, 2021

Understanding BigBird's Block Sparse Attention
By March 31, 2021 guest

The Partnership: Amazon SageMaker and Hugging Face
By March 23, 2021
My Journey to a serverless transformers pipeline on Google Cloud
By March 18, 2021 guest

Fine-Tune Wav2Vec2 for English ASR with 🤗 Transformers
By March 12, 2021

Hugging Face Reads, Feb. 2021 - Long-range Transformers
By March 9, 2021

Simple considerations for simple people building fancy neural networks
By February 25, 2021

Retrieval Augmented Generation with Huggingface Transformers and Ray
By February 10, 2021 guest

Hugging Face on PyTorch / XLA TPUs
By February 9, 2021 guest

Faster TensorFlow models in Hugging Face Transformers
By January 26, 2021

Fit More and Train Faster With ZeRO via DeepSpeed and FairScale
By January 19, 2021

How we sped up transformer inference 100x for 🤗 API customers
By January 18, 2021

Leveraging Pre-trained Language Model Checkpoints for Encoder-Decoder Models
By November 9, 2020