Back to home
Posts, articles, and discussions

How to train a new language model from scratch using Transformers and Tokenizers

February-May, 2020

How to generate text: using different decoding methods for language generation with Transformers

March, 2020

Zero Shot Topic Classification

May 29, 2020 joeddav

How Big Should My Language Model Be?

June 8, 2020 teven

Long Form Question Answering with ELI5

June 17, 2020 yjernite

The Reformer - Pushing the limits of language modeling

July 3, 2020

Block Sparse Matrices for Smaller and Faster Language Models

Sep 10, 2020

Retrieval Augmented Generation (RAG)

September 28, 2020 yjernite

Transformer-based Encoder-Decoder Models

October 10, 2020

Hyperparameter Search with Transformers and Ray Tune

November 2, 2020 ray-project guest

Porting fairseq wmt19 translation system to transformers

November 3, 2020 stas guest

Leveraging Pre-trained Language Model Checkpoints for Encoder-Decoder Models

November 09, 2020