# A Swedish Bert model

## Model description

This model follows the Bert Large model architecture as implemented in Megatron-LM framework. It was trained with a batch size of 512 in 600k steps. The model contains following parameters:

Mask token: undefined