Adds missing tokenizer configuration file

#8
by lysandre HF staff - opened

This repository is missing the tokenizer configuration file, and is instead relying on some attributes set within
the transformers library directly in order to correctly tokenize inputs.

In order to ensure repositories don't depend on internal configuration changes, we're removing these attribute maps
in transformers#29112.

In doing so, we see that the following attributes are currently missing from the configuration and would be
ill-configured without this PR:

{'model_max_length': 512}

This PR aims to add these attributes and their values to the tokenizer config file.
This will proceed to make this repository more robust by ensuring that:

  • the repository does not depend on intra-library code
  • clones of this repository continue working as expected even without the correct repository name
  • other libraries that would like to leverage this repository do not depend on code within the transformers library

Thanks πŸ€—

lysandre changed pull request status to open
wissamantoun changed pull request status to merged

Sign up or log in to comment