Updates incorrect tokenizer configuration file

#2
by lysandre HF staff - opened

This repository contains in an incorrect tokenizer configuration file, and is instead relying on some attributes set within
the transformers library directly in order to correctly tokenize inputs.

In order to ensure repositories don't depend on internal configuration changes, we're removing these attribute maps
in transformers#29112.

In doing so, we see that the following attributes are currently missing from the configuration and would be
ill-configured without this PR:

{'do_lower_case': False, 'word_tokenizer_type': 'mecab', 'subword_tokenizer_type': 'character', 'model_max_length': 512}

This PR aims to add these attributes and their values to the tokenizer config file.
This will proceed to make this repository more robust by ensuring that:

  • the repository does not depend on intra-library code
  • clones of this repository continue working as expected even without the correct repository name
  • other libraries that would like to leverage this repository do not depend on code within the transformers library

Thanks 🤗

lysandre changed pull request status to open
singletongue changed pull request status to merged

Sign up or log in to comment