Yuta Hayashibe
Update README.md
c501405
|
raw
history blame
No virus
1.41 kB
metadata
language: ja
tags:
  - t5
  - text2text-generation
  - seq2seq
license: apache-2.0
datasets:
  - mc4
  - wiki40b

t5-base-japanese-web (with Byte-fallback)

Description

megagonlabs/t5-base-japanese-web is a T5 (Text-to-Text Transfer Transformer) model pre-trained on Japanese web texts.
Training codes are available on GitHub.

Corpus

Tokenizer

SentencePiece trained on Japanese Wikipedia

Parameters

It took about 126 hours with TPU v3-8

Related models

License

Apache License 2.0