Full-text search
+ 1,000 results
nbroad / muril-bigbird-base
README.md
model
3 matches
Norod78 / distilgpt2-base-pretrained-he
README.md
model
3 matches
tags:
transformers, pytorch, tf, jax, coreml, onnx, safetensors, gpt2, text-generation, he, license:mit, autotrain_compatible, endpoints_compatible, has_space, text-generation-inference, region:us
16
17
18
19
20
the [TPU Research Cloud](https://sites.research.google/trc/) Program. Then was further fine-tuned on GPU.
## Dataset
### oscar (unshuffled deduplicated he) - [Homepage](https://oscar-corpus.com) | [Dataset Permalink](https://huggingface.co/datasets/viewer/?dataset=oscar&config=unshuffled_deduplicated_he)
adalbertojunior / image_captioning_portuguese
README.md
model
3 matches
jcrbsa / pt-gpt2vit
README.md
model
3 matches
gwlms / teams-base-dewiki-v1-generator
README.md
model
3 matches
tags:
transformers, pytorch, tensorboard, safetensors, electra, fill-mask, de, dataset:gwlms/dewiki-20230701-chunks, license:mit, autotrain_compatible, endpoints_compatible, region:us
15
16
e's [TPU Research Cloud](https://sites.research.google/trc/about/) (TRC).
Many Thanks for providing access to the TPUs ❤️
gwlms / teams-base-dewiki-v1-discriminator
README.md
model
3 matches
gwlms / t5-efficient-small-dewiki-v1
README.md
model
3 matches
tags:
transformers, pytorch, tensorboard, safetensors, t5, text2text-generation, de, dataset:gwlms/dewiki-20230701-chunks, license:mit, autotrain_compatible, endpoints_compatible, text-generation-inference, region:us
15
16
e's [TPU Research Cloud](https://sites.research.google/trc/about/) (TRC).
Many Thanks for providing access to the TPUs ❤️
gwlms / bert-base-dewiki-v1
README.md
model
3 matches
tags:
transformers, pytorch, tensorboard, safetensors, bert, fill-mask, de, dataset:gwlms/dewiki-20230701-chunks, license:mit, autotrain_compatible, endpoints_compatible, region:us
15
16
e's [TPU Research Cloud](https://sites.research.google/trc/about/) (TRC).
Many Thanks for providing access to the TPUs ❤️
gwlms / t5-efficient-large-dewiki-v1
README.md
model
3 matches
tags:
transformers, pytorch, jax, tensorboard, safetensors, t5, text2text-generation, de, dataset:gwlms/dewiki-20230701-chunks, license:mit, autotrain_compatible, endpoints_compatible, text-generation-inference, region:us
15
16
e's [TPU Research Cloud](https://sites.research.google/trc/about/) (TRC).
Many Thanks for providing access to the TPUs ❤️
gwlms / t5-efficient-base-dewiki-v1
README.md
model
3 matches
tags:
transformers, pytorch, jax, tensorboard, safetensors, t5, text2text-generation, de, dataset:gwlms/dewiki-20230701-chunks, license:mit, autotrain_compatible, endpoints_compatible, text-generation-inference, region:us
15
16
e's [TPU Research Cloud](https://sites.research.google/trc/about/) (TRC).
Many Thanks for providing access to the TPUs ❤️
gwlms / bert-base-token-dropping-dewiki-v1
README.md
model
3 matches
tags:
transformers, pytorch, tensorboard, safetensors, bert, fill-mask, de, dataset:gwlms/dewiki-20230701-chunks, license:mit, autotrain_compatible, endpoints_compatible, region:us
15
16
e's [TPU Research Cloud](https://sites.research.google/trc/about/) (TRC).
Many Thanks for providing access to the TPUs ❤️
gwlms / byt5-small-dewiki-v1
README.md
model
3 matches
tags:
transformers, pytorch, safetensors, t5, text2text-generation, de, dataset:gwlms/dewiki-20230701-chunks, license:mit, autotrain_compatible, endpoints_compatible, text-generation-inference, region:us
15
16
e's [TPU Research Cloud](https://sites.research.google/trc/about/) (TRC).
Many Thanks for providing access to the TPUs ❤️
sr5434 / AlphaZero-Kuhn-Poker
README.md
model
3 matches
team-lucid / hubert-large-korean
README.md
model
3 matches
tags:
transformers, pytorch, jax, safetensors, hubert, feature-extraction, speech, audio, automatic-speech-recognition, custom_code, ko, arxiv:2106.07447, license:apache-2.0, endpoints_compatible, has_space, region:us
19
20
21
22
23
구글의 TPU Research Cloud(TRC)를 통해 지원받은 Cloud TPU로 학습되었습니다.
### Model Description
<table>
team-lucid / deberta-v3-xlarge-korean
README.md
model
3 matches
AIBunCho / japanese-novel-gpt-j-6b
README.md
model
3 matches
yhavinga / t5-base-dutch
README.md
model
3 matches
tags:
transformers, pytorch, tf, jax, tensorboard, t5, text2text-generation, seq2seq, nl, dataset:yhavinga/mc4_nl_cleaned, arxiv:1910.10683, arxiv:2109.10686, license:apache-2.0, autotrain_compatible, has_space, text-generation-inference, region:us
193
194
195
196
197
[TPU Research Cloud](https://sites.research.google/trc/). The HuggingFace 🤗 ecosystem was instrumental in all parts
of the training. Weights & Biases made it possible to keep track of many training sessions
and orchestrate hyper-parameter sweeps with insightful visualizations.
The following repositories where helpful in setting up the TPU-VM,
and getting an idea what sensible hyper-parameters are for training gpt2 from scratch:
team-lucid / deberta-v3-base-korean
README.md
model
3 matches
Galuh / id-journal-gpt2
README.md
model
6 matches
tags:
transformers, pytorch, jax, tensorboard, gpt2, text-generation, id, autotrain_compatible, endpoints_compatible, has_space, text-generation-inference, region:us
8
9
10
11
12
by [TPU Research Cloud](https://sites.research.google/trc/).
The demo can be found [here](https://huggingface.co/spaces/flax-community/gpt2-indonesian).
## How to use
Norod78 / hebrew-gpt_neo-small
README.md
model
3 matches
tags:
transformers, pytorch, jax, onnx, safetensors, gpt_neo, text-generation, he, license:mit, autotrain_compatible, endpoints_compatible, has_space, region:us
16
17
18
19
20
the [TPU Research Cloud](https://sites.research.google/trc/) Program.
## Datasets
1. An assortment of various Hebrew corpuses - I have made it available [here](https://mega.nz/folder/CodSSA4R#4INvMes-56m_WUi7jQMbJQ)