This repository contains quantized T5-large model. This model is 5x lesser than the T5-large model and also the inference time has been reduced by 3x on CPU, whilst giving impressive results.

Downloads last month
12
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.