This repository contains quantized T5-large model. This model is 5x lesser than the T5-large model and also the inference time has been reduced by 3x on CPU, whilst giving impressive results.

Downloads last month
26
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support