This is just the encoder weights from "google/t5-v1_1-xl" It takes 11GB down to 4GB.

The script to do the extraction is included here as transform.py

Edit: Now that I have this in a convenient form... I got a chance to test t5-xxl projected down to 2048, vs this t5-xl Surprisingly, even with an untrained projection layer, trivial embedding diversity scores rate the projected xxl version higher than native xl at 2048.

So, while this model will continue to exist as a convenient way to compare.. and possibly as something to use if you are really, really REALLY tight on memory... you are probably best off using t5-xxl whenever you can.

Downloads last month
3
Safetensors
Model size
1.22B params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for opendiffusionai/t5-v1_1-xl-encoder-only

Base model

google/t5-v1_1-xl
Finetuned
(1)
this model