Cursor: Hold my beer.
Me: *Slacking off with colleagues*
Cursor: Ping.
Me: π€―
hf_xet
installed alongside the latest huggingface_hub
WOOHOO!!
medium
size is now available as a power-user featurelarge
(70GB VRAM)βbut this paves the way for:medium
will offer significantly more usage than large
)xlarge
size (141GB VRAM)auto
(future default)medium
large
(current default)large
medium
did you get it to work since?
which provider do you use?
We'll ship a provider="auto"
in the coming days BTW, cc
@sbrandeis
@Wauplin
@celinah
In the meantime, the model is served by those providers, you can use one of them, for instance, add provider="novita"
to your code:
Hey, things have been in flux somewhat, but they should stabilize now. Sorry about the moving parts!
More details, from @michellehbn :
In February, Inference billing usage had been a fixed rate while we added pay-as-you-go support so now, usage in March on takes into account compute time x price of the hardware. We're really sorry for any confusion or scare! We have more information about Inference Providers here: https://huggingface.co/docs/inference-providers/en/index
it's definitely the future:)