Celina's picture

Celina

celinah

AI & ML interests

inference, on-device and image generation

Recent Activity

reacted to m-ric's post with πŸ”₯ 5 days ago
After 6 years, BERT, the workhorse of encoder models, finally gets a replacement: π—ͺ𝗲𝗹𝗰𝗼𝗺𝗲 π— π—Όπ—±π—²π—Ώπ—»π—•π—˜π—₯𝗧! πŸ€— We talk a lot about ✨Generative AI✨, meaning "Decoder version of the Transformers architecture", but this is only one of the ways to build LLMs: encoder models, that turn a sentence in a vector, are maybe even more widely used in industry than generative models. The workhorse for this category has been BERT since its release in 2018 (that's prehistory for LLMs). It's not a fancy 100B parameters supermodel (just a few hundred millions), but it's an excellent workhorse, kind of a Honda Civic for LLMs. Many applications use BERT-family models - the top models in this category cumulate millions of downloads on the Hub. ➑️ Now a collaboration between Answer.AI and LightOn just introduced BERT's replacement: ModernBERT. π—§π—Ÿ;𝗗π—₯: πŸ›οΈ Architecture changes: β‡’ First, standard modernizations: - Rotary positional embeddings (RoPE) - Replace GeLU with GeGLU, - Use Flash Attention 2 ✨ The team also introduced innovative techniques like alternating attention instead of full attention, and sequence packing to get rid of padding overhead. πŸ₯‡ As a result, the model tops the game of encoder models: It beats previous standard DeBERTaV3 for 1/5th the memory footprint, and runs 4x faster! Read the blog post πŸ‘‰ https://huggingface.co/blog/modernbert
upvoted a paper 5 days ago
Qwen2.5 Technical Report
View all activity

Organizations

Hugging Face's profile picture Hugging Face OSS Metrics's profile picture Blog-explorers's profile picture Hugging Face for Computer Vision's profile picture MLX Community's profile picture Social Post Explorers's profile picture open/ acc's profile picture DDUF's profile picture

Posts 2

view post
Post
567
πŸš€ We've just dropped a new release v0.27.0 of the πš‘πšžπšπšπš’πš—πšπšπšŠπšŒπšŽ_πš‘πšžπš‹ Python library!

This release includes:
- πŸ’Ύ New torch model loading utilities in the serialization module β€” providing a standardized way to save and load torch models with built-in support for sharding and safe serialization.
- πŸ“¦ Tooling for something exciting β€” if you like single-file formats for models like GGUF, you'll love what we're cooking up πŸ‘€ More coming soon!
- πŸ› οΈ Loads of quality-of-life improvements and bug fixes!

release notes and full details here πŸ‘‡
Wauplin/huggingface_hub#10

$ pip install -U huggingface_hub
view post
Post
1100
πŸ“£ πš‘πšžπšπšπš’πš—πšπšπšŠπšŒπšŽ_πš‘πšžπš‹ v0.26.0 is out with some new features and improvements!

✨ 𝗧𝗼𝗽 π—›π—Άπ—΄π—΅π—Ήπ—Άπ—΄π—΅π˜π˜€:
- πŸ”Β Multiple access tokens support: Easily manage multiple access tokens with new CLI commands. Perfect for handling multiple tokens with specific permissions in production or when collaborating with external teams.
- πŸ–ΌοΈ Conversational VLMs inference is now supported withΒ InferenceClient's chat completion!
- πŸ“„ Daily Papers API: Seamlessly search and retrieve detailed paper information from the Hub!

We’ve also introduced multiple bug fixes and quality-of-life improvements - thanks to the awesome contributions from our community! πŸ€—

Check out the release notes here: Wauplin/huggingface_hub#9

and you can try it out now πŸ‘‡
pip install huggingface_hub==0.26.0

models

None public yet