Titus von Koeller

Titus-von-Koeller

AI & ML interests

NN Quantization, Generative AI, LLMs, alignment, algorithms for social justice, ethical humanism, mitigating gender bias, audio compression, AGI

Articles

Organizations

Hugging Face's profile picture Hugging Face OSS Metrics's profile picture Social Post Explorers's profile picture blhf's profile picture Hugging Face Party @ PyTorch Conference's profile picture

Posts 3

view post
Post
1963
๐Ÿ”ฅ Level up your model training w/ GaLore + Transformers for SOTA results on consumer-grade hardware!

โฌ‡๏ธ 82.5% less optimizer state memory footprint without performance degradation by expressing the gradient weight matrix as low rank.

๐Ÿ‘ฉ๐Ÿฟโ€๐Ÿ’ป Install via pip install transformers>=4.39.0 galore-torch. #ProudlyGpuPoor

The integration of GaLore into the training of large language models (LLMs) marks a significant advancement in the field of deep learning, particularly in terms of memory efficiency and the democratization of AI research. By allowing for the training of billion-parameter models on consumer-grade hardware, reducing memory footprint in optimizer states, and leveraging advanced projection matrix techniques, GaLore opens new horizons for researchers and practitioners with limited access to high-end computational resources.

๐Ÿ”ฌ Find out more about GaLore and investigate lots of juicy technical details: https://huggingface.co/blog/galore

๐Ÿค— Huge thanks to everyone involved โค๏ธ:

โ€ข authors: @jiaweizhao @Kyriection @beidic Zhangyang Wang @animakumar @tydsh
โ€ข community contributors: @hiyouga @mdouglas and others!
โ€ข @ybelkada for taking such swift action in composing and coordinating necessary PRs to get this live at โšก speed!

๐Ÿ—๏ธ๐Ÿ“ˆ Super rewarding to see how @timdettmers work with optimizers is being built upon to achieve even greater heights!

๐Ÿšง Actually, there are ongoing works to integrate GaLore into bitsandbytes and optimize memory efficiency even further ๐Ÿ’ช. We'll keep you posted!
view post
Post
We just released bitsandbytes==0.43.0 ๐Ÿ“ฆ , with these significant new additions:

โ€ฃ ๐Ÿ›ซ FSDP+QLoRA support (alpha release)
โ—ฆ now anyone with 2 powerful gaming GPUs can fine-tune 70B param models at home!
โ—ฆ in collab with Jeremy Howard + team @ answer.ai
โ—ฆ answer.ai blogpost: https://www.answer.ai/posts/2024-03-06-fsdp-qlora.html
โ—ฆ example repo: https://github.com/AnswerDotAI/fsdp_qlora/

โ€ฃ ๐ŸŒˆโŠž Official Windows support
โ—ฆ now via simple pip install bitsandbytes>=0.43.0

โ€ฃ ๐Ÿ“„ Huge docs update:
โ—ฆ https://huggingface.co/docs/bitsandbytes/main
โ—ฆ Be sure to check out the optimizers and the API docs
โ—ฆ ... even more upcoming ...

Under the hood there we have many other improvements, due to extensive maintenance activity, community contributions by super active + knowledgable volunteers โœจ ๐Ÿš€ and the official sponsorship by Hugging Face that makes all this possible ๐Ÿค— โค๏ธ ๐ŸŒ

We would greatly appreciate any further community contributions, be it to help with refactorings, exterminating flaky tests, writing doc-strings, tutorials, new features. Don't be shy, just contact us and we see where this leads us:
https://github.com/TimDettmers/bitsandbytes/discussions

Have a great weekend everyone!

models

None public yet

datasets

None public yet