AI & ML interests

'an LLM is only as good as the dataset it was trained on' - Sun Tzu

Recent Activity

pszemrajย  updated a dataset about 1 month ago
BEE-spoke-data/govdocs1-pdf-source
pszemrajย  updated a dataset about 1 month ago
BEE-spoke-data/govdocs1-by-extension
amazingvinceย  updated a dataset about 2 months ago
BEE-spoke-data/SurvivorLib-Nanonets-OCR-s
View all activity

qnguyen3ย 
posted an update about 1 year ago
qnguyen3ย 
posted an update over 1 year ago
view post
Post
6161
๐ŸŽ‰ Introducing nanoLLaVA, a powerful multimodal AI model that packs the capabilities of a 1B parameter vision language model into just 5GB of VRAM. ๐Ÿš€ This makes it an ideal choice for edge devices, bringing cutting-edge visual understanding and generation to your devices like never before. ๐Ÿ“ฑ๐Ÿ’ป

Model: qnguyen3/nanoLLaVA ๐Ÿ”
Spaces: qnguyen3/nanoLLaVA (thanks to @merve )

Under the hood, nanoLLaVA is based on the powerful vilm/Quyen-SE-v0.1 (my Qwen1.5-0.5B finetune) and Google's impressive google/siglip-so400m-patch14-384. ๐Ÿง  The model is trained using a data-centric approach to ensure optimal performance. ๐Ÿ“Š

In the spirit of transparency and collaboration, all code and model weights are open-sourced under the Apache 2.0 license. ๐Ÿค
  • 1 reply
ยท