Nick Doiron

monsoon-nlp

AI & ML interests

biology and multilingual models

Recent Activity

liked a dataset about 20 hours ago
Datadog/BOOM
liked a model 5 days ago
nari-labs/Dia-1.6B
View all activity

Organizations

BigScience Workshop's profile picture Spaces-explorers's profile picture BigCode's profile picture Blog-explorers's profile picture Scary Snake's profile picture Hugging Face Discord Community's profile picture Hugging Face MCP Course's profile picture

monsoon-nlp's activity

reacted to seawolf2357's post with πŸ‘€ 8 days ago
view post
Post
5757
Samsung Hacking Incident: Samsung Electronics' Official Hugging Face Account Compromised
Samsung Electronics' official Hugging Face account has been hacked. Approximately 17 hours ago, two new language models (LLMs) were registered under Samsung Electronics' official Hugging Face account. These models are:

https://huggingface.co/Samsung/MuTokenZero2-32B
https://huggingface.co/Samsung/MythoMax-L2-13B

The model descriptions contain absurd and false claims, such as being trained on "1 million W200 GPUs," hardware that doesn't even exist.
Moreover, community participants on Hugging Face who have noticed this issue are continuously posting that Samsung Electronics' account has been compromised.
There is concern about potential secondary and tertiary damage if users download these LLMs released under the Samsung Electronics account, trusting Samsung's reputation without knowing about the hack.
Samsung Electronics appears to be unaware of this situation, as they have not taken any visible measures yet, such as changing the account password.
Source: https://discord.gg/openfreeai
  • 2 replies
Β·
updated a Space about 1 month ago
published a Space about 1 month ago
New activity in monsoon-nlp/code-refusal-for-abliteration about 1 month ago

Integrate m/re spurces

1
#1 opened 6 months ago by
monsoon-nlp
updated a dataset about 1 month ago
upvoted an article about 2 months ago
view article
Article

Welcome Llama 4 Maverick & Scout on Hugging Face!

By burtenshaw and 6 others β€’
β€’ 144
reacted to merterbak's post with πŸ”₯ about 2 months ago
view post
Post
3022
Meta has unveiled its Llama 4 πŸ¦™ family of models, featuring native multimodality and mixture-of-experts architecture. Two model families are available now:
ModelsπŸ€—: meta-llama/llama-4-67f0c30d9fe03840bc9d0164
Blog Post: https://ai.meta.com/blog/llama-4-multimodal-intelligence/
HF's Blog Post: https://huggingface.co/blog/llama4-release

- 🧠 Native Multimodality - Process text and images in a unified architecture
- πŸ” Mixture-of-Experts - First Llama models using MoE for incredible efficiency
- πŸ“ Super Long Context - Up to 10M tokens
- 🌐 Multilingual Power - Trained on 200 languages with 10x more multilingual tokens than Llama 3 (including over 100 languages with over 1 billion tokens each)

πŸ”Ή Llama 4 Scout
- 17B active parameters (109B total)
- 16 experts architecture
- 10M context window
- Fits on a single H100 GPU
- Beats Gemma 3, Gemini 2.0 Flash-Lite, and Mistral 3.1

πŸ”Ή Llama 4 Maverick
- 17B active parameters (400B total)
- 128 experts architecture
- It can fit perfectly on DGX H100(8x H100)
- 1M context window
- Outperforms GPT-4o and Gemini 2.0 Flash
- ELO score of 1417 on LMArena currently second best model on arena

πŸ”Ή Llama 4 Behemoth (Coming Soon)
- 288B active parameters (2T total)
- 16 experts architecture
- Teacher model for Scout and Maverick
- Outperforms GPT-4.5, Claude Sonnet 3.7, and Gemini 2.0 Pro on STEM benchmarks