Victor Mustar's picture

Victor Mustar PRO

victor

AI & ML interests

Building the UX of this website

Recent Activity

liked a model about 8 hours ago
ResembleAI/chatterbox
liked a model about 8 hours ago
unsloth/DeepSeek-R1-0528-GGUF
liked a Space about 9 hours ago
ResembleAI/Chatterbox_TTS_Demo
View all activity

Organizations

Hugging Face's profile picture Google's profile picture Safetensors's profile picture Competitions's profile picture 21 RNN's profile picture Spaces-explorers's profile picture Text Generation Inference's profile picture Spaces Examples's profile picture CVPR Demo Track's profile picture Hugging Chat's profile picture Webhooks Explorers (BETA)'s profile picture lora concepts library's profile picture Huggingface Projects's profile picture Scanned Tokens's profile picture hf admins's profile picture Hugging Face OSS Metrics's profile picture Stable Diffusion Dreambooth Concepts Library's profile picture Core ML Projects's profile picture temp-org's profile picture Blog-explorers's profile picture Mustarz's profile picture Open LLM Leaderboard's profile picture Enterprise Explorers's profile picture The Collectionists's profile picture ZeroGPU Explorers's profile picture Hugging Face Tools's profile picture TstOrg141's profile picture Stable Video benchmark's profile picture Social Post Explorers's profile picture Dev Mode Explorers's profile picture LLHF's profile picture SLLHF's profile picture Self-serve FTW's profile picture Inference Explorers's profile picture hf-inference's profile picture Transformers Community's profile picture Changelog's profile picture Tiny Agents's profile picture

victor's activity

reacted to jeffboudier's post with ๐Ÿš€ about 21 hours ago
reacted to openfree's post with ๐Ÿ”ฅ 1 day ago
view post
Post
1610
๐Ÿง  AI Brand Naming with 15 Specialized Theories

๐ŸŽฏ Core Features
15 Expert Theories for professional brand naming
Bilingual Support Korean/English for global brands
Unified Evaluation System creativity/memorability/relevance scores
Real-time Visualization theory-specific custom designs

openfree/Naming

๐Ÿ”ฌ Applied Theories
Cognitive Theories (4)
๐ŸŸฆ Square Theory - Semantic square structure with 4-word relationships
๐Ÿ”Š Sound Symbolism - Psychological connections between phonemes and meaning
๐Ÿง  Cognitive Load - Minimized processing for instant recognition
๐Ÿ‘๏ธ Gestalt Theory - Perceptual principles where whole exceeds parts

Creative Theories (3)
๐Ÿ”€ Conceptual Blending - Merging concepts to create new meanings
๐Ÿ”ง SCAMPER Method - 7 creative transformation techniques
๐ŸŒฟ Biomimicry - Nature-inspired wisdom from 3.8 billion years of evolution

Strategic Theories (2)
โœ… Jobs-to-be-Done - Customer-centric problem-solving focus
๐Ÿ’ญ Design Thinking - Human-centered innovation methodology

Cultural Theories (3)
๐ŸŽญ Jung's Archetype - 12 universal archetypes for emotional connection
๐ŸŒ Linguistic Relativity - Cross-cultural thinking patterns consideration
๐Ÿงฌ Memetics - Cultural transmission and evolutionary potential

Differentiation Theories (3)
โšก Von Restorff Effect - Uniqueness for 30x better recall
๐ŸŽจ Color Psychology - Emotional associations and color meanings
๐ŸŒ Network Effects - Value maximization through network structures

๐Ÿ’ซ Special Features
Each theory provides unique visualizations and customized analysis:

Square Theory โ†’ 4-corner relationship diagram
Blending โ†’ Concept fusion flowchart
Color โ†’ Interactive color palette display
Theory-specific insights for each approach

๐ŸŽจ Output Information
Core: Brand name, slogan, values, emotions, personality
Visual: Colors, concepts, typography styles
Linguistic: Pronunciation, etymology, global adaptability
Strategic: Differentiation, positioning, growth potential
Theory-specific...
reacted to fdaudens's post with ๐Ÿค— 1 day ago
view post
Post
1716
๐ŸŽต Dream come true for content creators! TIGER AI can extract voice, effects & music from ANY audio file ๐Ÿคฏ
This lightweight model uses frequency band-split technology to separate speech like magic. Kudos to @fffiloni for the amazing demo! fffiloni/TIGER-audio-extraction
reacted to ProCreations's post with ๐Ÿš€ 2 days ago
view post
Post
2750
Eyyyy 50 followers ๐Ÿคฏ
  • 1 reply
ยท
reacted to clem's post with ๐Ÿค— 2 days ago
view post
Post
3072
It's just become easier to share your apps on the biggest AI app store (aka HF spaces) for unlimited storage, more visibility and community interactions.

Just pick a React, Svelte, or Vue template when you create your space or add app_build_command: npm run build in your README's YAML and app_file: build/index.html in your README's YAML block.

Or follow this link: https://huggingface.co/new-space?sdk=static

Let's build!
  • 1 reply
ยท
reacted to fdaudens's post with โค๏ธ 2 days ago
view post
Post
3595
Just completed the AI Agents course and wow, that capstone project really makes you understand how to build agents that can handle real-world complexity!

The final project uses the GAIA dataset - your agent has to solve tasks like analyzing Excel files, processing audio recordings, answering questions about YouTube videos, and diving into research papers. This isn't toy examples, it's the messy, multimodal stuff agents need to handle in practice.

Whether youโ€™re just getting started with agents or want to go deeper with tools like LangChain, LlamaIndex, and SmolAgents, this course has tons of useful stuff. A few key insights:
- Code agents are incredibly versatile once you get the architecture right
- The sweet spot is finding the right balance of guidance vs autonomy for each use case
- Once the logic clicks, the possibilities really are endless - it's like letting LLMs break free from the chatbox

The course is free and the certification deadline is July 1st, 2025.

The Hugging Face team built something special here. If you're tired of AI that impresses in demos but fails in practice, this is your path to building agents that actually deliver. https://huggingface.co/learn/agents-course/unit0/introduction

Best part? There's the MCP course next!
reacted to Jofthomas's post with ๐Ÿ”ฅ 6 days ago
view post
Post
2487
Meet our new agentic model : ๐——๐—ฒ๐˜ƒ๐˜€๐˜๐—ฟ๐—ฎ๐—น

Devstral is an open-source LLM built software engineering tasks built under a collaboration between Mistral AI and All Hands AI ๐Ÿ™Œ.

๐—ž๐—ฒ๐˜† ๐—ณ๐—ฒ๐—ฎ๐˜๐˜‚๐—ฟ๐—ฒ๐˜€ :
โ€ข ๐Ÿค– ๐—”๐—ด๐—ฒ๐—ป๐˜๐˜€ : perfect for Agentic coding
โ€ข ๐Ÿƒ ๐—น๐—ถ๐—ด๐—ต๐˜๐˜„๐—ฒ๐—ถ๐—ด๐—ต๐˜: Devstral is a ๐Ÿฎ๐Ÿฐ๐—• parameter based on Mistral small.
โ€ข ยฉ๏ธ ๐—”๐—ฝ๐—ฎ๐—ฐ๐—ต๐—ฒ ๐Ÿฎ.๐Ÿฌ, meaning fully open-source !
โ€ข ๐Ÿ“„ A ๐Ÿญ๐Ÿฎ๐Ÿด๐—ธ context window.

๐Ÿ“šBlog : https://mistral.ai/news/devstral
โšกAPI : The model is also available on our API under the name ๐—ฑ๐—ฒ๐˜ƒ๐˜€๐˜๐—ฟ๐—ฎ๐—น-๐˜€๐—บ๐—ฎ๐—น๐—น-๐Ÿฎ๐Ÿฑ๐Ÿฌ๐Ÿฑ
๐Ÿค— repo : mistralai/Devstral-Small-2505

Can't wait to see what you will build with it !
  • 1 reply
ยท
reacted to clem's post with ๐Ÿ”ฅ 8 days ago
view post
Post
3278
Playing with Veo3 this morning. Share your prompt if you want me to create videos for you (bonus point if they funnily reference HF/open-source). These videos are "a cat on the moon rapping "I love Hugging Face""!
ยท
reacted to ProCreations's post with ๐Ÿค— 10 days ago
view post
Post
3168
Eyyy thank you guys for 40 followers!
reacted to nicolay-r's post with ๐Ÿ”ฅ 10 days ago
view post
Post
2337
๐Ÿš€ For those who interested in minimalistic integration of LLMs inferece with predefined reasoning shema, excited to share the latest bulk chain 1.1.0. It represents a no-string solution for deploying your LLM for efficient inference over data iterators.
โœจ Key Features:
- Full async inference support + Including streaming mode for real-time output
- simplified inference API
๐Ÿ”— Check out the repo: https://github.com/nicolay-r/bulk-chain

๐Ÿ’ก Special thanks to @RicardoLee for his work on effective async LLaMA-3 deployment that helped shape this release:
https://github.com/RicardoLeeV587/Llama3-FastInference
reacted to Jaward's post with ๐Ÿ‘ 10 days ago
reacted to cbensimon's post with ๐Ÿ”ฅ 13 days ago
view post
Post
5651
๐Ÿš€ ZeroGPU medium size is now available as a power-user feature

Nothing too fancy for nowโ€”ZeroGPU Spaces still default to large (70GB VRAM)โ€”but this paves the way for:
- ๐Ÿ’ฐ size-based quotas / pricing (medium will offer significantly more usage than large)
- ๐Ÿฆฃ the upcoming xlarge size (141GB VRAM)

You can as of now control GPU size via a Space variable. Accepted values:
- auto (future default)
- medium
- large (current default)

The auto mode checks total CUDA tensor size during startup:
- More than 30GB โ†’ large
- Otherwise โ†’ medium
ยท
reacted to RiverZ's post with ๐Ÿค— 24 days ago
view post
Post
3078
๐Ÿš€ Excited to Share Our Latest Work: In-Context Edit: Enabling Instructional Image Editing with In-Context Generation in Large Scale Diffusion Transformer๏ฝž

๐ŸŽจ Daily Paper:
In-Context Edit: Enabling Instructional Image Editing with In-Context Generation in Large Scale Diffusion Transformer (2504.20690)


๐Ÿ”“ Code is now open source!
๐Ÿ”ฅ Huggingface DEMO:
RiverZ/ICEdit

๐ŸŒ Project Website: https://river-zhang.github.io/ICEdit-gh-pages/
๐Ÿ  GitHub Repository: https://github.com/River-Zhang/ICEdit/blob/main/scripts/gradio_demo.py
๐Ÿค— Huggingface:
sanaka87/ICEdit-MoE-LoRA

๐Ÿ“„ arxiv Paper:
In-Context Edit: Enabling Instructional Image Editing with In-Context Generation in Large Scale Diffusion Transformer (2504.20690)


๐Ÿ”ฅ Why itโ€™s cool:
- Achieves high-quality, multi-task image editing.
- Uses only 1% of the training parameters and 0.1% of the training data compared to existing methods โ€” extremely efficient
- Beats several commercial models on background preservation, ID control, and consistency
- Open-source, low-cost, faster, and stronger โ€” think of it as the โ€œDeepSeek of image editingโ€ ๐Ÿ‘€

We also implemented a Gradio demo app, available directly in our GitHub repo! And we made a flashy demo video โ€” happy to send it your way!
reacted to prithivMLmods's post with ๐Ÿ”ฅ 28 days ago
view post
Post
1933
Dropping downstream tasks using newly initialized parameters and weights supports domain-specific image classification post-training, based on the SigLIP-2 models: Patch-16/224, Patch-16/256, and Patch-32/256. For more details, please refer to the respective model cards : ๐Ÿค—

+ watermark detection : prithivMLmods/Watermark-Detection-SigLIP2
+ resisc45 : prithivMLmods/RESISC45-SigLIP2
+ pacs dg : prithivMLmods/PACS-DG-SigLIP2
+ 3d printed or not : prithivMLmods/3D-Printed-Or-Not-SigLIP2
+ formula or text : prithivMLmods/Formula-Text-Detection

Categorizing Un-Safe Content :
- explicit content patch16 256 : prithivMLmods/siglip2-x256-explicit-content
- explicit content patch32 256 : prithivMLmods/siglip2-x256p32-explicit-content

Collection :
> SigLIP2 Content Filters 042025 Final : https://huggingface.co/collections/prithivMLmods/siglip2-content-filters-04202-final-680fe4aa1a9d589bf2c915ff
> SigLIP2 : google/siglip2-67b5dcef38c175486e240107
> SigLIP2 Multilingual Vision-Language Encoders : https://arxiv.org/pdf/2502.14786
reacted to MikeDoes's post with ๐Ÿš€ 28 days ago
view post
Post
1519
PII-Masking-1M Final Day (7/7)! ๐Ÿš€ Today, we unveil 5 NEW Enterprise PII (E-PII) Dataset PREVIEWS!

Standard PII tools often miss sensitive *business* data. That's why we built E-PII previews for the data that powers your operations and compliance needs.

Get a first look (representing 100,000 samples each!) into datasets designed for real-world enterprise security across these categories:

๐Ÿฅ **PHI Preview**: For Healthcare Data
๐Ÿ’ณ **PFI Preview:** For Financial Data
๐Ÿข **PWI Preview:** For Workplace Data
๐Ÿ’ป **PDI Preview:** For Digital Activity Data
๐Ÿ“ **PLI Preview:** For Location Data


That wraps up our #PIIMasking1M 7 days announcement! HUGE thanks for following along and for your engagement.
Explore ALL our releases, including these E-PII previews, in the Ai4Privacy Hugging Face Collection & show some love โค๏ธ if you find them useful!
๐Ÿ”— Visit the Collection:https://huggingface.co/ai4privacy

Let's keep building safer AI, together!
replied to onekq's post 28 days ago
view reply

it's trained to think - probably with the idea to punctually use /no_think for some messages in a conversation where you don't want it to :) (The no think is probably more a product feature than something means to be used as default)

reacted to merterbak's post with ๐Ÿ”ฅ 29 days ago
view post
Post
4861
Qwen 3 models released๐Ÿ”ฅ
It offers 2 MoE and 6 dense models with following parameter sizes: 0.6B, 1.7B, 4B, 8B, 14B, 30B(MoE), 32B, and 235B(MoE).
Models: Qwen/qwen3-67dd247413f0e2e4f653967f
Blog: https://qwenlm.github.io/blog/qwen3/
Demo: Qwen/Qwen3-Demo
GitHub: https://github.com/QwenLM/Qwen3

โœ… Pre-trained 119 languages(36 trillion tokens) and dialects with strong translation and instruction following abilities. (Qwen2.5 was pre-trained on 18 trillion tokens.)
โœ…Qwen3 dense models match the performance of larger Qwen2.5 models. For example, Qwen3-1.7B/4B/8B/14B/32B perform like Qwen2.5-3B/7B/14B/32B/72B.
โœ… Three stage done while pretraining:
โ€ข Stage 1: General language learning and knowledge building.
โ€ข Stage 2: Reasoning boost with STEM, coding, and logic skills.
โ€ข Stage 3: Long context training
โœ… It supports MCP in the model
โœ… Strong agent skills
โœ… Supports seamless between thinking mode (for hard tasks like math and coding) and non-thinking mode (for fast chatting) inside chat template.
โœ… Better human alignment for creative writing, roleplay, multi-turn conversations, and following detailed instructions.
reacted to AdinaY's post with ๐Ÿš€ 29 days ago
view post
Post
3011
DeepSeek, Alibaba, Skywork, Xiaomi, Bytedance.....
And thatโ€™s just part of the companies from the Chinese community that released open models in April ๐Ÿคฏ

zh-ai-community/april-2025-open-releases-from-the-chinese-community-67ea699965f6e4c135cab10f

๐ŸŽฌ Video
> MAGI-1 by SandAI
> SkyReels-A2 & SkyReels-V2 by Skywork
> Wan2.1-FLF2V by Alibaba-Wan

๐ŸŽจ Image
> HiDream-I1 by Vivago AI
> Kimi-VL by Moonshot AI
> InstantCharacter by InstantX & Tencent-Hunyuan
> Step1X-Edit by StepFun
> EasyControl by Shanghai Jiaotong University

๐Ÿง  Reasoning
> MiMo by Xiaomi
> Skywork-R1V 2.0 by Skywork
> ChatTS by ByteDance
> Kimina by Moonshot AI & Numina
> GLM-Z1 by Zhipu AI
> Skywork OR1 by Skywork
> Kimi-VL-Thinking by Moonshot AI

๐Ÿ”Š Audio
> Kimi-Audio by Moonshot AI
> IndexTTS by BiliBili
> MegaTTS3 by ByteDance
> Dolphin by DataOceanAI

๐Ÿ”ข Math
> DeepSeek Prover V2 by Deepseek

๐ŸŒ LLM
> Qwen by Alibaba-Qwen
> InternVL3 by Shanghai AI lab
> Ernie4.5 (demo) by Baidu

๐Ÿ“Š Dataset
> PHYBench by Eureka-Lab
> ChildMandarin & Seniortalk by BAAI

Please feel free to add if I missed anything!
reacted to Xenova's post with ๐Ÿ”ฅ 29 days ago
view post
Post
7524
Introducing the ONNX model explorer: Browse, search, and visualize neural networks directly in your browser. ๐Ÿคฏ A great tool for anyone studying Machine Learning! We're also releasing the entire dataset of graphs so you can use them in your own projects! ๐Ÿค—

Check it out! ๐Ÿ‘‡
Demo: onnx-community/model-explorer
Dataset: onnx-community/model-explorer
Source code: https://github.com/xenova/model-explorer
reacted to abidlabs's post with ๐Ÿค— 29 days ago
view post
Post
2669
Hi folks! Excited to share a new feature from the Gradio team along with a tutorial.

If you don't already know, Gradio is an open-source Python library used to build interfaces for machine learning models. Beyond just creating UIs, Gradio also exposes API capabilities and now, Gradio apps can be launched Model Context Protocol (MCP) servers for LLMs.

If you already know how to use Gradio, there are only two additional things you need to do:
* Add standard docstrings to your function (these will be used to generate the descriptions for your tools for the LLM)
* Set mcp_server=True in launch()


Here's a complete example (make sure you already have the latest version of Gradio installed):


import gradio as gr

def letter_counter(word, letter):
    """Count the occurrences of a specific letter in a word.
    
    Args:
        word: The word or phrase to analyze
        letter: The letter to count occurrences of
        
    Returns:
        The number of times the letter appears in the word
    """
    return word.lower().count(letter.lower())

demo = gr.Interface(
    fn=letter_counter,
    inputs=["text", "text"],
    outputs="number",
    title="Letter Counter",
    description="Count how many times a letter appears in a word"
)

demo.launch(mcp_server=True)



This is a very simple example, but you can add the ability to generate Ghibli images or speak emotions to any LLM that supports MCP. Once you have an MCP running locally, you can copy-paste the same app to host it on [Hugging Face Spaces](https://huggingface.co/spaces/) as well.

All free and open-source of course! Full tutorial: https://www.gradio.app/guides/building-mcp-server-with-gradio
  • 2 replies
ยท