I updated the LLM Scientist roadmap and added a ton of new information and references. It covers training, datasets, evaluation, quantization, and new trends like test-time compute scaling.
The LLM Course has been incredibly popular (41.3k stars!) and I've been touched to receive many, many messages about how it helped people in their careers.
I know how difficult this stuff can be, so I'm super proud of the impact it had. I want to keep updating it in 2025, especially with the LLM Engineer roadmap.
❤️🔥Stranger Zone's MidJourney Mix Model Adapter is trending on the Very Model Page, with over 45,000+ downloads. Additionally, the Super Realism Model Adapter has over 52,000+ downloads, remains the top two adapter on Stranger Zone! strangerzonehf/Flux-Midjourney-Mix2-LoRA, strangerzonehf/Flux-Super-Realism-LoRA
We’re launching a FREE and CERTIFIED course on Agents!
We're thrilled to announce the launch of the Hugging Face Agents course on Learn! This interactive, certified course will guide you through building and deploying your own AI agents.
Here's what you'll learn:
- Understanding Agents: We'll break down the fundamentals of AI agents, showing you how they use LLMs to perceive their environment (observations), reason about it (thoughts), and take actions. Think of a smart assistant that can book appointments, answer emails, or even write code based on your instructions. - Building with Frameworks: You'll dive into popular agent frameworks like LangChain, LlamaIndex and smolagents. These tools provide the building blocks for creating complex agent behaviors. - Real-World Applications: See how agents are used in practice, from automating SQL queries to generating code and summarizing complex documents. - Certification: Earn a certification by completing the course modules, implementing a use case, and passing a benchmark assessment. This proves your skills in building and deploying AI agents. Audience
This course is designed for anyone interested in the future of AI. Whether you're a developer, data scientist, or simply curious about AI, this course will equip you with the knowledge and skills to build your own intelligent agents.
Enroll today and start building the next generation of AI agent applications!
📚 Overview A cutting-edge AI system that combines transformer architecture with citation pattern analysis to predict research impact. Our model, trained on 120,000+ CS papers, analyzes innovation potential, methodological robustness, and future impact, providing researchers with valuable insights before publication. 🧠 Scientific Foundation
Pre-submission impact assessment Research direction optimization Time-saving paper evaluation Competitive edge in academia Trend identification advantage
🎯 Key Features
One-click arXiv paper analysis Real-time impact scoring (0-1) 9-tier grading system (AAA-C) Smart input validation Instant visual feedback
🌟 Unique Benefits "Don't wait years to know your paper's impact. Get instant, AI-powered insights to strengthen your research strategy and maximize your academic influence." Perfect for:
Research authors PhD students Journal editors Research institutions Grant committees
A collection of 1,892,816 artwork posts featuring: - High-quality art pieces with various styles and techniques - Complete metadata including artist IDs, titles, and moderation flags - Content from Artfol social media platform
The dataset contains: - Public domain artwork posts - Artist attribution and identifiers - Direct image URLs and web page links - Content safety flags (NSFW, gore) - Post titles and descriptions
All content is available under CC0 license, allowing unrestricted use including commercial applications.
I am pleased to announce that I have founded the University of Glasgow organization on Huggingface. If you are affiliated with the University of Glasgow or have a relative who is, you can log in through the relevant link.
We are thrilled to present the improved "ClearerVoice-Studio", an open-source platform designed to make speech processing easy use for everyone! Whether you’re working on speech enhancement, speech separation, speech super-resolution, or target speaker extraction, this unified platform has you covered.
** Why Choose ClearerVoice-Studio?**
- Pre-Trained Models: Includes cutting-edge pre-trained models, fine-tuned on extensive, high-quality datasets. No need to start from scratch! - Ease of Use: Designed for seamless integration with your projects, offering a simple yet flexible interface for inference and training.
- Enhance noisy speech recordings to achieve crystal-clear quality. - Separate speech from complex audio mixtures with ease. - Transform low-resolution audio into high-resolution audio. A full upscaled LJSpeech-1.1-48kHz dataset can be downloaded from alibabasglab/LJSpeech-1.1-48kHz . - Extract target speaker voices with precision using audio-visual models.
**Join Us in Growing ClearerVoice-Studio!**
We believe in the power of open-source collaboration. By starring our GitHub repository and sharing ClearerVoice-Studio with your network, you can help us grow this community-driven platform.
**Support us by:**
- Starring it on GitHub. - Exploring and contributing to our codebase . - Sharing your feedback and use cases to make the platform even better. - Joining our community discussions to exchange ideas and innovations. - Together, let’s push the boundaries of speech processing! Thank you for your support! :sparkling_heart:
🙋🏻♂️Hey there folks , Open LLM Europe just released Lucie 7B-Instruct model , a billingual instruct model trained on open data ! You can check out my unofficial demo here while we wait for the official inference api from the group : Tonic/Lucie-7B hope you like it 🚀
I'm excited to announce significant improvements to my HF Daily Paper Newsletter Bot! Here are the key updates:
🖼️ Enhanced Poster Generation - Implemented dynamic height adjustment for daily paper posters - Added support for displaying complete paper content without truncation - Improved Chinese font rendering and text layout - Integrated Hugging Face logo for better branding - Enhanced visual aesthetics with optimized card layouts
📝 Content Improvements - Removed paper count limitations (previously capped at 5 papers) - Enhanced title and summary extraction algorithms - Improved text wrapping and spacing for better readability - Added proper handling of long content with automatic layout adjustments
🛠️ Technical Enhancements - Implemented better font loading mechanism with fallback options - Added support for multiple Chinese font paths - Improved error handling and logging - Enhanced memory management for image processing - Added detailed debugging information
🌟 Visual Design Updates - Refined color scheme with HF brand colors - Improved card spacing and padding - Enhanced typography with better font sizing - Added smooth transitions between paper cards - Optimized overall layout for better visual hierarchy
🔧 Infrastructure Updates - Improved GitHub Actions workflow reliability - Enhanced error notification system - Added automatic retries for API calls - Improved logging and debugging capabilities
The bot now generates more professional and visually appealing daily paper summaries while ensuring complete content display. These updates make the newsletter more readable and informative for our users.
Try it out and let me know what you think! Your feedback helps me make continuous improvements to better serve the AI research community.
Can it run DeepSeek V3 671B is the new 'can it run Doom'.
How minimalistic can I go with on device AI with behemoth models - here I'm running DeepSeek V3 MoE on a single A6000 GPU.
Not great, not terrible, for this minimalistic setup. I love the Mixture of Experts architectures. Typically I'm running my core LLM distributed over the 4 GPUs.
Make sure you own your AI. AI in the cloud is not aligned with you; it's aligned with the company that owns it.
We had a few people asking about the differences and methodologies of our addition to the open-image-preferences dataset. So my colleague and I wrote a blog post about it with the new huggingface blog functionality: https://huggingface.co/blog/RapidataAI/more-image-preferences
Major update on the Talking to Chatbots dataset! Expanded the 'wrapped' dataset (one row per chat) to 2.86k records, and the 'unwrapped' version (one row per conversation turn) to 11k records. The main source is my ChatGPT archive with nearly 2 years of chats. It is still a work in progress as I incorporate chats from other sources and qualitative metrics (SCBN) for responses.