AI & ML interests

Exploring Extreme Quantization techniques !

Abhaykoulย 
posted an update 3 days ago
view post
Post
3515
๐Ÿš€ Dhanishtha-2.0-preview-0825 Is Here

The Intermediate Thinking Model just leveled up again.

With sharper reasoning, better tool use, and expanded capabilities, Dhanishtha-2.0-preview-0825 is now live and ready to impress.

๐Ÿง  What Makes Dhanishtha Special?
Unlike typical CoT models that only thinks one time, Dhanishtha thinks iteratively:

> Think โ†’ Answer โ†’ Rethink โ†’ Improve โ†’ Rethink again if needed.

๐Ÿ”— Try it now: HelpingAI/Dhanishtha-2.0-preview-0825

๐Ÿ”ž Dhanishtha NSFW Preview

For those exploring more expressive and immersive roleplay scenarios, weโ€™re also releasing:

HelpingAI/Dhanishtha-nsfw
A specialized version tuned for adult-themed interactions and character-driven roleplay.

๐Ÿ”— Explore it here: HelpingAI/Dhanishtha-nsfw

๐Ÿ’ฌ You can also try all of these live at chat.helpingai.co
ยท
Parveshiiiiย 
posted an update 5 days ago
view post
Post
968
๐Ÿš€ Launch Alert: Dev-Stack-Agents
Meet your 50-agent senior AI team โ€” principal-level experts in engineering, AI, DevOps, security, product, and more โ€” all bundled into one modular repo.

+ Code. Optimize. Scale. Secure.
- Full-stack execution, Claude-powered. No human bottlenecks.


๐Ÿ”ง Built for Claude Code
Seamlessly plug into Claudeโ€™s dev environment:

* ๐Ÿง  Each .md file = a fully defined expert persona
* โš™๏ธ Claude indexes them as agents with roles, skills & strategy
* ๐Ÿค– You chat โ†’ Claude auto-routes to the right agent(s)
* โœ๏ธ Want precision? Just call @agent-name directly
* ๐Ÿ‘ฅ Complex task? Mention multiple agents for team execution

Examples:

"@security-auditor please review auth flow for risks"
"@cloud-architect + @devops-troubleshooter โ†’ design a resilient multi-region setup"
"@ai-engineer + @legal-advisor โ†’ build a privacy-safe RAG pipeline"


๐Ÿ”— https://github.com/Parveshiiii/Dev-Stack-Agents
MIT License | Claude-Ready | PRs Welcome

  • 1 reply
ยท
erikkaumย 
posted an update 18 days ago
view post
Post
2544
ZML just released a technical preview of their new Inference Engine: LLMD.

- Just 2.4GB container, which means fast startup times and efficient autoscaling
- Cross-Platform GPU Support: works on both NVIDIA and AMD GPUs.
- written in Zig

I just tried it out and deployed it on Hugging Face Inference Endpoints and wrote a quick guide ๐Ÿ‘‡ You can try it in like 5 minutes!

https://huggingface.co/blog/erikkaum/test-driving-llmd-inference-engine
  • 1 reply
ยท
erikkaumย 
posted an update 19 days ago
view post
Post
1994
We just released native support for @SGLang and @vllm-project in Inference Endpoints ๐Ÿ”ฅ

Inference Endpoints is becoming the central place where you deploy high performance Inference Engines.

And that provides the managed infra for it. Instead of spending weeks configuring infrastructure, managing servers, and debugging deployment issues, you can focus on what matters most: your AI model and your users ๐Ÿ™Œ
Abhaykoulย 
posted an update 20 days ago
view post
Post
3002
๐ŸŽ‰ Dhanishtha-2.0-preview-0725 is Now Live

The Intermediate Thinking Model just got even better.
With the new update, Dhanishtha is now sharper, smarter, and trained further on tool use

๐Ÿง  What Makes Dhanishtha Different?
Unlike standard COT models that give one-shot responses, Dhanishtha thinks in layers:

> Think โ†’ Answer โ†’ Rethink โ†’ Improve โ†’ Rethink again if needed.

HelpingAI/Dhanishtha-2.0-preview-0725
Parveshiiiiย 
posted an update 29 days ago
view post
Post
2655
๐Ÿง  Glimpses of AGI โ€” A Vision for All Humanity
What if AGI wasnโ€™t just a distant dreamโ€”but a blueprint already unfolding?

Iโ€™ve just published a deep dive called Glimpses of AGI, exploring how scalable intelligence, synthetic reasoning, and alignment strategies are paving a new path forward. This isnโ€™t your average tech commentaryโ€”itโ€™s a bold vision for conscious AI systems that reason, align, and adapt beyond narrow tasks.

๐Ÿ” Read it, upvote it if it sparks something, and letโ€™s ignite a collective conversation about the future of AGI.

https://huggingface.co/blog/Parveshiiii/glimpses-of-agi


Parveshiiiiย 
posted an update about 1 month ago
view post
Post
2805
๐Ÿง  MathX-5M by XenArcAI โ€” Scalable Math Reasoning for Smarter LLMs

Introducing MathX-5M, a high-quality, instruction-tuned dataset built to supercharge mathematical reasoning in large language models. With 5 million rigorously filtered examples, it spans everything from basic arithmetic to advanced calculusโ€”curated from public sources and enhanced with synthetic data.

๐Ÿ” Key Highlights:
- Step-by-step reasoning with verified answers
- Covers algebra, geometry, calculus, logic, and more
- RL-validated correctness and multi-stage filtering
- Ideal for fine-tuning, benchmarking, and educational AI

๐Ÿ“‚ - XenArcAI/MathX-5M


  • 1 reply
ยท
Abhaykoulย 
posted an update about 1 month ago
view post
Post
2980
๐ŸŽ‰ Dhanishtha 2.0 Preview is Now Open Source!

The world's first Intermediate Thinking Model is now available to everyone!

Dhanishtha 2.0 Preview brings revolutionary intermediate thinking capabilities to the open-source community. Unlike traditional reasoning models that think once, Dhanishtha can think, answer, rethink, answer again, and continue rethinking as needed using multiple blocks between responses.

๐Ÿš€ Key Features
- Intermediate thinking: Think โ†’ Answer โ†’ Rethink โ†’ Answer โ†’ Rethink if needed...
- Token efficient: Uses up to 79% fewer tokens than DeepSeek R1 on similar queries
- Transparent thinking: See the model's reasoning process in real-time
- Open source: Freely available for research and development


HelpingAI/Dhanishtha-2.0-preview
https://helpingai.co/chat
  • 1 reply
ยท
Abhaykoulย 
posted an update about 1 month ago
view post
Post
4385
Introducing Dhanishtha 2.0: World's first Intermediate Thinking Model

Dhanishtha 2.0 is the world's first LLM designed to think between the responses. Unlike other Reasoning LLMs, which think just once.

Dhanishtha can think, rethink, self-evaluate, and refine in between responses using multiple <think> blocks.
This technique makes it Hinghlt Token efficient it Uses up to 79% fewer tokens than DeepSeek R1
---

You can try our model from: https://helpingai.co/chat
Also, we're gonna Open-Source Dhanistha on July 1st.

---
For Devs:
๐Ÿ”‘ Get your API key at https://helpingai.co/dashboard
from HelpingAI import HAI  # pip install HelpingAI==1.1.1
from rich import print

hai = HAI(api_key="hl-***********************")

response = hai.chat.completions.create(
    model="Dhanishtha-2.0-preview",
    messages=[{"role": "user", "content": "What is the value of โˆซ0โˆž๐‘ฅ3/๐‘ฅโˆ’1๐‘‘๐‘ฅ ?"}],
    stream=True,
    hide_think=False # Hide or show models thinking
)

for chunk in response:
    print(chunk.choices[0].delta.content, end="", flush=True)
  • 2 replies
ยท
alielfilali01ย 
posted an update 3 months ago
alielfilali01ย 
posted an update 6 months ago
view post
Post
1064
๐Ÿšจ Arabic LLM Evaluation ๐Ÿšจ

Few models join the ranking of https://huggingface.co/spaces/inceptionai/AraGen-Leaderboard Today.

The new MistralAI model, Saba, is quite impressive, Top10 ! Well done @arthurmensch and team.

Sadly Mistral did not follow its strategy about public weights this time, we hope this changes soon and we get the model with a permissive license.

We added other Mistral models and apparently, we have been sleeping on mistralai/Mistral-Large-Instruct-2411 !

Another impressive model that joined the ranking today is ALLaM-AI/ALLaM-7B-Instruct-preview. After a long wait finally ALLaM is here and it is IMPRESSIVE given its size !

ALLaM is ranked on OALL/Open-Arabic-LLM-Leaderboard as well.
Abhaykoulย 
posted an update 6 months ago
view post
Post
4778
๐Ÿ”ฅ THE WAIT IS OVER... HAI-SER IS HERE! ๐Ÿ”ฅ

Yo fam, this ain't just another AI dropโ€” this is the FUTURE of emotional intelligence! ๐Ÿš€

Introducing HAI-SER, powered by Structured Emotional Reasoning (SER), the next-level AI that doesnโ€™t just understand your wordsโ€”it feels you, analyzes your emotions, and helps you navigate lifeโ€™s toughest moments. ๐Ÿ’ก

๐Ÿ’ฅ What makes HAI-SER a game-changer?
๐Ÿ”น Emotional Vibe Check โ€“ Gets the mood, energy, and whatโ€™s really going on ๐ŸŽญ
๐Ÿ”น Mind-State Analysis โ€“ Breaks down your thoughts, beliefs, and patterns ๐Ÿคฏ
๐Ÿ”น Root Cause Deep-Dive โ€“ Unpacks the WHY behind your emotions ๐Ÿ’ก
๐Ÿ”น Impact Check โ€“ Sees how itโ€™s affecting your life and mental health ๐Ÿ’”
๐Ÿ”น Safety Check โ€“ Prioritizes your well-being and crisis management ๐Ÿšจ
๐Ÿ”น Healing Game Plan โ€“ Custom strategies to help you bounce back ๐Ÿ’ช
๐Ÿ”น Growth Potential โ€“ Turns struggles into opportunities for self-improvement ๐Ÿ“ˆ
๐Ÿ”น How to Approach โ€“ Teaches you and others how to communicate and heal ๐Ÿค
๐Ÿ”น Personalized Response โ€“ Not just generic adviceโ€”real talk, tailored to YOU ๐Ÿ’ฏ

No more robotic AI responses. No more surface-level advice. HAI-SER gets deep, analyzing emotions with precision and giving real, actionable support.

This ainโ€™t just AIโ€”this is your digital therapist, life coach, and hype squad all in one. Whether itโ€™s mental health, career struggles, relationships, or personal growth, HAI-SER has your back.

๐Ÿš€ The future of emotionally intelligent AI is HERE.
Are you ready? ๐Ÿ”ฅ๐Ÿ’ฏ

HelpingAI/HAI-SER
ยท
alielfilali01ย 
posted an update 7 months ago
view post
Post
2147
3C3H AraGen Leaderboard welcomes today deepseek-ai/DeepSeek-V3 and 12 other models (including the late gpt-3.5 ๐Ÿ’€) to the ranking of best LLMs in Arabic !


Observations:
- DeepSeek-v3 ranked 3rd and only Open model among the top 5 !

- A 14B open model ( Qwen/Qwen2.5-14B-Instruct) outperforms gpt-3.5-turbo-0125 (from last year). This shows how much we came in advancing and supporting Arabic presence within the LLM ecosystem !

- Contrary to what observed in likelihood-acc leaderboards (like OALL/Open-Arabic-LLM-Leaderboard) further finetuned models like maldv/Qwentile2.5-32B-Instruct actually decreased the performance compared to the original model Qwen/Qwen2.5-32B-Instruct.
It's worth to note that the decrease is statiscally insignificant which imply that at best, the out-domain finetuning do not really hurts the model original capabilities acquired during pretraining.
Previous work addressed this (finetuning VS pretraining) but more investigation in this regard is required (any PhDs here ? This could be your question ...)


Check out the latest rankings: https://huggingface.co/spaces/inceptionai/AraGen-Leaderboard
alielfilali01ย 
posted an update 7 months ago
view post
Post
2058
~75% on the challenging GPQA with only 40M parameters ๐Ÿ”ฅ๐Ÿฅณ

GREAT ACHIEVEMENT ! Or is it ?

This new Work, "Data Laundering: Artificially Boosting Benchmark Results through Knowledge Distillation", take out the mystery about many models i personally suspected their results. Speacially on leaderboards other than the english one, Like the Open Arabic LLM Leaderbaord OALL/Open-Arabic-LLM-Leaderboard.

The authors of this work, first started by training a model on the GPQA data, which, unsurprisingly, led to the model achieving 100% performance.

Afterward, they trained what they referred to as a 'legitimate' model on legitimate data (MedMCQA). However, they introduced a distillation loss from the earlier, 'cheated' model.

What they discovered was fascinating: the knowledge of GPQA leaked through this distillation loss, even though the legitimate model was never explicitly trained on GPQA during this stage.

This raises important questions about the careful use of distillation in model training, especially when the training data is opaque. As they demonstrated, itโ€™s apparently possible to (intentionally or unintentionally) leak test data through this method.

Find out more: Data Laundering: Artificially Boosting Benchmark Results through Knowledge Distillation (2412.15255)
  • 1 reply
ยท
Abhaykoulย 
posted an update 8 months ago
view post
Post
2208
๐Ÿ”ฅ BIG ANNOUNCEMENT: THE HELPINGAI API IS LIVE! ๐Ÿ”ฅ

Yo, the moment youโ€™ve all been waiting for is here! ๐Ÿš€ The HelpingAI API is now LIVE and ready to level up your projects! ๐Ÿ”ฅ Weโ€™re bringing that next-level AI goodness straight to your fingertips. ๐Ÿ’ฏ

No more waitingโ€” itโ€™s time to build something epic! ๐Ÿ™Œ

From now on, you can integrate our cutting-edge AI models into your own applications, workflows, and everything in between. Whether youโ€™re a developer, a creator, or just someone looking to make some serious moves, this is your chance to unlock the full potential of emotional intelligence and adaptive AI.

Check out the docs ๐Ÿ”ฅ and letโ€™s get to work! ๐Ÿš€

๐Ÿ‘‰ Check out the docs and start building (https://helpingai.co/docs)
๐Ÿ‘‰ Visit the HelpingAI website (https://helpingai.co/)
ยท