AI & ML interests

sharing knowledge

Recent Activity

HuggingFaceDocBuilderย  updated a dataset about 3 hours ago
hf-doc-build/doc-build-dev
HuggingFaceDocBuilderย  updated a dataset about 18 hours ago
hf-doc-build/doc-build
View all activity

hf-doc-build's activity

regisssย 
posted an update 7 days ago
abhishekย 
posted an update 26 days ago
view post
Post
1617
๐ŸŽ‰ SUPER BLACK FRIDAY DEAL ๐ŸŽ‰

Train almost any model on a variety of tasks such as llm finetuning, text classification/regression, summarization, question answering, image classification/regression, object detection, tabular data, etc for FREE using AutoTrain locally. ๐Ÿ”ฅ
https://github.com/huggingface/autotrain-advanced
abhishekย 
posted an update about 2 months ago
view post
Post
5410
INTRODUCING Hugging Face AutoTrain Client ๐Ÿ”ฅ
Fine-tuning models got even easier!!!!
Now you can fine-tune SOTA models on all compatible dataset-model pairs on Hugging Face Hub using Python on Hugging Face Servers. Choose from a number of GPU flavors, millions of models and dataset pairs and 10+ tasks ๐Ÿค—

To try, install autotrain-advanced using pip. You can ignore dependencies and install without --no-deps and then you'd need to install some dependencies by hand.

"pip install autotrain-advanced"

Github repo: https://github.com/huggingface/autotrain-advanced
  • 6 replies
ยท
abhishekย 
posted an update 2 months ago
regisssย 
posted an update 2 months ago
view post
Post
1378
Interested in performing inference with an ONNX model?โšก๏ธ

The Optimum docs about model inference with ONNX Runtime is now much clearer and simpler!

You want to deploy your favorite model on the hub but you don't know how to export it to the ONNX format? You can do it in one line of code as follows:
from optimum.onnxruntime import ORTModelForSequenceClassification

# Load the model from the hub and export it to the ONNX format
model_id = "distilbert-base-uncased-finetuned-sst-2-english"
model = ORTModelForSequenceClassification.from_pretrained(model_id, export=True)

Check out the whole guide ๐Ÿ‘‰ https://huggingface.co/docs/optimum/onnxruntime/usage_guides/models
Wauplinย 
posted an update 3 months ago
view post
Post
2790
What a great milestone to celebrate! The huggingface_hub library is slowly becoming a cornerstone of the Python ML ecosystem when it comes to interacting with the @huggingface Hub. It wouldn't be there without the hundreds of community contributions and feedback! No matter if you are loading a model, sharing a dataset, running remote inference or starting jobs on our infra, you are for sure using it! And this is only the beginning so give a star if you wanna follow the project ๐Ÿ‘‰ https://github.com/huggingface/huggingface_hub
  • 1 reply
ยท
Wauplinย 
posted an update 3 months ago
view post
Post
4573
๐Ÿš€ Exciting News! ๐Ÿš€

We've just released ๐š‘๐šž๐š๐š๐š’๐š—๐š๐š๐šŠ๐šŒ๐šŽ_๐š‘๐šž๐š‹ v0.25.0 and it's packed with powerful new features and improvements!

โœจ ๐—ง๐—ผ๐—ฝ ๐—›๐—ถ๐—ด๐—ต๐—น๐—ถ๐—ด๐—ต๐˜๐˜€:

โ€ข ๐Ÿ“ ๐—จ๐—ฝ๐—น๐—ผ๐—ฎ๐—ฑ ๐—น๐—ฎ๐—ฟ๐—ด๐—ฒ ๐—ณ๐—ผ๐—น๐—ฑ๐—ฒ๐—ฟ๐˜€ with ease using huggingface-cli upload-large-folder. Designed for your massive models and datasets. Much recommended if you struggle to upload your Llama 70B fine-tuned model ๐Ÿคก
โ€ข ๐Ÿ”Ž ๐—ฆ๐—ฒ๐—ฎ๐—ฟ๐—ฐ๐—ต ๐—”๐—ฃ๐—œ: new search filters (gated status, inference status) and fetch trending score.
โ€ข โšก๐—œ๐—ป๐—ณ๐—ฒ๐—ฟ๐—ฒ๐—ป๐—ฐ๐—ฒ๐—–๐—น๐—ถ๐—ฒ๐—ป๐˜: major improvements simplifying chat completions and handling async tasks better.

Weโ€™ve also introduced tons of bug fixes and quality-of-life improvements - thanks to the awesome contributions from our community! ๐Ÿ’ช

๐Ÿ’ก Check out the release notes: Wauplin/huggingface_hub#8

Want to try it out? Install the release with:

pip install huggingface_hub==0.25.0

  • 1 reply
ยท
abhishekย 
posted an update 4 months ago
abhishekย 
posted an update 4 months ago
view post
Post
1854
๐Ÿšจ NEW TASK ALERT ๐Ÿšจ
Extractive Question Answering: because sometimes generative is not all you need ๐Ÿ˜‰
AutoTrain is the only open-source, no code solution to offer so many tasks across different modalities. Current task count: 23 ๐Ÿš€
Check out the blog post on getting started with this task: https://huggingface.co/blog/abhishek/extractive-qa-autotrain
Wauplinย 
posted an update 5 months ago
view post
Post
1984
๐Ÿš€ Just released version 0.24.0 of the ๐š‘๐šž๐š๐š๐š’๐š—๐š๐š๐šŠ๐šŒ๐šŽ_๐š‘๐šž๐š‹ Python library!

Exciting updates include:
โšก InferenceClient is now a drop-in replacement for OpenAI's chat completion!

โœจ Support for response_format, adapter_id , truncate, and more in InferenceClient

๐Ÿ’พ Serialization module with a save_torch_model helper that handles shared layers, sharding, naming convention, and safe serialization. Basically a condensed version of logic scattered across safetensors, transformers , accelerate

๐Ÿ“ Optimized HfFileSystem to avoid getting rate limited when browsing HuggingFaceFW/fineweb

๐Ÿ”จ HfApi & CLI improvements: prevent empty commits, create repo inside resource group, webhooks API, more options in the Search API, etc.

Check out the full release notes for more details:
Wauplin/huggingface_hub#7
๐Ÿ‘€
ยท
Wauplinย 
posted an update 6 months ago
view post
Post
3355
๐Ÿš€ I'm excited to announce that huggingface_hub's InferenceClient now supports OpenAI's Python client syntax! For developers integrating AI into their codebases, this means you can switch to open-source models with just three lines of code. Here's a quick example of how easy it is.

Why use the InferenceClient?
๐Ÿ”„ Seamless transition: keep your existing code structure while leveraging LLMs hosted on the Hugging Face Hub.
๐Ÿค— Direct integration: easily launch a model to run inference using our Inference Endpoint service.
๐Ÿš€ Stay Updated: always be in sync with the latest Text-Generation-Inference (TGI) updates.

More details in https://huggingface.co/docs/huggingface_hub/main/en/guides/inference#openai-compatibility
ยท
abhishekย 
posted an update 7 months ago
abhishekย 
posted an update 8 months ago
view post
Post
2935
๐Ÿšจ NEW TASK ALERT ๐Ÿšจ
๐ŸŽ‰ AutoTrain now supports Object Detection! ๐ŸŽ‰
Transform your projects with these powerful new features:
๐Ÿ”น Fine-tune any supported model from the Hugging Face Hub
๐Ÿ”น Seamless logging with TensorBoard or W&B
๐Ÿ”น Support for local and hub datasets
๐Ÿ”น Configurable training for tailored results
๐Ÿ”น Train locally or leverage Hugging Face Spaces
๐Ÿ”น Deployment-ready with API inference or Hugging Face endpoints
AutoTrain: https://hf.co/autotrain
abhishekย 
posted an update 8 months ago
view post
Post
3062
๐Ÿš€๐Ÿš€๐Ÿš€๐Ÿš€ Introducing AutoTrain Configs! ๐Ÿš€๐Ÿš€๐Ÿš€๐Ÿš€
Now you can train models using yaml config files! ๐Ÿ’ฅ These configs are easy to understand and are not at all overwhelming. So, even a person with almost zero knowledge of machine learning can train state of the art models without writing any code. Check out example configs in the config directory of autotrain-advanced github repo and feel free to share configs by creating a pull request ๐Ÿค—
Github repo: https://github.com/huggingface/autotrain-advanced
  • 2 replies
ยท
Wauplinย 
posted an update 8 months ago
view post
Post
1827
๐Ÿš€ Just released version 0.23.0 of the huggingface_hub Python library!

Exciting updates include:
๐Ÿ“ Seamless download to local dir!
๐Ÿ’ก Grammar and Tools in InferenceClient!
๐ŸŒ Documentation full translated to Korean!
๐Ÿ‘ฅ User API: get likes, upvotes, nb of repos, etc.!
๐Ÿงฉ Better model cards and encoding for ModelHubMixin!

Check out the full release notes for more details:
Wauplin/huggingface_hub#6
๐Ÿ‘€
abhishekย 
posted an update 8 months ago
abhishekย 
posted an update 8 months ago
view post
Post
2371
Trained another version of llama3-8b-instruct which beats the base model. This time without losing too many points on gsm8k benchmark. Again, using AutoTrain ๐Ÿ’ฅ pip install autotrain-advanced
Trained model: abhishek/autotrain-llama3-orpo-v2
  • 1 reply
ยท