Radamรฉs Ajna's picture

Radamรฉs Ajna

radames

AI & ML interests

None yet

Recent Activity

View all activity

Organizations

Spaces-explorers's profile picture CVPR Demo Track's profile picture MONAI's profile picture Gradio-Blocks-Party's profile picture Webhooks Explorers (BETA)'s profile picture Open Access AI Collective's profile picture The Team Ten's profile picture Open-Source AI Meetup's profile picture Stable Diffusion concepts library's profile picture Stable Diffusion Dreambooth Concepts Library's profile picture Daily's profile picture DragGan's profile picture meta-private's profile picture temp-org's profile picture Blog-explorers's profile picture Editing Images's profile picture leditsplusplus's profile picture sci-blender's profile picture Lilac AI's profile picture Latent Consistency's profile picture rtemp's profile picture ZeroGPU Explorers's profile picture cvmistralhackathon's profile picture Shizuku's profile picture Journalists on Hugging Face's profile picture Hugging Face - Visual Blocks's profile picture Social Post Explorers's profile picture +RAIN film festival's profile picture Agents-MCP-Hackathon's profile picture

radames's activity

reacted to cbensimon's post with ๐Ÿ‘๐Ÿ‘€๐Ÿค—๐Ÿš€๐Ÿ”ฅ 2 days ago
view post
Post
5654
๐Ÿš€ ZeroGPU medium size is now available as a power-user feature

Nothing too fancy for nowโ€”ZeroGPU Spaces still default to large (70GB VRAM)โ€”but this paves the way for:
- ๐Ÿ’ฐ size-based quotas / pricing (medium will offer significantly more usage than large)
- ๐Ÿฆฃ the upcoming xlarge size (141GB VRAM)

You can as of now control GPU size via a Space variable. Accepted values:
- auto (future default)
- medium
- large (current default)

The auto mode checks total CUDA tensor size during startup:
- More than 30GB โ†’ large
- Otherwise โ†’ medium
ยท
reacted to prithivMLmods's post with โค๏ธ 5 months ago
reacted to ucsahin's post with ๐Ÿ”ฅ๐Ÿš€ 7 months ago
view post
Post
4246
Florence-2 has a great capability of detecting various objects in a zero-shot setting with the task prompt "<OD>". However, if you want to detect specific objects that the base model is not able to in its current form, you can easily finetune it for this particular task. Below I show how to finetune the model to detect tables in a given image, but a similar process can be applied to detect any objects. Thanks to @andito , @merve , and @SkalskiP for sharing the fix for finetuning the Florence-2 model. Please also check their great blog post at https://huggingface.co/blog/finetune-florence2.

Colab notebook: https://colab.research.google.com/drive/1Y8GVjwzBIgfmfD3ZypDX5H1JA_VG0YDL?usp=sharing
Finetuned model: ucsahin/Florence-2-large-TableDetection
ยท
reacted to prithivMLmods's post with โค๏ธ 7 months ago
view post
Post
5978
New Style, New Mix, New Drop ๐Ÿงค

๐ŸงจFlux LoRA DLC: prithivMLmods/FLUX-LoRA-DLC

๐ŸŽ†Glowing-Body: prithivMLmods/Glowing-Body-Flux-LoRA
๐ŸŽ†Electric-Blue: prithivMLmods/Electric-Blue-Flux-LoRA
๐ŸŽ†Intense-Red: prithivMLmods/Intense-Red-Flux-LoRA
๐ŸŽ†Clouds-Illusion: prithivMLmods/Clouds-Illusion-Flux-LoRA
๐ŸŽ†Digital-Yellow: prithivMLmods/Digital-Yellow-Flux-LoRA

๐ŸงจFlux LoRA Collection: prithivMLmods/flux-lora-collections-66dd5908be2206cfaa8519be

.
.
.
@prithivMLmods
reacted to gokaygokay's post with ๐Ÿ”ฅ 10 months ago
view post
Post
15776
I've built a space for creating prompts for FLUX

gokaygokay/FLUX-Prompt-Generator

You can create long prompts from images or simple words. Enhance your short prompts with prompt enhancer. You can configure various settings such as artform, photo type, character details, scene details, style, and artist to create tailored prompts.

And you can combine all of them with custom prompts using llms (Mixtral, Mistral, Llama 3, and Mistral-Nemo).

The UI is a bit complex, but it includes almost everything you need. Choosing random option is the most fun!

And i've created some other spaces for using FLUX models with captioners and enhancers.

- gokaygokay/FLUX.1-dev-with-Captioner
ยท
reacted to sayakpaul's post with ๐Ÿ”ฅ 10 months ago
view post
Post
4536
Flux.1-Dev like images but in fewer steps.

Merging code (very simple), inference code, merged params: sayakpaul/FLUX.1-merged

Enjoy the Monday ๐Ÿค—
ยท
reacted to sayakpaul's post with โค๏ธ 11 months ago
view post
Post
3174
What is your favorite part of our Diffusers integration of Stable Diffusion 3?

My personal favorite is the ability to run it on a variety of different GPUs with minimal code changes.

Learn more about them here:
https://huggingface.co/blog/sd3
reacted to sayakpaul's post with ๐Ÿ”ฅ 11 months ago
reacted to merve's post with ๐Ÿคฏ๐Ÿ‘€ 11 months ago
view post
Post
3616
EPFL and Apple (at @EPFL-VILAB ) just released 4M-21: single any-to-any model that can do anything from text-to-image generation to generating depth masks! ๐Ÿ™€
4M is a multimodal training framework introduced by Apple and EPFL.
Resulting model takes image and text and output image and text ๐Ÿคฉ

Models: EPFL-VILAB/4m-models-660193abe3faf4b4d98a2742
Demo: EPFL-VILAB/4M
Paper: 4M-21: An Any-to-Any Vision Model for Tens of Tasks and Modalities (2406.09406)

This model consists of transformer encoder and decoder, where the key to multimodality lies in input and output data:

input and output tokens are decoded to generate bounding boxes, generated image's pixels, captions and more!

This model also learnt to generate canny maps, SAM edges and other things for steerable text-to-image generation ๐Ÿ–ผ๏ธ

The authors only added image-to-all capabilities for the demo, but you can try to use this model for text-to-image generation as well โ˜บ๏ธ
reacted to merve's post with โค๏ธ๐Ÿ”ฅ 11 months ago
view post
Post
4234
I love Depth Anything V2 ๐Ÿ˜
Itโ€™s Depth Anything, but scaled with both larger teacher model and a gigantic dataset!

Here's a small TLDR of paper with a lot of findings, experiments and more.
I have also created a collection that has the models, the dataset, the demo and CoreML converted model ๐Ÿ˜š merve/depth-anything-v2-release-6671902e798cd404513ffbf5

The authors have analyzed Marigold, a diffusion based model against Depth Anything and found out whatโ€™s up with using synthetic images vs real images for MDE:

๐Ÿ”– Real data has a lot of label noise, inaccurate depth maps (caused by depth sensors missing transparent objects etc) and there are many details overlooked

๐Ÿ”– Synthetic data have more precise and detailed depth labels and they are truly ground-truth, but thereโ€™s a distribution shift between real and synthetic images, and they have restricted scene coverage

The authors train different image encoders only on synthetic images and find out unless the encoder is very large the model canโ€™t generalize well (but large models generalize inherently anyway) ๐Ÿง
But they still fail encountering real images that have wide distribution in labels (e.g. diverse instances of objects) ๐Ÿฅฒ

Depth Anything v2 framework is to..

๐Ÿฆ– Train a teacher model based on DINOv2-G based on 595K synthetic images
๐Ÿท๏ธ Label 62M real images using teacher model
๐Ÿฆ• Train a student model using the real images labelled by teacher
Result: 10x faster and more accurate than Marigold!

The authors also construct a new benchmark called DA-2K that is less noisy, highly detailed and more diverse!
reacted to m-ric's post with ๐Ÿ‘ 11 months ago
view post
Post
3141
๐Ÿ’ฐ ๐—š๐—ฒ๐˜ ๐˜๐—ต๐—ฒ ๐—ฝ๐—ฟ๐—ถ๐—ฐ๐—ฒ ๐—ผ๐—ณ ๐—ฎ๐—ป๐˜† ๐—Ÿ๐—Ÿ๐—  ๐—”๐—ฃ๐—œ ๐—ฟ๐—ฒ๐—พ๐˜‚๐—ฒ๐˜€๐˜ โ‡’ ๐˜๐—ผ๐—ธ๐—ฒ๐—ป๐—ฐ๐—ผ๐˜€๐˜

I've just found out about ๐™ฐ๐š๐šŽ๐š—๐š๐™พ๐š™๐šœ-๐™ฐ๐™ธ/๐š๐š˜๐š”๐šŽ๐š—๐šŒ๐š˜๐šœ๐š (https://github.com/AgentOps-AI/tokencost).
๐—ง๐—ต๐—ถ๐˜€ ๐—น๐—ถ๐—ฏ๐—ฟ๐—ฎ๐—ฟ๐˜† ๐—ด๐—ถ๐˜ƒ๐—ฒ๐˜€ ๐˜†๐—ผ๐˜‚ ๐˜๐—ต๐—ฒ ๐—ฝ๐—ฟ๐—ถ๐—ฐ๐—ฒ ๐—ผ๐—ณ ๐˜†๐—ผ๐˜‚๐—ฟ ๐—ฐ๐—ฎ๐—น๐—น๐˜€ ๐˜๐—ผ ๐—ฎ๐—ป๐˜† ๐—Ÿ๐—Ÿ๐—  ๐—”๐—ฃ๐—œ: OpenAI, Anthropic, Mistral, AWS or Databricks...

For any model, you can use as input either string prompts or messages, and get as outputs either the price or token count.

Congrats to the AgentOps-AI team: this will be very useful when trying to get a ballpark estimate of a project's price, to compare APIs, or for precise monitoring of usage!

โœจ Daily reminder: ๐—ฟ๐˜‚๐—ป๐—ป๐—ถ๐—ป๐—ด ๐—ฎ๐—ป ๐—”๐Ÿญ๐Ÿฌ๐Ÿฌ ๐—ฐ๐—ผ๐˜€๐˜๐˜€ ๐˜†๐—ผ๐˜‚ ๐—ฒ๐˜…๐—ฎ๐—ฐ๐˜๐—น๐˜† $๐Ÿฌ.๐Ÿฌ๐Ÿฌ/๐—ต๐—ผ๐˜‚๐—ฟ (or 0.00โ‚ฌ in current exchange rates) on a HF space with ZeroGPU!
Learn more on ZeroGPU ๐Ÿ‘‰ https://www.datacenterdynamics.com/en/news/hugging-face-launches-zerogpu-project-to-democratize-ai-gives-away-10-million-worth-of-compute/
ยท
reacted to flozi00's post with โค๏ธ 11 months ago
view post
Post
2583
๐ŸŒŸ Progress in the German FineWeb edu reproduction ๐ŸŒŸ

We're delighted to share the launch of our new Data Quality Classification Model, designed specifically for evaluating educational content in German. This tool uses advanced machine learning techniques to assess texts across all educational levels, from primary school to university.

๐Ÿ” Inspired by Huggingface's fine web edu dataset, we've worked hard to refine data classification methods ensuring educators and learners access top-quality resources.
We're excited about the future as we continue improving our models and expanding our datasets.

Access the model here: pL-Community/GermanEduScorer-Qwen2-1.5b

๐Ÿ™ A huge thank you to David and Daryoush from Vago Solutions; Bjรถrn and Jan from Ellamind / DiscoResearch for their expert insights throughout this project. Your support has been crucial.
This project was made possible by the support of PrimeLine AI.
  • 2 replies
ยท
replied to dvilasuero's post 12 months ago