Croissant: A Metadata Format for ML-Ready Datasets
Paper
•
2403.19546
•
Published
Document and User Interface Parsing, Understanding, Q&A.
timm
release, v 1.0.12, with a focus on optimizers. The optimizer factory has been refactored, there's now a timm.optim.list_optimizers()
and new way to register optimizers and their attributes. As always you can use an timm
optimizer like a torch
one, just replace torch.optim
with timm.optim
adfactorbv
adopt
/ adoptw
(decoupled decay)mars
laprop
c
as well as cadamw
, cnadamw
, csgdw
, clamb
, crmsproptf
timm
, OpenCLIP
, and hopefully more.timm
scripts soon:
timm/plant-pathology-2021 timm
support for object detection, eventually segmentation, is finally under development :Otimm
model to use before commiting to download or training with a large dataset? Try mini-imagenet:
timm/mini-imagenet timm
release (1.0.11) is out now. A also wrote an article on one of the included models: https://huggingface.co/blog/rwightman/mambaouttimm
leaderboard
timm/leaderboard has been updated with the ability to select different hardware benchmark sets: RTX4090, RTX3090, two different CPUs along with some NCHW / NHWC layout and torch.compile (dynamo) variations. timm
(https://github.com/huggingface/pytorch-image-models) the go-to library for efficient image encoders for your mobile and edge devices, I've started working on an implementation of the new MobileNet-V4 model. Take a look at a short article I wrote about the model: https://huggingface.co/blog/rwightman/mobilenetv4timm
1.0 is finally out. The big feature that I wanted to complete before doing this? Having the unified feature map extraciton interface (features_only=True) supporting almost all models (97%) 🎉 See docs at https://huggingface.co/docs/timm/en/feature_extractionoutlines
library with transformers
to to define a JSON schema that the generation has to follow. It uses a Finite State Machine with token_id
as transitions.from datasets import load_dataset
pdfa_dataset = load_dataset('pixparse/pdfa-eng-wds', streaming=True)
IDL_dataset = load_dataset('pixparse/idl-wds', streaming=True)
import chug
task_cfg = chug.DataTaskDocReadCfg(
page_sampling='all',
)
data_cfg = chug.DataCfg(
source='pixparse/pdfa-eng-wds',
split='train',
batch_size=None,
format='hfids',
num_workers=0,
)
data_loader = chug.create_loader(
data_cfg,
task_cfg,
)
sample = next(iter(data_loader))