Model Card for hgnet_v2_b2_danube-delta
A HGNet v2 image classification model. This model was trained on the danube-delta
dataset (all the relevant bird species found int the Danube Delta region).
The species list is derived from data available at https://www.discoverdanubedelta.com/wp-content/uploads/2023/01/BirdsList-ian-2023.pdf.
Note: this is a subset of the eu-common
dataset.
Model Details
- Model Type: Image classification and detection backbone
- Model Stats:
- Params (M): 9.9
- Input image size: 256 x 256
- Dataset: danube-delta (368 classes)
Model Usage
Image Classification
import birder
from birder.inference.classification import infer_image
(net, model_info) = birder.load_pretrained_model("hgnet_v2_b2_danube-delta", inference=True)
# Get the image size the model was trained on
size = birder.get_size_from_signature(model_info.signature)
# Create an inference transform
transform = birder.classification_transform(size, model_info.rgb_stats)
image = "path/to/image.jpeg" # or a PIL image, must be loaded in RGB format
(out, _) = infer_image(net, image, transform)
# out is a NumPy array with shape of (1, 368), representing class probabilities.
Image Embeddings
import birder
from birder.inference.classification import infer_image
(net, model_info) = birder.load_pretrained_model("hgnet_v2_b2_danube-delta", inference=True)
# Get the image size the model was trained on
size = birder.get_size_from_signature(model_info.signature)
# Create an inference transform
transform = birder.classification_transform(size, model_info.rgb_stats)
image = "path/to/image.jpeg" # or a PIL image
(out, embedding) = infer_image(net, image, transform, return_embedding=True)
# embedding is a NumPy array with shape of (1, 2048)
Detection Feature Map
from PIL import Image
import birder
(net, model_info) = birder.load_pretrained_model("hgnet_v2_b2_danube-delta", inference=True)
# Get the image size the model was trained on
size = birder.get_size_from_signature(model_info.signature)
# Create an inference transform
transform = birder.classification_transform(size, model_info.rgb_stats)
image = Image.open("path/to/image.jpeg")
features = net.detection_features(transform(image).unsqueeze(0))
# features is a dict (stage name -> torch.Tensor)
print([(k, v.size()) for k, v in features.items()])
# Output example:
# [('stage1', torch.Size([1, 96, 64, 64])),
# ('stage2', torch.Size([1, 384, 32, 32])),
# ('stage3', torch.Size([1, 768, 17, 17])),
# ('stage4', torch.Size([1, 1536, 10, 10]))]
- Downloads last month
- -
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support