Spaces:
Runtime error
Runtime error
| # Copyright (c) Facebook, Inc. and its affiliates. | |
| from abc import ABCMeta, abstractmethod | |
| from typing import Dict | |
| import torch.nn as nn | |
| from detectron2.layers import ShapeSpec | |
| __all__ = ["Backbone"] | |
| class Backbone(nn.Module, metaclass=ABCMeta): | |
| """ | |
| Abstract base class for network backbones. | |
| """ | |
| def __init__(self): | |
| """ | |
| The `__init__` method of any subclass can specify its own set of arguments. | |
| """ | |
| super().__init__() | |
| def forward(self): | |
| """ | |
| Subclasses must override this method, but adhere to the same return type. | |
| Returns: | |
| dict[str->Tensor]: mapping from feature name (e.g., "res2") to tensor | |
| """ | |
| pass | |
| def size_divisibility(self) -> int: | |
| """ | |
| Some backbones require the input height and width to be divisible by a | |
| specific integer. This is typically true for encoder / decoder type networks | |
| with lateral connection (e.g., FPN) for which feature maps need to match | |
| dimension in the "bottom up" and "top down" paths. Set to 0 if no specific | |
| input size divisibility is required. | |
| """ | |
| return 0 | |
| def padding_constraints(self) -> Dict[str, int]: | |
| """ | |
| This property is a generalization of size_divisibility. Some backbones and training | |
| recipes require specific padding constraints, such as enforcing divisibility by a specific | |
| integer (e.g., FPN) or padding to a square (e.g., ViTDet with large-scale jitter | |
| in :paper:vitdet). `padding_constraints` contains these optional items like: | |
| { | |
| "size_divisibility": int, | |
| "square_size": int, | |
| # Future options are possible | |
| } | |
| `size_divisibility` will read from here if presented and `square_size` indicates the | |
| square padding size if `square_size` > 0. | |
| TODO: use type of Dict[str, int] to avoid torchscipt issues. The type of padding_constraints | |
| could be generalized as TypedDict (Python 3.8+) to support more types in the future. | |
| """ | |
| return {} | |
| def output_shape(self): | |
| """ | |
| Returns: | |
| dict[str->ShapeSpec] | |
| """ | |
| # this is a backward-compatible default | |
| return { | |
| name: ShapeSpec( | |
| channels=self._out_feature_channels[name], stride=self._out_feature_strides[name] | |
| ) | |
| for name in self._out_features | |
| } | |