from typing import Dict, Optional from torch import Tensor, nn from transformers import PreTrainedModel from .configuration_dptdepth import DPTDepthConfig from .models import DPTDepthModel as DPTDepth class DPTDepthModel(PreTrainedModel): """ The line that sets the config_class is not mandatory, unless you want to register your model with the auto classes """ config_class = DPTDepthConfig def __init__(self, config: DPTDepthConfig): super().__init__(config) self.model = DPTDepth() self.loss = nn.L1Loss() """ You can have your model return anything you want, but returning a dictionary with the loss included when labels are passed, will make your model directly usable inside the Trainer class. Using another output format is fine as long as you are planning on using your own training loop or another library for training. """ def forward(self, rgbs: Tensor, gts: Optional[Tensor] = None) -> Dict[str, Tensor]: logits = self.model(rgbs) if gts is not None: loss = self.loss(logits, gts) return {"loss": loss, "logits": logits} return {"logits": logits}