Spaces:
Runtime error
Runtime error
# Copyright 2024 EPFL and Apple Inc. | |
# | |
# Licensed under the Apache License, Version 2.0 (the "License"); | |
# you may not use this file except in compliance with the License. | |
# You may obtain a copy of the License at | |
# | |
# http://www.apache.org/licenses/LICENSE-2.0 | |
# | |
# Unless required by applicable law or agreed to in writing, software | |
# distributed under the License is distributed on an "AS IS" BASIS, | |
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |
# See the License for the specific language governing permissions and | |
# limitations under the License. | |
from typing import List, Union | |
import torch | |
import torch.nn as nn | |
import torch.nn.functional as F | |
import torchvision.transforms as transforms | |
import torchvision.transforms.functional as TF | |
from torchvision.models.feature_extraction import create_feature_extractor | |
from einops import rearrange | |
import timm | |
class TimmPerceptualLoss(nn.Module): | |
"""Perceptual loss module using features from arbitrary timm models. | |
Args: | |
model_id: timm model id. E.g. 'vit_base_patch14_dinov2.lvd142m' | |
feature_ids: List or hyphen-separated string of feature names to extract from the model. | |
For example, 'blocks.2-blocks.5-blocks.8-blocks.11'. To list all available features, use: | |
```python | |
from torchvision.models.feature_extraction import get_graph_node_names | |
nodes, _ = get_graph_node_names(model) | |
``` | |
feature_loss: Feature loss to use. One of ['cosine' or 'cos', 'l1' or 'mae']. Default: 'cosine'. | |
If 'l1' or 'mae' is used, the features will be normalized first. If 'cosine' or 'cos' is | |
used, the features will not be normalized, but the cosine similarity will be computed, | |
which is equivalent to normalization + MSE up to a factor of 2. | |
""" | |
def __init__(self, | |
model_id: str, | |
feature_ids: Union[str, List[str]], | |
feature_loss: str = 'cosine'): | |
super().__init__() | |
feature_ids = feature_ids.split('-') if isinstance(feature_ids, str) else feature_ids | |
self.feature_ids = feature_ids | |
self.feature_loss = feature_loss | |
self.model = timm.create_model(model_id, pretrained=True) | |
self.feature_extractor = create_feature_extractor(self.model, return_nodes=self.feature_ids) | |
# Transforms to preprocess inputs to the model | |
self.data_config = timm.data.resolve_model_data_config(self.model) | |
self.percept_transform = transforms.Compose([ | |
transforms.Normalize((-1.0, -1.0, -1.0), (2.0, 2.0, 2.0)), # [-1, 1] -> [0, 1] | |
transforms.Normalize(self.data_config['mean'], self.data_config['std']), # [0, 1] -> standardize with pre-computed statistics | |
transforms.Resize(self.data_config['input_size'][-2:], interpolation=TF.InterpolationMode.BILINEAR, antialias=True), | |
]) | |
def forward(self, preds: torch.Tensor, targets: torch.Tensor, preprocess_inputs=False) -> torch.Tensor: | |
""" | |
Compute perceptual loss between predictions and targets. If | |
preprocess_inputs is True, it is assumed that the targets are | |
scaled to the [-1, 1] range. Predictions will be scaled | |
assuming the same input range. | |
Args: | |
preds: Predictions tensor of shape (B, C, H, W) | |
targets: Targets tensor of shape (B, C, H, W) | |
preprocess_inputs: If inputs are scaled to [-1, 1], enable | |
this to apply model specific preprocessing. Default: False. | |
Returns: | |
Perceptual loss between predictions and targets. | |
""" | |
# Preprocess predictions and targets for the given feature extractor | |
if preprocess_inputs: | |
preds = self.percept_transform(preds) | |
targets = self.percept_transform(targets) | |
# Extract features from predictions and targets | |
# Each is a dict of feature_name: feature_tensor | |
feats_preds = self.feature_extractor(preds) | |
feats_targets = self.feature_extractor(targets) | |
loss = 0 | |
for feat_name in feats_preds.keys(): | |
# Get individual feature map and reshape from (B, C, H, W) to (B, N, C) if needed | |
feat_preds = feats_preds[feat_name] | |
feat_targets = feats_targets[feat_name] | |
if feat_preds.ndim == 4: | |
feat_preds = rearrange(feat_preds, 'b c h w -> b (h w) c') | |
feat_targets = rearrange(feat_targets, 'b c h w -> b (h w) c') | |
# Compute feature-wise loss | |
if self.feature_loss in ['l1', 'mae']: | |
feat_preds = F.normalize(feat_preds, dim=-1) | |
feat_targets = F.normalize(feat_targets, dim=-1) | |
loss += F.l1_loss(feat_preds, feat_targets, reduction='none').sum(-1).mean(-1) | |
elif self.feature_loss in ['cosine', 'cos']: | |
loss += 1 - F.cosine_similarity(feat_preds, feat_targets, dim=-1).mean(dim=-1) | |
else: | |
raise ValueError(f'Unknown feature loss: {self.feature_loss}') | |
loss /= preds.shape[0] | |
return loss |