diff --git a/dinov2/models/vision_transformer.py b/dinov2/models/vision_transformer.py index 13b44ae3c..74df767eb 100644 --- a/dinov2/models/vision_transformer.py +++ b/dinov2/models/vision_transformer.py @@ -12,6 +12,7 @@ import logging from typing import Sequence, Tuple, Union, Callable +import numpy as np import torch import torch.nn as nn import torch.utils.checkpoint @@ -116,7 +117,7 @@ def __init__( if drop_path_uniform is True: dpr = [drop_path_rate] * depth else: - dpr = [x.item() for x in torch.linspace(0, drop_path_rate, depth)] # stochastic depth decay rule + dpr = np.linspace(0, drop_path_rate, depth).tolist() # stochastic depth decay rule if ffn_layer == "mlp": logger.info("using MLP layer as FFN")