Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
Expand Up @@ -39,12 +39,7 @@
valid_images,
validate_preprocess_arguments,
)
from ...utils import (
TensorType,
filter_out_non_signature_kwargs,
is_vision_available,
logging,
)
from ...utils import TensorType, filter_out_non_signature_kwargs, is_vision_available, logging


if is_vision_available():
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -21,6 +21,7 @@
from typing import Optional, Union

import torch
from torchvision.transforms.v2 import functional as F

from ...image_processing_utils_fast import (
BaseImageProcessorFast,
Expand All @@ -39,13 +40,7 @@
pil_torch_interpolation_mapping,
)
from ...processing_utils import Unpack
from ...utils import TensorType, auto_docstring, is_torchvision_v2_available


if is_torchvision_v2_available():
from torchvision.transforms.v2 import functional as F
else:
from torchvision.transforms import functional as F
from ...utils import TensorType, auto_docstring


class DeepseekVLHybridFastImageProcessorKwargs(DefaultFastImageProcessorKwargs):
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -29,11 +29,7 @@
from ...modeling_outputs import ModelOutput
from ...modeling_utils import PreTrainedModel
from ...processing_utils import Unpack
from ...utils import (
TransformersKwargs,
auto_docstring,
can_return_tuple,
)
from ...utils import TransformersKwargs, auto_docstring, can_return_tuple
from ..auto import AutoModel
from .configuration_deepseek_vl_hybrid import DeepseekVLHybridConfig

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -16,6 +16,7 @@

import torch
import torch.nn as nn
from torchvision.transforms.v2 import functional as F

from ...cache_utils import Cache
from ...image_processing_utils_fast import (
Expand Down Expand Up @@ -53,7 +54,6 @@
auto_docstring,
can_return_tuple,
filter_out_non_signature_kwargs,
is_torchvision_v2_available,
logging,
)
from ..auto import CONFIG_MAPPING, AutoConfig, AutoModel
Expand All @@ -70,12 +70,6 @@
from ..sam.modeling_sam import SamLayerNorm, SamVisionNeck


if is_torchvision_v2_available():
from torchvision.transforms.v2 import functional as F
else:
from torchvision.transforms import functional as F


logger = logging.get_logger(__name__)


Expand Down
8 changes: 2 additions & 6 deletions src/transformers/models/dpt/image_processing_dpt_fast.py
Original file line number Diff line number Diff line change
Expand Up @@ -25,6 +25,7 @@
from typing import TYPE_CHECKING, Optional, Union

import torch
from torchvision.transforms.v2 import functional as F

from ...image_processing_base import BatchFeature
from ...image_processing_utils_fast import BaseImageProcessorFast, DefaultFastImageProcessorKwargs
Expand All @@ -39,17 +40,12 @@
is_torch_tensor,
)
from ...processing_utils import Unpack
from ...utils import TensorType, auto_docstring, is_torchvision_v2_available, requires_backends
from ...utils import TensorType, auto_docstring, requires_backends


if TYPE_CHECKING:
from ...modeling_outputs import DepthEstimatorOutput

if is_torchvision_v2_available():
from torchvision.transforms.v2 import functional as F
else:
from torchvision.transforms import functional as F


class DPTFastImageProcessorKwargs(DefaultFastImageProcessorKwargs):
"""
Expand Down
6 changes: 1 addition & 5 deletions src/transformers/models/dpt/modular_dpt.py
Original file line number Diff line number Diff line change
Expand Up @@ -32,7 +32,6 @@
from ...utils import (
TensorType,
auto_docstring,
is_torchvision_v2_available,
requires_backends,
)
from ..beit.image_processing_beit_fast import BeitImageProcessorFast
Expand All @@ -41,10 +40,7 @@
if TYPE_CHECKING:
from ...modeling_outputs import DepthEstimatorOutput

if is_torchvision_v2_available():
from torchvision.transforms.v2 import functional as F
else:
from torchvision.transforms import functional as F
from torchvision.transforms.v2 import functional as F


def get_resize_output_image_size(
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -22,6 +22,7 @@
from typing import Optional, Union

import torch
from torchvision.transforms.v2 import functional as F

from ...image_processing_utils import BatchFeature, get_patch_output_size, select_best_resolution
from ...image_processing_utils_fast import (
Expand All @@ -41,13 +42,7 @@
get_image_size,
)
from ...processing_utils import Unpack
from ...utils import TensorType, auto_docstring, is_torchvision_v2_available


if is_torchvision_v2_available():
from torchvision.transforms.v2 import functional as F
else:
from torchvision.transforms import functional as F
from ...utils import TensorType, auto_docstring


class LlavaOnevisionFastImageProcessorKwargs(DefaultFastImageProcessorKwargs):
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -18,6 +18,7 @@

import torch
from torch import nn
from torchvision.transforms.v2 import functional as F

from transformers.models.llava_next.image_processing_llava_next_fast import LlavaNextImageProcessorFast
from transformers.models.llava_next_video.modeling_llava_next_video import (
Expand Down Expand Up @@ -50,16 +51,10 @@
TensorType,
auto_docstring,
can_return_tuple,
is_torchvision_v2_available,
logging,
)


if is_torchvision_v2_available():
from torchvision.transforms.v2 import functional as F
else:
from torchvision.transforms import functional as F

logger = logging.get_logger(__name__)


Expand Down
9 changes: 2 additions & 7 deletions src/transformers/models/owlv2/image_processing_owlv2_fast.py
Original file line number Diff line number Diff line change
Expand Up @@ -23,6 +23,7 @@
from typing import TYPE_CHECKING, Optional, Union

import torch
from torchvision.transforms.v2 import functional as F

from ...image_processing_utils_fast import BaseImageProcessorFast, BatchFeature, DefaultFastImageProcessorKwargs
from ...image_transforms import center_to_corners_format, group_images_by_shape, reorder_images
Expand All @@ -35,16 +36,10 @@
SizeDict,
)
from ...processing_utils import Unpack
from ...utils import TensorType, auto_docstring, is_torchvision_v2_available
from ...utils import TensorType, auto_docstring
from .image_processing_owlv2 import _scale_boxes, box_iou


if is_torchvision_v2_available():
from torchvision.transforms.v2 import functional as F
else:
from torchvision.transforms import functional as F


if TYPE_CHECKING:
from .modeling_owlv2 import Owlv2ObjectDetectionOutput

Expand Down
8 changes: 1 addition & 7 deletions src/transformers/models/owlv2/modular_owlv2.py
Original file line number Diff line number Diff line change
Expand Up @@ -18,6 +18,7 @@
from typing import Optional, Union

import torch
from torchvision.transforms.v2 import functional as F

from ...image_processing_utils_fast import (
BaseImageProcessorFast,
Expand All @@ -37,17 +38,10 @@
from ...utils import (
TensorType,
auto_docstring,
is_torchvision_v2_available,
)
from ..owlvit.image_processing_owlvit_fast import OwlViTImageProcessorFast


if is_torchvision_v2_available():
from torchvision.transforms.v2 import functional as F
else:
from torchvision.transforms import functional as F


class Owlv2FastImageProcessorKwargs(DefaultFastImageProcessorKwargs): ...


Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -8,6 +8,7 @@
from typing import Any, Optional, Union

import torch
from torchvision.transforms.v2 import functional as F

from ...image_processing_utils import BatchFeature
from ...image_processing_utils_fast import (
Expand Down Expand Up @@ -36,12 +37,6 @@
from .image_processing_rt_detr import get_size_with_aspect_ratio


if is_torchvision_v2_available():
from torchvision.transforms.v2 import functional as F
else:
from torchvision.transforms import functional as F


class RTDetrFastImageProcessorKwargs(DefaultFastImageProcessorKwargs):
r"""
format (`str`, *optional*, defaults to `AnnotationFormat.COCO_DETECTION`):
Expand Down
8 changes: 1 addition & 7 deletions src/transformers/models/rt_detr/modular_rt_detr.py
Original file line number Diff line number Diff line change
Expand Up @@ -2,6 +2,7 @@
from typing import Optional, Union

import torch
from torchvision.transforms.v2 import functional as F

from transformers.models.detr.image_processing_detr_fast import DetrFastImageProcessorKwargs, DetrImageProcessorFast

Expand All @@ -22,18 +23,11 @@
from ...processing_utils import Unpack
from ...utils import (
TensorType,
is_torchvision_v2_available,
logging,
requires_backends,
)


if is_torchvision_v2_available():
from torchvision.transforms.v2 import functional as F
else:
from torchvision.transforms import functional as F


logger = logging.get_logger(__name__)

SUPPORTED_ANNOTATION_FORMATS = (AnnotationFormat.COCO_DETECTION,)
Expand Down
5 changes: 1 addition & 4 deletions src/transformers/models/sam2_video/modeling_sam2_video.py
Original file line number Diff line number Diff line change
Expand Up @@ -39,10 +39,7 @@
from ...modeling_utils import ALL_ATTENTION_FUNCTIONS, PreTrainedModel
from ...processing_utils import Unpack
from ...pytorch_utils import compile_compatible_method_lru_cache
from ...utils import (
ModelOutput,
auto_docstring,
)
from ...utils import ModelOutput, auto_docstring
from ...utils.generic import OutputRecorder, TransformersKwargs
from ..auto import AutoModel
from .configuration_sam2_video import Sam2VideoConfig, Sam2VideoMaskDecoderConfig, Sam2VideoPromptEncoderConfig
Expand Down
8 changes: 0 additions & 8 deletions src/transformers/models/sam2_video/modular_sam2_video.py
Original file line number Diff line number Diff line change
Expand Up @@ -36,8 +36,6 @@
from ...utils import (
ModelOutput,
auto_docstring,
is_torchvision_available,
is_torchvision_v2_available,
logging,
)
from ...utils.generic import OutputRecorder, TransformersKwargs
Expand All @@ -59,12 +57,6 @@
from ..sam2.processing_sam2 import Sam2Processor


if is_torchvision_available() and is_torchvision_v2_available():
from torchvision.transforms.v2 import functional as F
else:
from torchvision.transforms import functional as F


logger = logging.get_logger(__name__)


Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -22,6 +22,7 @@
from typing import Optional, Union

import torch
from torchvision.transforms.v2 import functional as F

from ...image_processing_utils import BatchFeature
from ...image_processing_utils_fast import (
Expand All @@ -40,13 +41,7 @@
is_torch_tensor,
)
from ...processing_utils import Unpack
from ...utils import TensorType, auto_docstring, is_torchvision_v2_available


if is_torchvision_v2_available():
from torchvision.transforms.v2 import functional as F
else:
from torchvision.transforms import functional as F
from ...utils import TensorType, auto_docstring


class SegformerFastImageProcessorKwargs(DefaultFastImageProcessorKwargs):
Expand Down Expand Up @@ -135,9 +130,7 @@ def _preprocess_image_like_inputs(
"do_normalize": False,
"do_rescale": False,
# Nearest interpolation is used for segmentation maps instead of BILINEAR.
"interpolation": F.InterpolationMode.NEAREST_EXACT
if is_torchvision_v2_available()
else F.InterpolationMode.NEAREST,
"interpolation": F.InterpolationMode.NEAREST_EXACT,
}
)
processed_segmentation_maps = self._preprocess(
Expand Down
12 changes: 2 additions & 10 deletions src/transformers/models/segformer/modular_segformer.py
Original file line number Diff line number Diff line change
Expand Up @@ -17,6 +17,7 @@
from typing import Optional, Union

import torch
from torchvision.transforms.v2 import functional as F

from transformers.models.beit.image_processing_beit_fast import BeitFastImageProcessorKwargs, BeitImageProcessorFast

Expand All @@ -36,16 +37,9 @@
from ...processing_utils import Unpack
from ...utils import (
TensorType,
is_torchvision_v2_available,
)


if is_torchvision_v2_available():
from torchvision.transforms.v2 import functional as F
else:
from torchvision.transforms import functional as F


class SegformerFastImageProcessorKwargs(BeitFastImageProcessorKwargs):
pass

Expand Down Expand Up @@ -96,9 +90,7 @@ def _preprocess_image_like_inputs(
"do_normalize": False,
"do_rescale": False,
# Nearest interpolation is used for segmentation maps instead of BILINEAR.
"interpolation": F.InterpolationMode.NEAREST_EXACT
if is_torchvision_v2_available()
else F.InterpolationMode.NEAREST,
"interpolation": F.InterpolationMode.NEAREST_EXACT,
}
)
processed_segmentation_maps = self._preprocess(
Expand Down
6 changes: 1 addition & 5 deletions src/transformers/utils/import_utils.py
Original file line number Diff line number Diff line change
Expand Up @@ -400,11 +400,7 @@ def is_torchvision_available() -> bool:


def is_torchvision_v2_available() -> bool:
if not is_torchvision_available():
return False

# NOTE: We require torchvision>=0.15 as v2 transforms are available from this version: https://pytorch.org/vision/stable/transforms.html#v1-or-v2-which-one-should-i-use
return version.parse(_torchvision_version) >= version.parse("0.15")
return is_torchvision_available()


def is_galore_torch_available() -> Union[tuple[bool, str], bool]:
Expand Down