diff --git a/src/diffusers/pipelines/prx/pipeline_prx.py b/src/diffusers/pipelines/prx/pipeline_prx.py index 0171c9a42a40..e14815b91c41 100644 --- a/src/diffusers/pipelines/prx/pipeline_prx.py +++ b/src/diffusers/pipelines/prx/pipeline_prx.py @@ -18,7 +18,6 @@ import urllib.parse as ul from typing import Callable -import ftfy import torch from transformers import ( AutoTokenizer, @@ -34,13 +33,13 @@ from diffusers.pipelines.pipeline_utils import DiffusionPipeline from diffusers.pipelines.prx.pipeline_output import PRXPipelineOutput from diffusers.schedulers import FlowMatchEulerDiscreteScheduler -from diffusers.utils import ( - logging, - replace_example_docstring, -) +from diffusers.utils import is_ftfy_available, logging, replace_example_docstring from diffusers.utils.torch_utils import randn_tensor +if is_ftfy_available(): + import ftfy + DEFAULT_RESOLUTION = 512 ASPECT_RATIO_256_BIN = { diff --git a/tests/pipelines/prx/test_pipeline_prx.py b/tests/pipelines/prx/test_pipeline_prx.py index 46c6a5760e22..1d04da1844fa 100644 --- a/tests/pipelines/prx/test_pipeline_prx.py +++ b/tests/pipelines/prx/test_pipeline_prx.py @@ -1,7 +1,6 @@ import unittest import numpy as np -import pytest import torch from transformers import AutoTokenizer from transformers.models.t5gemma.configuration_t5gemma import T5GemmaConfig, T5GemmaModuleConfig @@ -11,17 +10,11 @@ from diffusers.models.transformers.transformer_prx import PRXTransformer2DModel from diffusers.pipelines.prx.pipeline_prx import PRXPipeline from diffusers.schedulers import FlowMatchEulerDiscreteScheduler -from diffusers.utils import is_transformers_version from ..pipeline_params import TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin -@pytest.mark.xfail( - condition=is_transformers_version(">", "4.57.1"), - reason="See https://github.com/huggingface/diffusers/pull/12456#issuecomment-3424228544", - strict=False, -) class PRXPipelineFastTests(PipelineTesterMixin, unittest.TestCase): pipeline_class = PRXPipeline params = TEXT_TO_IMAGE_PARAMS - {"cross_attention_kwargs"}