Skip to content

Commit a91e8ed

Browse files
yiyixuxua-r-r-o-w
authored andcommitted
Revert "[LoRA] fix: lora loading when using with a device_mapped mode… (#9823)
Revert "[LoRA] fix: lora loading when using with a device_mapped model. (#9449)" This reverts commit 41e4779.
1 parent 3ef6487 commit a91e8ed

22 files changed

+8
-546
lines changed

docs/source/en/training/distributed_inference.md

-2
Original file line numberDiff line numberDiff line change
@@ -237,5 +237,3 @@ with torch.no_grad():
237237
```
238238

239239
By selectively loading and unloading the models you need at a given stage and sharding the largest models across multiple GPUs, it is possible to run inference with large models on consumer GPUs.
240-
241-
This workflow is also compatible with LoRAs via [`~loaders.StableDiffusionLoraLoaderMixin.load_lora_weights`]. However, only LoRAs without text encoder components are currently supported in this workflow.

src/diffusers/loaders/lora_base.py

+1-11
Original file line numberDiff line numberDiff line change
@@ -31,7 +31,6 @@
3131
delete_adapter_layers,
3232
deprecate,
3333
is_accelerate_available,
34-
is_accelerate_version,
3534
is_peft_available,
3635
is_transformers_available,
3736
logging,
@@ -215,18 +214,9 @@ def _optionally_disable_offloading(cls, _pipeline):
215214
is_model_cpu_offload = False
216215
is_sequential_cpu_offload = False
217216

218-
def model_has_device_map(model):
219-
if not is_accelerate_available() or is_accelerate_version("<", "0.14.0"):
220-
return False
221-
return getattr(model, "hf_device_map", None) is not None
222-
223217
if _pipeline is not None and _pipeline.hf_device_map is None:
224218
for _, component in _pipeline.components.items():
225-
if (
226-
isinstance(component, nn.Module)
227-
and hasattr(component, "_hf_hook")
228-
and not model_has_device_map(component)
229-
):
219+
if isinstance(component, nn.Module) and hasattr(component, "_hf_hook"):
230220
if not is_model_cpu_offload:
231221
is_model_cpu_offload = isinstance(component._hf_hook, CpuOffload)
232222
if not is_sequential_cpu_offload:

src/diffusers/loaders/unet.py

+1-11
Original file line numberDiff line numberDiff line change
@@ -39,7 +39,6 @@
3939
get_adapter_name,
4040
get_peft_kwargs,
4141
is_accelerate_available,
42-
is_accelerate_version,
4342
is_peft_version,
4443
is_torch_version,
4544
logging,
@@ -399,18 +398,9 @@ def _optionally_disable_offloading(cls, _pipeline):
399398
is_model_cpu_offload = False
400399
is_sequential_cpu_offload = False
401400

402-
def model_has_device_map(model):
403-
if not is_accelerate_available() or is_accelerate_version("<", "0.14.0"):
404-
return False
405-
return getattr(model, "hf_device_map", None) is not None
406-
407401
if _pipeline is not None and _pipeline.hf_device_map is None:
408402
for _, component in _pipeline.components.items():
409-
if (
410-
isinstance(component, nn.Module)
411-
and hasattr(component, "_hf_hook")
412-
and not model_has_device_map(component)
413-
):
403+
if isinstance(component, nn.Module) and hasattr(component, "_hf_hook"):
414404
if not is_model_cpu_offload:
415405
is_model_cpu_offload = isinstance(component._hf_hook, CpuOffload)
416406
if not is_sequential_cpu_offload:

src/diffusers/pipelines/pipeline_loading_utils.py

-7
Original file line numberDiff line numberDiff line change
@@ -36,7 +36,6 @@
3636
deprecate,
3737
get_class_from_dynamic_module,
3838
is_accelerate_available,
39-
is_accelerate_version,
4039
is_peft_available,
4140
is_transformers_available,
4241
logging,
@@ -948,9 +947,3 @@ def _get_ignore_patterns(
948947
)
949948

950949
return ignore_patterns
951-
952-
953-
def model_has_device_map(model):
954-
if not is_accelerate_available() or is_accelerate_version("<", "0.14.0"):
955-
return False
956-
return getattr(model, "hf_device_map", None) is not None

src/diffusers/pipelines/pipeline_utils.py

-31
Original file line numberDiff line numberDiff line change
@@ -85,7 +85,6 @@
8585
_update_init_kwargs_with_connected_pipeline,
8686
load_sub_model,
8787
maybe_raise_or_warn,
88-
model_has_device_map,
8988
variant_compatible_siblings,
9089
warn_deprecated_model_variant,
9190
)
@@ -407,16 +406,6 @@ def module_is_offloaded(module):
407406

408407
return hasattr(module, "_hf_hook") and isinstance(module._hf_hook, accelerate.hooks.CpuOffload)
409408

410-
# device-mapped modules should not go through any device placements.
411-
device_mapped_components = [
412-
key for key, component in self.components.items() if model_has_device_map(component)
413-
]
414-
if device_mapped_components:
415-
raise ValueError(
416-
"The following pipeline components have been found to use a device map: "
417-
f"{device_mapped_components}. This is incompatible with explicitly setting the device using `to()`."
418-
)
419-
420409
# .to("cuda") would raise an error if the pipeline is sequentially offloaded, so we raise our own to make it clearer
421410
pipeline_is_sequentially_offloaded = any(
422411
module_is_sequentially_offloaded(module) for _, module in self.components.items()
@@ -1013,16 +1002,6 @@ def enable_model_cpu_offload(self, gpu_id: Optional[int] = None, device: Union[t
10131002
The PyTorch device type of the accelerator that shall be used in inference. If not specified, it will
10141003
default to "cuda".
10151004
"""
1016-
# device-mapped modules should not go through any device placements.
1017-
device_mapped_components = [
1018-
key for key, component in self.components.items() if model_has_device_map(component)
1019-
]
1020-
if device_mapped_components:
1021-
raise ValueError(
1022-
"The following pipeline components have been found to use a device map: "
1023-
f"{device_mapped_components}. This is incompatible with `enable_model_cpu_offload()`."
1024-
)
1025-
10261005
is_pipeline_device_mapped = self.hf_device_map is not None and len(self.hf_device_map) > 1
10271006
if is_pipeline_device_mapped:
10281007
raise ValueError(
@@ -1125,16 +1104,6 @@ def enable_sequential_cpu_offload(self, gpu_id: Optional[int] = None, device: Un
11251104
The PyTorch device type of the accelerator that shall be used in inference. If not specified, it will
11261105
default to "cuda".
11271106
"""
1128-
# device-mapped modules should not go through any device placements.
1129-
device_mapped_components = [
1130-
key for key, component in self.components.items() if model_has_device_map(component)
1131-
]
1132-
if device_mapped_components:
1133-
raise ValueError(
1134-
"The following pipeline components have been found to use a device map: "
1135-
f"{device_mapped_components}. This is incompatible with `enable_sequential_cpu_offload()`."
1136-
)
1137-
11381107
if is_accelerate_available() and is_accelerate_version(">=", "0.14.0"):
11391108
from accelerate import cpu_offload
11401109
else:

tests/pipelines/audioldm2/test_audioldm2.py

-5
Original file line numberDiff line numberDiff line change
@@ -506,14 +506,9 @@ def test_to_dtype(self):
506506
model_dtypes = {key: component.dtype for key, component in components.items() if hasattr(component, "dtype")}
507507
self.assertTrue(all(dtype == torch.float16 for dtype in model_dtypes.values()))
508508

509-
@unittest.skip("Test currently not supported.")
510509
def test_sequential_cpu_offload_forward_pass(self):
511510
pass
512511

513-
@unittest.skip("Test currently not supported.")
514-
def test_calling_mco_raises_error_device_mapped_components(self):
515-
pass
516-
517512

518513
@nightly
519514
class AudioLDM2PipelineSlowTests(unittest.TestCase):

tests/pipelines/controlnet/test_controlnet.py

-24
Original file line numberDiff line numberDiff line change
@@ -514,18 +514,6 @@ def test_inference_multiple_prompt_input(self):
514514

515515
assert image.shape == (4, 64, 64, 3)
516516

517-
@unittest.skip("Test not supported.")
518-
def test_calling_mco_raises_error_device_mapped_components(self):
519-
pass
520-
521-
@unittest.skip("Test not supported.")
522-
def test_calling_to_raises_error_device_mapped_components(self):
523-
pass
524-
525-
@unittest.skip("Test not supported.")
526-
def test_calling_sco_raises_error_device_mapped_components(self):
527-
pass
528-
529517

530518
class StableDiffusionMultiControlNetOneModelPipelineFastTests(
531519
IPAdapterTesterMixin, PipelineTesterMixin, PipelineKarrasSchedulerTesterMixin, unittest.TestCase
@@ -709,18 +697,6 @@ def test_save_pretrained_raise_not_implemented_exception(self):
709697
except NotImplementedError:
710698
pass
711699

712-
@unittest.skip("Test not supported.")
713-
def test_calling_mco_raises_error_device_mapped_components(self):
714-
pass
715-
716-
@unittest.skip("Test not supported.")
717-
def test_calling_to_raises_error_device_mapped_components(self):
718-
pass
719-
720-
@unittest.skip("Test not supported.")
721-
def test_calling_sco_raises_error_device_mapped_components(self):
722-
pass
723-
724700

725701
@slow
726702
@require_torch_gpu

tests/pipelines/controlnet/test_controlnet_img2img.py

-12
Original file line numberDiff line numberDiff line change
@@ -389,18 +389,6 @@ def test_save_pretrained_raise_not_implemented_exception(self):
389389
except NotImplementedError:
390390
pass
391391

392-
@unittest.skip("Test not supported.")
393-
def test_calling_mco_raises_error_device_mapped_components(self):
394-
pass
395-
396-
@unittest.skip("Test not supported.")
397-
def test_calling_to_raises_error_device_mapped_components(self):
398-
pass
399-
400-
@unittest.skip("Test not supported.")
401-
def test_calling_sco_raises_error_device_mapped_components(self):
402-
pass
403-
404392

405393
@slow
406394
@require_torch_gpu

tests/pipelines/controlnet/test_controlnet_inpaint.py

-12
Original file line numberDiff line numberDiff line change
@@ -441,18 +441,6 @@ def test_save_pretrained_raise_not_implemented_exception(self):
441441
except NotImplementedError:
442442
pass
443443

444-
@unittest.skip("Test not supported.")
445-
def test_calling_mco_raises_error_device_mapped_components(self):
446-
pass
447-
448-
@unittest.skip("Test not supported.")
449-
def test_calling_to_raises_error_device_mapped_components(self):
450-
pass
451-
452-
@unittest.skip("Test not supported.")
453-
def test_calling_sco_raises_error_device_mapped_components(self):
454-
pass
455-
456444

457445
@slow
458446
@require_torch_gpu

tests/pipelines/controlnet/test_controlnet_sdxl.py

-24
Original file line numberDiff line numberDiff line change
@@ -683,18 +683,6 @@ def test_inference_batch_single_identical(self):
683683
def test_save_load_optional_components(self):
684684
return self._test_save_load_optional_components()
685685

686-
@unittest.skip("Test not supported.")
687-
def test_calling_mco_raises_error_device_mapped_components(self):
688-
pass
689-
690-
@unittest.skip("Test not supported.")
691-
def test_calling_to_raises_error_device_mapped_components(self):
692-
pass
693-
694-
@unittest.skip("Test not supported.")
695-
def test_calling_sco_raises_error_device_mapped_components(self):
696-
pass
697-
698686

699687
class StableDiffusionXLMultiControlNetOneModelPipelineFastTests(
700688
PipelineKarrasSchedulerTesterMixin, PipelineTesterMixin, SDXLOptionalComponentsTesterMixin, unittest.TestCase
@@ -899,18 +887,6 @@ def test_negative_conditions(self):
899887

900888
self.assertTrue(np.abs(image_slice_without_neg_cond - image_slice_with_neg_cond).max() > 1e-2)
901889

902-
@unittest.skip("Test not supported.")
903-
def test_calling_mco_raises_error_device_mapped_components(self):
904-
pass
905-
906-
@unittest.skip("Test not supported.")
907-
def test_calling_to_raises_error_device_mapped_components(self):
908-
pass
909-
910-
@unittest.skip("Test not supported.")
911-
def test_calling_sco_raises_error_device_mapped_components(self):
912-
pass
913-
914890

915891
@slow
916892
@require_torch_gpu

0 commit comments

Comments
 (0)