diff --git a/helpers/training/default_settings/safety_check.py b/helpers/training/default_settings/safety_check.py index 01444a32..518afd63 100644 --- a/helpers/training/default_settings/safety_check.py +++ b/helpers/training/default_settings/safety_check.py @@ -126,3 +126,9 @@ def safety_check(args, accelerator): f"--enable_xformers_memory_efficient_attention is only compatible with --attention_mechanism=diffusers. Please set --attention_mechanism=diffusers to enable this feature or disable xformers to use alternative attention mechanisms." ) sys.exit(1) + + if "nf4" in args.base_model_precision: + logger.error( + f"{args.base_model_precision} is not supported with SageAttention. Please select from int8 or fp8, or, disable quantisation to use SageAttention." + ) + sys.exit(1)