diff --git a/comfy/model_management.py b/comfy/model_management.py index b899f8b8..bcc93779 100644 --- a/comfy/model_management.py +++ b/comfy/model_management.py @@ -369,7 +369,7 @@ def minimum_inference_memory(): EXTRA_RESERVED_VRAM = 200 * 1024 * 1024 if any(platform.win32_ver()): - EXTRA_RESERVED_VRAM = 400 * 1024 * 1024 #Windows is higher because of the shared vram issue + EXTRA_RESERVED_VRAM = 500 * 1024 * 1024 #Windows is higher because of the shared vram issue if args.reserve_vram is not None: EXTRA_RESERVED_VRAM = args.reserve_vram * 1024 * 1024 * 1024 diff --git a/comfy/ops.py b/comfy/ops.py index 6eb8eb3d..d1038edb 100644 --- a/comfy/ops.py +++ b/comfy/ops.py @@ -250,7 +250,6 @@ def fp8_linear(self, input): return None if len(input.shape) == 3: - out = torch.empty((input.shape[0], input.shape[1], self.weight.shape[0]), device=input.device, dtype=input.dtype) inn = input.view(-1, input.shape[2]).to(dtype) non_blocking = comfy.model_management.device_supports_non_blocking(input.device) w = cast_to(self.weight, device=input.device, non_blocking=non_blocking).t()