From 03ec517afba721fbd13af5d55408489470b07e43 Mon Sep 17 00:00:00 2001 From: comfyanonymous Date: Wed, 21 Aug 2024 00:47:19 -0400 Subject: [PATCH] Remove useless line, adjust windows default reserved vram. --- comfy/model_management.py | 2 +- comfy/ops.py | 1 - 2 files changed, 1 insertion(+), 2 deletions(-) diff --git a/comfy/model_management.py b/comfy/model_management.py index b899f8b8..bcc93779 100644 --- a/comfy/model_management.py +++ b/comfy/model_management.py @@ -369,7 +369,7 @@ def minimum_inference_memory(): EXTRA_RESERVED_VRAM = 200 * 1024 * 1024 if any(platform.win32_ver()): - EXTRA_RESERVED_VRAM = 400 * 1024 * 1024 #Windows is higher because of the shared vram issue + EXTRA_RESERVED_VRAM = 500 * 1024 * 1024 #Windows is higher because of the shared vram issue if args.reserve_vram is not None: EXTRA_RESERVED_VRAM = args.reserve_vram * 1024 * 1024 * 1024 diff --git a/comfy/ops.py b/comfy/ops.py index 6eb8eb3d..d1038edb 100644 --- a/comfy/ops.py +++ b/comfy/ops.py @@ -250,7 +250,6 @@ def fp8_linear(self, input): return None if len(input.shape) == 3: - out = torch.empty((input.shape[0], input.shape[1], self.weight.shape[0]), device=input.device, dtype=input.dtype) inn = input.view(-1, input.shape[2]).to(dtype) non_blocking = comfy.model_management.device_supports_non_blocking(input.device) w = cast_to(self.weight, device=input.device, non_blocking=non_blocking).t()