mirror of
https://github.com/comfyanonymous/ComfyUI.git
synced 2025-01-11 10:25:16 +00:00
Make VAE use common function to get free memory.
This commit is contained in:
parent
798c90e1c0
commit
cc8baf1080
@ -7,6 +7,7 @@ from einops import rearrange
|
|||||||
from typing import Optional, Any
|
from typing import Optional, Any
|
||||||
|
|
||||||
from ldm.modules.attention import MemoryEfficientCrossAttention
|
from ldm.modules.attention import MemoryEfficientCrossAttention
|
||||||
|
import model_management
|
||||||
|
|
||||||
try:
|
try:
|
||||||
import xformers
|
import xformers
|
||||||
@ -199,12 +200,7 @@ class AttnBlock(nn.Module):
|
|||||||
|
|
||||||
r1 = torch.zeros_like(k, device=q.device)
|
r1 = torch.zeros_like(k, device=q.device)
|
||||||
|
|
||||||
stats = torch.cuda.memory_stats(q.device)
|
mem_free_total = model_management.get_free_memory(q.device)
|
||||||
mem_active = stats['active_bytes.all.current']
|
|
||||||
mem_reserved = stats['reserved_bytes.all.current']
|
|
||||||
mem_free_cuda, _ = torch.cuda.mem_get_info(torch.cuda.current_device())
|
|
||||||
mem_free_torch = mem_reserved - mem_active
|
|
||||||
mem_free_total = mem_free_cuda + mem_free_torch
|
|
||||||
|
|
||||||
gb = 1024 ** 3
|
gb = 1024 ** 3
|
||||||
tensor_size = q.shape[0] * q.shape[1] * k.shape[2] * q.element_size()
|
tensor_size = q.shape[0] * q.shape[1] * k.shape[2] * q.element_size()
|
||||||
|
Loading…
Reference in New Issue
Block a user