Skip to content

Commit b9d909d

Browse files
authored
Merge branch 'comfyanonymous:master' into master
2 parents a9b3e20 + f1c2301 commit b9d909d

3 files changed

Lines changed: 36 additions & 10 deletions

File tree

.github/workflows/stale-issues.yml

Lines changed: 21 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,21 @@
1+
name: 'Close stale issues'
2+
on:
3+
schedule:
4+
# Run daily at 430 am PT
5+
- cron: '30 11 * * *'
6+
permissions:
7+
issues: write
8+
9+
jobs:
10+
stale:
11+
runs-on: ubuntu-latest
12+
steps:
13+
- uses: actions/stale@v9
14+
with:
15+
stale-issue-message: "This issue is being marked stale because it has not had any activity for 30 days. Reply below within 7 days if your issue still isn't solved, and it will be left open. Otherwise, the issue will be closed automatically."
16+
days-before-stale: 30
17+
days-before-close: 7
18+
stale-issue-label: 'Stale'
19+
only-labels: 'User Support'
20+
exempt-all-assignees: true
21+
exempt-all-milestones: true

README.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -135,17 +135,17 @@ Put your VAE in: models/vae
135135
### AMD GPUs (Linux only)
136136
AMD users can install rocm and pytorch with pip if you don't have it already installed, this is the command to install the stable version:
137137

138-
```pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/rocm6.0```
138+
```pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/rocm6.1```
139139

140-
This is the command to install the nightly with ROCm 6.0 which might have some performance improvements:
140+
This is the command to install the nightly with ROCm 6.2 which might have some performance improvements:
141141

142-
```pip install --pre torch torchvision torchaudio --index-url https://download.pytorch.org/whl/nightly/rocm6.1```
142+
```pip install --pre torch torchvision torchaudio --index-url https://download.pytorch.org/whl/nightly/rocm6.2```
143143

144144
### NVIDIA
145145

146146
Nvidia users should install stable pytorch using this command:
147147

148-
```pip install torch torchvision torchaudio --extra-index-url https://download.pytorch.org/whl/cu121```
148+
```pip install torch torchvision torchaudio --extra-index-url https://download.pytorch.org/whl/cu124```
149149

150150
This is the command to install pytorch nightly instead which might have performance improvements:
151151

comfy/model_management.py

Lines changed: 11 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -370,12 +370,11 @@ def offloaded_memory(loaded_models, device):
370370
offloaded_mem += m.model_offloaded_memory()
371371
return offloaded_mem
372372

373-
def minimum_inference_memory():
374-
return (1024 * 1024 * 1024) * 1.2
373+
WINDOWS = any(platform.win32_ver())
375374

376-
EXTRA_RESERVED_VRAM = 200 * 1024 * 1024
377-
if any(platform.win32_ver()):
378-
EXTRA_RESERVED_VRAM = 500 * 1024 * 1024 #Windows is higher because of the shared vram issue
375+
EXTRA_RESERVED_VRAM = 400 * 1024 * 1024
376+
if WINDOWS:
377+
EXTRA_RESERVED_VRAM = 600 * 1024 * 1024 #Windows is higher because of the shared vram issue
379378

380379
if args.reserve_vram is not None:
381380
EXTRA_RESERVED_VRAM = args.reserve_vram * 1024 * 1024 * 1024
@@ -384,6 +383,9 @@ def minimum_inference_memory():
384383
def extra_reserved_memory():
385384
return EXTRA_RESERVED_VRAM
386385

386+
def minimum_inference_memory():
387+
return (1024 * 1024 * 1024) * 0.8 + extra_reserved_memory()
388+
387389
def unload_model_clones(model, unload_weights_only=True, force_unload=True):
388390
to_unload = []
389391
for i in range(len(current_loaded_models)):
@@ -1002,7 +1004,10 @@ def should_use_fp16(device=None, model_params=0, prioritize_performance=True, ma
10021004
nvidia_10_series = ["1080", "1070", "titan x", "p3000", "p3200", "p4000", "p4200", "p5000", "p5200", "p6000", "1060", "1050", "p40", "p100", "p6", "p4"]
10031005
for x in nvidia_10_series:
10041006
if x in props.name.lower():
1005-
return True
1007+
if WINDOWS or manual_cast:
1008+
return True
1009+
else:
1010+
return False #weird linux behavior where fp32 is faster
10061011

10071012
if manual_cast:
10081013
free_model_memory = maximum_vram_for_weights(device)

0 commit comments

Comments
 (0)