Commit 2eaef42
committed
Disable dropout to workaround PyTorch 2.11 checkpoint recomputation bug
Disable dropout (resid_pdrop=0, attn_pdrop=0, embd_pdrop=0) in the
run_training_ac function to avoid SystemError from _VF.dropout returning
NULL during backward recomputation of GPT2Block. Dropout is irrelevant
to the memory profiling purpose of this tutorial.
Issue: #37741 parent f92d01b commit 2eaef42
1 file changed
Lines changed: 7 additions & 1 deletion
| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
306 | 306 | | |
307 | 307 | | |
308 | 308 | | |
309 | | - | |
| 309 | + | |
| 310 | + | |
| 311 | + | |
| 312 | + | |
| 313 | + | |
| 314 | + | |
| 315 | + | |
310 | 316 | | |
311 | 317 | | |
312 | 318 | | |
| |||
0 commit comments