Skip to content

Latest commit

 

History

History
13 lines (7 loc) · 722 Bytes

qa.md

File metadata and controls

13 lines (7 loc) · 722 Bytes

Frequently asked questions

Q1) SophiaH, AdaHessian optimizers give RuntimeError: ~ tensors does not require grad and does not have a grad_fn in compute_hutchinson_hessian().

create_graph must be set True when calling backward(). here's an example.

Q2) Memory leak happens when using SophiaH, AdaHessian optimizers.

torch.autograd.grad with complex gradient flows sometimes leads memory leak issues, and you might encounter OOM issue. related issue

Q3) How to run visualizations?

Run python3 -m examples.visualize_optimizers on the project root.