About 2,400,000 results
Open links in new tab
  1. Difference between autograd.grad and autograd.backward?

    Sep 12, 2021 · Descriptions The torch.autograd module is the automatic differentiation package for PyTorch. As described in the documentation it only requires minimal change to the code base to be …

  2. python - Autograd.grad () for Tensor in pytorch - Stack Overflow

    Feb 19, 2019 · As mentioned in the docs, the output of torch.autograd.grad is related to derivatives but it's not actually dy/dx. For example, assume you have a neural network that inputs a tensor of shape …

  3. python - How to Use AutoGrad Packages? - Stack Overflow

    How to Use AutoGrad Packages? Asked 7 years, 10 months ago Modified 5 years, 5 months ago Viewed 1k times

  4. python - Pytorch .backward () error when trying to compute policy ...

    Nov 27, 2024 · I am consistently getting a backward pass error or an inplace autograd error whenever trying to use a memory buffer to train my policy. Here is the function that causes the issues:

  5. Understanding pytorch autograd - Stack Overflow

    Jan 27, 2020 · Understanding pytorch autograd Asked 5 years, 11 months ago Modified 4 years, 11 months ago Viewed 767 times

  6. How to find and understand the autograd source code in PyTorch

    Nov 28, 2017 · Try to understand the autograd variable is probably the first thing, what you can do. From my understanding is autograd only a naming for the modules, which containing classes with …

  7. numpy - Python Autograd Not working "ValueError: setting an array ...

    Jan 20, 2022 · I am trying to compute the Jacobian of this function but when I run the code below I get "ValueError: setting an array element with a sequence." import autograd.numpy as np …

  8. python 3.x - Pytorch autograd.grad how to write the parameters for ...

    Sep 23, 2019 · In the documentation of torch.autograd.grad, it is stated that, for parameters, parameters: outputs (sequence of Tensor) – outputs of the differentiated function. inputs (sequence of T...

  9. Fast way to calculate Hessian matrix of model parameters in PyTorch

    Dec 23, 2022 · I want to calculate the Hessian matrix of a loss w.r.t. model parameters in PyTorch, but using torch.autograd.functional.hessian is not an option for me since it recomputes the model output …

  10. python - accessing autograd arraybox values - Stack Overflow

    May 24, 2019 · I am trying to use the python package autograd and I want to use values of an autograd arraybox in an interpolation np.interp(x, x_values, y_values) where y_values are store in an autograd …