blob: a79906e2cb485c0dce20d39dfde2b8a4056e605f [file] [log] [blame]
# Owner(s): ["module: autograd"]
import logging
import torch
from torch.testing._internal.logging_utils import LoggingTestCase, make_logging_test
class TestAutogradLogging(LoggingTestCase):
@make_logging_test(autograd=logging.DEBUG)
def test_logging(self, records):
a = torch.rand(10, requires_grad=True)
b = a.mul(2).div(3).sum()
c = b.clone()
torch.autograd.backward((b, c))
self.assertEqual(len(records), 5)
expected = [
"CloneBackward0",
"SumBackward0",
"DivBackward0",
"MulBackward0",
"AccumulateGrad",
]
for i, record in enumerate(records):
self.assertIn(expected[i], record.getMessage())
if __name__ == "__main__":
from torch._dynamo.test_case import run_tests
run_tests()