From: Zhang Dong Date: Wed, 20 Mar 2019 19:41:17 +0000 (-0700) Subject: fixed typo in shape_analysis.cpp (#18227) X-Git-Tag: accepted/tizen/6.5/unified/20211028.231830~725 X-Git-Url: http://review.tizen.org/git/?a=commitdiff_plain;h=1442808fcdb3fba7eae1122ad4fc6f4e2cb8df15;p=platform%2Fupstream%2Fpytorch.git fixed typo in shape_analysis.cpp (#18227) Summary: cc: VitalyFedyunin Pull Request resolved: https://github.com/pytorch/pytorch/pull/18227 Differential Revision: D14541764 Pulled By: VitalyFedyunin fbshipit-source-id: 9477deb1a99e6581f15a4de4d7631d747f56f3a6 --- diff --git a/torch/csrc/jit/passes/shape_analysis.cpp b/torch/csrc/jit/passes/shape_analysis.cpp index 6d771b4..43377e5 100644 --- a/torch/csrc/jit/passes/shape_analysis.cpp +++ b/torch/csrc/jit/passes/shape_analysis.cpp @@ -1026,7 +1026,7 @@ class ShapePropagator { // with ops that have a single output, we will fix up its return right // below. auto output_types = multidim_reduce_with_postprocess( - node, /*num_reduce_dim=*/1, /*integer_upcast=*/false); + node, /*num_reduced_dim=*/1, /*upcast_integer=*/false); if (!output_types.empty() && node->outputs().size() == 2) { output_types.push_back( output_types.back()->toScalarType(at::kLong)); @@ -1072,7 +1072,7 @@ class ShapePropagator { [](Node* node) -> type_vec_t { if (auto dim = node->get>(attr::dim)) { return multidim_reduce_with_postprocess( - node, /*num_reduce_dim=*/dim->size(), /*integer_upcast=*/false); + node, /*num_reduced_dim=*/dim->size(), /*upcast_integer=*/false); } return {}; }}; @@ -1093,7 +1093,7 @@ class ShapePropagator { if (auto dim = node->get>(attr::dim)) { // TODO: can dim contain duplicates? return multidim_reduce_with_postprocess( - node, /*num_reduce_dim=*/dim->size(), /*integer_upcast=*/true); + node, /*num_reduced_dim=*/dim->size(), /*upcast_integer=*/true); } return {}; }};