@@ -1406,17 +1406,17 @@ def f(x):
14061406 """\
14071407 def forward(self, arg0_1: "Sym(s0)", arg1_1: "f32[s0][1]cpu"):
14081408 clone: "f32[s0][1]cpu" = torch.ops.aten.clone.default(arg1_1)
1409- nonzero: "i64[u0, 1][1, 1 ]cpu" = torch.ops.aten.nonzero.default(clone); clone = None
1409+ nonzero: "i64[u0, 1][1, u0 ]cpu" = torch.ops.aten.nonzero.default(clone); clone = None
14101410 sym_size_int_1: "Sym(u0)" = torch.ops.aten.sym_size.int(nonzero, 0)
14111411 ge_1: "Sym(u0 >= 0)" = sym_size_int_1 >= 0; sym_size_int_1 = None
14121412 _assert_scalar = torch.ops.aten._assert_scalar.default(ge_1, "Runtime assertion failed for expression u0 >= 0 on node 'ge'"); ge_1 = _assert_scalar = None
1413- _to_copy: "f32[u0, 1][1, 1 ]cpu" = torch.ops.aten._to_copy.default(nonzero, dtype = torch.float32); nonzero = None
1413+ _to_copy: "f32[u0, 1][1, u0 ]cpu" = torch.ops.aten._to_copy.default(nonzero, dtype = torch.float32); nonzero = None
14141414 auto_functionalized_v2 = torch.ops.higher_order.auto_functionalized_v2(torch.ops.mylib.foo.default, _x_base_index = 0, _x_alias = True, _y_base_index = 1, _y_alias = True, _all_bases = [arg1_1, _to_copy]); _to_copy = None
14151415 getitem_1: "f32[s0][1]cpu" = auto_functionalized_v2[1]
1416- getitem_2: "f32[u0, 1][1, 1 ]cpu" = auto_functionalized_v2[2]; auto_functionalized_v2 = None
1416+ getitem_2: "f32[u0, 1][1, u0 ]cpu" = auto_functionalized_v2[2]; auto_functionalized_v2 = None
14171417 copy_: "f32[s0][1]cpu" = torch.ops.aten.copy_.default(arg1_1, getitem_1); arg1_1 = copy_ = None
14181418 alias_1: "f32[s0][1]cpu" = torch.ops.aten.alias.default(getitem_1); getitem_1 = None
1419- slice_2: "f32[u0, 1][1, 1 ]cpu" = torch.ops.aten.slice.Tensor(getitem_2); getitem_2 = None
1419+ slice_2: "f32[u0, 1][1, u0 ]cpu" = torch.ops.aten.slice.Tensor(getitem_2); getitem_2 = None
14201420 return (alias_1, slice_2)""" , # noqa: B950
14211421 ignore_comments = True ,
14221422 ignore_empty_lines = True ,
@@ -1427,19 +1427,19 @@ def forward(self, arg0_1: "Sym(s0)", arg1_1: "f32[s0][1]cpu"):
14271427 """\
14281428 def forward(self, arg0_1: "f32[2][1]cpu"):
14291429 clone: "f32[2][1]cpu" = torch.ops.aten.clone.default(arg0_1)
1430- nonzero: "i64[u0, 1][1, 1 ]cpu" = torch.ops.aten.nonzero.default(clone); clone = None
1430+ nonzero: "i64[u0, 1][1, u0 ]cpu" = torch.ops.aten.nonzero.default(clone); clone = None
14311431 sym_size_int: "Sym(u0)" = torch.ops.aten.sym_size.int(nonzero, 0)
14321432 ge_1: "Sym(u0 >= 0)" = sym_size_int >= 0
14331433 _assert_scalar = torch.ops.aten._assert_scalar.default(ge_1, "Runtime assertion failed for expression u0 >= 0 on node 'ge'"); ge_1 = _assert_scalar = None
14341434 le: "Sym(u0 <= 2)" = sym_size_int <= 2; sym_size_int = None
14351435 _assert_scalar_1 = torch.ops.aten._assert_scalar.default(le, "Runtime assertion failed for expression u0 <= 2 on node 'le'"); le = _assert_scalar_1 = None
1436- _to_copy: "f32[u0, 1][1, 1 ]cpu" = torch.ops.aten._to_copy.default(nonzero, dtype = torch.float32); nonzero = None
1436+ _to_copy: "f32[u0, 1][1, u0 ]cpu" = torch.ops.aten._to_copy.default(nonzero, dtype = torch.float32); nonzero = None
14371437 auto_functionalized_v2 = torch.ops.higher_order.auto_functionalized_v2(torch.ops.mylib.foo.default, _x_base_index = 0, _x_alias = True, _y_base_index = 1, _y_alias = True, _all_bases = [arg0_1, _to_copy]); _to_copy = None
14381438 getitem_1: "f32[2][1]cpu" = auto_functionalized_v2[1]
1439- getitem_2: "f32[u0, 1][1, 1 ]cpu" = auto_functionalized_v2[2]; auto_functionalized_v2 = None
1439+ getitem_2: "f32[u0, 1][1, u0 ]cpu" = auto_functionalized_v2[2]; auto_functionalized_v2 = None
14401440 copy_: "f32[2][1]cpu" = torch.ops.aten.copy_.default(arg0_1, getitem_1); arg0_1 = copy_ = None
14411441 alias_1: "f32[2][1]cpu" = torch.ops.aten.alias.default(getitem_1); getitem_1 = None
1442- slice_2: "f32[u0, 1][1, 1 ]cpu" = torch.ops.aten.slice.Tensor(getitem_2); getitem_2 = None
1442+ slice_2: "f32[u0, 1][1, u0 ]cpu" = torch.ops.aten.slice.Tensor(getitem_2); getitem_2 = None
14431443 return (alias_1, slice_2)""" , # noqa: B950
14441444 ignore_comments = True ,
14451445 ignore_empty_lines = True ,
@@ -1452,16 +1452,16 @@ def forward(self, arg0_1: "f32[2][1]cpu"):
14521452 graph_inductor ,
14531453 """\
14541454 def forward(self, arg0_1: "Sym(s0)", arg1_1: "f32[s0][1]cpu"):
1455- nonzero: "i64[u0, 1][1, 1 ]cpu" = torch.ops.aten.nonzero.default(arg1_1)
1455+ nonzero: "i64[u0, 1][1, u0 ]cpu" = torch.ops.aten.nonzero.default(arg1_1)
14561456 sym_size_int_1: "Sym(u0)" = torch.ops.aten.sym_size.int(nonzero, 0)
14571457 ge_1: "Sym(u0 >= 0)" = sym_size_int_1 >= 0; sym_size_int_1 = None
14581458 _assert_scalar = torch.ops.aten._assert_scalar.default(ge_1, "Runtime assertion failed for expression u0 >= 0 on node 'ge'"); ge_1 = _assert_scalar = None
1459- convert_element_type: "f32[u0, 1][1, 1 ]cpu" = torch.ops.prims.convert_element_type.default(nonzero, torch.float32); nonzero = None
1459+ convert_element_type: "f32[u0, 1][1, u0 ]cpu" = torch.ops.prims.convert_element_type.default(nonzero, torch.float32); nonzero = None
14601460 alias_default: "f32[s0][1]cpu" = torch.ops.aten.alias.default(arg1_1)
1461- alias_default_1: "f32[u0, 1][1, 1 ]cpu" = torch.ops.aten.alias.default(convert_element_type)
1461+ alias_default_1: "f32[u0, 1][1, u0 ]cpu" = torch.ops.aten.alias.default(convert_element_type)
14621462 foo_default = torch.ops.mylib.foo.default(alias_default, alias_default_1); alias_default = alias_default_1 = foo_default = None
14631463 copy_: "f32[s0][1]cpu" = torch.ops.aten.copy_.default(arg1_1, arg1_1); copy_ = None
1464- slice_2: "f32[u0, 1][1, 1 ]cpu" = torch.ops.aten.slice.Tensor(convert_element_type); convert_element_type = None
1464+ slice_2: "f32[u0, 1][1, u0 ]cpu" = torch.ops.aten.slice.Tensor(convert_element_type); convert_element_type = None
14651465 return (arg1_1, slice_2)""" , # noqa: B950
14661466 ignore_comments = True ,
14671467 ignore_empty_lines = True ,
@@ -1471,18 +1471,18 @@ def forward(self, arg0_1: "Sym(s0)", arg1_1: "f32[s0][1]cpu"):
14711471 graph_inductor ,
14721472 """\
14731473 def forward(self, arg0_1: "f32[2][1]cpu"):
1474- nonzero: "i64[u0, 1][1, 1 ]cpu" = torch.ops.aten.nonzero.default(arg0_1)
1474+ nonzero: "i64[u0, 1][1, u0 ]cpu" = torch.ops.aten.nonzero.default(arg0_1)
14751475 sym_size_int: "Sym(u0)" = torch.ops.aten.sym_size.int(nonzero, 0)
14761476 ge_1: "Sym(u0 >= 0)" = sym_size_int >= 0
14771477 _assert_scalar = torch.ops.aten._assert_scalar.default(ge_1, "Runtime assertion failed for expression u0 >= 0 on node 'ge'"); ge_1 = _assert_scalar = None
14781478 le: "Sym(u0 <= 2)" = sym_size_int <= 2; sym_size_int = None
14791479 _assert_scalar_1 = torch.ops.aten._assert_scalar.default(le, "Runtime assertion failed for expression u0 <= 2 on node 'le'"); le = _assert_scalar_1 = None
1480- convert_element_type: "f32[u0, 1][1, 1 ]cpu" = torch.ops.prims.convert_element_type.default(nonzero, torch.float32); nonzero = None
1480+ convert_element_type: "f32[u0, 1][1, u0 ]cpu" = torch.ops.prims.convert_element_type.default(nonzero, torch.float32); nonzero = None
14811481 alias_default: "f32[2][1]cpu" = torch.ops.aten.alias.default(arg0_1)
1482- alias_default_1: "f32[u0, 1][1, 1 ]cpu" = torch.ops.aten.alias.default(convert_element_type)
1482+ alias_default_1: "f32[u0, 1][1, u0 ]cpu" = torch.ops.aten.alias.default(convert_element_type)
14831483 foo_default = torch.ops.mylib.foo.default(alias_default, alias_default_1); alias_default = alias_default_1 = foo_default = None
14841484 copy_: "f32[2][1]cpu" = torch.ops.aten.copy_.default(arg0_1, arg0_1); copy_ = None
1485- slice_2: "f32[u0, 1][1, 1 ]cpu" = torch.ops.aten.slice.Tensor(convert_element_type); convert_element_type = None
1485+ slice_2: "f32[u0, 1][1, u0 ]cpu" = torch.ops.aten.slice.Tensor(convert_element_type); convert_element_type = None
14861486 return (arg0_1, slice_2)""" , # noqa: B950
14871487 ignore_comments = True ,
14881488 ignore_empty_lines = True ,
0 commit comments