@@ -3902,7 +3902,7 @@ def codegen(self, wrapper):
3902
3902
self .output_view ,
3903
3903
self .codegen_reference (),
3904
3904
args ,
3905
- self .cpp_kernel if V .graph .cpp_wrapper else self .kernel ,
3905
+ self .cpp_kernel_name if V .graph .cpp_wrapper else self .kernel ,
3906
3906
)
3907
3907
3908
3908
def __init__ (
@@ -3913,7 +3913,7 @@ def __init__(
3913
3913
kwargs = None ,
3914
3914
output_view = None ,
3915
3915
kernel = None ,
3916
- cpp_kernel = None ,
3916
+ cpp_kernel_name = None ,
3917
3917
ordered_kwargs_for_cpp_kernel = (),
3918
3918
):
3919
3919
super ().__init__ (
@@ -3922,7 +3922,7 @@ def __init__(
3922
3922
self .output_view = output_view
3923
3923
self .name = V .graph .register_buffer (self )
3924
3924
self .kernel = kernel
3925
- self .cpp_kernel = cpp_kernel
3925
+ self .cpp_kernel_name = cpp_kernel_name
3926
3926
self .ordered_kwargs_for_cpp_kernel = ordered_kwargs_for_cpp_kernel
3927
3927
3928
3928
def should_allocate (self ):
@@ -3941,13 +3941,13 @@ def __init__(self, count: int, device: torch.device):
3941
3941
inputs = [],
3942
3942
constant_args = [limits .min , limits .max , [count ]],
3943
3943
kernel = "aten.randint.low_out" ,
3944
- cpp_kernel = "at::randint_out" ,
3944
+ cpp_kernel_name = "at::randint_out" ,
3945
3945
)
3946
3946
3947
3947
3948
3948
class ExternKernelAlloc (ExternKernel ):
3949
3949
def codegen_kernel_name (self ):
3950
- return self .cpp_kernel if V .graph .cpp_wrapper else self .kernel
3950
+ return self .cpp_kernel_name if V .graph .cpp_wrapper else self .kernel
3951
3951
3952
3952
def codegen (self , wrapper ):
3953
3953
self .codegen_comment (wrapper )
@@ -3963,15 +3963,15 @@ def __init__(
3963
3963
constant_args = (),
3964
3964
kwargs = None ,
3965
3965
kernel = None ,
3966
- cpp_kernel = None ,
3966
+ cpp_kernel_name = None ,
3967
3967
ordered_kwargs_for_cpp_kernel = (),
3968
3968
):
3969
3969
super ().__init__ (
3970
3970
None , layout , self .unwrap_storage (inputs ), constant_args , kwargs or {}
3971
3971
)
3972
3972
self .name = V .graph .register_buffer (self )
3973
3973
self .kernel = kernel
3974
- self .cpp_kernel = cpp_kernel
3974
+ self .cpp_kernel_name = cpp_kernel_name
3975
3975
self .ordered_kwargs_for_cpp_kernel = ordered_kwargs_for_cpp_kernel
3976
3976
3977
3977
def should_allocate (self ):
@@ -4182,7 +4182,7 @@ def codegen(self, wrapper):
4182
4182
get_operator_enum = {"add" : "sum" , "multiply" : "prod" }
4183
4183
if reduce in get_operator_enum :
4184
4184
reduce = get_operator_enum [reduce ]
4185
- self .cpp_kernel = self .get_cpp_kernel (self .fn , reduce )
4185
+ self .cpp_kernel_name = self .get_cpp_kernel (self .fn , reduce )
4186
4186
4187
4187
if self .src_is_tensor :
4188
4188
(x , index , src ) = (t .codegen_reference () for t in self .inputs )
@@ -4192,7 +4192,7 @@ def codegen(self, wrapper):
4192
4192
wrapper .generate_scatter_fallback (
4193
4193
x ,
4194
4194
[x , self .constant_args [0 ], index , src ],
4195
- self .cpp_kernel if V .graph .cpp_wrapper else self .kernel ,
4195
+ self .cpp_kernel_name if V .graph .cpp_wrapper else self .kernel ,
4196
4196
self .fn ,
4197
4197
self .src_is_tensor ,
4198
4198
reduce ,
@@ -4281,7 +4281,7 @@ def codegen(self, wrapper):
4281
4281
args = [x , indices_str , values , * self .codegen_const_args ()]
4282
4282
wrapper .writeline (
4283
4283
wrapper .wrap_kernel_call (
4284
- self .cpp_kernel if V .graph .cpp_wrapper else self .kernel , args
4284
+ self .cpp_kernel_name if V .graph .cpp_wrapper else self .kernel , args
4285
4285
)
4286
4286
)
4287
4287
@@ -4305,7 +4305,7 @@ def __init__(self, x, indices, values, accumulate):
4305
4305
(accumulate ,),
4306
4306
)
4307
4307
self .name = V .graph .register_buffer (self )
4308
- self .cpp_kernel = "at::index_put_"
4308
+ self .cpp_kernel_name = "at::index_put_"
4309
4309
self .kernel = "aten.index_put_"
4310
4310
mark_node_as_mutating (self , x )
4311
4311
@@ -4452,10 +4452,10 @@ def is_not_write(arg):
4452
4452
is_not_write (x ) for x in kernel ._schema .returns
4453
4453
), f"{ kernel .__name__ } with alias_info returns is not supported with cpp_wrapper"
4454
4454
4455
- self .cpp_kernel = kernel ._schema .name
4455
+ self .cpp_kernel_name = kernel ._schema .name
4456
4456
self .cpp_kernel_overload_name = kernel ._schema .overload_name
4457
4457
self .cpp_kernel_key = (
4458
- f"{ self .cpp_kernel .replace ('::' , '_' )} _{ self .cpp_kernel_overload_name } "
4458
+ f"{ self .cpp_kernel_name .replace ('::' , '_' )} _{ self .cpp_kernel_overload_name } "
4459
4459
)
4460
4460
4461
4461
self .cpp_op_schema = get_cpp_op_schema (kernel )
@@ -4489,14 +4489,14 @@ def sdpa_ver_fn():
4489
4489
self .get_kwargs_value (arg_name ) is None
4490
4490
for arg_name in self .ordered_kwargs_for_cpp_kernel
4491
4491
):
4492
- return f"{ self .cpp_kernel } _v2"
4492
+ return f"{ self .cpp_kernel_name } _v2"
4493
4493
else :
4494
- return self .cpp_kernel
4494
+ return self .cpp_kernel_name
4495
4495
4496
4496
kernel_to_ver = {"at::_scaled_dot_product_flash_attention" : sdpa_ver_fn }
4497
- if (ver_fn := kernel_to_ver .get (self .cpp_kernel , None )) is not None :
4497
+ if (ver_fn := kernel_to_ver .get (self .cpp_kernel_name , None )) is not None :
4498
4498
return ver_fn ()
4499
- return self .cpp_kernel
4499
+ return self .cpp_kernel_name
4500
4500
4501
4501
def codegen_args (self ):
4502
4502
@dataclasses .dataclass
@@ -4667,7 +4667,7 @@ def codegen(self, wrapper):
4667
4667
# repeat_interleave(const at::Tensor & repeats, c10::optional<int64_t> output_size=c10::nullopt)
4668
4668
# repeat_interleave(const at::Tensor & self, int64_t repeats,
4669
4669
# c10::optional<int64_t> dim=c10::nullopt, c10::optional<int64_t> output_size=c10::nullopt)
4670
- self .cpp_kernel = (
4670
+ self .cpp_kernel_name = (
4671
4671
f"at::{ op_base_name } "
4672
4672
if kernel ._overloadname == "default"
4673
4673
else f"at::_ops::{ kernel .__name__ .replace ('.' , '_' )} ::call"
@@ -5121,7 +5121,7 @@ def __init__(
5121
5121
constant_args ,
5122
5122
None ,
5123
5123
kernel = "torch.ops.mkldnn._convolution_pointwise" ,
5124
- cpp_kernel = "mkldnn::_convolution_pointwise" ,
5124
+ cpp_kernel_name = "mkldnn::_convolution_pointwise" ,
5125
5125
)
5126
5126
self .cpp_kernel_key = "convolution_pointwise"
5127
5127
self .cpp_op_schema = """
@@ -5140,7 +5140,7 @@ def __init__(
5140
5140
def codegen (self , wrapper ):
5141
5141
wrapper .generate_extern_kernel_alloc_and_find_schema_if_needed (
5142
5142
self .get_name (),
5143
- self .cpp_kernel if V .graph .cpp_wrapper else self .kernel ,
5143
+ self .cpp_kernel_name if V .graph .cpp_wrapper else self .kernel ,
5144
5144
self .codegen_args (),
5145
5145
self .cpp_op_schema ,
5146
5146
self .cpp_kernel_key ,
@@ -5191,7 +5191,7 @@ def __init__(
5191
5191
constant_args ,
5192
5192
None ,
5193
5193
kernel = "torch.ops.mkldnn._convolution_pointwise.binary" ,
5194
- cpp_kernel = "mkldnn::_convolution_pointwise" ,
5194
+ cpp_kernel_name = "mkldnn::_convolution_pointwise" ,
5195
5195
)
5196
5196
self .cpp_kernel_overload_name = "binary"
5197
5197
self .cpp_kernel_key = "convolution_pointwise_binary"
@@ -5281,7 +5281,7 @@ def __init__(
5281
5281
constant_args ,
5282
5282
None ,
5283
5283
kernel = "torch.ops.mkldnn._convolution_pointwise_.binary" ,
5284
- cpp_kernel = "mkldnn::_convolution_pointwise_" ,
5284
+ cpp_kernel_name = "mkldnn::_convolution_pointwise_" ,
5285
5285
)
5286
5286
self .cpp_kernel_overload_name = "binary"
5287
5287
self .cpp_kernel_key = "convolution_pointwise_binary_"
@@ -5377,7 +5377,7 @@ def __init__(
5377
5377
constant_args ,
5378
5378
None ,
5379
5379
kernel = "torch.ops.mkl._mkl_linear" ,
5380
- cpp_kernel = "mkl::_mkl_linear" ,
5380
+ cpp_kernel_name = "mkl::_mkl_linear" ,
5381
5381
)
5382
5382
self .cpp_kernel_key = "mkl_linear"
5383
5383
self .cpp_op_schema = """
@@ -5430,7 +5430,7 @@ def __init__(
5430
5430
constant_args ,
5431
5431
None ,
5432
5432
kernel = "torch.ops.mkldnn._linear_pointwise" ,
5433
- cpp_kernel = "mkldnn::_linear_pointwise" ,
5433
+ cpp_kernel_name = "mkldnn::_linear_pointwise" ,
5434
5434
)
5435
5435
self .cpp_kernel_key = "linear_pointwise"
5436
5436
self .cpp_op_schema = """
@@ -5495,7 +5495,7 @@ def __init__(
5495
5495
constant_args ,
5496
5496
None ,
5497
5497
kernel = "torch.ops.mkldnn._linear_pointwise.binary" ,
5498
- cpp_kernel = "mkldnn::_linear_pointwise" ,
5498
+ cpp_kernel_name = "mkldnn::_linear_pointwise" ,
5499
5499
)
5500
5500
self .cpp_kernel_overload_name = "binary"
5501
5501
self .cpp_kernel_key = "linear_pointwise_binary"
@@ -5562,7 +5562,7 @@ def __init__(
5562
5562
constant_args ,
5563
5563
None ,
5564
5564
kernel = "torch.ops.mkldnn._convolution_transpose_pointwise" ,
5565
- cpp_kernel = "mkldnn::_convolution_transpose_pointwise" ,
5565
+ cpp_kernel_name = "mkldnn::_convolution_transpose_pointwise" ,
5566
5566
)
5567
5567
self .cpp_kernel_key = "convolution_transpose_pointwise"
5568
5568
self .cpp_op_schema = """
@@ -5646,7 +5646,7 @@ def __init__(
5646
5646
constant_args ,
5647
5647
None ,
5648
5648
kernel = "aten.mkldnn_rnn_layer" ,
5649
- cpp_kernel = "at::mkldnn_rnn_layer" ,
5649
+ cpp_kernel_name = "at::mkldnn_rnn_layer" ,
5650
5650
)
5651
5651
5652
5652
@classmethod
@@ -5766,7 +5766,7 @@ def __init__(
5766
5766
constant_args ,
5767
5767
None ,
5768
5768
kernel = "torch.ops.onednn.qconv2d_pointwise" ,
5769
- cpp_kernel = "onednn::qconv2d_pointwise" ,
5769
+ cpp_kernel_name = "onednn::qconv2d_pointwise" ,
5770
5770
)
5771
5771
self .cpp_kernel_key = "qconv2d_pointwise"
5772
5772
self .cpp_op_schema = """
@@ -5936,7 +5936,7 @@ def __init__(
5936
5936
constant_args ,
5937
5937
None ,
5938
5938
kernel = "torch.ops.onednn.qconv2d_pointwise.binary" ,
5939
- cpp_kernel = "onednn::qconv2d_pointwise" ,
5939
+ cpp_kernel_name = "onednn::qconv2d_pointwise" ,
5940
5940
)
5941
5941
self .cpp_kernel_overload_name = "binary"
5942
5942
self .cpp_kernel_key = "qconv2d_pointwise_binary"
@@ -6136,7 +6136,7 @@ def __init__(
6136
6136
constant_args ,
6137
6137
None ,
6138
6138
kernel = "torch.ops.onednn.qlinear_pointwise" ,
6139
- cpp_kernel = "onednn::qlinear_pointwise" ,
6139
+ cpp_kernel_name = "onednn::qlinear_pointwise" ,
6140
6140
)
6141
6141
self .cpp_kernel_key = "qlinear_pointwise"
6142
6142
self .cpp_op_schema = """
@@ -7223,10 +7223,10 @@ def has_side_effects(self):
7223
7223
def set_cpp_kernel (self , kernel ):
7224
7224
from .codegen .wrapper import get_cpp_op_schema
7225
7225
7226
- self .cpp_kernel = kernel ._schema .name
7226
+ self .cpp_kernel_name = kernel ._schema .name
7227
7227
self .cpp_kernel_overload_name = kernel ._schema .overload_name
7228
7228
self .cpp_kernel_key = (
7229
- f"{ self .cpp_kernel .replace ('::' , '_' )} _{ self .cpp_kernel_overload_name } "
7229
+ f"{ self .cpp_kernel_name .replace ('::' , '_' )} _{ self .cpp_kernel_overload_name } "
7230
7230
)
7231
7231
7232
7232
self .cpp_op_schema = get_cpp_op_schema (kernel )
0 commit comments