@@ -547,7 +547,7 @@ at::Tensor conv_forward_gather_scatter_cuda_latest(
547
547
548
548
// all gather
549
549
AT_DISPATCH_FLOATING_TYPES_AND_HALF (
550
- in_feat.type (), " conv_forward_gather_scatter_cuda" , ([&] {
550
+ in_feat.scalar_type (), " conv_forward_gather_scatter_cuda" , ([&] {
551
551
gather_all_kernel_pad_sep_with_mask<scalar_t >
552
552
<<<ceil((double )(n_in_feats * n_in_channels) /
553
553
(256 << (sizeof (scalar_t ) == 2 ) + 2 )),
@@ -779,7 +779,7 @@ at::Tensor conv_forward_gather_scatter_cuda_fallback(
779
779
// gather n_active_feats dense features from N sparse input features with c
780
780
// feature dimensions
781
781
AT_DISPATCH_FLOATING_TYPES_AND_HALF (
782
- in_feat.type (), " conv_forward_gather_scatter_cuda" , ([&] {
782
+ in_feat.scalar_type (), " conv_forward_gather_scatter_cuda" , ([&] {
783
783
gather_kernel<scalar_t >
784
784
<<<ceil((double )(n_active_feats * n_in_channels) / 256 ), 256 >>> (
785
785
n_active_feats, n_in_feats, n_in_channels,
@@ -796,7 +796,7 @@ at::Tensor conv_forward_gather_scatter_cuda_fallback(
796
796
// scatter n_active_feats dense features into n_out_feats output features of
797
797
// dimension n_out_channels
798
798
AT_DISPATCH_FLOATING_TYPES_AND_HALF (
799
- in_feat.type (), " conv_forward_gather_scatter_cuda" , ([&] {
799
+ in_feat.scalar_type (), " conv_forward_gather_scatter_cuda" , ([&] {
800
800
scatter_kernel<scalar_t >
801
801
<<<ceil((double )(n_active_feats * n_out_channels) / 256 ), 256 >>> (
802
802
n_active_feats, n_out_feats, n_out_channels,
@@ -877,7 +877,7 @@ void conv_backward_gather_scatter_cuda(at::Tensor in_feat, at::Tensor grad_in_fe
877
877
}
878
878
// gather
879
879
AT_DISPATCH_FLOATING_TYPES_AND_HALF (
880
- in_feat.type (), " conv_forward_gather_scatter_cuda" , ([&] {
880
+ in_feat.scalar_type (), " conv_forward_gather_scatter_cuda" , ([&] {
881
881
gather_kernel<scalar_t >
882
882
<<<ceil((double )(n_active_feats * n_out_channels) / 256 ), 256 >>> (
883
883
n_active_feats, n_out_feats, n_out_channels,
@@ -886,7 +886,7 @@ void conv_backward_gather_scatter_cuda(at::Tensor in_feat, at::Tensor grad_in_fe
886
886
neighbor_map.data_ptr <int >() + cur_offset, !transpose);
887
887
}));
888
888
AT_DISPATCH_FLOATING_TYPES_AND_HALF (
889
- in_feat.type (), " conv_forward_gather_scatter_cuda" , ([&] {
889
+ in_feat.scalar_type (), " conv_forward_gather_scatter_cuda" , ([&] {
890
890
gather_kernel<scalar_t >
891
891
<<<ceil((double )(n_active_feats * n_in_channels) / 256 ), 256 >>> (
892
892
n_active_feats, n_in_feats, n_in_channels,
@@ -902,7 +902,7 @@ void conv_backward_gather_scatter_cuda(at::Tensor in_feat, at::Tensor grad_in_fe
902
902
out_grad_buffer_activated);
903
903
// scatter
904
904
AT_DISPATCH_FLOATING_TYPES_AND_HALF (
905
- in_feat.type (), " conv_forward_gather_scatter_cuda" , ([&] {
905
+ in_feat.scalar_type (), " conv_forward_gather_scatter_cuda" , ([&] {
906
906
scatter_kernel<scalar_t >
907
907
<<<ceil((double )(n_active_feats * n_in_channels) / 256 ), 256 >>> (
908
908
n_active_feats, n_in_feats, n_in_channels,
0 commit comments