Skip to content

Commit f0f2d87

Browse files
committed
issue/1081 - fix hygon nv_bf16
1 parent 05a4e27 commit f0f2d87

File tree

6 files changed

+5
-6
lines changed

6 files changed

+5
-6
lines changed

src/infiniop/ops/addcmul/cuda/kernel.cuh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -22,7 +22,7 @@ public:
2222
float f_t2 = __half2float(t2);
2323
return __float2half(f_input + v * f_t1 * f_t2);
2424

25-
} else if constexpr (std::is_same_v<T, nv_bfloat16>) {
25+
} else if constexpr (std::is_same_v<T, cuda_bfloat16>) {
2626
float f_input = __bfloat162float(input);
2727
float f_t1 = __bfloat162float(t1);
2828
float f_t2 = __bfloat162float(t2);

src/infiniop/ops/addcmul/nvidia/addcmul_nvidia.cu

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -166,7 +166,7 @@ infiniStatus_t Descriptor::calculate(
166166
case INFINI_DTYPE_F16:
167167
return launch_addcmul_kernel<half>(this, output, inputs, stream);
168168
case INFINI_DTYPE_BF16:
169-
return launch_addcmul_kernel<nv_bfloat16>(this, output, inputs, stream);
169+
return launch_addcmul_kernel<cuda_bfloat16>(this, output, inputs, stream);
170170
case INFINI_DTYPE_F32:
171171
return launch_addcmul_kernel<float>(this, output, inputs, stream);
172172
case INFINI_DTYPE_F64:

src/infiniop/ops/atanh/cuda/kernel.cuh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -21,7 +21,7 @@ public:
2121
} else if constexpr (std::is_same_v<T, half>) {
2222
// half 类型先转为 float 计算再转回
2323
return __float2half(atanhf(__half2float(a)));
24-
} else if constexpr (std::is_same_v<T, nv_bfloat16>) {
24+
} else if constexpr (std::is_same_v<T, cuda_bfloat16>) {
2525
// bfloat16 类型处理同上
2626
return __float2bfloat16(atanhf(__bfloat162float(a)));
2727
} else if constexpr (std::is_same_v<T, float>) {

src/infiniop/ops/atanh/nvidia/atanh_nvidia.cu

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -44,7 +44,7 @@ infiniStatus_t Descriptor::calculate(
4444
case INFINI_DTYPE_F16:
4545
return _device_info->calculate<256, cuda::AtanhOp, half>(_info, workspace, output, inputs, stream);
4646
case INFINI_DTYPE_BF16:
47-
return _device_info->calculate<256, cuda::AtanhOp, nv_bfloat16>(_info, workspace, output, inputs, stream);
47+
return _device_info->calculate<256, cuda::AtanhOp, cuda_bfloat16>(_info, workspace, output, inputs, stream);
4848
case INFINI_DTYPE_F32:
4949
return _device_info->calculate<256, cuda::AtanhOp, float>(_info, workspace, output, inputs, stream);
5050
case INFINI_DTYPE_F64:

src/infiniop/ops/avg_pool1d/cuda/kernel.cuh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -46,7 +46,7 @@ __device__ void avgPool1dKernel(
4646
}
4747
}
4848

49-
#if defined(ENABLE_ILUVATAR_API)
49+
#if defined(ENABLE_ILUVATAR_API) || defined(ENABLE_HYGON_API)
5050
// Iluvatar __half doesn't accept size_t directly.
5151
y[y_offset] = sum / static_cast<T>(static_cast<double>(kernel_size));
5252
#else

src/infiniop/ops/var_mean/nvidia/var_mean_nvidia.cu

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -44,7 +44,6 @@ infiniStatus_t launchKernel(
4444
bool unbiased, bool keepdim,
4545
cudaStream_t stream, void *workspace, size_t workspace_size) {
4646
size_t input_ndim = info.permuted_input_shape.size();
47-
size_t output_ndim = info.output_shape.size();
4847
size_t input_size = info.input_size;
4948
size_t output_size = info.output_size;
5049
size_t reduce_num = info.reduce_num;

0 commit comments

Comments
 (0)