@@ -39,7 +39,7 @@ __AMDGCN_CLC_SUBGROUP_SUB_I32(unsigned short);
39
39
// _Z28__spirv_SubgroupShuffleINTELIhET_S0_j - unsigned char
40
40
// _Z28__spirv_SubgroupShuffleINTELIsET_S0_j - long
41
41
// _Z28__spirv_SubgroupShuffleINTELItET_S0_j - unsigned long
42
- // _Z28__spirv_SubgroupShuffleINTELIDF16_ET_S0_j - half
42
+ // _Z28__spirv_SubgroupShuffleINTELIDhET_S0_j - half
43
43
#define __AMDGCN_CLC_SUBGROUP_SUB_I32 (TYPE , MANGLED_TYPE_NAME ) \
44
44
_CLC_DEF TYPE _Z28__spirv_SubgroupShuffleINTELI##MANGLED_TYPE_NAME##ET_S0_j( \
45
45
TYPE Data, unsigned int InvocationId) { \
@@ -58,7 +58,7 @@ __spirv_SubgroupShuffleINTEL(half Data, unsigned int InvocationId) {
58
58
tmp = __spirv_SubgroupShuffleINTEL (tmp , InvocationId );
59
59
return __clc_as_half (tmp );
60
60
}
61
- _CLC_DEF half _Z28__spirv_SubgroupShuffleINTELIDF16_ET_S0_j (
61
+ _CLC_DEF half _Z28__spirv_SubgroupShuffleINTELIDhET_S0_j (
62
62
half Data , unsigned int InvocationId ) {
63
63
return __spirv_SubgroupShuffleINTEL (Data , InvocationId );
64
64
}
@@ -227,10 +227,10 @@ __AMDGCN_CLC_SUBGROUP_TO_VEC(ulong8, m, 8)
227
227
__AMDGCN_CLC_SUBGROUP_TO_VEC (ulong16 , m , 16 )
228
228
// half
229
229
#ifdef cl_khr_fp16
230
- __AMDGCN_CLC_SUBGROUP_TO_VEC (half2 , DF16_ , 2 )
231
- __AMDGCN_CLC_SUBGROUP_TO_VEC (half4 , DF16_ , 4 )
232
- __AMDGCN_CLC_SUBGROUP_TO_VEC (half8 , DF16_ , 8 )
233
- __AMDGCN_CLC_SUBGROUP_TO_VEC (half16 , DF16_ , 16 )
230
+ __AMDGCN_CLC_SUBGROUP_TO_VEC (half2 , Dh , 2 )
231
+ __AMDGCN_CLC_SUBGROUP_TO_VEC (half4 , Dh , 4 )
232
+ __AMDGCN_CLC_SUBGROUP_TO_VEC (half8 , Dh , 8 )
233
+ __AMDGCN_CLC_SUBGROUP_TO_VEC (half16 , Dh , 16 )
234
234
#endif // cl_khr_fp16
235
235
// float
236
236
__AMDGCN_CLC_SUBGROUP_TO_VEC (float2 , f , 2 )
@@ -271,8 +271,8 @@ __AMDGCN_CLC_SUBGROUP_XOR_SUB_I32(unsigned char);
271
271
__AMDGCN_CLC_SUBGROUP_XOR_SUB_I32 (short );
272
272
__AMDGCN_CLC_SUBGROUP_XOR_SUB_I32 (unsigned short );
273
273
#ifdef cl_khr_fp16
274
- _CLC_OVERLOAD _CLC_DEF half
275
- __spirv_SubgroupShuffleXorINTEL ( half Data , unsigned int InvocationId ) {
274
+ _CLC_OVERLOAD _CLC_DEF half __spirv_SubgroupShuffleXorINTEL (
275
+ half Data , unsigned int InvocationId ) {
276
276
unsigned short tmp = __clc_as_ushort (Data );
277
277
tmp = (unsigned short )__spirv_SubgroupShuffleXorINTEL (tmp , InvocationId );
278
278
return __clc_as_half (tmp );
@@ -284,7 +284,7 @@ __spirv_SubgroupShuffleXorINTEL(half Data, unsigned int InvocationId) {
284
284
// _Z31__spirv_SubgroupShuffleXorINTELIhET_S0_j - unsigned char
285
285
// _Z31__spirv_SubgroupShuffleXorINTELIsET_S0_j - short
286
286
// _Z31__spirv_SubgroupShuffleXorINTELItET_S0_j - unsigned short
287
- // _Z31__spirv_SubgroupShuffleXorINTELIDF16_ET_S0_j - half
287
+ // _Z31__spirv_SubgroupShuffleXorINTELIDhET_S0_j - half
288
288
#define __AMDGCN_CLC_SUBGROUP_XOR_SUB_I32 (TYPE , MANGLED_TYPE_NAME ) \
289
289
_CLC_DEF TYPE \
290
290
_Z31__spirv_SubgroupShuffleXorINTELI##MANGLED_TYPE_NAME##ET_S0_j( \
@@ -296,7 +296,7 @@ __AMDGCN_CLC_SUBGROUP_XOR_SUB_I32(unsigned char, h);
296
296
__AMDGCN_CLC_SUBGROUP_XOR_SUB_I32 (short , s );
297
297
__AMDGCN_CLC_SUBGROUP_XOR_SUB_I32 (unsigned short, t );
298
298
#ifdef cl_khr_fp16
299
- _CLC_DEF half _Z31__spirv_SubgroupShuffleXorINTELIDF16_ET_S0_j (
299
+ _CLC_DEF half _Z31__spirv_SubgroupShuffleXorINTELIDhET_S0_j (
300
300
half Data , unsigned int InvocationId ) {
301
301
return __spirv_SubgroupShuffleXorINTEL (Data , InvocationId );
302
302
}
@@ -470,10 +470,10 @@ __AMDGCN_CLC_SUBGROUP_XOR_TO_VEC(float8, f, 8)
470
470
__AMDGCN_CLC_SUBGROUP_XOR_TO_VEC (float16 , f , 16 )
471
471
// half
472
472
#ifdef cl_khr_fp16
473
- __AMDGCN_CLC_SUBGROUP_XOR_TO_VEC (half2 , DF16_ , 2 )
474
- __AMDGCN_CLC_SUBGROUP_XOR_TO_VEC (half4 , DF16_ , 4 )
475
- __AMDGCN_CLC_SUBGROUP_XOR_TO_VEC (half8 , DF16_ , 8 )
476
- __AMDGCN_CLC_SUBGROUP_XOR_TO_VEC (half16 , DF16_ , 16 )
473
+ __AMDGCN_CLC_SUBGROUP_XOR_TO_VEC (half2 , Dh , 2 )
474
+ __AMDGCN_CLC_SUBGROUP_XOR_TO_VEC (half4 , Dh , 4 )
475
+ __AMDGCN_CLC_SUBGROUP_XOR_TO_VEC (half8 , Dh , 8 )
476
+ __AMDGCN_CLC_SUBGROUP_XOR_TO_VEC (half16 , Dh , 16 )
477
477
#endif // cl_khr_fp16
478
478
// double
479
479
__AMDGCN_CLC_SUBGROUP_XOR_TO_VEC (double2 , d , 2 )
@@ -521,11 +521,11 @@ __AMDGCN_CLC_SUBGROUP_UP_SUB_I32(char);
521
521
__AMDGCN_CLC_SUBGROUP_UP_SUB_I32 (unsigned char );
522
522
__AMDGCN_CLC_SUBGROUP_UP_SUB_I32 (short );
523
523
__AMDGCN_CLC_SUBGROUP_UP_SUB_I32 (unsigned short );
524
+
524
525
// half
525
526
#ifdef cl_khr_fp16
526
- _CLC_OVERLOAD _CLC_DEF half __spirv_SubgroupShuffleUpINTEL (half previous ,
527
- half current ,
528
- unsigned int delta ) {
527
+ _CLC_OVERLOAD _CLC_DEF half __spirv_SubgroupShuffleUpINTEL (
528
+ half previous , half current , unsigned int delta ) {
529
529
unsigned short tmpP = __clc_as_ushort (previous );
530
530
unsigned short tmpC = __clc_as_ushort (current );
531
531
tmpC = __spirv_SubgroupShuffleUpINTEL (tmpP , tmpC , delta );
@@ -538,7 +538,7 @@ _CLC_OVERLOAD _CLC_DEF half __spirv_SubgroupShuffleUpINTEL(half previous,
538
538
// _Z30__spirv_SubgroupShuffleUpINTELIhET_S0_S0_j - unsigned char
539
539
// _Z30__spirv_SubgroupShuffleUpINTELIsET_S0_S0_j - short
540
540
// _Z30__spirv_SubgroupShuffleUpINTELItET_S0_S0_j - unsigned short
541
- // _Z30__spirv_SubgroupShuffleUpINTELIDF16_ET_S0_S0_j - half
541
+ // _Z30__spirv_SubgroupShuffleUpINTELIDhET_S0_S0_j - half
542
542
#define __AMDGCN_CLC_SUBGROUP_UP_SUB_I32 (TYPE , MANGLED_TYPE_NAME ) \
543
543
_CLC_DEF TYPE \
544
544
_Z30__spirv_SubgroupShuffleUpINTELI##MANGLED_TYPE_NAME##ET_S0_S0_j( \
@@ -551,7 +551,7 @@ __AMDGCN_CLC_SUBGROUP_UP_SUB_I32(short, s);
551
551
__AMDGCN_CLC_SUBGROUP_UP_SUB_I32 (unsigned short, t );
552
552
// half
553
553
#ifdef cl_khr_fp16
554
- _CLC_DEF half _Z30__spirv_SubgroupShuffleUpINTELIDF16_ET_S0_S0_j (
554
+ _CLC_DEF half _Z30__spirv_SubgroupShuffleUpINTELIDhET_S0_S0_j (
555
555
half previous , half current , unsigned int delta ) {
556
556
return __spirv_SubgroupShuffleUpINTEL (previous , current , delta );
557
557
}
@@ -724,10 +724,10 @@ __AMDGCN_CLC_SUBGROUP_UP_TO_VEC(ulong8, m, 8)
724
724
__AMDGCN_CLC_SUBGROUP_UP_TO_VEC (ulong16 , m , 16 )
725
725
// half
726
726
#ifdef cl_khr_fp16
727
- __AMDGCN_CLC_SUBGROUP_UP_TO_VEC (half2 , DF16_ , 2 )
728
- __AMDGCN_CLC_SUBGROUP_UP_TO_VEC (half4 , DF16_ , 4 )
729
- __AMDGCN_CLC_SUBGROUP_UP_TO_VEC (half8 , DF16_ , 8 )
730
- __AMDGCN_CLC_SUBGROUP_UP_TO_VEC (half16 , DF16_ , 16 )
727
+ __AMDGCN_CLC_SUBGROUP_UP_TO_VEC (half2 , Dh , 2 )
728
+ __AMDGCN_CLC_SUBGROUP_UP_TO_VEC (half4 , Dh , 4 )
729
+ __AMDGCN_CLC_SUBGROUP_UP_TO_VEC (half8 , Dh , 8 )
730
+ __AMDGCN_CLC_SUBGROUP_UP_TO_VEC (half16 , Dh , 16 )
731
731
#endif // cl_khr_fp16
732
732
// float
733
733
__AMDGCN_CLC_SUBGROUP_UP_TO_VEC (float2 , f , 2 )
@@ -782,8 +782,8 @@ __AMDGCN_CLC_SUBGROUP_DOWN_TO_I32(short);
782
782
__AMDGCN_CLC_SUBGROUP_DOWN_TO_I32 (unsigned short );
783
783
// half
784
784
#ifdef cl_khr_fp16
785
- _CLC_OVERLOAD _CLC_DEF half
786
- __spirv_SubgroupShuffleDownINTEL ( half current , half next , unsigned int delta ) {
785
+ _CLC_OVERLOAD _CLC_DEF half __spirv_SubgroupShuffleDownINTEL (
786
+ half current , half next , unsigned int delta ) {
787
787
unsigned short tmpC = __clc_as_ushort (current );
788
788
unsigned short tmpN = __clc_as_ushort (next );
789
789
tmpC = __spirv_SubgroupShuffleDownINTEL (tmpC , tmpN , delta );
@@ -796,7 +796,7 @@ __spirv_SubgroupShuffleDownINTEL(half current, half next, unsigned int delta) {
796
796
// _Z32__spirv_SubgroupShuffleDownINTELIhET_S0_S0_j - unsigned char
797
797
// _Z32__spirv_SubgroupShuffleDownINTELIsET_S0_S0_j - short
798
798
// _Z32__spirv_SubgroupShuffleDownINTELItET_S0_S0_j - unsigned short
799
- // _Z32__spirv_SubgroupShuffleDownINTELIDF16_ET_S0_S0_j - half
799
+ // _Z32__spirv_SubgroupShuffleDownINTELIDhET_S0_S0_j - half
800
800
#define __AMDGCN_CLC_SUBGROUP_DOWN_TO_I32 (TYPE , MANGLED_TYPE_NAME ) \
801
801
_CLC_DEF TYPE \
802
802
_Z32__spirv_SubgroupShuffleDownINTELI##MANGLED_TYPE_NAME##ET_S0_S0_j( \
@@ -809,7 +809,7 @@ __AMDGCN_CLC_SUBGROUP_DOWN_TO_I32(short, s);
809
809
__AMDGCN_CLC_SUBGROUP_DOWN_TO_I32 (unsigned short, t );
810
810
// half
811
811
#ifdef cl_khr_fp16
812
- _CLC_DEF half _Z32__spirv_SubgroupShuffleDownINTELIDF16_ET_S0_S0_j (
812
+ _CLC_DEF half _Z32__spirv_SubgroupShuffleDownINTELIDhET_S0_S0_j (
813
813
half current , half next , unsigned int delta ) {
814
814
return __spirv_SubgroupShuffleDownINTEL (current , next , delta );
815
815
}
@@ -980,10 +980,10 @@ __AMDGCN_CLC_SUBGROUP_DOWN_TO_VEC(ulong8, m, 8)
980
980
__AMDGCN_CLC_SUBGROUP_DOWN_TO_VEC (ulong16 , m , 16 )
981
981
// half
982
982
#ifdef cl_khr_fp16
983
- __AMDGCN_CLC_SUBGROUP_DOWN_TO_VEC (half2 , DF16_ , 2 )
984
- __AMDGCN_CLC_SUBGROUP_DOWN_TO_VEC (half4 , DF16_ , 4 )
985
- __AMDGCN_CLC_SUBGROUP_DOWN_TO_VEC (half8 , DF16_ , 8 )
986
- __AMDGCN_CLC_SUBGROUP_DOWN_TO_VEC (half16 , DF16_ , 16 )
983
+ __AMDGCN_CLC_SUBGROUP_DOWN_TO_VEC (half2 , Dh , 2 )
984
+ __AMDGCN_CLC_SUBGROUP_DOWN_TO_VEC (half4 , Dh , 4 )
985
+ __AMDGCN_CLC_SUBGROUP_DOWN_TO_VEC (half8 , Dh , 8 )
986
+ __AMDGCN_CLC_SUBGROUP_DOWN_TO_VEC (half16 , Dh , 16 )
987
987
#endif // cl_khr_fp16
988
988
// float
989
989
__AMDGCN_CLC_SUBGROUP_DOWN_TO_VEC (float2 , f , 2 )
0 commit comments