@@ -1164,7 +1164,7 @@ simde_mm256_cmpgt_epi8 (simde__m256i a, simde__m256i b) {
11641164 #if SIMDE_NATURAL_VECTOR_SIZE_LE (128 )
11651165 r_ .m128i [0 ] = simde_mm_cmpgt_epi8 (a_ .m128i [0 ], b_ .m128i [0 ]);
11661166 r_ .m128i [1 ] = simde_mm_cmpgt_epi8 (a_ .m128i [1 ], b_ .m128i [1 ]);
1167- #elif defined(SIMDE_VECTOR_SUBSCRIPT_OPS ) && !defined(SIMDE_SKIP_EXTENDED_E2K_VECTOR_OPS )
1167+ #elif defined(SIMDE_VECTOR_SUBSCRIPT_OPS ) && !defined(SIMDE_BUG_LCC_TOO_STRICT_VECTOR_SHIFTS_AND_COMPARES )
11681168 r_ .i8 = HEDLEY_STATIC_CAST (__typeof__ (r_ .i8 ), a_ .i8 > b_ .i8 );
11691169 #else
11701170 SIMDE_VECTORIZE
@@ -4377,7 +4377,7 @@ simde_mm_sllv_epi32 (simde__m128i a, simde__m128i b) {
43774377 #if defined(SIMDE_ARM_NEON_A32V7_NATIVE )
43784378 r_ .neon_u32 = vshlq_u32 (a_ .neon_u32 , vreinterpretq_s32_u32 (b_ .neon_u32 ));
43794379 r_ .neon_u32 = vandq_u32 (r_ .neon_u32 , vcltq_u32 (b_ .neon_u32 , vdupq_n_u32 (32 )));
4380- #elif defined(SIMDE_VECTOR_SUBSCRIPT_SCALAR ) && !defined(SIMDE_SKIP_EXTENDED_E2K_VECTOR_OPS )
4380+ #elif defined(SIMDE_VECTOR_SUBSCRIPT_SCALAR ) && !defined(SIMDE_BUG_LCC_TOO_STRICT_VECTOR_SHIFTS_AND_COMPARES )
43814381 r_ .u32 = HEDLEY_STATIC_CAST (__typeof__ (r_ .u32 ), (b_ .u32 < 32 ) & (a_ .u32 << b_ .u32 ));
43824382 #else
43834383 SIMDE_VECTORIZE
@@ -4407,7 +4407,7 @@ simde_mm256_sllv_epi32 (simde__m256i a, simde__m256i b) {
44074407 #if SIMDE_NATURAL_VECTOR_SIZE_LE (128 )
44084408 r_ .m128i [0 ] = simde_mm_sllv_epi32 (a_ .m128i [0 ], b_ .m128i [0 ]);
44094409 r_ .m128i [1 ] = simde_mm_sllv_epi32 (a_ .m128i [1 ], b_ .m128i [1 ]);
4410- #elif defined(SIMDE_VECTOR_SUBSCRIPT_SCALAR ) && !defined(SIMDE_SKIP_EXTENDED_E2K_VECTOR_OPS )
4410+ #elif defined(SIMDE_VECTOR_SUBSCRIPT_SCALAR ) && !defined(SIMDE_BUG_LCC_TOO_STRICT_VECTOR_SHIFTS_AND_COMPARES )
44114411 r_ .u32 = HEDLEY_STATIC_CAST (__typeof__ (r_ .u32 ), (b_ .u32 < 32 ) & (a_ .u32 << b_ .u32 ));
44124412 #else
44134413 SIMDE_VECTORIZE
@@ -4437,7 +4437,7 @@ simde_mm_sllv_epi64 (simde__m128i a, simde__m128i b) {
44374437 #if defined(SIMDE_ARM_NEON_A64V8_NATIVE )
44384438 r_ .neon_u64 = vshlq_u64 (a_ .neon_u64 , vreinterpretq_s64_u64 (b_ .neon_u64 ));
44394439 r_ .neon_u64 = vandq_u64 (r_ .neon_u64 , vcltq_u64 (b_ .neon_u64 , vdupq_n_u64 (64 )));
4440- #elif defined(SIMDE_VECTOR_SUBSCRIPT_SCALAR ) && !defined(SIMDE_SKIP_EXTENDED_E2K_VECTOR_OPS )
4440+ #elif defined(SIMDE_VECTOR_SUBSCRIPT_SCALAR ) && !defined(SIMDE_BUG_LCC_TOO_STRICT_VECTOR_SHIFTS_AND_COMPARES )
44414441 r_ .u64 = HEDLEY_STATIC_CAST (__typeof__ (r_ .u64 ), (b_ .u64 < 64 ) & (a_ .u64 << b_ .u64 ));
44424442 #else
44434443 SIMDE_VECTORIZE
@@ -4467,7 +4467,7 @@ simde_mm256_sllv_epi64 (simde__m256i a, simde__m256i b) {
44674467 #if SIMDE_NATURAL_VECTOR_SIZE_LE (128 )
44684468 r_ .m128i [0 ] = simde_mm_sllv_epi64 (a_ .m128i [0 ], b_ .m128i [0 ]);
44694469 r_ .m128i [1 ] = simde_mm_sllv_epi64 (a_ .m128i [1 ], b_ .m128i [1 ]);
4470- #elif defined(SIMDE_VECTOR_SUBSCRIPT_SCALAR ) && !defined(SIMDE_SKIP_EXTENDED_E2K_VECTOR_OPS )
4470+ #elif defined(SIMDE_VECTOR_SUBSCRIPT_SCALAR ) && !defined(SIMDE_BUG_LCC_TOO_STRICT_VECTOR_SHIFTS_AND_COMPARES )
44714471 r_ .u64 = HEDLEY_STATIC_CAST (__typeof__ (r_ .u64 ), (b_ .u64 < 64 ) & (a_ .u64 << b_ .u64 ));
44724472 #else
44734473 SIMDE_VECTORIZE
@@ -4964,7 +4964,7 @@ simde_mm_srlv_epi32 (simde__m128i a, simde__m128i b) {
49644964 b_ = simde__m128i_to_private (b ),
49654965 r_ ;
49664966
4967- #if defined(SIMDE_VECTOR_SUBSCRIPT_SCALAR ) && !defined(SIMDE_SKIP_EXTENDED_E2K_VECTOR_OPS )
4967+ #if defined(SIMDE_VECTOR_SUBSCRIPT_SCALAR ) && !defined(SIMDE_BUG_LCC_TOO_STRICT_VECTOR_SHIFTS_AND_COMPARES )
49684968 r_ .u32 = HEDLEY_STATIC_CAST (__typeof__ (r_ .u32 ), (b_ .u32 < 32 ) & (a_ .u32 >> b_ .u32 ));
49694969 #else
49704970 SIMDE_VECTORIZE
@@ -4991,7 +4991,7 @@ simde_mm256_srlv_epi32 (simde__m256i a, simde__m256i b) {
49914991 b_ = simde__m256i_to_private (b ),
49924992 r_ ;
49934993
4994- #if defined(SIMDE_VECTOR_SUBSCRIPT_SCALAR ) && !defined(SIMDE_SKIP_EXTENDED_E2K_VECTOR_OPS )
4994+ #if defined(SIMDE_VECTOR_SUBSCRIPT_SCALAR ) && !defined(SIMDE_BUG_LCC_TOO_STRICT_VECTOR_SHIFTS_AND_COMPARES )
49954995 r_ .u32 = HEDLEY_STATIC_CAST (__typeof__ (r_ .u32 ), (b_ .u32 < 32 ) & (a_ .u32 >> b_ .u32 ));
49964996 #else
49974997 SIMDE_VECTORIZE
@@ -5018,7 +5018,7 @@ simde_mm_srlv_epi64 (simde__m128i a, simde__m128i b) {
50185018 b_ = simde__m128i_to_private (b ),
50195019 r_ ;
50205020
5021- #if defined(SIMDE_VECTOR_SUBSCRIPT_SCALAR ) && !defined(SIMDE_SKIP_EXTENDED_E2K_VECTOR_OPS )
5021+ #if defined(SIMDE_VECTOR_SUBSCRIPT_SCALAR ) && !defined(SIMDE_BUG_LCC_TOO_STRICT_VECTOR_SHIFTS_AND_COMPARES )
50225022 r_ .u64 = HEDLEY_STATIC_CAST (__typeof__ (r_ .u64 ), (b_ .u64 < 64 ) & (a_ .u64 >> b_ .u64 ));
50235023 #else
50245024 SIMDE_VECTORIZE
@@ -5045,7 +5045,7 @@ simde_mm256_srlv_epi64 (simde__m256i a, simde__m256i b) {
50455045 b_ = simde__m256i_to_private (b ),
50465046 r_ ;
50475047
5048- #if defined(SIMDE_VECTOR_SUBSCRIPT_SCALAR ) && !defined(SIMDE_SKIP_EXTENDED_E2K_VECTOR_OPS )
5048+ #if defined(SIMDE_VECTOR_SUBSCRIPT_SCALAR ) && !defined(SIMDE_BUG_LCC_TOO_STRICT_VECTOR_SHIFTS_AND_COMPARES )
50495049 r_ .u64 = HEDLEY_STATIC_CAST (__typeof__ (r_ .u64 ), (b_ .u64 < 64 ) & (a_ .u64 >> b_ .u64 ));
50505050 #else
50515051 SIMDE_VECTORIZE
0 commit comments