Ruby 3.5.0dev (2025-01-10 revision 5fab31b15e32622c4b71d1d347a41937e9f9c212)
parser_bits.h (5fab31b15e32622c4b71d1d347a41937e9f9c212)
1#ifndef INTERNAL_BITS2_H /*-*-C-*-vi:se ft=c:*/
2#define INTERNAL_BITS2_H
28#include "ruby/internal/config.h"
29#include <limits.h> /* for CHAR_BITS */
30#include <stdint.h> /* for uintptr_t */
31#include "internal/compilers.h" /* for MSC_VERSION_SINCE */
32
33#if MSC_VERSION_SINCE(1310)
34# include <stdlib.h> /* for _byteswap_uint64 */
35#endif
36
37#if defined(HAVE_X86INTRIN_H)
38# include <x86intrin.h> /* for _lzcnt_u64 */
39#elif MSC_VERSION_SINCE(1310)
40# include <intrin.h> /* for the following intrinsics */
41#endif
42
43#if defined(_MSC_VER) && defined(__AVX__)
44# pragma intrinsic(__popcnt)
45# pragma intrinsic(__popcnt64)
46#endif
47
48#if defined(_MSC_VER) && defined(__AVX2__)
49# pragma intrinsic(__lzcnt)
50# pragma intrinsic(__lzcnt64)
51#endif
52
53#if MSC_VERSION_SINCE(1310)
54# pragma intrinsic(_rotl)
55# pragma intrinsic(_rotr)
56# ifdef _WIN64
57# pragma intrinsic(_rotl64)
58# pragma intrinsic(_rotr64)
59# endif
60#endif
61
62#if MSC_VERSION_SINCE(1400)
63# pragma intrinsic(_BitScanForward)
64# pragma intrinsic(_BitScanReverse)
65# ifdef _WIN64
66# pragma intrinsic(_BitScanForward64)
67# pragma intrinsic(_BitScanReverse64)
68# endif
69#endif
70
71#include "parser_value.h" /* for VALUE */
72#include "internal/static_assert.h" /* for STATIC_ASSERT */
73
74/* The most significant bit of the lower part of half-long integer.
75 * If sizeof(long) == 4, this is 0x8000.
76 * If sizeof(long) == 8, this is 0x80000000.
77 */
78#define HALF_LONG_MSB ((SIGNED_VALUE)1<<((SIZEOF_LONG*CHAR_BIT-1)/2))
79
80#define SIGNED_INTEGER_TYPE_P(T) (0 > ((T)0)-1)
81
82#define SIGNED_INTEGER_MIN(T) \
83 ((sizeof(T) == sizeof(int8_t)) ? ((T)INT8_MIN) : \
84 ((sizeof(T) == sizeof(int16_t)) ? ((T)INT16_MIN) : \
85 ((sizeof(T) == sizeof(int32_t)) ? ((T)INT32_MIN) : \
86 ((sizeof(T) == sizeof(int64_t)) ? ((T)INT64_MIN) : \
87 0))))
88
89#define SIGNED_INTEGER_MAX(T) ((T)(SIGNED_INTEGER_MIN(T) ^ ((T)~(T)0)))
90
91#define UNSIGNED_INTEGER_MAX(T) ((T)~(T)0)
92
93#if __has_builtin(__builtin_mul_overflow_p)
94# define MUL_OVERFLOW_P(a, b) \
95 __builtin_mul_overflow_p((a), (b), (__typeof__(a * b))0)
96#elif __has_builtin(__builtin_mul_overflow)
97# define MUL_OVERFLOW_P(a, b) \
98 __extension__ ({ __typeof__(a) c; __builtin_mul_overflow((a), (b), &c); })
99#endif
100
101#define MUL_OVERFLOW_SIGNED_INTEGER_P(a, b, min, max) ( \
102 (a) == 0 ? 0 : \
103 (a) == -1 ? (b) < -(max) : \
104 (a) > 0 ? \
105 ((b) > 0 ? (max) / (a) < (b) : (min) / (a) > (b)) : \
106 ((b) > 0 ? (min) / (a) < (b) : (max) / (a) > (b)))
107
108#if __has_builtin(__builtin_mul_overflow_p)
109/* __builtin_mul_overflow_p can take bitfield */
110/* and GCC permits bitfields for integers other than int */
111# define MUL_OVERFLOW_FIXNUM_P(a, b) \
112 __extension__ ({ \
113 struct { long fixnum : sizeof(long) * CHAR_BIT - 1; } c = { 0 }; \
114 __builtin_mul_overflow_p((a), (b), c.fixnum); \
115 })
116#else
117# define MUL_OVERFLOW_FIXNUM_P(a, b) \
118 MUL_OVERFLOW_SIGNED_INTEGER_P(a, b, FIXNUM_MIN, FIXNUM_MAX)
119#endif
120
121#ifdef MUL_OVERFLOW_P
122# define MUL_OVERFLOW_LONG_LONG_P(a, b) MUL_OVERFLOW_P(a, b)
123# define MUL_OVERFLOW_LONG_P(a, b) MUL_OVERFLOW_P(a, b)
124# define MUL_OVERFLOW_INT_P(a, b) MUL_OVERFLOW_P(a, b)
125#else
126# define MUL_OVERFLOW_LONG_LONG_P(a, b) MUL_OVERFLOW_SIGNED_INTEGER_P(a, b, LLONG_MIN, LLONG_MAX)
127# define MUL_OVERFLOW_LONG_P(a, b) MUL_OVERFLOW_SIGNED_INTEGER_P(a, b, LONG_MIN, LONG_MAX)
128# define MUL_OVERFLOW_INT_P(a, b) MUL_OVERFLOW_SIGNED_INTEGER_P(a, b, INT_MIN, INT_MAX)
129#endif
130
131#ifdef HAVE_UINT128_T
132# define bit_length(x) \
133 (unsigned int) \
134 (sizeof(x) <= sizeof(int32_t) ? 32 - nlz_int32((uint32_t)(x)) : \
135 sizeof(x) <= sizeof(int64_t) ? 64 - nlz_int64((uint64_t)(x)) : \
136 128 - nlz_int128((uint128_t)(x)))
137#else
138# define bit_length(x) \
139 (unsigned int) \
140 (sizeof(x) <= sizeof(int32_t) ? 32 - nlz_int32((uint32_t)(x)) : \
141 64 - nlz_int64((uint64_t)(x)))
142#endif
143
144#ifndef swap16
145# define swap16 ruby_swap16
146#endif
147
148#ifndef swap32
149# define swap32 ruby_swap32
150#endif
151
152#ifndef swap64
153# define swap64 ruby_swap64
154#endif
155
156static inline uint16_t ruby_swap16(uint16_t);
157static inline uint32_t ruby_swap32(uint32_t);
158static inline uint64_t ruby_swap64(uint64_t);
159static inline unsigned nlz_int(unsigned x);
160static inline unsigned nlz_long(unsigned long x);
161static inline unsigned nlz_long_long(unsigned long long x);
162static inline unsigned nlz_intptr(uintptr_t x);
163static inline unsigned nlz_int32(uint32_t x);
164static inline unsigned nlz_int64(uint64_t x);
165#ifdef HAVE_UINT128_T
166static inline unsigned nlz_int128(uint128_t x);
167#endif
168static inline unsigned rb_popcount32(uint32_t x);
169static inline unsigned rb_popcount64(uint64_t x);
170static inline unsigned rb_popcount_intptr(uintptr_t x);
171static inline int ntz_int32(uint32_t x);
172static inline int ntz_int64(uint64_t x);
173static inline int ntz_intptr(uintptr_t x);
174static inline VALUE RUBY_BIT_ROTL(VALUE, int);
175static inline VALUE RUBY_BIT_ROTR(VALUE, int);
176
177static inline uint16_t
178ruby_swap16(uint16_t x)
179{
180#if __has_builtin(__builtin_bswap16)
181 return __builtin_bswap16(x);
182
183#elif MSC_VERSION_SINCE(1310)
184 return _byteswap_ushort(x);
185
186#else
187 return (x << 8) | (x >> 8);
188
189#endif
190}
191
192static inline uint32_t
193ruby_swap32(uint32_t x)
194{
195#if __has_builtin(__builtin_bswap32)
196 return __builtin_bswap32(x);
197
198#elif MSC_VERSION_SINCE(1310)
199 return _byteswap_ulong(x);
200
201#else
202 x = ((x & 0x0000FFFF) << 16) | ((x & 0xFFFF0000) >> 16);
203 x = ((x & 0x00FF00FF) << 8) | ((x & 0xFF00FF00) >> 8);
204 return x;
205
206#endif
207}
208
209static inline uint64_t
210ruby_swap64(uint64_t x)
211{
212#if __has_builtin(__builtin_bswap64)
213 return __builtin_bswap64(x);
214
215#elif MSC_VERSION_SINCE(1310)
216 return _byteswap_uint64(x);
217
218#else
219 x = ((x & 0x00000000FFFFFFFFULL) << 32) | ((x & 0xFFFFFFFF00000000ULL) >> 32);
220 x = ((x & 0x0000FFFF0000FFFFULL) << 16) | ((x & 0xFFFF0000FFFF0000ULL) >> 16);
221 x = ((x & 0x00FF00FF00FF00FFULL) << 8) | ((x & 0xFF00FF00FF00FF00ULL) >> 8);
222 return x;
223
224#endif
225}
226
227static inline unsigned int
228nlz_int32(uint32_t x)
229{
230#if defined(_MSC_VER) && defined(__AVX2__)
231 /* Note: It seems there is no such thing like __LZCNT__ predefined in MSVC.
232 * AMD CPUs have had this instruction for decades (since K10) but for
233 * Intel, Haswell is the oldest one. We need to use __AVX2__ for maximum
234 * safety. */
235 return (unsigned int)__lzcnt(x);
236
237#elif defined(__x86_64__) && defined(__LZCNT__)
238 return (unsigned int)_lzcnt_u32(x);
239
240#elif MSC_VERSION_SINCE(1400) /* &&! defined(__AVX2__) */
241 unsigned long r;
242 return _BitScanReverse(&r, x) ? (31 - (int)r) : 32;
243
244#elif __has_builtin(__builtin_clz)
245 STATIC_ASSERT(sizeof_int, sizeof(int) * CHAR_BIT == 32);
246 return x ? (unsigned int)__builtin_clz(x) : 32;
247
248#else
249 uint32_t y;
250 unsigned n = 32;
251 y = x >> 16; if (y) {n -= 16; x = y;}
252 y = x >> 8; if (y) {n -= 8; x = y;}
253 y = x >> 4; if (y) {n -= 4; x = y;}
254 y = x >> 2; if (y) {n -= 2; x = y;}
255 y = x >> 1; if (y) {return n - 2;}
256 return (unsigned int)(n - x);
257#endif
258}
259
260static inline unsigned int
261nlz_int64(uint64_t x)
262{
263#if defined(_MSC_VER) && defined(__AVX2__)
264 return (unsigned int)__lzcnt64(x);
265
266#elif defined(__x86_64__) && defined(__LZCNT__)
267 return (unsigned int)_lzcnt_u64(x);
268
269#elif defined(_WIN64) && MSC_VERSION_SINCE(1400) /* &&! defined(__AVX2__) */
270 unsigned long r;
271 return _BitScanReverse64(&r, x) ? (63u - (unsigned int)r) : 64;
272
273#elif __has_builtin(__builtin_clzl)
274 if (x == 0) {
275 return 64;
276 }
277 else if (sizeof(long) * CHAR_BIT == 64) {
278 return (unsigned int)__builtin_clzl((unsigned long)x);
279 }
280 else if (sizeof(long long) * CHAR_BIT == 64) {
281 return (unsigned int)__builtin_clzll((unsigned long long)x);
282 }
283 else {
284 /* :FIXME: Is there a way to make this branch a compile-time error? */
286 }
287
288#else
289 uint64_t y;
290 unsigned int n = 64;
291 y = x >> 32; if (y) {n -= 32; x = y;}
292 y = x >> 16; if (y) {n -= 16; x = y;}
293 y = x >> 8; if (y) {n -= 8; x = y;}
294 y = x >> 4; if (y) {n -= 4; x = y;}
295 y = x >> 2; if (y) {n -= 2; x = y;}
296 y = x >> 1; if (y) {return n - 2;}
297 return (unsigned int)(n - x);
298
299#endif
300}
301
302#ifdef HAVE_UINT128_T
303static inline unsigned int
304nlz_int128(uint128_t x)
305{
306 uint64_t y = (uint64_t)(x >> 64);
307
308 if (x == 0) {
309 return 128;
310 }
311 else if (y == 0) {
312 return (unsigned int)nlz_int64(x) + 64;
313 }
314 else {
315 return (unsigned int)nlz_int64(y);
316 }
317}
318#endif
319
320static inline unsigned int
321nlz_int(unsigned int x)
322{
323 if (sizeof(unsigned int) * CHAR_BIT == 32) {
324 return nlz_int32((uint32_t)x);
325 }
326 else if (sizeof(unsigned int) * CHAR_BIT == 64) {
327 return nlz_int64((uint64_t)x);
328 }
329 else {
331 }
332}
333
334static inline unsigned int
335nlz_long(unsigned long x)
336{
337 if (sizeof(unsigned long) * CHAR_BIT == 32) {
338 return nlz_int32((uint32_t)x);
339 }
340 else if (sizeof(unsigned long) * CHAR_BIT == 64) {
341 return nlz_int64((uint64_t)x);
342 }
343 else {
345 }
346}
347
348static inline unsigned int
349nlz_long_long(unsigned long long x)
350{
351 if (sizeof(unsigned long long) * CHAR_BIT == 64) {
352 return nlz_int64((uint64_t)x);
353 }
354#ifdef HAVE_UINT128_T
355 else if (sizeof(unsigned long long) * CHAR_BIT == 128) {
356 return nlz_int128((uint128_t)x);
357 }
358#endif
359 else {
361 }
362}
363
364static inline unsigned int
365nlz_intptr(uintptr_t x)
366{
367 if (sizeof(uintptr_t) == sizeof(unsigned int)) {
368 return nlz_int((unsigned int)x);
369 }
370 if (sizeof(uintptr_t) == sizeof(unsigned long)) {
371 return nlz_long((unsigned long)x);
372 }
373 if (sizeof(uintptr_t) == sizeof(unsigned long long)) {
374 return nlz_long_long((unsigned long long)x);
375 }
376 else {
378 }
379}
380
381static inline unsigned int
382rb_popcount32(uint32_t x)
383{
384#if defined(_MSC_VER) && defined(__AVX__)
385 /* Note: CPUs since Nehalem and Barcelona have had this instruction so SSE
386 * 4.2 should suffice, but it seems there is no such thing like __SSE_4_2__
387 * predefined macro in MSVC. They do have __AVX__ so use it instead. */
388 return (unsigned int)__popcnt(x);
389
390#elif __has_builtin(__builtin_popcount)
391 STATIC_ASSERT(sizeof_int, sizeof(int) * CHAR_BIT >= 32);
392 return (unsigned int)__builtin_popcount(x);
393
394#else
395 x = (x & 0x55555555) + (x >> 1 & 0x55555555);
396 x = (x & 0x33333333) + (x >> 2 & 0x33333333);
397 x = (x & 0x0f0f0f0f) + (x >> 4 & 0x0f0f0f0f);
398 x = (x & 0x001f001f) + (x >> 8 & 0x001f001f);
399 x = (x & 0x0000003f) + (x >>16 & 0x0000003f);
400 return (unsigned int)x;
401
402#endif
403}
404
405static inline unsigned int
406rb_popcount64(uint64_t x)
407{
408#if defined(_MSC_VER) && defined(__AVX__)
409 return (unsigned int)__popcnt64(x);
410
411#elif __has_builtin(__builtin_popcount)
412 if (sizeof(long) * CHAR_BIT == 64) {
413 return (unsigned int)__builtin_popcountl((unsigned long)x);
414 }
415 else if (sizeof(long long) * CHAR_BIT == 64) {
416 return (unsigned int)__builtin_popcountll((unsigned long long)x);
417 }
418 else {
419 /* :FIXME: Is there a way to make this branch a compile-time error? */
421 }
422
423#else
424 x = (x & 0x5555555555555555) + (x >> 1 & 0x5555555555555555);
425 x = (x & 0x3333333333333333) + (x >> 2 & 0x3333333333333333);
426 x = (x & 0x0707070707070707) + (x >> 4 & 0x0707070707070707);
427 x = (x & 0x001f001f001f001f) + (x >> 8 & 0x001f001f001f001f);
428 x = (x & 0x0000003f0000003f) + (x >>16 & 0x0000003f0000003f);
429 x = (x & 0x000000000000007f) + (x >>32 & 0x000000000000007f);
430 return (unsigned int)x;
431
432#endif
433}
434
435static inline unsigned int
436rb_popcount_intptr(uintptr_t x)
437{
438 if (sizeof(uintptr_t) * CHAR_BIT == 64) {
439 return rb_popcount64((uint64_t)x);
440 }
441 else if (sizeof(uintptr_t) * CHAR_BIT == 32) {
442 return rb_popcount32((uint32_t)x);
443 }
444 else {
446 }
447}
448
449static inline int
450ntz_int32(uint32_t x)
451{
452#if defined(__x86_64__) && defined(__BMI__)
453 return (unsigned)_tzcnt_u32(x);
454
455#elif MSC_VERSION_SINCE(1400)
456 /* :FIXME: Is there any way to issue TZCNT instead of BSF, apart from using
457 * assembly? Because issuing LZCNT seems possible (see nlz.h). */
458 unsigned long r;
459 return _BitScanForward(&r, x) ? (int)r : 32;
460
461#elif __has_builtin(__builtin_ctz)
462 STATIC_ASSERT(sizeof_int, sizeof(int) * CHAR_BIT == 32);
463 return x ? (unsigned)__builtin_ctz(x) : 32;
464
465#else
466 return rb_popcount32((~x) & (x-1));
467
468#endif
469}
470
471static inline int
472ntz_int64(uint64_t x)
473{
474#if defined(__x86_64__) && defined(__BMI__)
475 return (unsigned)_tzcnt_u64(x);
476
477#elif defined(_WIN64) && MSC_VERSION_SINCE(1400)
478 unsigned long r;
479 return _BitScanForward64(&r, x) ? (int)r : 64;
480
481#elif __has_builtin(__builtin_ctzl)
482 if (x == 0) {
483 return 64;
484 }
485 else if (sizeof(long) * CHAR_BIT == 64) {
486 return (unsigned)__builtin_ctzl((unsigned long)x);
487 }
488 else if (sizeof(long long) * CHAR_BIT == 64) {
489 return (unsigned)__builtin_ctzll((unsigned long long)x);
490 }
491 else {
492 /* :FIXME: Is there a way to make this branch a compile-time error? */
494 }
495
496#else
497 return rb_popcount64((~x) & (x-1));
498
499#endif
500}
501
502static inline int
503ntz_intptr(uintptr_t x)
504{
505 if (sizeof(uintptr_t) * CHAR_BIT == 64) {
506 return ntz_int64((uint64_t)x);
507 }
508 else if (sizeof(uintptr_t) * CHAR_BIT == 32) {
509 return ntz_int32((uint32_t)x);
510 }
511 else {
513 }
514}
515
516static inline VALUE
517RUBY_BIT_ROTL(VALUE v, int n)
518{
519#if __has_builtin(__builtin_rotateleft32) && (SIZEOF_VALUE * CHAR_BIT == 32)
520 return __builtin_rotateleft32(v, n);
521
522#elif __has_builtin(__builtin_rotateleft64) && (SIZEOF_VALUE * CHAR_BIT == 64)
523 return __builtin_rotateleft64(v, n);
524
525#elif MSC_VERSION_SINCE(1310) && (SIZEOF_VALUE * CHAR_BIT == 32)
526 return _rotl(v, n);
527
528#elif MSC_VERSION_SINCE(1310) && (SIZEOF_VALUE * CHAR_BIT == 64)
529 return _rotl64(v, n);
530
531#elif defined(_lrotl) && (SIZEOF_VALUE == SIZEOF_LONG)
532 return _lrotl(v, n);
533
534#else
535 const int m = (sizeof(VALUE) * CHAR_BIT) - 1;
536 return (v << (n & m)) | (v >> (-n & m));
537#endif
538}
539
540static inline VALUE
541RUBY_BIT_ROTR(VALUE v, int n)
542{
543#if __has_builtin(__builtin_rotateright32) && (SIZEOF_VALUE * CHAR_BIT == 32)
544 return __builtin_rotateright32(v, n);
545
546#elif __has_builtin(__builtin_rotateright64) && (SIZEOF_VALUE * CHAR_BIT == 64)
547 return __builtin_rotateright64(v, n);
548
549#elif MSC_VERSION_SINCE(1310) && (SIZEOF_VALUE * CHAR_BIT == 32)
550 return _rotr(v, n);
551
552#elif MSC_VERSION_SINCE(1310) && (SIZEOF_VALUE * CHAR_BIT == 64)
553 return _rotr64(v, n);
554
555#elif defined(_lrotr) && (SIZEOF_VALUE == SIZEOF_LONG)
556 return _lrotr(v, n);
557
558#else
559 const int m = (sizeof(VALUE) * CHAR_BIT) - 1;
560 return (v << (-n & m)) | (v >> (n & m));
561#endif
562}
563
564#endif /* INTERNAL_BITS2_H */
#define UNREACHABLE_RETURN
Old name of RBIMPL_UNREACHABLE_RETURN.
Definition assume.h:29
uintptr_t VALUE
Type that represents a Ruby object.
Definition value.h:40