• Home
  • Raw
  • Download

Lines Matching refs:_Mask

164 unsigned char _BitScanForward(unsigned long *_Index, unsigned long _Mask);
166 unsigned char _BitScanReverse(unsigned long *_Index, unsigned long _Mask);
182 long _InterlockedAnd(long volatile *_Value, long _Mask);
184 short _InterlockedAnd16(short volatile *_Value, short _Mask);
186 char _InterlockedAnd8(char volatile *_Value, char _Mask);
236 long _InterlockedOr(long volatile *_Value, long _Mask);
238 short _InterlockedOr16(short volatile *_Value, short _Mask);
240 char _InterlockedOr8(char volatile *_Value, char _Mask);
242 long _InterlockedXor(long volatile *_Value, long _Mask);
244 short _InterlockedXor16(short volatile *_Value, short _Mask);
246 char _InterlockedXor8(char volatile *_Value, char _Mask);
341 unsigned char _BitScanForward64(unsigned long *_Index, unsigned __int64 _Mask);
343 unsigned char _BitScanReverse64(unsigned long *_Index, unsigned __int64 _Mask);
353 long _InterlockedAnd_np(long volatile *_Value, long _Mask);
354 short _InterlockedAnd16_np(short volatile *_Value, short _Mask);
355 __int64 _InterlockedAnd64_np(__int64 volatile *_Value, __int64 _Mask);
356 char _InterlockedAnd8_np(char volatile *_Value, char _Mask);
391 long _InterlockedOr_np(long volatile *_Value, long _Mask);
392 short _InterlockedOr16_np(short volatile *_Value, short _Mask);
394 __int64 _InterlockedOr64(__int64 volatile *_Value, __int64 _Mask);
395 __int64 _InterlockedOr64_np(__int64 volatile *_Value, __int64 _Mask);
396 char _InterlockedOr8_np(char volatile *_Value, char _Mask);
397 long _InterlockedXor_np(long volatile *_Value, long _Mask);
398 short _InterlockedXor16_np(short volatile *_Value, short _Mask);
400 __int64 _InterlockedXor64(__int64 volatile *_Value, __int64 _Mask);
401 __int64 _InterlockedXor64_np(__int64 volatile *_Value, __int64 _Mask);
402 char _InterlockedXor8_np(char volatile *_Value, char _Mask);
504 _BitScanForward(unsigned long *_Index, unsigned long _Mask) { in _BitScanForward() argument
505 if (!_Mask) in _BitScanForward()
507 *_Index = __builtin_ctzl(_Mask); in _BitScanForward()
511 _BitScanReverse(unsigned long *_Index, unsigned long _Mask) { in _BitScanReverse() argument
512 if (!_Mask) in _BitScanReverse()
514 *_Index = 31 - __builtin_clzl(_Mask); in _BitScanReverse()
554 _BitScanForward64(unsigned long *_Index, unsigned __int64 _Mask) { in _BitScanForward64() argument
555 if (!_Mask) in _BitScanForward64()
557 *_Index = __builtin_ctzll(_Mask); in _BitScanForward64()
561 _BitScanReverse64(unsigned long *_Index, unsigned __int64 _Mask) { in _BitScanReverse64() argument
562 if (!_Mask) in _BitScanReverse64()
564 *_Index = 63 - __builtin_clzll(_Mask); in _BitScanReverse64()
669 _InterlockedAnd8(char volatile *_Value, char _Mask) { in _InterlockedAnd8() argument
670 return __atomic_and_fetch(_Value, _Mask, __ATOMIC_SEQ_CST); in _InterlockedAnd8()
673 _InterlockedAnd16(short volatile *_Value, short _Mask) { in _InterlockedAnd16() argument
674 return __atomic_and_fetch(_Value, _Mask, __ATOMIC_SEQ_CST); in _InterlockedAnd16()
677 _InterlockedAnd(long volatile *_Value, long _Mask) { in _InterlockedAnd() argument
678 return __atomic_and_fetch(_Value, _Mask, __ATOMIC_SEQ_CST); in _InterlockedAnd()
682 _InterlockedAnd64(__int64 volatile *_Value, __int64 _Mask) { in _InterlockedAnd64() argument
683 return __atomic_and_fetch(_Value, _Mask, __ATOMIC_SEQ_CST); in _InterlockedAnd64()
690 _InterlockedOr8(char volatile *_Value, char _Mask) { in _InterlockedOr8() argument
691 return __atomic_or_fetch(_Value, _Mask, __ATOMIC_SEQ_CST); in _InterlockedOr8()
694 _InterlockedOr16(short volatile *_Value, short _Mask) { in _InterlockedOr16() argument
695 return __atomic_or_fetch(_Value, _Mask, __ATOMIC_SEQ_CST); in _InterlockedOr16()
698 _InterlockedOr(long volatile *_Value, long _Mask) { in _InterlockedOr() argument
699 return __atomic_or_fetch(_Value, _Mask, __ATOMIC_SEQ_CST); in _InterlockedOr()
703 _InterlockedOr64(__int64 volatile *_Value, __int64 _Mask) { in _InterlockedOr64() argument
704 return __atomic_or_fetch(_Value, _Mask, __ATOMIC_SEQ_CST); in _InterlockedOr64()
711 _InterlockedXor8(char volatile *_Value, char _Mask) { in _InterlockedXor8() argument
712 return __atomic_xor_fetch(_Value, _Mask, __ATOMIC_SEQ_CST); in _InterlockedXor8()
715 _InterlockedXor16(short volatile *_Value, short _Mask) { in _InterlockedXor16() argument
716 return __atomic_xor_fetch(_Value, _Mask, __ATOMIC_SEQ_CST); in _InterlockedXor16()
719 _InterlockedXor(long volatile *_Value, long _Mask) { in _InterlockedXor() argument
720 return __atomic_xor_fetch(_Value, _Mask, __ATOMIC_SEQ_CST); in _InterlockedXor()
724 _InterlockedXor64(__int64 volatile *_Value, __int64 _Mask) { in _InterlockedXor64() argument
725 return __atomic_xor_fetch(_Value, _Mask, __ATOMIC_SEQ_CST); in _InterlockedXor64()