Lines Matching refs:_Mask
167 unsigned char _BitScanForward(unsigned long *_Index, unsigned long _Mask);
169 unsigned char _BitScanReverse(unsigned long *_Index, unsigned long _Mask);
198 long _InterlockedAnd(long volatile *_Value, long _Mask);
200 short _InterlockedAnd16(short volatile *_Value, short _Mask);
202 char _InterlockedAnd8(char volatile *_Value, char _Mask);
252 long _InterlockedOr(long volatile *_Value, long _Mask);
254 short _InterlockedOr16(short volatile *_Value, short _Mask);
256 char _InterlockedOr8(char volatile *_Value, char _Mask);
258 long _InterlockedXor(long volatile *_Value, long _Mask);
260 short _InterlockedXor16(short volatile *_Value, short _Mask);
262 char _InterlockedXor8(char volatile *_Value, char _Mask);
371 unsigned char _BitScanForward64(unsigned long *_Index, unsigned __int64 _Mask);
373 unsigned char _BitScanReverse64(unsigned long *_Index, unsigned __int64 _Mask);
396 long _InterlockedAnd_np(long volatile *_Value, long _Mask);
397 short _InterlockedAnd16_np(short volatile *_Value, short _Mask);
398 __int64 _InterlockedAnd64_np(__int64 volatile *_Value, __int64 _Mask);
399 char _InterlockedAnd8_np(char volatile *_Value, char _Mask);
434 long _InterlockedOr_np(long volatile *_Value, long _Mask);
435 short _InterlockedOr16_np(short volatile *_Value, short _Mask);
437 __int64 _InterlockedOr64(__int64 volatile *_Value, __int64 _Mask);
438 __int64 _InterlockedOr64_np(__int64 volatile *_Value, __int64 _Mask);
439 char _InterlockedOr8_np(char volatile *_Value, char _Mask);
440 long _InterlockedXor_np(long volatile *_Value, long _Mask);
441 short _InterlockedXor16_np(short volatile *_Value, short _Mask);
443 __int64 _InterlockedXor64(__int64 volatile *_Value, __int64 _Mask);
444 __int64 _InterlockedXor64_np(__int64 volatile *_Value, __int64 _Mask);
445 char _InterlockedXor8_np(char volatile *_Value, char _Mask);
535 _BitScanForward(unsigned long *_Index, unsigned long _Mask) { in _BitScanForward() argument
536 if (!_Mask) in _BitScanForward()
538 *_Index = __builtin_ctzl(_Mask); in _BitScanForward()
542 _BitScanReverse(unsigned long *_Index, unsigned long _Mask) { in _BitScanReverse() argument
543 if (!_Mask) in _BitScanReverse()
545 *_Index = 31 - __builtin_clzl(_Mask); in _BitScanReverse()
598 _BitScanForward64(unsigned long *_Index, unsigned __int64 _Mask) { in _BitScanForward64() argument
599 if (!_Mask) in _BitScanForward64()
601 *_Index = __builtin_ctzll(_Mask); in _BitScanForward64()
605 _BitScanReverse64(unsigned long *_Index, unsigned __int64 _Mask) { in _BitScanReverse64() argument
606 if (!_Mask) in _BitScanReverse64()
608 *_Index = 63 - __builtin_clzll(_Mask); in _BitScanReverse64()
724 _InterlockedAnd8(char volatile *_Value, char _Mask) { in _InterlockedAnd8() argument
725 return __atomic_and_fetch(_Value, _Mask, 0); in _InterlockedAnd8()
728 _InterlockedAnd16(short volatile *_Value, short _Mask) { in _InterlockedAnd16() argument
729 return __atomic_and_fetch(_Value, _Mask, 0); in _InterlockedAnd16()
732 _InterlockedAnd(long volatile *_Value, long _Mask) { in _InterlockedAnd() argument
733 return __atomic_and_fetch(_Value, _Mask, 0); in _InterlockedAnd()
737 _InterlockedAnd64(__int64 volatile *_Value, __int64 _Mask) { in _InterlockedAnd64() argument
738 return __atomic_and_fetch(_Value, _Mask, 0); in _InterlockedAnd64()
745 _InterlockedOr8(char volatile *_Value, char _Mask) { in _InterlockedOr8() argument
746 return __atomic_or_fetch(_Value, _Mask, 0); in _InterlockedOr8()
749 _InterlockedOr16(short volatile *_Value, short _Mask) { in _InterlockedOr16() argument
750 return __atomic_or_fetch(_Value, _Mask, 0); in _InterlockedOr16()
753 _InterlockedOr(long volatile *_Value, long _Mask) { in _InterlockedOr() argument
754 return __atomic_or_fetch(_Value, _Mask, 0); in _InterlockedOr()
758 _InterlockedOr64(__int64 volatile *_Value, __int64 _Mask) { in _InterlockedOr64() argument
759 return __atomic_or_fetch(_Value, _Mask, 0); in _InterlockedOr64()
766 _InterlockedXor8(char volatile *_Value, char _Mask) { in _InterlockedXor8() argument
767 return __atomic_xor_fetch(_Value, _Mask, 0); in _InterlockedXor8()
770 _InterlockedXor16(short volatile *_Value, short _Mask) { in _InterlockedXor16() argument
771 return __atomic_xor_fetch(_Value, _Mask, 0); in _InterlockedXor16()
774 _InterlockedXor(long volatile *_Value, long _Mask) { in _InterlockedXor() argument
775 return __atomic_xor_fetch(_Value, _Mask, 0); in _InterlockedXor()
779 _InterlockedXor64(__int64 volatile *_Value, __int64 _Mask) { in _InterlockedXor64() argument
780 return __atomic_xor_fetch(_Value, _Mask, 0); in _InterlockedXor64()