Lines Matching refs:_BitPos
526 _bittest(long const *_BitBase, long _BitPos) { in _bittest() argument
527 return (*_BitBase >> _BitPos) & 1; in _bittest()
530 _bittestandcomplement(long *_BitBase, long _BitPos) { in _bittestandcomplement() argument
531 unsigned char _Res = (*_BitBase >> _BitPos) & 1; in _bittestandcomplement()
532 *_BitBase = *_BitBase ^ (1 << _BitPos); in _bittestandcomplement()
536 _bittestandreset(long *_BitBase, long _BitPos) { in _bittestandreset() argument
537 unsigned char _Res = (*_BitBase >> _BitPos) & 1; in _bittestandreset()
538 *_BitBase = *_BitBase & ~(1 << _BitPos); in _bittestandreset()
542 _bittestandset(long *_BitBase, long _BitPos) { in _bittestandset() argument
543 unsigned char _Res = (*_BitBase >> _BitPos) & 1; in _bittestandset()
544 *_BitBase = *_BitBase | (1 << _BitPos); in _bittestandset()
548 _interlockedbittestandset(long volatile *_BitBase, long _BitPos) { in _interlockedbittestandset() argument
549 long _PrevVal = __atomic_fetch_or(_BitBase, 1l << _BitPos, __ATOMIC_SEQ_CST); in _interlockedbittestandset()
550 return (_PrevVal >> _BitPos) & 1; in _interlockedbittestandset()
573 _bittest64(__int64 const *_BitBase, __int64 _BitPos) { in _bittest64() argument
574 return (*_BitBase >> _BitPos) & 1; in _bittest64()
577 _bittestandcomplement64(__int64 *_BitBase, __int64 _BitPos) { in _bittestandcomplement64() argument
578 unsigned char _Res = (*_BitBase >> _BitPos) & 1; in _bittestandcomplement64()
579 *_BitBase = *_BitBase ^ (1ll << _BitPos); in _bittestandcomplement64()
583 _bittestandreset64(__int64 *_BitBase, __int64 _BitPos) { in _bittestandreset64() argument
584 unsigned char _Res = (*_BitBase >> _BitPos) & 1; in _bittestandreset64()
585 *_BitBase = *_BitBase & ~(1ll << _BitPos); in _bittestandreset64()
589 _bittestandset64(__int64 *_BitBase, __int64 _BitPos) { in _bittestandset64() argument
590 unsigned char _Res = (*_BitBase >> _BitPos) & 1; in _bittestandset64()
591 *_BitBase = *_BitBase | (1ll << _BitPos); in _bittestandset64()
595 _interlockedbittestandset64(__int64 volatile *_BitBase, __int64 _BitPos) { in _interlockedbittestandset64() argument
597 __atomic_fetch_or(_BitBase, 1ll << _BitPos, __ATOMIC_SEQ_CST); in _interlockedbittestandset64()
598 return (_PrevVal >> _BitPos) & 1; in _interlockedbittestandset64()