Lines Matching refs:_BitPos
525 _bittest(long const *_BitBase, long _BitPos) { in _bittest() argument
526 return (*_BitBase >> _BitPos) & 1; in _bittest()
529 _bittestandcomplement(long *_BitBase, long _BitPos) { in _bittestandcomplement() argument
530 unsigned char _Res = (*_BitBase >> _BitPos) & 1; in _bittestandcomplement()
531 *_BitBase = *_BitBase ^ (1 << _BitPos); in _bittestandcomplement()
535 _bittestandreset(long *_BitBase, long _BitPos) { in _bittestandreset() argument
536 unsigned char _Res = (*_BitBase >> _BitPos) & 1; in _bittestandreset()
537 *_BitBase = *_BitBase & ~(1 << _BitPos); in _bittestandreset()
541 _bittestandset(long *_BitBase, long _BitPos) { in _bittestandset() argument
542 unsigned char _Res = (*_BitBase >> _BitPos) & 1; in _bittestandset()
543 *_BitBase = *_BitBase | (1 << _BitPos); in _bittestandset()
547 _interlockedbittestandset(long volatile *_BitBase, long _BitPos) { in _interlockedbittestandset() argument
548 long _PrevVal = __atomic_fetch_or(_BitBase, 1l << _BitPos, __ATOMIC_SEQ_CST); in _interlockedbittestandset()
549 return (_PrevVal >> _BitPos) & 1; in _interlockedbittestandset()
572 _bittest64(__int64 const *_BitBase, __int64 _BitPos) { in _bittest64() argument
573 return (*_BitBase >> _BitPos) & 1; in _bittest64()
576 _bittestandcomplement64(__int64 *_BitBase, __int64 _BitPos) { in _bittestandcomplement64() argument
577 unsigned char _Res = (*_BitBase >> _BitPos) & 1; in _bittestandcomplement64()
578 *_BitBase = *_BitBase ^ (1ll << _BitPos); in _bittestandcomplement64()
582 _bittestandreset64(__int64 *_BitBase, __int64 _BitPos) { in _bittestandreset64() argument
583 unsigned char _Res = (*_BitBase >> _BitPos) & 1; in _bittestandreset64()
584 *_BitBase = *_BitBase & ~(1ll << _BitPos); in _bittestandreset64()
588 _bittestandset64(__int64 *_BitBase, __int64 _BitPos) { in _bittestandset64() argument
589 unsigned char _Res = (*_BitBase >> _BitPos) & 1; in _bittestandset64()
590 *_BitBase = *_BitBase | (1ll << _BitPos); in _bittestandset64()
594 _interlockedbittestandset64(__int64 volatile *_BitBase, __int64 _BitPos) { in _interlockedbittestandset64() argument
596 __atomic_fetch_or(_BitBase, 1ll << _BitPos, __ATOMIC_SEQ_CST); in _interlockedbittestandset64()
597 return (_PrevVal >> _BitPos) & 1; in _interlockedbittestandset64()