Lines Matching refs:U32
278 typedef uint32_t U32; typedef
288 typedef unsigned int U32; typedef
327 … const union { U32 u; BYTE c[4]; } one = { 1 }; /* don't use static : performance detrimental */ in LZ4_isLittleEndian()
336 static U32 LZ4_read32(const void* memPtr) { return *(const U32*) memPtr; } in LZ4_read32()
340 static void LZ4_write32(void* memPtr, U32 value) { *(U32*)memPtr = value; } in LZ4_write32()
346 typedef union { U16 u16; U32 u32; reg_t uArch; } __attribute__((packed)) unalign;
349 static U32 LZ4_read32(const void* ptr) { return ((const unalign*)ptr)->u32; } in LZ4_read32()
353 static void LZ4_write32(void* memPtr, U32 value) { ((unalign*)memPtr)->u32 = value; } in LZ4_write32()
362 static U32 LZ4_read32(const void* memPtr) in LZ4_read32()
364 U32 val; LZ4_memcpy(&val, memPtr, sizeof(val)); return val; in LZ4_read32()
377 static void LZ4_write32(void* memPtr, U32 value) in LZ4_write32()
537 _BitScanForward(&r, (U32)val); in LZ4_NbCommonBytes()
542 return (unsigned)__builtin_ctz((U32)val) >> 3; in LZ4_NbCommonBytes()
544 const U32 m = 0x01010101; in LZ4_NbCommonBytes()
575 static const U32 by32 = sizeof(val)*4; /* 32 on 64 bits (goal), 16 on 32 bits. in LZ4_NbCommonBytes()
589 return (unsigned)__builtin_clz((U32)val) >> 3; in LZ4_NbCommonBytes()
634 static const U32 LZ4_skipTrigger = 6; /* Increase this value ==> compression run slower on incompr…
698 LZ4_FORCE_INLINE U32 LZ4_hash4(U32 sequence, tableType_t const tableType) in LZ4_hash4()
706 LZ4_FORCE_INLINE U32 LZ4_hash5(U64 sequence, tableType_t const tableType) in LZ4_hash5()
708 const U32 hashLog = (tableType == byU16) ? LZ4_HASHLOG+1 : LZ4_HASHLOG; in LZ4_hash5()
711 return (U32)(((sequence << 24) * prime5bytes) >> (64 - hashLog)); in LZ4_hash5()
714 return (U32)(((sequence >> 24) * prime8bytes) >> (64 - hashLog)); in LZ4_hash5()
718 LZ4_FORCE_INLINE U32 LZ4_hashPosition(const void* const p, tableType_t const tableType) in LZ4_hashPosition()
724 LZ4_FORCE_INLINE void LZ4_clearHash(U32 h, void* tableBase, tableType_t const tableType) in LZ4_clearHash()
731 case byU32: { U32* hashTable = (U32*) tableBase; hashTable[h] = 0; return; } in LZ4_clearHash()
736 LZ4_FORCE_INLINE void LZ4_putIndexOnHash(U32 idx, U32 h, void* tableBase, tableType_t const tableTy… in LZ4_putIndexOnHash()
743 case byU32: { U32* hashTable = (U32*) tableBase; hashTable[h] = idx; return; } in LZ4_putIndexOnHash()
748 LZ4_FORCE_INLINE void LZ4_putPositionOnHash(const BYTE* p, U32 h, in LZ4_putPositionOnHash()
756 case byU32: { U32* hashTable = (U32*) tableBase; hashTable[h] = (U32)(p-srcBase); return; } in LZ4_putPositionOnHash()
763 U32 const h = LZ4_hashPosition(p, tableType); in LZ4_putPosition()
773 LZ4_FORCE_INLINE U32 LZ4_getIndexOnHash(U32 h, const void* tableBase, tableType_t tableType) in LZ4_getIndexOnHash()
777 const U32* const hashTable = (const U32*) tableBase; in LZ4_getIndexOnHash()
789 static const BYTE* LZ4_getPositionOnHash(U32 h, const void* tableBase, tableType_t tableType, const… in LZ4_getPositionOnHash()
792 …if (tableType == byU32) { const U32* const hashTable = (const U32*) tableBase; return hashTable[h]… in LZ4_getPositionOnHash()
801 U32 const h = LZ4_hashPosition(p, tableType); in LZ4_getPosition()
824 cctx->tableType = (U32)clearedTable; in LZ4_prepareTable()
867 U32 const startIndex = cctx->currentOffset; in LZ4_compress_generic_validated()
874 const U32 dictSize = in LZ4_compress_generic_validated()
876 …const U32 dictDelta = (dictDirective == usingDictCtx) ? startIndex - dictCtx->currentOffset : 0; … in LZ4_compress_generic_validated()
879 U32 const prefixIdxLimit = startIndex - dictSize; /* used when dictDirective == dictSmall */ in LZ4_compress_generic_validated()
895 U32 offset = 0; in LZ4_compress_generic_validated()
896 U32 forwardH; in LZ4_compress_generic_validated()
914 cctx->dictSize = (U32)inputSize; in LZ4_compress_generic_validated()
916 cctx->dictSize += (U32)inputSize; in LZ4_compress_generic_validated()
918 cctx->currentOffset += (U32)inputSize; in LZ4_compress_generic_validated()
919 cctx->tableType = (U32)tableType; in LZ4_compress_generic_validated()
939 U32 const h = forwardH; in LZ4_compress_generic_validated()
960 U32 const h = forwardH; in LZ4_compress_generic_validated()
961 U32 const current = (U32)(forwardIp - base); in LZ4_compress_generic_validated()
962 U32 matchIndex = LZ4_getIndexOnHash(h, cctx->hashTable, tableType); in LZ4_compress_generic_validated()
1070 DEBUGLOG(6, " with offset=%u (same segment)", (U32)(ip - match)); in LZ4_compress_generic_validated()
1101 …U32 newMatchCode = 15 /* in token */ - 1 /* to avoid needing a zero byte */ + ((U32)(olimit - op) … in LZ4_compress_generic_validated()
1112 DEBUGLOG(5, "Clearing %u positions", (U32)(filledIp - ip)); in LZ4_compress_generic_validated()
1114 U32 const h = LZ4_hashPosition(ptr, tableType); in LZ4_compress_generic_validated()
1159 U32 const h = LZ4_hashPosition(ip, tableType); in LZ4_compress_generic_validated()
1160 U32 const current = (U32)(ip-base); in LZ4_compress_generic_validated()
1161 U32 matchIndex = LZ4_getIndexOnHash(h, cctx->hashTable, tableType); in LZ4_compress_generic_validated()
1262 …if ((U32)srcSize > (U32)LZ4_MAX_INPUT_SIZE) { return 0; } /* Unsupported srcSize, too large (or n… in LZ4_compress_generic()
1506 dict->dictSize = (U32)(dictEnd - p); in LZ4_loadDict()
1507 dict->tableType = (U32)tableType; in LZ4_loadDict()
1550 U32 const delta = LZ4_dict->currentOffset - 64 KB; in LZ4_renormDictT()
1592 streamPtr->dictSize = (U32)(dictEnd - sourceEnd); in LZ4_compress_fast_continue()
1634 streamPtr->dictSize = (U32)inputSize; in LZ4_compress_fast_continue()
1655 streamPtr->dictSize = (U32)srcSize; in LZ4_compress_forceExtDict()
1673 if ((U32)dictSize > 64 KB) { dictSize = 64 KB; } /* useless to define a dictionary > 64 KB */ in LZ4_saveDict()
1674 if ((U32)dictSize > dict->dictSize) { dictSize = (int)dict->dictSize; } in LZ4_saveDict()
1681 dict->dictSize = (U32)dictSize; in LZ4_saveDict()
1712 U32 length = 0; in read_variable_length()
1713 U32 s; in read_variable_length()