Home
last modified time | relevance | path

Searched refs:relRow (Results 1 – 1 of 1) sorted by relevance

/linux/lib/zstd/compress/
H A Dzstd_lazy.c817 FORCE_INLINE_TEMPLATE void ZSTD_row_prefetch(U32 const* hashTable, BYTE const* tagTable, U32 const relRow, U32 const rowLog) { in ZSTD_row_prefetch() argument
818 PREFETCH_L1(hashTable + relRow); in ZSTD_row_prefetch()
820 PREFETCH_L1(hashTable + relRow + 16); in ZSTD_row_prefetch()
823 PREFETCH_L1(tagTable + relRow); in ZSTD_row_prefetch()
825 PREFETCH_L1(tagTable + relRow + 32); in ZSTD_row_prefetch()
828 assert(ZSTD_isAligned(hashTable + relRow, 64)); /* prefetched hash row always 64-byte aligned */ in ZSTD_row_prefetch()
829 assert(ZSTD_isAligned(tagTable + relRow, (size_t)1 << rowLog)); /* prefetched tagRow sits on correct multiple of bytes (32,64,128) */ in ZSTD_row_prefetch()
900 U32 const relRow = (hash >> ZSTD_ROW_HASH_TAG_BITS) << rowLog; in ZSTD_row_update_internalImpl() local
901 U32* const row = hashTable + relRow; in ZSTD_row_update_internalImpl()
902 BYTE* tagRow = tagTable + relRow; in ZSTD_row_update_internalImpl()
1219 U32 const relRow = (hash >> ZSTD_ROW_HASH_TAG_BITS) << rowLog; ZSTD_RowFindBestMatch() local
[all...]