| /src/contrib/llvm-project/clang/lib/AST/Interp/ |
| H A D | InterpStack.cpp | 26 if (Chunk && Chunk->Next) in clear() 27 std::free(Chunk->Next); in clear() 28 if (Chunk) in clear() 29 std::free(Chunk); in clear() 30 Chunk = nullptr; in clear() 40 if (!Chunk || sizeof(StackChunk) + Chunk->size() + Size > ChunkSize) { in grow() 41 if (Chunk && Chunk->Next) { in grow() 42 Chunk = Chunk->Next; in grow() 44 StackChunk *Next = new (std::malloc(ChunkSize)) StackChunk(Chunk); in grow() 45 if (Chunk) in grow() [all …]
|
| /src/contrib/libucl/src/ |
| H A D | ucl_parser.c | 46 * @return new position in chunk 48 #define ucl_chunk_skipc(chunk, p) \ argument 50 if (p == chunk->end) { \ 54 (chunk)->line ++; \ 55 (chunk)->column = 0; \ 57 else (chunk)->column ++; \ 59 (chunk)->pos ++; \ 60 (chunk)->remain --; \ 67 struct ucl_chunk *chunk = parser->chunks; in ucl_set_err() local 76 if (chunk->pos < chunk->end) { in ucl_set_err() [all …]
|
| /src/sys/dev/mlx4/mlx4_core/ |
| H A D | mlx4_icm.c | 48 * per chunk. 55 static void mlx4_free_icm_pages(struct mlx4_dev *dev, struct mlx4_icm_chunk *chunk) in mlx4_free_icm_pages() argument 59 if (chunk->nsg > 0) in mlx4_free_icm_pages() 60 pci_unmap_sg(dev->persist->pdev, chunk->mem, chunk->npages, in mlx4_free_icm_pages() 63 for (i = 0; i < chunk->npages; ++i) in mlx4_free_icm_pages() 64 __free_pages(sg_page(&chunk->mem[i]), in mlx4_free_icm_pages() 65 get_order(chunk->mem[i].length)); in mlx4_free_icm_pages() 68 static void mlx4_free_icm_coherent(struct mlx4_dev *dev, struct mlx4_icm_chunk *chunk) in mlx4_free_icm_coherent() argument 72 for (i = 0; i < chunk->npages; ++i) in mlx4_free_icm_coherent() 74 chunk->mem[i].length, in mlx4_free_icm_coherent() [all …]
|
| /src/contrib/llvm-project/lld/COFF/ |
| H A D | DLL.h | 20 // call create() to populate the chunk vectors. 29 std::vector<Chunk *> dirs; 30 std::vector<Chunk *> lookups; 31 std::vector<Chunk *> addresses; 32 std::vector<Chunk *> hints; 33 std::vector<Chunk *> dllNames; 44 std::vector<Chunk *> getChunks(); 45 std::vector<Chunk *> getDataChunks(); 46 ArrayRef<Chunk *> getCodeChunks() { return thunks; } in getCodeChunks() 47 ArrayRef<Chunk *> getCodePData() { return pdata; } in getCodePData() [all …]
|
| H A D | Chunks.h | 52 // A Chunk represents a chunk of data that will occupy space in the 56 class Chunk { 66 // Returns the size of this chunk (even if this is a common or BSS.) 69 // Returns chunk alignment in power of two form. Value values are powers of 73 // Update the chunk section alignment measured in bytes. Internally alignment 84 // Write this chunk to a mmap'ed file, assuming Buf is pointing to 91 // larger than 2GB. Chunks are always laid as part of the image, so Chunk RVAs 102 // Returns the section name if this is a section chunk. 107 // chunk has a back pointer to an output section. 115 // Returns a human-readable name of this chunk. Chunks are unnamed chunks of [all …]
|
| H A D | DLL.cpp | 39 // A chunk for the import descriptor table. 61 // A chunk for the import descriptor table. 64 explicit LookupChunk(COFFLinkerContext &ctx, Chunk *c) in LookupChunk() 77 Chunk *hintName; 83 // A chunk for the import descriptor table. 84 // This chunk represent import-by-ordinal symbols. 110 // A chunk for the import descriptor table. 113 explicit ImportDirectoryChunk(Chunk *n) : dllName(n) { setAlignment(4); } in ImportDirectoryChunk() 125 Chunk *dllName; 126 Chunk *lookupTab; [all …]
|
| /src/sys/cddl/dev/kinst/ |
| H A D | trampoline.c | 75 struct trampchunk *chunk; in kinst_trampchunk_alloc() local 93 * Allocate virtual memory for the trampoline chunk. The returned in kinst_trampchunk_alloc() 101 KINST_LOG("trampoline chunk allocation failed: %d", error); in kinst_trampchunk_alloc() 111 /* Allocate a tracker for this chunk. */ in kinst_trampchunk_alloc() 112 chunk = malloc(sizeof(*chunk), M_KINST, M_WAITOK); in kinst_trampchunk_alloc() 113 chunk->addr = (void *)trampaddr; in kinst_trampchunk_alloc() 114 BIT_FILL(KINST_TRAMPS_PER_CHUNK, &chunk->free); in kinst_trampchunk_alloc() 116 TAILQ_INSERT_HEAD(&kinst_trampchunks, chunk, next); in kinst_trampchunk_alloc() 118 return (chunk); in kinst_trampchunk_alloc() 122 kinst_trampchunk_free(struct trampchunk *chunk) in kinst_trampchunk_free() argument [all …]
|
| /src/crypto/libecc/src/hash/ |
| H A D | hash_algs.c | 155 ATTRIBUTE_WARN_UNUSED_RET int _sha224_update(hash_context * hctx, const unsigned char *chunk, u32 c… in _sha224_update() argument 157 return sha224_update((sha224_context*)hctx, chunk, chunklen); in _sha224_update() 169 ATTRIBUTE_WARN_UNUSED_RET int _sha256_update(hash_context * hctx, const unsigned char *chunk, u32 c… in _sha256_update() argument 171 return sha256_update((sha256_context*)hctx, chunk, chunklen); in _sha256_update() 183 ATTRIBUTE_WARN_UNUSED_RET int _sha384_update(hash_context * hctx, const unsigned char *chunk, u32 c… in _sha384_update() argument 185 return sha384_update((sha384_context*)hctx, chunk, chunklen); in _sha384_update() 197 ATTRIBUTE_WARN_UNUSED_RET int _sha512_update(hash_context * hctx, const unsigned char *chunk, u32 c… in _sha512_update() argument 199 return sha512_update((sha512_context*)hctx, chunk, chunklen); in _sha512_update() 211 ATTRIBUTE_WARN_UNUSED_RET int _sha512_224_update(hash_context * hctx, const unsigned char *chunk, u… in _sha512_224_update() argument 213 return sha512_224_update((sha512_224_context*)hctx, chunk, chunklen); in _sha512_224_update() [all …]
|
| /src/sys/contrib/openzfs/module/zfs/ |
| H A D | zap_leaf.c | 31 * chunk number n means l_chunk[n], even though the header precedes it. 48 #define CHAIN_END 0xffff /* end of the chunk chain */ 180 uint_t chunk = zap_leaf_phys(l)->l_hdr.lh_freelist; in zap_leaf_chunk_alloc() local 181 ASSERT3U(chunk, <, ZAP_LEAF_NUMCHUNKS(l)); in zap_leaf_chunk_alloc() 182 ASSERT3U(ZAP_LEAF_CHUNK(l, chunk).l_free.lf_type, ==, ZAP_CHUNK_FREE); in zap_leaf_chunk_alloc() 185 ZAP_LEAF_CHUNK(l, chunk).l_free.lf_next; in zap_leaf_chunk_alloc() 189 return (chunk); in zap_leaf_chunk_alloc() 193 zap_leaf_chunk_free(zap_leaf_t *l, uint16_t chunk) in zap_leaf_chunk_free() argument 195 struct zap_leaf_free *zlf = &ZAP_LEAF_CHUNK(l, chunk).l_free; in zap_leaf_chunk_free() 197 ASSERT3U(chunk, <, ZAP_LEAF_NUMCHUNKS(l)); in zap_leaf_chunk_free() [all …]
|
| /src/contrib/libdiff/lib/ |
| H A D | diff_main.c | 39 diff_chunk_type(const struct diff_chunk *chunk) in diff_chunk_type() argument 41 if (!chunk->left_count && !chunk->right_count) in diff_chunk_type() 43 if (!chunk->solved) in diff_chunk_type() 45 if (!chunk->right_count) in diff_chunk_type() 47 if (!chunk->left_count) in diff_chunk_type() 49 if (chunk->left_count != chunk->right_count) in diff_chunk_type() 234 const struct diff_chunk *chunk) in diff_state_add_solved_chunk() argument 242 … to solved chunks; make sure that adjacent chunks of same type are combined, and that a minus chunk in diff_state_add_solved_chunk() 243 * never directly follows a plus chunk. */ in diff_state_add_solved_chunk() 248 new_t = diff_chunk_type(chunk); in diff_state_add_solved_chunk() [all …]
|
| /src/sys/dev/mthca/ |
| H A D | mthca_memfree.c | 47 * per chunk. 63 static void mthca_free_icm_pages(struct mthca_dev *dev, struct mthca_icm_chunk *chunk) in mthca_free_icm_pages() argument 67 if (chunk->nsg > 0) in mthca_free_icm_pages() 68 pci_unmap_sg(dev->pdev, chunk->mem, chunk->npages, in mthca_free_icm_pages() 71 for (i = 0; i < chunk->npages; ++i) in mthca_free_icm_pages() 72 __free_pages(sg_page(&chunk->mem[i]), in mthca_free_icm_pages() 73 get_order(chunk->mem[i].length)); in mthca_free_icm_pages() 76 static void mthca_free_icm_coherent(struct mthca_dev *dev, struct mthca_icm_chunk *chunk) in mthca_free_icm_coherent() argument 80 for (i = 0; i < chunk->npages; ++i) { in mthca_free_icm_coherent() 81 dma_free_coherent(&dev->pdev->dev, chunk->mem[i].length, in mthca_free_icm_coherent() [all …]
|
| /src/contrib/libdiff/test/ |
| H A D | expect004.diff | 4 +int Chunk_bounds_check(Chunk *chunk, size_t start, size_t n) 6 + if (chunk == NULL) return 0; 8 + return start <= chunk->length && n <= chunk->length - start; 11 void Chunk_copy(Chunk *src, size_t src_start, Chunk *dst, size_t dst_start, size_t n) 19 -int Chunk_bounds_check(Chunk *chunk, size_t start, size_t n) 21 - if (chunk == NULL) return 0; 23 - return start <= chunk->length && n <= chunk->length - start;
|
| H A D | expect104.diff | 4 +int Chunk_bounds_check(Chunk *chunk, size_t start, size_t n) 6 + if (chunk == NULL) return 0; 8 + return start <= chunk->length && n <= chunk->length - start; 11 void Chunk_copy(Chunk *src, size_t src_start, Chunk *dst, size_t dst_start, size_t n) 19 -int Chunk_bounds_check(Chunk *chunk, size_t start, size_t n) 21 - if (chunk == NULL) return 0; 23 - return start <= chunk->length && n <= chunk->length - start;
|
| /src/crypto/openssl/engines/asm/ |
| H A D | e_padlock-x86_64.pl | 47 $chunk="%rbx"; 318 mov \$$PADLOCK_CHUNK,$chunk 321 cmp $chunk,$len 322 cmovc $len,$chunk # chunk=len>PADLOCK_CHUNK?PADLOCK_CHUNK:len 323 and $chunk,%rax # out_misaligned?chunk:0 324 mov $len,$chunk 326 and \$$PADLOCK_CHUNK-1,$chunk # chunk%=PADLOCK_CHUNK 329 cmovz %rax,$chunk # chunk=chunk?:PADLOCK_CHUNK 337 mov \$$PADLOCK_CHUNK,$chunk 339 cmovz $chunk,%rax [all …]
|
| /src/usr.sbin/makefs/zfs/ |
| H A D | objset.c | 80 struct objset_dnode_chunk *chunk; in objset_alloc() local 92 chunk = ecalloc(1, sizeof(*chunk)); in objset_alloc() 93 chunk->nextfree = 1; in objset_alloc() 94 STAILQ_INSERT_HEAD(&os->dnodechunks, chunk, next); in objset_alloc() 114 struct objset_dnode_chunk *chunk, *tmp; in _objset_write() local 122 STAILQ_FOREACH_SAFE(chunk, &os->dnodechunks, next, tmp) { in _objset_write() 125 assert(chunk->nextfree > 0); in _objset_write() 126 assert(chunk->nextfree <= os->dnodecount); in _objset_write() 127 assert(chunk->nextfree <= DNODES_PER_CHUNK); in _objset_write() 129 for (i = 0; i < chunk->nextfree; i += DNODES_PER_BLOCK) { in _objset_write() [all …]
|
| /src/contrib/llvm-project/compiler-rt/lib/scudo/standalone/ |
| H A D | combined.h | 14 #include "chunk.h" 69 // Chunk recycling function, returns a quarantined chunk to the backend, 72 Chunk::UnpackedHeader Header; in recycle() 73 Chunk::loadHeader(Allocator.Cookie, Ptr, &Header); in recycle() 74 if (UNLIKELY(Header.State != Chunk::State::Quarantined)) in recycle() 77 Header.State = Chunk::State::Available; in recycle() 78 Chunk::storeHeader(Allocator.Cookie, Ptr, &Header); in recycle() 91 sizeof(QuarantineBatch) + Chunk::getHeaderSize()); in allocate() 98 Chunk::getHeaderSize()); in allocate() 99 Chunk::UnpackedHeader Header = {}; in allocate() [all …]
|
| H A D | wrappers_cpp.cpp | 39 void *Ptr = Allocator.allocate(size, scudo::Chunk::Origin::New); in operator new() 44 void *Ptr = Allocator.allocate(size, scudo::Chunk::Origin::NewArray); in operator new[]() 50 void *Ptr = Allocator.allocate(size, scudo::Chunk::Origin::New); in operator new() 56 void *Ptr = Allocator.allocate(size, scudo::Chunk::Origin::NewArray); in operator new[]() 61 void *Ptr = Allocator.allocate(size, scudo::Chunk::Origin::New, in operator new() 67 void *Ptr = Allocator.allocate(size, scudo::Chunk::Origin::NewArray, in operator new[]() 74 void *Ptr = Allocator.allocate(size, scudo::Chunk::Origin::New, in operator new() 81 void *Ptr = Allocator.allocate(size, scudo::Chunk::Origin::NewArray, in operator new[]() 89 Allocator.deallocate(ptr, scudo::Chunk::Origin::New); in operator delete() 93 Allocator.deallocate(ptr, scudo::Chunk::Origin::NewArray); in operator delete[]() [all …]
|
| /src/sys/dev/irdma/ |
| H A D | irdma_pble.c | 51 struct irdma_chunk *chunk; in irdma_destroy_pble_prm() local 55 chunk = (struct irdma_chunk *)(&pinfo->clist)->next; in irdma_destroy_pble_prm() 56 list_del(&chunk->list); in irdma_destroy_pble_prm() 57 if (chunk->type == PBLE_SD_PAGED) in irdma_destroy_pble_prm() 58 irdma_pble_free_paged_mem(chunk); in irdma_destroy_pble_prm() 59 bitmap_free(chunk->bitmapbuf); in irdma_destroy_pble_prm() 60 kfree(chunk->chunkmem.va); in irdma_destroy_pble_prm() 128 struct irdma_chunk *chunk = info->chunk; in add_sd_direct() local 141 chunk->type = PBLE_SD_CONTIGOUS; in add_sd_direct() 145 chunk->size = info->pages << HMC_PAGED_BP_SHIFT; in add_sd_direct() [all …]
|
| /src/contrib/llvm-project/openmp/runtime/src/ |
| H A D | kmp_sched.cpp | 79 typename traits_t<T>::signed_t chunk in __kmp_for_static_init() argument 141 " %%%s, %%%s) incr=%%%s chunk=%%%s signed?<%s>\n", in __kmp_for_static_init() 145 *pstride, incr, chunk)); in __kmp_for_static_init() 364 KMP_DEBUG_ASSERT(chunk != 0); in __kmp_for_static_init() 365 if (chunk < 1) in __kmp_for_static_init() 366 chunk = 1; in __kmp_for_static_init() 367 else if ((UT)chunk > trip_count) in __kmp_for_static_init() 368 chunk = trip_count; in __kmp_for_static_init() 369 nchunks = (trip_count) / (UT)chunk + (trip_count % (UT)chunk ? 1 : 0); in __kmp_for_static_init() 370 span = chunk * incr; in __kmp_for_static_init() [all …]
|
| /src/crypto/openssl/providers/implementations/ciphers/ |
| H A D | cipher_des_hw.c | 110 size_t chunk = MAXCHUNK; in cipher_hw_des_cfb64_cipher() local 114 if (len < chunk) in cipher_hw_des_cfb64_cipher() 115 chunk = len; in cipher_hw_des_cfb64_cipher() 116 while (len > 0 && len >= chunk) { in cipher_hw_des_cfb64_cipher() 117 DES_cfb64_encrypt(in, out, (long)chunk, key, (DES_cblock *)ctx->iv, in cipher_hw_des_cfb64_cipher() 119 len -= chunk; in cipher_hw_des_cfb64_cipher() 120 in += chunk; in cipher_hw_des_cfb64_cipher() 121 out += chunk; in cipher_hw_des_cfb64_cipher() 122 if (len < chunk) in cipher_hw_des_cfb64_cipher() 123 chunk = len; in cipher_hw_des_cfb64_cipher() [all …]
|
| H A D | ciphercommon_hw.c | 148 size_t chunk = MAXCHUNK; in ossl_cipher_hw_chunked_cfb8() local 150 if (inl < chunk) in ossl_cipher_hw_chunked_cfb8() 151 chunk = inl; in ossl_cipher_hw_chunked_cfb8() 152 while (inl > 0 && inl >= chunk) { in ossl_cipher_hw_chunked_cfb8() 154 inl -= chunk; in ossl_cipher_hw_chunked_cfb8() 155 in += chunk; in ossl_cipher_hw_chunked_cfb8() 156 out += chunk; in ossl_cipher_hw_chunked_cfb8() 157 if (inl < chunk) in ossl_cipher_hw_chunked_cfb8() 158 chunk = inl; in ossl_cipher_hw_chunked_cfb8() 166 size_t chunk = MAXCHUNK; in ossl_cipher_hw_chunked_cfb128() local [all …]
|
| /src/contrib/llvm-project/llvm/lib/Target/AArch64/ |
| H A D | AArch64ExpandImm.cpp | 20 /// Helper function which extracts the specified 16-bit chunk from a 23 assert(ChunkIdx < 4 && "Out of range chunk index specified!"); in getChunk() 28 /// Check whether the given 16-bit chunk replicated to full 64-bit width 30 static bool canUseOrr(uint64_t Chunk, uint64_t &Encoding) { in canUseOrr() argument 31 Chunk = (Chunk << 48) | (Chunk << 32) | (Chunk << 16) | Chunk; in canUseOrr() 33 return AArch64_AM::processLogicalImmediate(Chunk, 64, Encoding); in canUseOrr() 49 // Scan the constant and count how often every chunk occurs. in tryToreplicateChunks() 54 for (const auto &Chunk : Counts) { in tryToreplicateChunks() local 55 const uint64_t ChunkVal = Chunk.first; in tryToreplicateChunks() 56 const unsigned Count = Chunk.second; in tryToreplicateChunks() [all …]
|
| /src/contrib/llvm-project/llvm/lib/Support/BLAKE3/ |
| H A D | blake3.c | 82 // Chaining values within a given chunk (specifically the compress_in_place 159 // Given some input larger than one chunk, return the number of bytes that 170 // on a single thread. Write out the chunk chaining values and return the 195 // Hash the remaining partial chunk, if there is one. Note that the empty in compress_chunks_parallel() 196 // chunk (meaning the empty message) is a different codepath. in compress_chunks_parallel() 263 // not used when the whole input is only 1 chunk long; that's a different 274 // Note that the single chunk case does *not* bump the SIMD degree up to 2 in blake3_compress_subtree_wide() 276 // this gives us the option of multi-threading even the 2-chunk case, which in blake3_compress_subtree_wide() 300 // sure there are two outputs. Except, as noted above, at the chunk in blake3_compress_subtree_wide() 301 // level, where we allow degree=1. (Note that the 1-chunk-input case is in blake3_compress_subtree_wide() [all …]
|
| /src/sys/contrib/openzfs/module/icp/algs/blake3/ |
| H A D | blake3.c | 127 * Chaining values within a given chunk (specifically the compress_in_place 212 * Given some input larger than one chunk, return the number of bytes that 229 * on a single thread. Write out the chunk chaining values and return the 251 * Hash the remaining partial chunk, if there is one. Note that the in compress_chunks_parallel() 252 * empty chunk (meaning the empty message) is a different codepath. in compress_chunks_parallel() 316 * not used when the whole input is only 1 chunk long; that's a different 328 * Note that the single chunk case does *not* bump the SIMD degree up in blake3_compress_subtree_wide() 331 * 2-chunk case, which can help performance on smaller platforms. in blake3_compress_subtree_wide() 364 * at the chunk level, where we allow degree=1. (Note that the in blake3_compress_subtree_wide() 365 * 1-chunk-input case is a different codepath.) in blake3_compress_subtree_wide() [all …]
|
| /src/lib/libc/aarch64/string/ |
| H A D | memccpy.S | 42 /* match in first chunk */ 57 ldr q3, [x10, #16] // load second string chunk 59 cmeq v1.16b, v3.16b, v0.16b // char found in second chunk? 61 /* process second chunk */ 67 /* match in second chunk */ 82 /* string didn't end in second chunk and neither did buffer */ 83 ldr q1, [x10, #32] // load next string chunk 87 str q3, [x0, #16] // deposit second chunk 97 cmeq v2.16b, v1.16b, v0.16b // char found in second chunk? 104 ldr q1, [x10, #16] // load next chunk [all …]
|