Home
last modified time | relevance | path

Searched full:chunk (Results 1 – 25 of 996) sorted by relevance

12345678910>>...40

/src/contrib/llvm-project/clang/lib/AST/Interp/
H A DInterpStack.cpp26 if (Chunk && Chunk->Next) in clear()
27 std::free(Chunk->Next); in clear()
28 if (Chunk) in clear()
29 std::free(Chunk); in clear()
30 Chunk = nullptr; in clear()
40 if (!Chunk || sizeof(StackChunk) + Chunk->size() + Size > ChunkSize) { in grow()
41 if (Chunk && Chunk->Next) { in grow()
42 Chunk = Chunk->Next; in grow()
44 StackChunk *Next = new (std::malloc(ChunkSize)) StackChunk(Chunk); in grow()
45 if (Chunk) in grow()
[all …]
/src/contrib/libucl/src/
H A Ducl_parser.c46 * @return new position in chunk
48 #define ucl_chunk_skipc(chunk, p) \ argument
50 if (p == chunk->end) { \
54 (chunk)->line ++; \
55 (chunk)->column = 0; \
57 else (chunk)->column ++; \
59 (chunk)->pos ++; \
60 (chunk)->remain --; \
67 struct ucl_chunk *chunk = parser->chunks; in ucl_set_err() local
76 if (chunk->pos < chunk->end) { in ucl_set_err()
[all …]
/src/sys/dev/mlx4/mlx4_core/
H A Dmlx4_icm.c48 * per chunk.
55 static void mlx4_free_icm_pages(struct mlx4_dev *dev, struct mlx4_icm_chunk *chunk) in mlx4_free_icm_pages() argument
59 if (chunk->nsg > 0) in mlx4_free_icm_pages()
60 pci_unmap_sg(dev->persist->pdev, chunk->mem, chunk->npages, in mlx4_free_icm_pages()
63 for (i = 0; i < chunk->npages; ++i) in mlx4_free_icm_pages()
64 __free_pages(sg_page(&chunk->mem[i]), in mlx4_free_icm_pages()
65 get_order(chunk->mem[i].length)); in mlx4_free_icm_pages()
68 static void mlx4_free_icm_coherent(struct mlx4_dev *dev, struct mlx4_icm_chunk *chunk) in mlx4_free_icm_coherent() argument
72 for (i = 0; i < chunk->npages; ++i) in mlx4_free_icm_coherent()
74 chunk->mem[i].length, in mlx4_free_icm_coherent()
[all …]
/src/contrib/llvm-project/lld/COFF/
H A DDLL.h20 // call create() to populate the chunk vectors.
29 std::vector<Chunk *> dirs;
30 std::vector<Chunk *> lookups;
31 std::vector<Chunk *> addresses;
32 std::vector<Chunk *> hints;
33 std::vector<Chunk *> dllNames;
44 std::vector<Chunk *> getChunks();
45 std::vector<Chunk *> getDataChunks();
46 ArrayRef<Chunk *> getCodeChunks() { return thunks; } in getCodeChunks()
47 ArrayRef<Chunk *> getCodePData() { return pdata; } in getCodePData()
[all …]
H A DChunks.h52 // A Chunk represents a chunk of data that will occupy space in the
56 class Chunk {
66 // Returns the size of this chunk (even if this is a common or BSS.)
69 // Returns chunk alignment in power of two form. Value values are powers of
73 // Update the chunk section alignment measured in bytes. Internally alignment
84 // Write this chunk to a mmap'ed file, assuming Buf is pointing to
91 // larger than 2GB. Chunks are always laid as part of the image, so Chunk RVAs
102 // Returns the section name if this is a section chunk.
107 // chunk has a back pointer to an output section.
115 // Returns a human-readable name of this chunk. Chunks are unnamed chunks of
[all …]
H A DDLL.cpp39 // A chunk for the import descriptor table.
61 // A chunk for the import descriptor table.
64 explicit LookupChunk(COFFLinkerContext &ctx, Chunk *c) in LookupChunk()
77 Chunk *hintName;
83 // A chunk for the import descriptor table.
84 // This chunk represent import-by-ordinal symbols.
110 // A chunk for the import descriptor table.
113 explicit ImportDirectoryChunk(Chunk *n) : dllName(n) { setAlignment(4); } in ImportDirectoryChunk()
125 Chunk *dllName;
126 Chunk *lookupTab;
[all …]
/src/sys/cddl/dev/kinst/
H A Dtrampoline.c75 struct trampchunk *chunk; in kinst_trampchunk_alloc() local
93 * Allocate virtual memory for the trampoline chunk. The returned in kinst_trampchunk_alloc()
101 KINST_LOG("trampoline chunk allocation failed: %d", error); in kinst_trampchunk_alloc()
111 /* Allocate a tracker for this chunk. */ in kinst_trampchunk_alloc()
112 chunk = malloc(sizeof(*chunk), M_KINST, M_WAITOK); in kinst_trampchunk_alloc()
113 chunk->addr = (void *)trampaddr; in kinst_trampchunk_alloc()
114 BIT_FILL(KINST_TRAMPS_PER_CHUNK, &chunk->free); in kinst_trampchunk_alloc()
116 TAILQ_INSERT_HEAD(&kinst_trampchunks, chunk, next); in kinst_trampchunk_alloc()
118 return (chunk); in kinst_trampchunk_alloc()
122 kinst_trampchunk_free(struct trampchunk *chunk) in kinst_trampchunk_free() argument
[all …]
/src/crypto/libecc/src/hash/
H A Dhash_algs.c155 ATTRIBUTE_WARN_UNUSED_RET int _sha224_update(hash_context * hctx, const unsigned char *chunk, u32 c… in _sha224_update() argument
157 return sha224_update((sha224_context*)hctx, chunk, chunklen); in _sha224_update()
169 ATTRIBUTE_WARN_UNUSED_RET int _sha256_update(hash_context * hctx, const unsigned char *chunk, u32 c… in _sha256_update() argument
171 return sha256_update((sha256_context*)hctx, chunk, chunklen); in _sha256_update()
183 ATTRIBUTE_WARN_UNUSED_RET int _sha384_update(hash_context * hctx, const unsigned char *chunk, u32 c… in _sha384_update() argument
185 return sha384_update((sha384_context*)hctx, chunk, chunklen); in _sha384_update()
197 ATTRIBUTE_WARN_UNUSED_RET int _sha512_update(hash_context * hctx, const unsigned char *chunk, u32 c… in _sha512_update() argument
199 return sha512_update((sha512_context*)hctx, chunk, chunklen); in _sha512_update()
211 ATTRIBUTE_WARN_UNUSED_RET int _sha512_224_update(hash_context * hctx, const unsigned char *chunk, u… in _sha512_224_update() argument
213 return sha512_224_update((sha512_224_context*)hctx, chunk, chunklen); in _sha512_224_update()
[all …]
/src/sys/contrib/openzfs/module/zfs/
H A Dzap_leaf.c31 * chunk number n means l_chunk[n], even though the header precedes it.
48 #define CHAIN_END 0xffff /* end of the chunk chain */
180 uint_t chunk = zap_leaf_phys(l)->l_hdr.lh_freelist; in zap_leaf_chunk_alloc() local
181 ASSERT3U(chunk, <, ZAP_LEAF_NUMCHUNKS(l)); in zap_leaf_chunk_alloc()
182 ASSERT3U(ZAP_LEAF_CHUNK(l, chunk).l_free.lf_type, ==, ZAP_CHUNK_FREE); in zap_leaf_chunk_alloc()
185 ZAP_LEAF_CHUNK(l, chunk).l_free.lf_next; in zap_leaf_chunk_alloc()
189 return (chunk); in zap_leaf_chunk_alloc()
193 zap_leaf_chunk_free(zap_leaf_t *l, uint16_t chunk) in zap_leaf_chunk_free() argument
195 struct zap_leaf_free *zlf = &ZAP_LEAF_CHUNK(l, chunk).l_free; in zap_leaf_chunk_free()
197 ASSERT3U(chunk, <, ZAP_LEAF_NUMCHUNKS(l)); in zap_leaf_chunk_free()
[all …]
/src/contrib/libdiff/lib/
H A Ddiff_main.c39 diff_chunk_type(const struct diff_chunk *chunk) in diff_chunk_type() argument
41 if (!chunk->left_count && !chunk->right_count) in diff_chunk_type()
43 if (!chunk->solved) in diff_chunk_type()
45 if (!chunk->right_count) in diff_chunk_type()
47 if (!chunk->left_count) in diff_chunk_type()
49 if (chunk->left_count != chunk->right_count) in diff_chunk_type()
234 const struct diff_chunk *chunk) in diff_state_add_solved_chunk() argument
242 … to solved chunks; make sure that adjacent chunks of same type are combined, and that a minus chunk in diff_state_add_solved_chunk()
243 * never directly follows a plus chunk. */ in diff_state_add_solved_chunk()
248 new_t = diff_chunk_type(chunk); in diff_state_add_solved_chunk()
[all …]
/src/sys/dev/mthca/
H A Dmthca_memfree.c47 * per chunk.
63 static void mthca_free_icm_pages(struct mthca_dev *dev, struct mthca_icm_chunk *chunk) in mthca_free_icm_pages() argument
67 if (chunk->nsg > 0) in mthca_free_icm_pages()
68 pci_unmap_sg(dev->pdev, chunk->mem, chunk->npages, in mthca_free_icm_pages()
71 for (i = 0; i < chunk->npages; ++i) in mthca_free_icm_pages()
72 __free_pages(sg_page(&chunk->mem[i]), in mthca_free_icm_pages()
73 get_order(chunk->mem[i].length)); in mthca_free_icm_pages()
76 static void mthca_free_icm_coherent(struct mthca_dev *dev, struct mthca_icm_chunk *chunk) in mthca_free_icm_coherent() argument
80 for (i = 0; i < chunk->npages; ++i) { in mthca_free_icm_coherent()
81 dma_free_coherent(&dev->pdev->dev, chunk->mem[i].length, in mthca_free_icm_coherent()
[all …]
/src/contrib/libdiff/test/
H A Dexpect004.diff4 +int Chunk_bounds_check(Chunk *chunk, size_t start, size_t n)
6 + if (chunk == NULL) return 0;
8 + return start <= chunk->length && n <= chunk->length - start;
11 void Chunk_copy(Chunk *src, size_t src_start, Chunk *dst, size_t dst_start, size_t n)
19 -int Chunk_bounds_check(Chunk *chunk, size_t start, size_t n)
21 - if (chunk == NULL) return 0;
23 - return start <= chunk->length && n <= chunk->length - start;
H A Dexpect104.diff4 +int Chunk_bounds_check(Chunk *chunk, size_t start, size_t n)
6 + if (chunk == NULL) return 0;
8 + return start <= chunk->length && n <= chunk->length - start;
11 void Chunk_copy(Chunk *src, size_t src_start, Chunk *dst, size_t dst_start, size_t n)
19 -int Chunk_bounds_check(Chunk *chunk, size_t start, size_t n)
21 - if (chunk == NULL) return 0;
23 - return start <= chunk->length && n <= chunk->length - start;
/src/crypto/openssl/engines/asm/
H A De_padlock-x86_64.pl47 $chunk="%rbx";
318 mov \$$PADLOCK_CHUNK,$chunk
321 cmp $chunk,$len
322 cmovc $len,$chunk # chunk=len>PADLOCK_CHUNK?PADLOCK_CHUNK:len
323 and $chunk,%rax # out_misaligned?chunk:0
324 mov $len,$chunk
326 and \$$PADLOCK_CHUNK-1,$chunk # chunk%=PADLOCK_CHUNK
329 cmovz %rax,$chunk # chunk=chunk?:PADLOCK_CHUNK
337 mov \$$PADLOCK_CHUNK,$chunk
339 cmovz $chunk,%rax
[all …]
/src/usr.sbin/makefs/zfs/
H A Dobjset.c80 struct objset_dnode_chunk *chunk; in objset_alloc() local
92 chunk = ecalloc(1, sizeof(*chunk)); in objset_alloc()
93 chunk->nextfree = 1; in objset_alloc()
94 STAILQ_INSERT_HEAD(&os->dnodechunks, chunk, next); in objset_alloc()
114 struct objset_dnode_chunk *chunk, *tmp; in _objset_write() local
122 STAILQ_FOREACH_SAFE(chunk, &os->dnodechunks, next, tmp) { in _objset_write()
125 assert(chunk->nextfree > 0); in _objset_write()
126 assert(chunk->nextfree <= os->dnodecount); in _objset_write()
127 assert(chunk->nextfree <= DNODES_PER_CHUNK); in _objset_write()
129 for (i = 0; i < chunk->nextfree; i += DNODES_PER_BLOCK) { in _objset_write()
[all …]
/src/contrib/llvm-project/compiler-rt/lib/scudo/standalone/
H A Dcombined.h14 #include "chunk.h"
69 // Chunk recycling function, returns a quarantined chunk to the backend,
72 Chunk::UnpackedHeader Header; in recycle()
73 Chunk::loadHeader(Allocator.Cookie, Ptr, &Header); in recycle()
74 if (UNLIKELY(Header.State != Chunk::State::Quarantined)) in recycle()
77 Header.State = Chunk::State::Available; in recycle()
78 Chunk::storeHeader(Allocator.Cookie, Ptr, &Header); in recycle()
91 sizeof(QuarantineBatch) + Chunk::getHeaderSize()); in allocate()
98 Chunk::getHeaderSize()); in allocate()
99 Chunk::UnpackedHeader Header = {}; in allocate()
[all …]
H A Dwrappers_cpp.cpp39 void *Ptr = Allocator.allocate(size, scudo::Chunk::Origin::New); in operator new()
44 void *Ptr = Allocator.allocate(size, scudo::Chunk::Origin::NewArray); in operator new[]()
50 void *Ptr = Allocator.allocate(size, scudo::Chunk::Origin::New); in operator new()
56 void *Ptr = Allocator.allocate(size, scudo::Chunk::Origin::NewArray); in operator new[]()
61 void *Ptr = Allocator.allocate(size, scudo::Chunk::Origin::New, in operator new()
67 void *Ptr = Allocator.allocate(size, scudo::Chunk::Origin::NewArray, in operator new[]()
74 void *Ptr = Allocator.allocate(size, scudo::Chunk::Origin::New, in operator new()
81 void *Ptr = Allocator.allocate(size, scudo::Chunk::Origin::NewArray, in operator new[]()
89 Allocator.deallocate(ptr, scudo::Chunk::Origin::New); in operator delete()
93 Allocator.deallocate(ptr, scudo::Chunk::Origin::NewArray); in operator delete[]()
[all …]
/src/sys/dev/irdma/
H A Dirdma_pble.c51 struct irdma_chunk *chunk; in irdma_destroy_pble_prm() local
55 chunk = (struct irdma_chunk *)(&pinfo->clist)->next; in irdma_destroy_pble_prm()
56 list_del(&chunk->list); in irdma_destroy_pble_prm()
57 if (chunk->type == PBLE_SD_PAGED) in irdma_destroy_pble_prm()
58 irdma_pble_free_paged_mem(chunk); in irdma_destroy_pble_prm()
59 bitmap_free(chunk->bitmapbuf); in irdma_destroy_pble_prm()
60 kfree(chunk->chunkmem.va); in irdma_destroy_pble_prm()
128 struct irdma_chunk *chunk = info->chunk; in add_sd_direct() local
141 chunk->type = PBLE_SD_CONTIGOUS; in add_sd_direct()
145 chunk->size = info->pages << HMC_PAGED_BP_SHIFT; in add_sd_direct()
[all …]
/src/contrib/llvm-project/openmp/runtime/src/
H A Dkmp_sched.cpp79 typename traits_t<T>::signed_t chunk in __kmp_for_static_init() argument
141 " %%%s, %%%s) incr=%%%s chunk=%%%s signed?<%s>\n", in __kmp_for_static_init()
145 *pstride, incr, chunk)); in __kmp_for_static_init()
364 KMP_DEBUG_ASSERT(chunk != 0); in __kmp_for_static_init()
365 if (chunk < 1) in __kmp_for_static_init()
366 chunk = 1; in __kmp_for_static_init()
367 else if ((UT)chunk > trip_count) in __kmp_for_static_init()
368 chunk = trip_count; in __kmp_for_static_init()
369 nchunks = (trip_count) / (UT)chunk + (trip_count % (UT)chunk ? 1 : 0); in __kmp_for_static_init()
370 span = chunk * incr; in __kmp_for_static_init()
[all …]
/src/crypto/openssl/providers/implementations/ciphers/
H A Dcipher_des_hw.c110 size_t chunk = MAXCHUNK; in cipher_hw_des_cfb64_cipher() local
114 if (len < chunk) in cipher_hw_des_cfb64_cipher()
115 chunk = len; in cipher_hw_des_cfb64_cipher()
116 while (len > 0 && len >= chunk) { in cipher_hw_des_cfb64_cipher()
117 DES_cfb64_encrypt(in, out, (long)chunk, key, (DES_cblock *)ctx->iv, in cipher_hw_des_cfb64_cipher()
119 len -= chunk; in cipher_hw_des_cfb64_cipher()
120 in += chunk; in cipher_hw_des_cfb64_cipher()
121 out += chunk; in cipher_hw_des_cfb64_cipher()
122 if (len < chunk) in cipher_hw_des_cfb64_cipher()
123 chunk = len; in cipher_hw_des_cfb64_cipher()
[all …]
H A Dciphercommon_hw.c148 size_t chunk = MAXCHUNK; in ossl_cipher_hw_chunked_cfb8() local
150 if (inl < chunk) in ossl_cipher_hw_chunked_cfb8()
151 chunk = inl; in ossl_cipher_hw_chunked_cfb8()
152 while (inl > 0 && inl >= chunk) { in ossl_cipher_hw_chunked_cfb8()
154 inl -= chunk; in ossl_cipher_hw_chunked_cfb8()
155 in += chunk; in ossl_cipher_hw_chunked_cfb8()
156 out += chunk; in ossl_cipher_hw_chunked_cfb8()
157 if (inl < chunk) in ossl_cipher_hw_chunked_cfb8()
158 chunk = inl; in ossl_cipher_hw_chunked_cfb8()
166 size_t chunk = MAXCHUNK; in ossl_cipher_hw_chunked_cfb128() local
[all …]
/src/contrib/llvm-project/llvm/lib/Target/AArch64/
H A DAArch64ExpandImm.cpp20 /// Helper function which extracts the specified 16-bit chunk from a
23 assert(ChunkIdx < 4 && "Out of range chunk index specified!"); in getChunk()
28 /// Check whether the given 16-bit chunk replicated to full 64-bit width
30 static bool canUseOrr(uint64_t Chunk, uint64_t &Encoding) { in canUseOrr() argument
31 Chunk = (Chunk << 48) | (Chunk << 32) | (Chunk << 16) | Chunk; in canUseOrr()
33 return AArch64_AM::processLogicalImmediate(Chunk, 64, Encoding); in canUseOrr()
49 // Scan the constant and count how often every chunk occurs. in tryToreplicateChunks()
54 for (const auto &Chunk : Counts) { in tryToreplicateChunks() local
55 const uint64_t ChunkVal = Chunk.first; in tryToreplicateChunks()
56 const unsigned Count = Chunk.second; in tryToreplicateChunks()
[all …]
/src/contrib/llvm-project/llvm/lib/Support/BLAKE3/
H A Dblake3.c82 // Chaining values within a given chunk (specifically the compress_in_place
159 // Given some input larger than one chunk, return the number of bytes that
170 // on a single thread. Write out the chunk chaining values and return the
195 // Hash the remaining partial chunk, if there is one. Note that the empty in compress_chunks_parallel()
196 // chunk (meaning the empty message) is a different codepath. in compress_chunks_parallel()
263 // not used when the whole input is only 1 chunk long; that's a different
274 // Note that the single chunk case does *not* bump the SIMD degree up to 2 in blake3_compress_subtree_wide()
276 // this gives us the option of multi-threading even the 2-chunk case, which in blake3_compress_subtree_wide()
300 // sure there are two outputs. Except, as noted above, at the chunk in blake3_compress_subtree_wide()
301 // level, where we allow degree=1. (Note that the 1-chunk-input case is in blake3_compress_subtree_wide()
[all …]
/src/sys/contrib/openzfs/module/icp/algs/blake3/
H A Dblake3.c127 * Chaining values within a given chunk (specifically the compress_in_place
212 * Given some input larger than one chunk, return the number of bytes that
229 * on a single thread. Write out the chunk chaining values and return the
251 * Hash the remaining partial chunk, if there is one. Note that the in compress_chunks_parallel()
252 * empty chunk (meaning the empty message) is a different codepath. in compress_chunks_parallel()
316 * not used when the whole input is only 1 chunk long; that's a different
328 * Note that the single chunk case does *not* bump the SIMD degree up in blake3_compress_subtree_wide()
331 * 2-chunk case, which can help performance on smaller platforms. in blake3_compress_subtree_wide()
364 * at the chunk level, where we allow degree=1. (Note that the in blake3_compress_subtree_wide()
365 * 1-chunk-input case is a different codepath.) in blake3_compress_subtree_wide()
[all …]
/src/lib/libc/aarch64/string/
H A Dmemccpy.S42 /* match in first chunk */
57 ldr q3, [x10, #16] // load second string chunk
59 cmeq v1.16b, v3.16b, v0.16b // char found in second chunk?
61 /* process second chunk */
67 /* match in second chunk */
82 /* string didn't end in second chunk and neither did buffer */
83 ldr q1, [x10, #32] // load next string chunk
87 str q3, [x0, #16] // deposit second chunk
97 cmeq v2.16b, v1.16b, v0.16b // char found in second chunk?
104 ldr q1, [x10, #16] // load next chunk
[all …]

12345678910>>...40