| /linux/fs/ext4/ |
| H A D | indirect.c | 252 Indirect *partial) in ext4_find_goal() argument 260 goal = ext4_find_near(inode, partial); in ext4_find_goal() 538 Indirect *partial; in ext4_ind_map_blocks() local 554 partial = ext4_get_branch(inode, depth, offsets, chain, &err); in ext4_ind_map_blocks() 557 if (!partial) { in ext4_ind_map_blocks() 586 for (i = partial - chain + 1; i < depth; i++) in ext4_ind_map_blocks() 620 ar.goal = ext4_find_goal(inode, map->m_lblk, partial); in ext4_ind_map_blocks() 623 indirect_blks = (chain + depth) - partial - 1; in ext4_ind_map_blocks() 629 ar.len = ext4_blks_to_allocate(partial, indirect_blks, in ext4_ind_map_blocks() 636 offsets + (partial - chain), partial); in ext4_ind_map_blocks() [all …]
|
| H A D | extents.c | 2480 struct partial_cluster *partial, in ext4_remove_blocks() argument 2511 trace_ext4_remove_blocks(inode, ex, from, to, partial); in ext4_remove_blocks() 2519 if (partial->state != initial && in ext4_remove_blocks() 2520 partial->pclu != EXT4_B2C(sbi, last_pblk)) { in ext4_remove_blocks() 2521 if (partial->state == tofree) { in ext4_remove_blocks() 2523 if (ext4_is_pending(inode, partial->lblk)) in ext4_remove_blocks() 2526 EXT4_C2B(sbi, partial->pclu), in ext4_remove_blocks() 2529 ext4_rereserve_cluster(inode, partial->lblk); in ext4_remove_blocks() 2531 partial->state = initial; in ext4_remove_blocks() 2548 (partial->state != nofree)) { in ext4_remove_blocks() [all …]
|
| /linux/fs/minix/ |
| H A D | itree_common.c | 158 Indirect *partial; in get_block() local 166 partial = get_branch(inode, depth, offsets, chain, &err); in get_block() 169 if (!partial) { in get_block() 173 partial = chain+depth-1; /* the whole chain */ in get_block() 180 while (partial > chain) { in get_block() 181 brelse(partial->bh); in get_block() 182 partial--; in get_block() 196 left = (chain + depth) - partial; in get_block() 197 err = alloc_branch(inode, left, offsets+(partial-chain), partial); in get_block() 201 if (splice_branch(inode, chain, partial, left) < 0) in get_block() [all …]
|
| /linux/fs/ext2/ |
| H A D | inode.c | 331 Indirect *partial) in ext2_find_goal() argument 346 return ext2_find_near(inode, partial); in ext2_find_goal() 632 Indirect *partial; in ext2_get_blocks() local 648 partial = ext2_get_branch(inode, depth, offsets, chain, &err); in ext2_get_blocks() 650 if (!partial) { in ext2_get_blocks() 666 partial = chain + depth - 1; in ext2_get_blocks() 696 if (err == -EAGAIN || !verify_chain(chain, partial)) { in ext2_get_blocks() 697 while (partial > chain) { in ext2_get_blocks() 698 brelse(partial->bh); in ext2_get_blocks() 699 partial--; in ext2_get_blocks() [all …]
|
| /linux/lib/crypto/ |
| H A D | sha1.c | 143 size_t partial = ctx->bytecount % SHA1_BLOCK_SIZE; in sha1_update() local 147 if (partial + len >= SHA1_BLOCK_SIZE) { in sha1_update() 150 if (partial) { in sha1_update() 151 size_t l = SHA1_BLOCK_SIZE - partial; in sha1_update() 153 memcpy(&ctx->buf[partial], data, l); in sha1_update() 167 partial = 0; in sha1_update() 170 memcpy(&ctx->buf[partial], data, len); in sha1_update() 177 size_t partial = ctx->bytecount % SHA1_BLOCK_SIZE; in __sha1_final() local 179 ctx->buf[partial++] = 0x80; in __sha1_final() 180 if (partial > SHA1_BLOCK_SIZE - 8) { in __sha1_final() [all …]
|
| H A D | md5.c | 146 size_t partial = ctx->bytecount % MD5_BLOCK_SIZE; in md5_update() local 150 if (partial + len >= MD5_BLOCK_SIZE) { in md5_update() 153 if (partial) { in md5_update() 154 size_t l = MD5_BLOCK_SIZE - partial; in md5_update() 156 memcpy(&ctx->buf[partial], data, l); in md5_update() 170 partial = 0; in md5_update() 173 memcpy(&ctx->buf[partial], data, len); in md5_update() 180 size_t partial = ctx->bytecount % MD5_BLOCK_SIZE; in __md5_final() local 182 ctx->buf[partial++] = 0x80; in __md5_final() 183 if (partial > MD5_BLOCK_SIZE - 8) { in __md5_final() [all …]
|
| H A D | sha512.c | 165 size_t partial = ctx->bytecount_lo % SHA512_BLOCK_SIZE; in __sha512_update() local 170 if (partial + len >= SHA512_BLOCK_SIZE) { in __sha512_update() 173 if (partial) { in __sha512_update() 174 size_t l = SHA512_BLOCK_SIZE - partial; in __sha512_update() 176 memcpy(&ctx->buf[partial], data, l); in __sha512_update() 190 partial = 0; in __sha512_update() 193 memcpy(&ctx->buf[partial], data, len); in __sha512_update() 202 size_t partial = ctx->bytecount_lo % SHA512_BLOCK_SIZE; in __sha512_final() local 204 ctx->buf[partial++] = 0x80; in __sha512_final() 205 if (partial > SHA512_BLOCK_SIZE - 16) { in __sha512_final() [all …]
|
| H A D | sha256.c | 187 size_t partial = ctx->bytecount % SHA256_BLOCK_SIZE; in __sha256_update() local 191 if (partial + len >= SHA256_BLOCK_SIZE) { in __sha256_update() 194 if (partial) { in __sha256_update() 195 size_t l = SHA256_BLOCK_SIZE - partial; in __sha256_update() 197 memcpy(&ctx->buf[partial], data, l); in __sha256_update() 211 partial = 0; in __sha256_update() 214 memcpy(&ctx->buf[partial], data, len); in __sha256_update() 222 size_t partial = ctx->bytecount % SHA256_BLOCK_SIZE; in __sha256_final() local 224 ctx->buf[partial++] = 0x80; in __sha256_final() 225 if (partial > SHA256_BLOCK_SIZE - 8) { in __sha256_final() [all …]
|
| H A D | polyval.c | 260 if (unlikely(ctx->partial)) { in polyval_update() 261 size_t n = min(len, POLYVAL_BLOCK_SIZE - ctx->partial); in polyval_update() 265 ctx->acc.bytes[ctx->partial++] ^= *data++; in polyval_update() 266 if (ctx->partial < POLYVAL_BLOCK_SIZE) in polyval_update() 279 ctx->partial = len; in polyval_update() 285 if (unlikely(ctx->partial)) in polyval_final()
|
| H A D | chacha20poly1305.c | 221 size_t partial = 0; in chacha20poly1305_crypt_sg_inplace() local 267 if (unlikely(partial)) { in chacha20poly1305_crypt_sg_inplace() 268 size_t l = min(length, CHACHA_BLOCK_SIZE - partial); in chacha20poly1305_crypt_sg_inplace() 270 crypto_xor(addr, b.chacha_stream + partial, l); in chacha20poly1305_crypt_sg_inplace() 271 partial = (partial + l) & (CHACHA_BLOCK_SIZE - 1); in chacha20poly1305_crypt_sg_inplace() 291 partial = length; in chacha20poly1305_crypt_sg_inplace()
|
| /linux/drivers/usb/storage/ |
| H A D | freecom.c | 227 unsigned int partial; in freecom_transport() local 266 FCM_STATUS_PACKET_LENGTH, &partial); in freecom_transport() 267 usb_stor_dbg(us, "foo Status result %d %u\n", result, partial); in freecom_transport() 271 US_DEBUG(pdump(us, (void *)fst, partial)); in freecom_transport() 308 FCM_STATUS_PACKET_LENGTH, &partial); in freecom_transport() 310 usb_stor_dbg(us, "bar Status result %d %u\n", result, partial); in freecom_transport() 314 US_DEBUG(pdump(us, (void *)fst, partial)); in freecom_transport() 317 if (partial != 4) in freecom_transport() 376 FCM_PACKET_LENGTH, &partial); in freecom_transport() 377 US_DEBUG(pdump(us, (void *)fst, partial)); in freecom_transport() [all …]
|
| H A D | initializers.c | 56 unsigned int partial; in usb_stor_ucr61s2b_init() local 70 US_BULK_CB_WRAP_LEN, &partial); in usb_stor_ucr61s2b_init() 76 US_BULK_CS_WRAP_LEN, &partial); in usb_stor_ucr61s2b_init()
|
| H A D | transport.c | 257 unsigned int length, int result, unsigned int partial) in interpret_urb_result() argument 260 result, partial, length); in interpret_urb_result() 265 if (partial != length) { in interpret_urb_result() 469 unsigned int partial; in usb_stor_bulk_srb() local 472 &partial); in usb_stor_bulk_srb() 474 scsi_set_resid(srb, scsi_bufflen(srb) - partial); in usb_stor_bulk_srb() 492 unsigned int partial; in usb_stor_bulk_transfer_sg() local 499 length_left, &partial); in usb_stor_bulk_transfer_sg() 500 length_left -= partial; in usb_stor_bulk_transfer_sg() 504 length_left, &partial); in usb_stor_bulk_transfer_sg() [all …]
|
| /linux/net/smc/ |
| H A D | smc_rx.c | 161 struct partial_page *partial; in smc_rx_splice() local 174 partial = kzalloc_objs(*partial, nr_pages); in smc_rx_splice() 175 if (!partial) in smc_rx_splice() 191 partial[0].offset = src - (char *)smc->conn.rmb_desc->cpu_addr; in smc_rx_splice() 192 partial[0].len = len; in smc_rx_splice() 193 partial[0].private = (unsigned long)priv[0]; in smc_rx_splice() 204 partial[i].offset = offset; in smc_rx_splice() 205 partial[i].len = size; in smc_rx_splice() 206 partial[i].private = (unsigned long)priv[i]; in smc_rx_splice() 215 spd.partial = partial; in smc_rx_splice() [all …]
|
| /linux/arch/x86/include/asm/ |
| H A D | unwind.h | 72 bool *partial) in unwind_get_entry_regs() argument 77 if (partial) { in unwind_get_entry_regs() 79 *partial = !state->full_regs; in unwind_get_entry_regs() 81 *partial = false; in unwind_get_entry_regs() 89 bool *partial) in unwind_get_entry_regs() argument
|
| H A D | unwind_hints.h | 23 .macro UNWIND_HINT_REGS base=%rsp offset=0 indirect=0 extra=1 partial=0 signal=1 44 .if \partial 57 UNWIND_HINT_REGS base=\base offset=\offset partial=1 signal=\signal
|
| /linux/arch/x86/kernel/ |
| H A D | dumpstack.c | 153 bool partial, const char *log_lvl) in show_regs_if_on_stack() argument 164 if (!partial && on_stack(info, regs, sizeof(*regs))) { in show_regs_if_on_stack() 167 } else if (partial && on_stack(info, (void *)regs + IRET_FRAME_OFFSET, in show_regs_if_on_stack() 191 bool partial = false; in __show_trace_log_lvl() local 197 regs = unwind_get_entry_regs(&state, &partial); in __show_trace_log_lvl() 237 show_regs_if_on_stack(&stack_info, regs, partial, log_lvl); in __show_trace_log_lvl() 296 regs = unwind_get_entry_regs(&state, &partial); in __show_trace_log_lvl() 298 show_regs_if_on_stack(&stack_info, regs, partial, log_lvl); in __show_trace_log_lvl()
|
| /linux/tools/mm/ |
| H A D | slabinfo.c | 36 unsigned long partial, objects, slabs, objects_partial, objects_total; member 579 s->slab_size, s->slabs - s->partial - s->cpu_slabs, in report() 582 page_size << s->order, s->partial, onoff(s->poison), in report() 627 s->partial, s->cpu_slabs); in slabcache() 670 s->slabs ? (s->partial * 100) / s->slabs : 100, in slabcache() 878 percentage_partial_slabs = s->partial * 100 / s->slabs; in totals() 890 if (s->partial < min_partial) in totals() 891 min_partial = s->partial; in totals() 915 if (s->partial > max_partial) in totals() 916 max_partial = s->partial; in totals() [all …]
|
| /linux/Documentation/ABI/testing/ |
| H A D | sysfs-kernel-slab | 96 allocation from a partial or new slab. It can be written to 179 The deactivate_to_head file shows how many times a partial cpu 180 slab was deactivated and added to the head of its node's partial 190 The deactivate_to_tail file shows how many times a partial cpu 191 slab was deactivated and added to the tail of its node's partial 212 partial list. It can be written to clear the current count. 255 its node's partial list. It can be written to clear the current 277 using the slow path (i.e. to a full or partial slab). It can 297 remain on a node's partial list to avoid the overhead of 326 objects are on partial slabs and from which nodes they are [all …]
|
| /linux/arch/powerpc/crypto/ |
| H A D | aes-gcm-p10.S | 383 # Process a normal partial block when we come here. 384 # Compute partial mask, Load and store partial block to stack. 392 # create partial mask 398 vslo 16, 16, 17 # partial block mask 400 lxvb16x 11, 0, 14 # load partial block 403 # AES crypt partial 422 # Write partial 427 mtctr 5 # partial block len 442 std 5, 56(7) # update partial 500 lxvb16x 32+19, 0, 14 # load partial block [all …]
|
| /linux/tools/testing/selftests/syscall_user_dispatch/ |
| H A D | sud_benchmark.c | 97 double partial = 0; in perf_syscall() local 100 partial += one_sysinfo_step()/(CALIBRATION_STEP*factor); in perf_syscall() 101 return partial; in perf_syscall()
|
| /linux/crypto/ |
| H A D | md5.c | 28 unsigned int partial; in __crypto_md5_export() local 31 partial = ctx.bytecount % MD5_BLOCK_SIZE; in __crypto_md5_export() 32 ctx.bytecount -= partial; in __crypto_md5_export() 35 *p = partial; in __crypto_md5_export()
|
| H A D | sha1.c | 31 unsigned int partial; in __crypto_sha1_export() local 34 partial = ctx.bytecount % SHA1_BLOCK_SIZE; in __crypto_sha1_export() 35 ctx.bytecount -= partial; in __crypto_sha1_export() 38 *p = partial; in __crypto_sha1_export()
|
| /linux/drivers/iommu/ |
| H A D | io-pgfault.c | 75 list_add(&iopf->list, &fault_param->partial); in report_partial_fault() 105 list_for_each_entry_safe(iopf, next, &iopf_param->partial, list) { in iopf_group_alloc() 366 list_for_each_entry_safe(iopf, next, &iopf_param->partial, in iopf_queue_discard_partial() 411 INIT_LIST_HEAD(&fault_param->partial); in iopf_queue_add_device() 468 list_for_each_entry_safe(partial_iopf, next, &fault_param->partial, list) in iopf_queue_remove_device()
|
| /linux/drivers/comedi/drivers/ni_routing/tools/ |
| H A D | Makefile | 69 clean-partial : 73 clean : clean-partial
|