| /linux/arch/arm64/crypto/ |
| H A D | sha3-ce-core.S | 15 .irp b,0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31 16 .set .Lv\b\().2d, \b 17 .set .Lv\b\().16b, \b 59 ld1 {v25.8b-v28.8b}, [x1], #32 60 ld1 {v29.8b-v31.8b}, [x1], #24 61 eor v0.8b, v0.8b, v25.8b 62 eor v1.8b, v1.8b, v26.8b 63 eor v2.8b, v2.8b, v27.8b 64 eor v3.8b, v3.8b, v28.8b 65 eor v4.8b, v4.8b, v29.8b [all …]
|
| H A D | aes-neon.S | 45 movi v12.16b, #0x1b 49 ld1 {v16.16b-v19.16b}, [\temp], #64 50 ld1 {v20.16b-v23.16b}, [\temp], #64 51 ld1 {v24.16b-v27.16b}, [\temp], #64 52 ld1 {v28.16b-v31.16b}, [\temp] 71 sub v9.16b, \in\().16b, v15.16b 72 tbl \in\().16b, {v16.16b-v19.16b}, \in\().16b 73 sub v10.16b, v9.16b, v15.16b 74 tbx \in\().16b, {v20.16b-v23.16b}, v9.16b 75 sub v11.16b, v10.16b, v15.16b [all …]
|
| H A D | aes-neonbs-core.S | 192 in_bs_ch \b0\().16b, \b1\().16b, \b2\().16b, \b3\().16b, \ 193 \b4\().16b, \b5\().16b, \b6\().16b, \b7\().16b 194 inv_gf256 \b6\().16b, \b5\().16b, \b0\().16b, \b3\().16b, \ 195 \b7\().16b, \b1\().16b, \b4\().16b, \b2\().16b, \ 196 \t0\().16b, \t1\().16b, \t2\().16b, \t3\().16b, \ 197 \s0\().16b, \s1\().16b, \s2\().16b, \s3\().16b 198 out_bs_ch \b7\().16b, \b1\().16b, \b4\().16b, \b2\().16b, \ 199 \b6\().16b, \b5\().16b, \b0\().16b, \b3\().16b 204 inv_in_bs_ch \b0\().16b, \b1\().16b, \b2\().16b, \b3\().16b, \ 205 \b4\().16b, \b5\().16b, \b6\().16b, \b7\().16b [all …]
|
| H A D | sm4-ce-core.S | 17 .irp b, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, \ 19 .set .Lv\b\().4s, \b 51 ld1 {v0.16b}, [x0]; 52 rev32 v0.16b, v0.16b; 53 ld1 {v1.16b}, [x3]; 55 ld1 {v24.16b-v27.16b}, [x4], #64; 56 ld1 {v28.16b-v31.16b}, [x4]; 59 eor v0.16b, v0.16b, v1.16b; 71 ld1 {v24.16b}, [x5] 73 st1 {v0.16b-v3.16b}, [x1], #64; [all …]
|
| H A D | ghash-ce-core.S | 72 ext t3.8b, \ad\().8b, \ad\().8b, #1 // A1 73 ext t5.8b, \ad\().8b, \ad\().8b, #2 // A2 74 ext t7.8b, \ad\().8b, \ad\().8b, #3 // A3 80 tbl t3.16b, {\ad\().16b}, perm1.16b // A1 81 tbl t5.16b, {\ad\().16b}, perm2.16b // A2 82 tbl t7.16b, {\ad\().16b}, perm3.16b // A3 88 __pmull_p8_tail \rq, \ad\().8b, SHASH.8b, 8b,, sh1, sh2, sh3, sh4 92 __pmull_p8_tail \rq, \ad\().8b, SHASH2.8b, 8b,, ss1, ss2, ss3, ss4 96 __pmull_p8_tail \rq, \ad\().16b, SHASH.16b, 16b, 2, sh1, sh2, sh3, sh4 109 eor t3.16b, t3.16b, t4.16b // L = E + F [all …]
|
| H A D | sm4-ce-gcm-core.S | 18 .irp b, 0, 1, 2, 3, 24, 25, 26, 27, 28, 29, 30, 31 19 .set .Lv\b\().4s, \b 40 ext T0.16b, m1.16b, m1.16b, #8; \ 45 eor T0.16b, T0.16b, T1.16b; \ 46 ext T1.16b, RZERO.16b, T0.16b, #8; \ 47 ext T0.16b, T0.16b, RZERO.16b, #8; \ 48 eor r0.16b, r0.16b, T1.16b; \ 49 eor r1.16b, r1.16b, T0.16b; 55 ext T0.16b, m1.16b, m1.16b, #8; \ 56 ext T2.16b, m3.16b, m3.16b, #8; \ [all …]
|
| H A D | sm4-neon-core.S | 35 ld1 {v16.16b-v19.16b}, [x5], #64; \ 36 ld1 {v20.16b-v23.16b}, [x5], #64; \ 37 ld1 {v24.16b-v27.16b}, [x5], #64; \ 38 ld1 {v28.16b-v31.16b}, [x5]; 99 eor RTMP1.16b, s2.16b, s3.16b; \ 100 eor RX0.16b, RX0.16b, s1.16b; \ 101 eor RX0.16b, RX0.16b, RTMP1.16b; \ 104 movi RTMP3.16b, #64; /* sizeof(sbox) / 4 */ \ 105 tbl RTMP0.16b, {v16.16b-v19.16b}, RX0.16b; \ 106 sub RX0.16b, RX0.16b, RTMP3.16b; \ [all …]
|
| H A D | aes-modes.S | 62 ld1 {v0.16b-v3.16b}, [x1], #64 /* get 4 pt blocks */ 64 ST5( ld1 {v4.16b}, [x1], #16 ) 66 st1 {v0.16b-v3.16b}, [x0], #64 67 ST5( st1 {v4.16b}, [x0], #16 ) 68 b .LecbencloopNx 73 ld1 {v0.16b}, [x1], #16 /* get next pt block */ 75 st1 {v0.16b}, [x0], #16 92 ld1 {v0.16b-v3.16b}, [x1], #64 /* get 4 ct blocks */ 94 ST5( ld1 {v4.16b}, [x1], #16 ) 96 st1 {v0.16b-v3.16b}, [x0], #64 [all …]
|
| H A D | sm4-ce-ccm-core.S | 17 .irp b, 0, 1, 8, 9, 10, 11, 12, 13, 14, 15, 16, 24, 25, 26, 27, 28, 29, 30, 31 18 .set .Lv\b\().4s, \b 35 rev64 vctr.16b, vctr.16b; \ 49 ld1 {RMAC.16b}, [x1] 57 ld1 {v0.16b-v3.16b}, [x2], #64 60 eor RMAC.16b, RMAC.16b, v0.16b 62 eor RMAC.16b, RMAC.16b, v1.16b 64 eor RMAC.16b, RMAC.16b, v2.16b 66 eor RMAC.16b, RMAC.16b, v3.16b 69 b .Lcbcmac_loop_4x [all …]
|
| H A D | polyval-ce-core.S | 96 ext v25.16b, X.16b, X.16b, #8 97 ext v26.16b, Y.16b, Y.16b, #8 98 eor v25.16b, v25.16b, X.16b 99 eor v26.16b, v26.16b, Y.16b 103 eor HI.16b, HI.16b, v28.16b 104 eor LO.16b, LO.16b, v29.16b 105 eor MI.16b, MI.16b, v27.16b 117 ext v25.16b, X.16b, X.16b, #8 118 ext v26.16b, Y.16b, Y.16b, #8 119 eor v25.16b, v25.16b, X.16b [all …]
|
| H A D | aes-ce-core.S | 13 ld1 {v0.16b}, [x2] 18 mov v3.16b, v1.16b 19 b 2f 20 0: mov v2.16b, v1.16b 22 1: aese v0.16b, v2.16b 23 aesmc v0.16b, v0.16b 25 aese v0.16b, v3.16b 26 aesmc v0.16b, v0.16b 29 aese v0.16b, v1.16b 30 aesmc v0.16b, v0.16b [all …]
|
| H A D | aes-ce-ccm-core.S | 27 aese \va\().16b, \vk\().16b 28 aesmc \va\().16b, \va\().16b 29 aese \vb\().16b, \vk\().16b 30 aesmc \vb\().16b, \vb\().16b 43 aese \va\().16b, v4.16b 44 aese \vb\().16b, v4.16b 50 ld1 {v0.16b}, [x5] /* load mac */ 55 ld1 {v1.8b}, [x6] /* load upper ctr */ 65 ld1 {v2.16b}, [x1], #16 /* load next input block */ 67 eor v2.16b, v2.16b, v5.16b /* final round enc+mac */ [all …]
|
| /linux/arch/x86/lib/ |
| H A D | usercopy_32.c | 45 _ASM_EXTABLE_TYPE_REG(0b, 2b, EX_TYPE_UCOPY_LEN4, %2) \ 46 _ASM_EXTABLE_UA(1b, 2b) \ 147 _ASM_EXTABLE_UA(1b, 100b) in __copy_user_intel() 148 _ASM_EXTABLE_UA(2b, 100b) in __copy_user_intel() 149 _ASM_EXTABLE_UA(3b, 100b) in __copy_user_intel() 150 _ASM_EXTABLE_UA(4b, 100b) in __copy_user_intel() 151 _ASM_EXTABLE_UA(5b, 100b) in __copy_user_intel() 152 _ASM_EXTABLE_UA(6b, 100b) in __copy_user_intel() 153 _ASM_EXTABLE_UA(7b, 100b) in __copy_user_intel() 154 _ASM_EXTABLE_UA(8b, 100b) in __copy_user_intel() [all …]
|
| /linux/drivers/iommu/ |
| H A D | msm_iommu_hw-8xxx.h | 20 #define SET_GLOBAL_REG_N(b, n, r, v) SET_GLOBAL_REG(b, ((r) + (n << 2)), (v)) argument 21 #define GET_GLOBAL_REG_N(b, n, r) GET_GLOBAL_REG(b, ((r) + (n << 2))) argument 24 #define GET_GLOBAL_FIELD(b, r, F) GET_FIELD(((b) + (r)), F##_MASK, F##_SHIFT) argument 25 #define GET_CONTEXT_FIELD(b, c, r, F) \ argument 26 GET_FIELD(((b) + (r) + ((c) << CTX_SHIFT)), F##_MASK, F##_SHIFT) 28 #define SET_GLOBAL_FIELD(b, r, F, v) \ argument 29 SET_FIELD(((b) + (r)), F##_MASK, F##_SHIFT, (v)) 30 #define SET_CONTEXT_FIELD(b, c, r, F, v) \ argument 31 SET_FIELD(((b) + (r) + ((c) << CTX_SHIFT)), F##_MASK, F##_SHIFT, (v)) 84 #define SET_M2VCBR_N(b, N, v) SET_GLOBAL_REG_N(M2VCBR_N, N, (b), (v)) argument [all …]
|
| /linux/lib/crc/arm64/ |
| H A D | crc-t10dif-core.S | 149 ext t7.16b, \b64\().16b, \b64\().16b, #1 150 tbl t5.16b, {\a16\().16b}, perm.16b 151 uzp1 t7.16b, \b64\().16b, t7.16b 153 ext \b64\().16b, t4.16b, t4.16b, #15 154 eor \c64\().16b, t8.16b, t5.16b 158 ext t6.16b, t5.16b, t5.16b, #8 160 pmull t3.8h, t7.8b, t5.8b 161 pmull t4.8h, t7.8b, t6.8b 162 pmull2 t5.8h, t7.16b, t5.16b 163 pmull2 t6.8h, t7.16b, t6.16b [all …]
|
| /linux/lib/crypto/arm64/ |
| H A D | chacha-neon-core.S | 47 eor v3.16b, v3.16b, v0.16b 52 eor v4.16b, v1.16b, v2.16b 58 eor v3.16b, v3.16b, v0.16b 59 tbl v3.16b, {v3.16b}, v12.16b 63 eor v4.16b, v1.16b, v2.16b 68 ext v1.16b, v1.16b, v1.16b, #4 70 ext v2.16b, v2.16b, v2.16b, #8 72 ext v3.16b, v3.16b, v3.16b, #12 76 eor v3.16b, v3.16b, v0.16b 81 eor v4.16b, v1.16b, v2.16b [all …]
|
| /linux/drivers/md/bcache/ |
| H A D | btree.c | 103 #define insert_lock(s, b) ((b)->level <= (s)->lock) argument 106 static inline struct bset *write_block(struct btree *b) in write_block() argument 108 return ((void *) btree_bset_first(b)) + b->written * block_bytes(b->c->cache); in write_block() 111 static void bch_btree_init_next(struct btree *b) in bch_btree_init_next() argument 114 if (b->level && b->keys.nsets) in bch_btree_init_next() 115 bch_btree_sort(&b->keys, &b->c->sort); in bch_btree_init_next() 117 bch_btree_sort_lazy(&b->keys, &b->c->sort); in bch_btree_init_next() 119 if (b->written < btree_blocks(b)) in bch_btree_init_next() 120 bch_bset_init_next(&b->keys, write_block(b), in bch_btree_init_next() 121 bset_magic(&b->c->cache->sb)); in bch_btree_init_next() [all …]
|
| /linux/drivers/gpu/drm/i915/gt/ |
| H A D | intel_breadcrumbs.c | 19 static bool irq_enable(struct intel_breadcrumbs *b) in irq_enable() argument 21 return intel_engine_irq_enable(b->irq_engine); in irq_enable() 24 static void irq_disable(struct intel_breadcrumbs *b) in irq_disable() argument 26 intel_engine_irq_disable(b->irq_engine); in irq_disable() 29 static void __intel_breadcrumbs_arm_irq(struct intel_breadcrumbs *b) in __intel_breadcrumbs_arm_irq() argument 37 wakeref = intel_gt_pm_get_if_awake(b->irq_engine->gt); in __intel_breadcrumbs_arm_irq() 47 WRITE_ONCE(b->irq_armed, wakeref); in __intel_breadcrumbs_arm_irq() 50 if (!b->irq_enabled++ && b->irq_enable(b)) in __intel_breadcrumbs_arm_irq() 51 irq_work_queue(&b->irq_work); in __intel_breadcrumbs_arm_irq() 54 static void intel_breadcrumbs_arm_irq(struct intel_breadcrumbs *b) in intel_breadcrumbs_arm_irq() argument [all …]
|
| /linux/Documentation/driver-api/media/drivers/ccs/ |
| H A D | ccs-regs.asc | 8 # - b bool bit 127 - b horizontal_mirror 0 128 - b vertical_flip 1 141 - b enable 0 142 - b ack 1 159 - b raw8_packing_for_raw16 0 160 - b raw10_packing_for_raw20 1 161 - b raw12_packing_for_raw24 2 166 - b enable 0 187 - b enabled 0 [all …]
|
| /linux/net/ceph/crush/ |
| H A D | crush.c | 27 int crush_get_bucket_item_weight(const struct crush_bucket *b, int p) in crush_get_bucket_item_weight() argument 29 if ((__u32)p >= b->size) in crush_get_bucket_item_weight() 32 switch (b->alg) { in crush_get_bucket_item_weight() 34 return ((struct crush_bucket_uniform *)b)->item_weight; in crush_get_bucket_item_weight() 36 return ((struct crush_bucket_list *)b)->item_weights[p]; in crush_get_bucket_item_weight() 38 return ((struct crush_bucket_tree *)b)->node_weights[crush_calc_tree_node(p)]; in crush_get_bucket_item_weight() 40 return ((struct crush_bucket_straw *)b)->item_weights[p]; in crush_get_bucket_item_weight() 42 return ((struct crush_bucket_straw2 *)b)->item_weights[p]; in crush_get_bucket_item_weight() 47 void crush_destroy_bucket_uniform(struct crush_bucket_uniform *b) in crush_destroy_bucket_uniform() argument 49 kfree(b->h.items); in crush_destroy_bucket_uniform() [all …]
|
| /linux/drivers/misc/ |
| H A D | vmw_balloon.c | 406 static inline void vmballoon_stats_op_inc(struct vmballoon *b, unsigned int op, in vmballoon_stats_op_inc() argument 410 atomic64_inc(&b->stats->ops[op][type]); in vmballoon_stats_op_inc() 413 static inline void vmballoon_stats_gen_inc(struct vmballoon *b, in vmballoon_stats_gen_inc() argument 417 atomic64_inc(&b->stats->general_stat[stat]); in vmballoon_stats_gen_inc() 420 static inline void vmballoon_stats_gen_add(struct vmballoon *b, in vmballoon_stats_gen_add() argument 425 atomic64_add(val, &b->stats->general_stat[stat]); in vmballoon_stats_gen_add() 428 static inline void vmballoon_stats_page_inc(struct vmballoon *b, in vmballoon_stats_page_inc() argument 433 atomic64_inc(&b->stats->page_stat[stat][size]); in vmballoon_stats_page_inc() 436 static inline void vmballoon_stats_page_add(struct vmballoon *b, in vmballoon_stats_page_add() argument 442 atomic64_add(val, &b->stats->page_stat[stat][size]); in vmballoon_stats_page_add() [all …]
|
| /linux/tools/perf/tests/ |
| H A D | time-utils-test.c | 159 u64 b = 1234567123456789ULL; in test__time_utils() local 162 .ptime = { {b, b + 1}, }, in test__time_utils() 164 .skip = { b - 1, b + 2, }, in test__time_utils() 165 .noskip = { b, b + 1, }, in test__time_utils() 172 u64 b = 1234567123456789ULL; in test__time_utils() local 179 .ptime = { {b, b + 1}, {c, c + 123}, {e, e + 5}, }, in test__time_utils() 181 .skip = { b - 1, b + 2, c - 1, c + 124, e - 1, e + 6 }, in test__time_utils() 182 .noskip = { b, b + 1, c, c + 123, e, e + 5 }, in test__time_utils() 189 u64 b = 7654321ULL * NSEC_PER_SEC; in test__time_utils() local 192 .first = b, in test__time_utils() [all …]
|
| /linux/drivers/block/drbd/ |
| H A D | drbd_bitmap.c | 115 struct drbd_bitmap *b = device->bitmap; in __bm_print_lock_info() local 120 func, b->bm_why ?: "?", in __bm_print_lock_info() 121 b->bm_task->comm, task_pid_nr(b->bm_task)); in __bm_print_lock_info() 126 struct drbd_bitmap *b = device->bitmap; in drbd_bm_lock() local 129 if (!b) { in drbd_bm_lock() 134 trylock_failed = !mutex_trylock(&b->bm_change); in drbd_bm_lock() 139 why, b->bm_why ?: "?", in drbd_bm_lock() 140 b->bm_task->comm, task_pid_nr(b->bm_task)); in drbd_bm_lock() 141 mutex_lock(&b->bm_change); in drbd_bm_lock() 143 if (BM_LOCKED_MASK & b->bm_flags) in drbd_bm_lock() [all …]
|
| /linux/arch/arm64/kernel/vdso/ |
| H A D | vgetrandom-chacha.S | 61 mov state0.16b, copy0.16b 62 mov state1.16b, copy1.16b 63 mov state2.16b, copy2.16b 64 mov state3.16b, copy3.16b 77 eor state3.16b, state3.16b, state0.16b 82 eor tmp.16b, state1.16b, state2.16b 88 eor tmp.16b, state3.16b, state0.16b 94 eor tmp.16b, state1.16b, state2.16b 99 ext state1.16b, state1.16b, state1.16b, #4 101 ext state2.16b, state2.16b, state2.16b, #8 [all …]
|
| /linux/drivers/media/platform/samsung/s5p-mfc/ |
| H A D | s5p_mfc_opr.c | 35 struct s5p_mfc_priv_buf *b) in s5p_mfc_alloc_priv_buf() argument 38 unsigned int count = b->size >> PAGE_SHIFT; in s5p_mfc_alloc_priv_buf() 42 mfc_debug(3, "Allocating priv: %zu\n", b->size); in s5p_mfc_alloc_priv_buf() 51 b->virt = dev->mem_virt + offset; in s5p_mfc_alloc_priv_buf() 52 b->dma = dev->mem_base + offset; in s5p_mfc_alloc_priv_buf() 57 b->ctx = mem_ctx; in s5p_mfc_alloc_priv_buf() 58 b->virt = dma_alloc_coherent(mem_dev, b->size, &b->dma, GFP_KERNEL); in s5p_mfc_alloc_priv_buf() 59 if (!b->virt) in s5p_mfc_alloc_priv_buf() 61 if (b->dma < base) { in s5p_mfc_alloc_priv_buf() 63 &b->dma, &base); in s5p_mfc_alloc_priv_buf() [all …]
|