Home
last modified time | relevance | path

Searched full:s0 (Results 1 – 25 of 374) sorted by relevance

12345678910>>...15

/linux-6.8/arch/powerpc/crypto/
Dsha256-spe-asm.S114 rotrwi rT0,a,2; /* 1: S0 = a rotr 2 */ \
116 rotrwi rT1,a,13; /* 1: S0' = a rotr 13 */ \
118 rotrwi rT3,a,22; /* 1: S0" = a rotr 22 */ \
119 xor rT0,rT0,rT1; /* 1: S0 = S0 xor S0' */ \
121 xor rT3,rT0,rT3; /* 1: S0 = S0 xor S0" */ \
129 add rT3,rT3,rT2; /* 1: temp2 = S0 + maj */ \
141 rotrwi rT0,h,2; /* 2: S0 = a rotr 2 */ \
143 rotrwi rT1,h,13; /* 2: S0' = a rotr 13 */ \
145 rotrwi rT3,h,22; /* 2: S0" = a rotr 22 */ \
146 xor rT0,rT0,rT1; /* 2: S0 = S0 xor S0' */ \
[all …]
/linux-6.8/arch/loongarch/kernel/
Drelocate_kernel.S25 move s0, a4
31 beqz s0, done
34 PTR_L s1, s0, 0
35 PTR_ADDI s0, s0, SZREG
45 /* indirection page, update s0 */
49 and s0, s1, t0
97 iocsrrd.w s0, t1 /* check PC as an indicator */
98 beqz s0, 1b
99 iocsrrd.d s0, t1 /* get PC via mailbox */
102 or s0, s0, t0 /* s0 = TO_CACHE(s0) */
[all …]
Dmcount.S21 PTR_S s0, sp, MCOUNT_S0_OFFSET
23 move s0, a0
27 move a0, s0
29 PTR_L s0, sp, MCOUNT_S0_OFFSET
42 move a1, s0 /* arg1: parent's return address */
74 move a2, s0 /* arg2: Callsite parent ra */
/linux-6.8/arch/riscv/kernel/
Dmcount.S19 REG_S s0, 0*SZREG(sp)
21 addi s0, sp, 16
30 REG_S s0, 2*SZREG(sp)
34 addi s0, sp, 4*SZREG
39 REG_L s0, 0*SZREG(sp)
45 REG_L s0, 2*SZREG(sp)
67 * s0 (frame pointer, if enabled) on entry and the sp (stack pointer) on return.
71 * value stored in -16(s0) on entry, and the s0 on return.
106 addi a0, s0, -SZREG
109 REG_L a2, -2*SZREG(s0)
[all …]
Dentry.S61 REG_L s0, TASK_TI_USER_SP(tp)
67 REG_S s0, PT_SP(sp)
122 REG_L s0, PT_STATUS(sp)
126 and s0, s0, t0
128 andi s0, s0, SR_SPP
130 bnez s0, 1f
133 addi s0, sp, PT_SIZE_ON_STACK
134 REG_S s0, TASK_TI_KERNEL_SP(tp)
207 REG_L s0, TASK_TI_KERNEL_SP(tp)
213 REG_S s0, PT_SP(sp)
[all …]
Dkexec_relocate.S16 * s0: Pointer to the current entry
24 mv s0, a0
59 REG_L t0, 0(s0) /* t0 = *image->entry */
60 addi s0, s0, RISCV_SZPTR /* image->entry++ */
72 andi s0, t0, ~0x2
113 mv s0, zero
153 * s0: (const) Phys address to jump to
157 mv s0, a1
168 mv a2, s0
177 mv s0, zero
/linux-6.8/arch/x86/crypto/
Dsha512-avx2-asm.S200 xor T1, y1 # y1 = (a>>39) ^ (a>>34) # S0
201 rorx $28, a, T1 # T1 = (a >> 28) # S0
204 xor T1, y1 # y1 = (a>>39) ^ (a>>34) ^ (a>>28) # S0
210 add y1, h # h = k + w + h + S0 # --
214 add y2, h # h = k + w + h + S0 + S1 + CH = t1 + S0# --
215 add y3, h # h = t1 + S0 + MAJ # --
227 vpxor YTMP1, YTMP3, YTMP1 # YTMP1 = s0
231 vpaddq YTMP1, YTMP0, YTMP0 # YTMP0 = W[-16] + W[-7] + s0
233 vperm2f128 $0x0, YTMP0, YTMP0, Y_0 # Y_0 = W[-16] + W[-7] + s0 {BABA}
235 vpand MASK_YMM_LO(%rip), YTMP0, YTMP0 # YTMP0 = W[-16] + W[-7] + s0 {DC00}
[all …]
Dsha256-avx2-asm.S179 xor T1, y1 # y1 = (a>>22) ^ (a>>13) # S0
180 rorx $2, a, T1 # T1 = (a >> 2) # S0
184 xor T1, y1 # y1 = (a>>22) ^ (a>>13) ^ (a>>2) # S0
191 add y1, h # h = k + w + h + S0 # --
197 add y2, h # h = k + w + h + S0 + S1 + CH = t1 + S0# --
198 add y3, h # h = t1 + S0 + MAJ # --
228 xor T1, y1 # y1 = (a>>22) ^ (a>>13) # S0
231 rorx $2, a, T1 # T1 = (a >> 2) # S0
235 xor T1, y1 # y1 = (a>>22) ^ (a>>13) ^ (a>>2) # S0
240 vpxor XTMP4, XTMP3, XTMP1 # XTMP1 = s0
[all …]
Dsm4-aesni-avx-asm_64.S188 #define ROUND(round, s0, s1, s2, s3) \ argument
201 vpxor RTMP0, s0, s0; /* s0 ^ x */ \
207 vpxor RTMP1, s0, s0; /* s0 ^ x ^ rol(x,24) */ \
210 vpxor RTMP0, s0, s0; \
211 /* s0 ^ x ^ rol(x,2) ^ rol(x,10) ^ rol(x,18) ^ rol(x,24) */ \
212 vpxor RTMP1, s0, s0;
276 #define ROUND(round, s0, s1, s2, s3, r0, r1, r2, r3) \ argument
301 vpxor RTMP0, s0, s0; /* s0 ^ x */ \
316 vpxor RTMP1, s0, s0; /* s0 ^ x ^ rol(x,24) */ \
317 /* s0 ^ x ^ rol(x,2) ^ rol(x,10) ^ rol(x,18) ^ rol(x,24) */ \
[all …]
Dsha256-avx-asm.S155 ## compute s0 four at a time and s1 two at a time
172 ## compute s0
177 MY_ROR 2, y1 # y1 = S0 = (a>>2) ^ (a>>13) ^ (a>>22)
189 add y1, h # h = h + S1 + CH + k + w + S0
192 add y0, h # h = h + S1 + CH + k + w + S0 + MAJ
215 MY_ROR 2, y1 # y1 = S0 = (a>>2) ^ (a>>13) ^ (a>>22)
220 vpxor XTMP4, XTMP3, XTMP1 # XTMP1 = s0
227 add y1, h # h = h + S1 + CH + k + w + S0
228 vpaddd XTMP1, XTMP0, XTMP0 # XTMP0 = W[-16] + W[-7] + s0
230 add y0, h # h = h + S1 + CH + k + w + S0 + MAJ
[all …]
Dsha256-ssse3-asm.S149 ## compute s0 four at a time and s1 two at a time
167 ## compute s0
173 ror $2, y1 # y1 = S0 = (a>>2) ^ (a>>13) ^ (a>>22)
186 add y1, h # h = h + S1 + CH + k + w + S0
189 add y0, h # h = h + S1 + CH + k + w + S0 + MAJ
215 ror $2, y1 # y1 = S0 = (a>>2) ^ (a>>13) ^ (a>>22)
220 pxor XTMP4, XTMP1 # XTMP1 = s0
227 add y1, h # h = h + S1 + CH + k + w + S0
228 paddd XTMP1, XTMP0 # XTMP0 = W[-16] + W[-7] + s0
230 add y0, h # h = h + S1 + CH + k + w + S0 + MAJ
[all …]
/linux-6.8/drivers/soc/bcm/brcmstb/pm/
Ds2-mips.S24 sw s0, 4(sp)
35 * s0: AON_CTRL base register
44 lw s0, 0(t0)
79 sw zero, AON_CTRL_PM_CTRL(s0)
80 lw zero, AON_CTRL_PM_CTRL(s0)
81 sw t0, AON_CTRL_PM_CTRL(s0)
82 lw t0, AON_CTRL_PM_CTRL(s0)
134 sw t1, AON_CTRL_HOST_MISC_CMDS(s0)
135 lw t1, AON_CTRL_HOST_MISC_CMDS(s0)
137 sw zero, AON_CTRL_PM_CTRL(s0)
[all …]
/linux-6.8/scripts/selinux/
Dinstall_policy.sh45 echo "__default__:user_u:s0" > /etc/selinux/dummy/seusers
46 echo "base_r:base_t:s0" > /etc/selinux/dummy/contexts/failsafe_context
47 echo "base_r:base_t:s0 base_r:base_t:s0" > /etc/selinux/dummy/default_contexts
49 client * user_u:base_r:base_t:s0
50 property * user_u:object_r:base_t:s0
51 extension * user_u:object_r:base_t:s0
52 selection * user_u:object_r:base_t:s0
53 event * user_u:object_r:base_t:s0
/linux-6.8/arch/arm/boot/dts/aspeed/
Daspeed-bmc-ampere-mtmitchell.dts111 channels = "s0", "s1";
121 channels = "s0", "s1";
131 channels = "s0", "s1";
141 channels = "s0", "s1";
151 channels = "s0", "s1";
161 channels = "s0", "s1";
171 channels = "s0", "s1";
181 channels = "s0", "s1";
191 channels = "s0", "s1";
201 channels = "s0", "s1";
[all …]
Daspeed-bmc-ampere-mtjade.dts112 channels = "s0", "s1";
121 channels = "s0", "s1";
130 channels = "s0", "s1";
139 channels = "s0", "s1";
148 channels = "s0", "s1";
157 channels = "s0", "s1";
166 channels = "s0", "s1";
175 channels = "s0", "s1";
184 channels = "s0", "s1";
193 channels = "s0", "s1";
[all …]
/linux-6.8/net/dccp/ccids/lib/
Dpacket_history.c154 u64 s0 = tfrc_rx_hist_loss_prev(h)->tfrchrx_seqno, in __do_track_loss() local
157 if (!dccp_loss_free(s0, s1, n1)) { /* gap between S0 and S1 */ in __do_track_loss()
165 u64 s0 = tfrc_rx_hist_loss_prev(h)->tfrchrx_seqno, in __one_after_loss() local
175 /* S0 < S2 < S1 */ in __one_after_loss()
177 if (dccp_loss_free(s0, s2, n2)) { in __one_after_loss()
181 /* hole is filled: S0, S2, and S1 are consecutive */ in __one_after_loss()
188 } else { /* gap between S0 and S2 */ in __one_after_loss()
190 * Reorder history to insert S2 between S0 and S1 in __one_after_loss()
202 u64 s0 = tfrc_rx_hist_loss_prev(h)->tfrchrx_seqno, in __two_after_loss() local
225 /* S0 < S3 < S1 */ in __two_after_loss()
[all …]
/linux-6.8/arch/arm64/crypto/
Dsm3-ce-core.S44 .macro round, ab, s0, t0, t1, i
49 sm3tt2\ab v9.4s, v5.4s, \s0\().4s, \i
52 .macro qround, ab, s0, s1, s2, s3, s4
55 ext v6.16b, \s0\().16b, \s1\().16b, #12
57 sm3partw1 \s4\().4s, \s0\().4s, \s3\().4s
60 eor v10.16b, \s0\().16b, \s1\().16b
62 round \ab, \s0, v11, v12, 0
63 round \ab, \s0, v12, v11, 1
64 round \ab, \s0, v11, v12, 2
65 round \ab, \s0, v12, v11, 3
Dsm4-neon-core.S40 #define transpose_4x4(s0, s1, s2, s3) \ argument
41 zip1 RTMP0.4s, s0.4s, s1.4s; \
43 zip2 RTMP2.4s, s0.4s, s1.4s; \
45 zip1 s0.2d, RTMP0.2d, RTMP1.2d; \
50 #define transpose_4x4_2x(s0, s1, s2, s3, s4, s5, s6, s7) \ argument
51 zip1 RTMP0.4s, s0.4s, s1.4s; \
53 zip2 RTMP2.4s, s0.4s, s1.4s; \
59 zip1 s0.2d, RTMP0.2d, RTMP1.2d; \
68 #define rotate_clockwise_4x4(s0, s1, s2, s3) \ argument
69 zip1 RTMP0.4s, s1.4s, s0.4s; \
[all …]
Dsha1-ce-core.S34 .macro add_only, op, ev, rc, s0, dg1
36 add t1.4s, v\s0\().4s, \rc\().4s
44 .ifnb \s0
45 add t0.4s, v\s0\().4s, \rc\().4s
52 .macro add_update, op, ev, rc, s0, s1, s2, s3, dg1
53 sha1su0 v\s0\().4s, v\s1\().4s, v\s2\().4s
55 sha1su1 v\s0\().4s, v\s3\().4s
Dsha2-ce-core.S29 .macro add_only, ev, rc, s0 argument
32 add t1.4s, v\s0\().4s, \rc\().4s
36 .ifnb \s0
37 add t0.4s, v\s0\().4s, \rc\().4s
44 .macro add_update, ev, rc, s0, s1, s2, s3
45 sha256su0 v\s0\().4s, v\s1\().4s
47 sha256su1 v\s0\().4s, v\s2\().4s, v\s3\().4s
/linux-6.8/arch/mips/kernel/
Drelocate_kernel.S22 PTR_L s0, kexec_indirection_page
26 PTR_L s2, (s0)
27 PTR_ADDIU s0, s0, SZREG
42 /* indirection page, update s0 */
45 and s0, s2, ~0x2
126 1: LONG_L s0, (t0)
127 bne s0, zero,1b
/linux-6.8/arch/arm/crypto/
Dsha2-ce-core.S32 .macro add_only, ev, s0 argument
34 .ifnb \s0
39 .ifnb \s0
40 vadd.u32 ta\ev, q\s0, k\ev
44 .macro add_update, ev, s0, s1, s2, s3
45 sha256su0.32 q\s0, q\s1
47 sha256su1.32 q\s0, q\s2, q\s3
Dsha1-ce-core.S36 .macro add_only, op, ev, rc, s0, dg1
37 .ifnb \s0
38 vadd.u32 tb\ev, q\s0, \rc
48 .macro add_update, op, ev, rc, s0, s1, s2, s3, dg1
49 sha1su0.32 q\s0, q\s1, q\s2
51 sha1su1.32 q\s0, q\s3
/linux-6.8/arch/mips/include/asm/mach-loongson64/
Dkernel-entry-init.h80 /* s0:prid s1:initfn */
87 mfc0 s0, CP0_PRID
88 andi s0, s0, (PRID_IMP_MASK | PRID_REV_MASK)
89 beq s0, (PRID_IMP_LOONGSON_64C | PRID_REV_LOONGSON3B_R1), 1f
90 beq s0, (PRID_IMP_LOONGSON_64C | PRID_REV_LOONGSON3B_R2), 1f
/linux-6.8/net/lapb/
Dlapb_in.c44 lapb_dbg(1, "(%p) S0 RX SABM(%d)\n", lapb->dev, frame->pf); in lapb_state0_machine()
46 lapb_dbg(1, "(%p) S0 TX DM(%d)\n", in lapb_state0_machine()
51 lapb_dbg(1, "(%p) S0 TX UA(%d)\n", in lapb_state0_machine()
53 lapb_dbg(0, "(%p) S0 -> S3\n", lapb->dev); in lapb_state0_machine()
69 lapb_dbg(1, "(%p) S0 RX SABME(%d)\n", lapb->dev, frame->pf); in lapb_state0_machine()
71 lapb_dbg(1, "(%p) S0 TX UA(%d)\n", in lapb_state0_machine()
73 lapb_dbg(0, "(%p) S0 -> S3\n", lapb->dev); in lapb_state0_machine()
86 lapb_dbg(1, "(%p) S0 TX DM(%d)\n", in lapb_state0_machine()
94 lapb_dbg(1, "(%p) S0 RX DISC(%d)\n", lapb->dev, frame->pf); in lapb_state0_machine()
95 lapb_dbg(1, "(%p) S0 TX UA(%d)\n", lapb->dev, frame->pf); in lapb_state0_machine()
[all …]

12345678910>>...15