/linux/drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/nvrm/ |
H A D | rpcfn.h | 9 /* Excerpt of RM headers from https://github.com/NVIDIA/open-gpu-kernel-modules/tree/570.144 */ 16 X(RM, NOP, 0) 17 X(RM, SET_GUEST_SYSTEM_INFO, 1) 18 X(RM, ALLOC_ROOT, 2) 19 X(RM, ALLOC_DEVICE, 3) // deprecated 20 X(RM, ALLOC_MEMORY, 4) 21 X(RM, ALLOC_CTX_DMA, 5) 22 X(RM, ALLOC_CHANNEL_DMA, 6) 23 X(RM, MAP_MEMORY, 7) 24 X(RM, BIND_CTX_DMA, 8) // deprecated [all …]
|
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/nvrm/ |
H A D | rpcfn.h | 9 /* Excerpt of RM headers from https://github.com/NVIDIA/open-gpu-kernel-modules/tree/535.113.01 */ 16 X(RM, NOP) // 0 17 X(RM, SET_GUEST_SYSTEM_INFO) // 1 18 X(RM, ALLOC_ROOT) // 2 19 X(RM, ALLOC_DEVICE) // 3 deprecated 20 X(RM, ALLOC_MEMORY) // 4 21 X(RM, ALLOC_CTX_DMA) // 5 22 X(RM, ALLOC_CHANNEL_DMA) // 6 23 X(RM, MAP_MEMORY) // 7 24 X(RM, BIND_CTX_DMA) // 8 deprecated [all …]
|
/linux/net/rds/ |
H A D | message.c | 51 void rds_message_addref(struct rds_message *rm) in rds_message_addref() argument 53 rdsdebug("addref rm %p ref %d\n", rm, refcount_read(&rm->m_refcount)); in rds_message_addref() 54 refcount_inc(&rm->m_refcount); in rds_message_addref() 130 static void rds_message_purge(struct rds_message *rm) in rds_message_purge() argument 135 if (unlikely(test_bit(RDS_MSG_PAGEVEC, &rm->m_flags))) in rds_message_purge() 138 spin_lock_irqsave(&rm->m_rs_lock, flags); in rds_message_purge() 139 if (rm->m_rs) { in rds_message_purge() 140 struct rds_sock *rs = rm->m_rs; in rds_message_purge() 142 if (rm->data.op_mmp_znotifier) { in rds_message_purge() 144 rds_rm_zerocopy_callback(rs, rm->data.op_mmp_znotifier); in rds_message_purge() [all …]
|
H A D | send.c | 67 struct rds_message *rm, *tmp; in rds_send_path_reset() local 71 rm = cp->cp_xmit_rm; in rds_send_path_reset() 77 rds_message_unmapped(rm); in rds_send_path_reset() 78 rds_message_put(rm); in rds_send_path_reset() 95 list_for_each_entry_safe(rm, tmp, &cp->cp_retrans, m_conn_item) { in rds_send_path_reset() 96 set_bit(RDS_MSG_ACK_REQUIRED, &rm->m_flags); in rds_send_path_reset() 97 set_bit(RDS_MSG_RETRANSMITTED, &rm->m_flags); in rds_send_path_reset() 139 struct rds_message *rm; in rds_send_xmit() local 201 rm = cp->cp_xmit_rm; in rds_send_xmit() 203 if (!rm) { in rds_send_xmit() [all …]
|
H A D | tcp_send.c | 70 int rds_tcp_xmit(struct rds_connection *conn, struct rds_message *rm, in rds_tcp_xmit() argument 73 struct rds_conn_path *cp = rm->m_inc.i_conn_path; in rds_tcp_xmit() 86 rm->m_ack_seq = tc->t_last_sent_nxt + in rds_tcp_xmit() 88 be32_to_cpu(rm->m_inc.i_hdr.h_len) - 1; in rds_tcp_xmit() 90 set_bit(RDS_MSG_HAS_ACK_SEQ, &rm->m_flags); in rds_tcp_xmit() 91 tc->t_last_expected_una = rm->m_ack_seq + 1; in rds_tcp_xmit() 93 if (test_bit(RDS_MSG_RETRANSMITTED, &rm->m_flags)) in rds_tcp_xmit() 94 rm->m_inc.i_hdr.h_flags |= RDS_FLAG_RETRANSMITTED; in rds_tcp_xmit() 96 rdsdebug("rm %p tcp nxt %u ack_seq %llu\n", in rds_tcp_xmit() 97 rm, rds_tcp_write_seq(tc), in rds_tcp_xmit() [all …]
|
H A D | ib_send.c | 48 static void rds_ib_send_complete(struct rds_message *rm, in rds_ib_send_complete() argument 50 void (*complete)(struct rds_message *rm, int status)) in rds_ib_send_complete() argument 70 complete(rm, notify_status); in rds_ib_send_complete() 146 * Returns the rm for no good reason other than it is unobtainable 154 struct rds_message *rm = NULL; in rds_ib_send_unmap_op() local 160 rm = container_of(send->s_op, struct rds_message, data); in rds_ib_send_unmap_op() 167 rm = container_of(send->s_op, struct rds_message, rdma); in rds_ib_send_unmap_op() 174 rm = container_of(send->s_op, struct rds_message, atomic); in rds_ib_send_unmap_op() 187 return rm; in rds_ib_send_unmap_op() 245 struct rds_message *rm = NULL; in rds_ib_send_cqe_handler() local [all …]
|
/linux/arch/arm/net/ |
H A D | bpf_jit_32.h | 165 #define _AL3_R(op, rd, rn, rm) ((op ## _R) | (rd) << 12 | (rn) << 16 | (rm)) argument 171 #define ARM_ADD_R(rd, rn, rm) _AL3_R(ARM_INST_ADD, rd, rn, rm) argument 172 #define ARM_ADDS_R(rd, rn, rm) _AL3_R(ARM_INST_ADDS, rd, rn, rm) argument 175 #define ARM_ADC_R(rd, rn, rm) _AL3_R(ARM_INST_ADC, rd, rn, rm) argument 178 #define ARM_AND_R(rd, rn, rm) _AL3_R(ARM_INST_AND, rd, rn, rm) argument 179 #define ARM_ANDS_R(rd, rn, rm) _AL3_R(ARM_INST_ANDS, rd, rn, rm) argument 182 #define ARM_BIC_R(rd, rn, rm) _AL3_R(ARM_INST_BIC, rd, rn, rm) argument 186 #define ARM_BX(rm) (ARM_INST_BX | (rm)) argument 187 #define ARM_BLX_R(rm) (ARM_INST_BLX_R | (rm)) argument 189 #define ARM_CMP_R(rn, rm) _AL3_R(ARM_INST_CMP, 0, rn, rm) argument [all …]
|
/linux/drivers/macintosh/ |
H A D | rack-meter.c | 51 struct rackmeter *rm; member 95 static void rackmeter_setup_i2s(struct rackmeter *rm) in rackmeter_setup_i2s() argument 97 struct macio_chip *macio = rm->mdev->bus->chip; in rackmeter_setup_i2s() 106 pmac_call_feature(PMAC_FTR_SOUND_CHIP_ENABLE, rm->i2s, 0, 1); in rackmeter_setup_i2s() 121 out_le32(rm->i2s_regs + 0x10, 0x01fa0000); in rackmeter_setup_i2s() 122 (void)in_le32(rm->i2s_regs + 0x10); in rackmeter_setup_i2s() 132 static void rackmeter_set_default_pattern(struct rackmeter *rm) in rackmeter_set_default_pattern() argument 138 rm->ubuf[i] = (i & 1) * 255; in rackmeter_set_default_pattern() 140 rm->ubuf[i] = ((~i) & 1) * 255; in rackmeter_set_default_pattern() 144 static void rackmeter_do_pause(struct rackmeter *rm, int pause) in rackmeter_do_pause() argument [all …]
|
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/ |
H A D | vmm.c | 30 if (vmm->rm.external) { in r535_mmu_vaspace_del() 33 ctrl = nvkm_gsp_rm_ctrl_get(&vmm->rm.device.object, in r535_mmu_vaspace_del() 37 ctrl->hVASpace = vmm->rm.object.handle; in r535_mmu_vaspace_del() 39 WARN_ON(nvkm_gsp_rm_ctrl_wr(&vmm->rm.device.object, ctrl)); in r535_mmu_vaspace_del() 42 vmm->rm.external = false; in r535_mmu_vaspace_del() 45 nvkm_gsp_rm_free(&vmm->rm.object); in r535_mmu_vaspace_del() 46 nvkm_gsp_device_dtor(&vmm->rm.device); in r535_mmu_vaspace_del() 47 nvkm_gsp_client_dtor(&vmm->rm.client); in r535_mmu_vaspace_del() 49 nvkm_vmm_put(vmm, &vmm->rm.rsvd); in r535_mmu_vaspace_del() 59 &vmm->rm.client, &vmm->rm.device); in r535_mmu_vaspace_new() [all …]
|
H A D | Kbuild | 5 nvkm-y += nvkm/subdev/gsp/rm/r535/rm.o 6 nvkm-y += nvkm/subdev/gsp/rm/r535/gsp.o 7 nvkm-y += nvkm/subdev/gsp/rm/r535/rpc.o 8 nvkm-y += nvkm/subdev/gsp/rm/r535/ctrl.o 9 nvkm-y += nvkm/subdev/gsp/rm/r535/alloc.o 10 nvkm-y += nvkm/subdev/gsp/rm/r535/client.o 11 nvkm-y += nvkm/subdev/gsp/rm/r535/device.o 13 nvkm-y += nvkm/subdev/gsp/rm/r535/bar.o 14 nvkm-y += nvkm/subdev/gsp/rm/r535/fbsr.o 15 nvkm-y += nvkm/subdev/gsp/rm/r535/vmm.o [all …]
|
H A D | disp.c | 37 #include <rm/gpu.h> 69 nvkm_gsp_rm_free(&chan->rm.object); in r535_chan_fini() 75 struct nvkm_gsp *gsp = disp->rm.objcom.client->gsp; in r535_disp_chan_set_pushbuf() 116 const struct nvkm_rm_api *rmapi = chan->disp->rm.objcom.client->gsp->rm->api; in r535_curs_init() 124 args = nvkm_gsp_rm_alloc_get(&chan->disp->rm.object, in r535_curs_init() 126 chan->object.oclass, sizeof(*args), &chan->rm.object); in r535_curs_init() 132 return nvkm_gsp_rm_alloc_wr(&chan->rm.object, args); in r535_curs_init() 154 (chan->disp->rm.client.object.handle & 0x3fff)); in r535_dmac_bind() 173 args = nvkm_gsp_rm_alloc_get(&disp->rm.object, (oclass << 16) | inst, oclass, in r535_dmac_alloc() 187 const struct nvkm_rm_api *rmapi = chan->disp->rm.objcom.client->gsp->rm->api; in r535_dmac_init() [all …]
|
H A D | fifo.c | 34 #include <rm/engine.h> 44 struct nvkm_gsp *gsp = chan->rm.object.client->gsp; in r535_chan_doorbell_handle() 46 return gsp->rm->gpu->fifo.chan.doorbell_handle(chan); in r535_chan_doorbell_handle() 64 nvkm_gsp_rm_free(&chan->rm.object); in r535_chan_ramfc_clear() 66 dma_free_coherent(fifo->engine.subdev.device->dev, fifo->rm.mthdbuf_size, in r535_chan_ramfc_clear() 67 chan->rm.mthdbuf.ptr, chan->rm.mthdbuf.addr); in r535_chan_ramfc_clear() 69 nvkm_cgrp_vctx_put(chan->cgrp, &chan->rm.grctx); in r535_chan_ramfc_clear() 120 args->hVASpace = vmm->rm.object.handle; in r535_chan_alloc() 139 args->mthdbufMem.size = fifo->rm.mthdbuf_size; in r535_chan_alloc() 159 const struct nvkm_rm_api *rmapi = device->gsp->rm->api; in r535_chan_ramfc_write() [all …]
|
H A D | bar.c | 98 vmm->rm.bar2_pdb = gsp->bar.rm_bar2_pdb; in r535_bar_bar2_init() 168 struct nvkm_bar_func *rm; in r535_bar_new_() local 172 if (!(rm = kzalloc(sizeof(*rm), GFP_KERNEL))) in r535_bar_new_() 175 rm->dtor = r535_bar_dtor; in r535_bar_new_() 176 rm->oneinit = hw->oneinit; in r535_bar_new_() 177 rm->bar1.init = r535_bar_bar1_init; in r535_bar_new_() 178 rm->bar1.fini = r535_bar_bar1_fini; in r535_bar_new_() 179 rm->bar1.wait = r535_bar_bar1_wait; in r535_bar_new_() 180 rm->bar1.vmm = hw->bar1.vmm; in r535_bar_new_() 181 rm->bar2.init = r535_bar_bar2_init; in r535_bar_new_() [all …]
|
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/ |
H A D | engine.c | 19 struct nvkm_gsp_object rm; member 27 nvkm_gsp_rm_free(&obj->rm); in nvkm_rm_engine_obj_dtor() 40 struct nvkm_rm *rm = chan->client->gsp->rm; in nvkm_rm_engine_obj_new() local 53 ret = rm->api->ce->alloc(chan, handle, class, inst, &obj->rm); in nvkm_rm_engine_obj_new() 56 ret = nvkm_gsp_rm_alloc(chan, handle, class, 0, &obj->rm); in nvkm_rm_engine_obj_new() 59 ret = rm->api->nvdec->alloc(chan, handle, class, inst, &obj->rm); in nvkm_rm_engine_obj_new() 62 ret = rm->api->nvenc->alloc(chan, handle, class, inst, &obj->rm); in nvkm_rm_engine_obj_new() 65 ret = rm->api->nvjpg->alloc(chan, handle, class, inst, &obj->rm); in nvkm_rm_engine_obj_new() 68 ret = rm->api->ofa->alloc(chan, handle, class, inst, &obj->rm); in nvkm_rm_engine_obj_new() 92 return nvkm_rm_engine_obj_new(&chan->rm.object, chan->id, oclass, pobject); in nvkm_rm_engine_obj_ctor() [all …]
|
H A D | Kbuild | 4 nvkm-y += nvkm/subdev/gsp/rm/client.o 5 nvkm-y += nvkm/subdev/gsp/rm/engine.o 6 nvkm-y += nvkm/subdev/gsp/rm/gr.o 7 nvkm-y += nvkm/subdev/gsp/rm/nvdec.o 8 nvkm-y += nvkm/subdev/gsp/rm/nvenc.o 10 nvkm-y += nvkm/subdev/gsp/rm/tu1xx.o 11 nvkm-y += nvkm/subdev/gsp/rm/ga100.o 12 nvkm-y += nvkm/subdev/gsp/rm/ga1xx.o 13 nvkm-y += nvkm/subdev/gsp/rm/ad10x.o 14 nvkm-y += nvkm/subdev/gsp/rm/gh100.o [all …]
|
H A D | gr.c | 16 return nvkm_rm_engine_obj_new(&chan->chan->rm.object, chan->chan->id, oclass, pobject); in nvkm_rm_gr_obj_ctor() 22 struct nvkm_rm *rm = base->engine.subdev.device->gsp->rm; in nvkm_rm_gr_fini() local 25 if (rm->api->gr->scrubber.fini) in nvkm_rm_gr_fini() 26 rm->api->gr->scrubber.fini(gr); in nvkm_rm_gr_fini() 34 struct nvkm_rm *rm = base->engine.subdev.device->gsp->rm; in nvkm_rm_gr_init() local 38 if (rm->api->gr->scrubber.init) { in nvkm_rm_gr_init() 39 ret = rm->api->gr->scrubber.init(gr); in nvkm_rm_gr_init() 48 nvkm_rm_gr_new(struct nvkm_rm *rm) in nvkm_rm_gr_new() argument 51 rm->gpu->gr.class.i2m, in nvkm_rm_gr_new() 52 rm->gpu->gr.class.twod, in nvkm_rm_gr_new() [all …]
|
/linux/drivers/gpu/drm/msm/disp/dpu1/ |
H A D | dpu_rm.c | 34 * @rm: DPU Resource Manager handle 41 struct dpu_rm *rm, in dpu_rm_init() argument 48 if (!rm || !cat || !mmio) { in dpu_rm_init() 54 memset(rm, 0, sizeof(*rm)); in dpu_rm_init() 56 rm->has_legacy_ctls = (cat->mdss_ver->core_major_ver < 5); in dpu_rm_init() 69 rm->mixer_blks[lm->id - LM_0] = &hw->base; in dpu_rm_init() 83 rm->merge_3d_blks[merge_3d->id - MERGE_3D_0] = &hw->base; in dpu_rm_init() 98 hw->merge_3d = to_dpu_hw_merge_3d(rm->merge_3d_blks[pp->merge_3d - MERGE_3D_0]); in dpu_rm_init() 99 rm->pingpong_blks[pp->id - PINGPONG_0] = &hw->base; in dpu_rm_init() 112 rm->hw_intf[intf->id - INTF_0] = hw; in dpu_rm_init() [all …]
|
H A D | dpu_rm.h | 70 struct dpu_rm *rm, 75 int dpu_rm_reserve(struct dpu_rm *rm, 83 struct dpu_hw_sspp *dpu_rm_reserve_sspp(struct dpu_rm *rm, 91 int dpu_rm_get_assigned_resources(struct dpu_rm *rm, 100 * @rm: DPU Resource Manager handle 103 static inline struct dpu_hw_intf *dpu_rm_get_intf(struct dpu_rm *rm, enum dpu_intf intf_idx) in dpu_rm_get_intf() argument 105 return rm->hw_intf[intf_idx - INTF_0]; in dpu_rm_get_intf() 110 * @rm: DPU Resource Manager handle 113 static inline struct dpu_hw_wb *dpu_rm_get_wb(struct dpu_rm *rm, enum dpu_wb wb_idx) in dpu_rm_get_wb() argument 115 return rm->hw_wb[wb_idx - WB_0]; in dpu_rm_get_wb() [all …]
|
/linux/sound/soc/codecs/ |
H A D | tscs42xx.h | 114 #define RM(m, b) ((m)<<(b)) macro 133 #define RM_HPVOLL RM(FM_HPVOLL, FB_HPVOLL) 156 #define RM_HPVOLR RM(FM_HPVOLR, FB_HPVOLR) 179 #define RM_SPKVOLL RM(FM_SPKVOLL, FB_SPKVOLL) 204 #define RM_SPKVOLR RM(FM_SPKVOLR, FB_SPKVOLR) 229 #define RM_DACVOLL RM(FM_DACVOLL, FB_DACVOLL) 254 #define RM_DACVOLR RM(FM_DACVOLR, FB_DACVOLR) 279 #define RM_ADCVOLL RM(FM_ADCVOLL, FB_ADCVOLL) 304 #define RM_ADCVOLR RM(FM_ADCVOLR, FB_ADCVOLR) 337 RM(FM_INVOLL_INMUTEL, FB_INVOLL_INMUTEL) [all …]
|
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/ |
H A D | disp.c | 5 #include <rm/rm.h> 20 args = nvkm_gsp_rm_alloc_get(&disp->rm.object, (oclass << 16) | inst, oclass, in r570_dmac_alloc() 35 struct nvkm_gsp *gsp = disp->rm.objcom.client->gsp; in r570_disp_chan_set_pushbuf() 86 ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom, in r570_dp_set_indexed_link_rates() 95 return nvkm_gsp_rm_ctrl_wr(&disp->rm.objcom, ctrl); in r570_dp_set_indexed_link_rates() 104 ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom, in r570_dp_get_caps() 111 ret = nvkm_gsp_rm_ctrl_push(&disp->rm.objcom, &ctrl, sizeof(*ctrl)); in r570_dp_get_caps() 113 nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl); in r570_dp_get_caps() 137 nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl); in r570_dp_get_caps() 149 ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom, cmd, sizeof(*ctrl)); in r570_bl_ctrl() [all …]
|
H A D | Kbuild | 2 nvkm-y += nvkm/subdev/gsp/rm/r570/rm.o 3 nvkm-y += nvkm/subdev/gsp/rm/r570/gsp.o 4 nvkm-y += nvkm/subdev/gsp/rm/r570/client.o 5 nvkm-y += nvkm/subdev/gsp/rm/r570/fbsr.o 6 nvkm-y += nvkm/subdev/gsp/rm/r570/disp.o 7 nvkm-y += nvkm/subdev/gsp/rm/r570/fifo.o 8 nvkm-y += nvkm/subdev/gsp/rm/r570/gr.o 9 nvkm-y += nvkm/subdev/gsp/rm/r570/ofa.o
|
/linux/tools/perf/arch/powerpc/util/ |
H A D | perf_regs.c | 139 regmatch_t rm[5]; in arch_sdt_arg_parse_op() local 152 if (!regexec(&sdt_op_regex1, old_op, 3, rm, 0)) { in arch_sdt_arg_parse_op() 156 new_len += (int)(rm[2].rm_eo - rm[2].rm_so); in arch_sdt_arg_parse_op() 163 (int)(rm[2].rm_eo - rm[2].rm_so), old_op + rm[2].rm_so); in arch_sdt_arg_parse_op() 164 } else if (!regexec(&sdt_op_regex2, old_op, 5, rm, 0)) { in arch_sdt_arg_parse_op() 169 prefix = (rm[1].rm_so == -1) ? '+' : '-'; in arch_sdt_arg_parse_op() 172 new_len += (int)(rm[2].rm_eo - rm[2].rm_so); in arch_sdt_arg_parse_op() 173 new_len += (int)(rm[4].rm_eo - rm[4].rm_so); in arch_sdt_arg_parse_op() 180 (int)(rm[2].rm_eo - rm[2].rm_so), old_op + rm[2].rm_so, in arch_sdt_arg_parse_op() 181 (int)(rm[4].rm_eo - rm[4].rm_so), old_op + rm[4].rm_so); in arch_sdt_arg_parse_op()
|
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/gsp/ |
H A D | base.c | 94 nvkm_firmware_put(gsp->fws.rm); in nvkm_gsp_dtor_fws() 95 gsp->fws.rm = NULL; in nvkm_gsp_dtor_fws() 107 kfree(gsp->rm); in nvkm_gsp_dtor() 146 if (fwif->rm) { in nvkm_gsp_new_() 147 nvkm_info(&gsp->subdev, "RM version: %s\n", fwif->ver); in nvkm_gsp_new_() 149 gsp->rm = kzalloc(sizeof(*gsp->rm), GFP_KERNEL); in nvkm_gsp_new_() 150 if (!gsp->rm) in nvkm_gsp_new_() 153 gsp->rm->device = device; in nvkm_gsp_new_() 154 gsp->rm->gpu = fwif->func->rm.gpu; in nvkm_gsp_new_() 155 gsp->rm->wpr = fwif->rm->wpr; in nvkm_gsp_new_() [all …]
|
/linux/tools/perf/arch/arm64/util/ |
H A D | perf_regs.c | 98 regmatch_t rm[5]; in arch_sdt_arg_parse_op() local 104 if (!regexec(&sdt_op_regex1, old_op, 3, rm, 0)) { in arch_sdt_arg_parse_op() 107 new_len += (int)(rm[1].rm_eo - rm[1].rm_so); in arch_sdt_arg_parse_op() 114 (int)(rm[1].rm_eo - rm[1].rm_so), old_op + rm[1].rm_so); in arch_sdt_arg_parse_op() 115 } else if (!regexec(&sdt_op_regex2, old_op, 5, rm, 0)) { in arch_sdt_arg_parse_op() 120 if (rm[2].rm_so == -1) in arch_sdt_arg_parse_op() 123 new_len += (int)(rm[2].rm_eo - rm[2].rm_so); in arch_sdt_arg_parse_op() 129 if (rm[2].rm_so == -1) in arch_sdt_arg_parse_op() 133 (int)(rm[2].rm_eo - rm[2].rm_so), in arch_sdt_arg_parse_op() 134 old_op + rm[2].rm_so); in arch_sdt_arg_parse_op()
|
/linux/arch/mips/math-emu/ |
H A D | dp_mul.c | 16 u64 rm; in ieee754dp_mul() local 104 /* rm = xm * ym, re = xe+ye basically */ in ieee754dp_mul() 116 * Multiply 64 bits xm, ym to give high 64 bits rm with stickness. in ieee754dp_mul() 143 rm = hrm | (lrm != 0); in ieee754dp_mul() 148 if ((s64) rm < 0) { in ieee754dp_mul() 149 rm = (rm >> (64 - (DP_FBITS + 1 + 3))) | in ieee754dp_mul() 150 ((rm << (DP_FBITS + 1 + 3)) != 0); in ieee754dp_mul() 153 rm = (rm >> (64 - (DP_FBITS + 1 + 3 + 1))) | in ieee754dp_mul() 154 ((rm << (DP_FBITS + 1 + 3 + 1)) != 0); in ieee754dp_mul() 156 assert(rm & (DP_HIDDEN_BIT << 3)); in ieee754dp_mul() [all …]
|