| /linux/lib/vdso/ |
| H A D | gettimeofday.c | 15 # define VDSO_DELTA_MASK(vd) ULLONG_MAX argument 17 # define VDSO_DELTA_MASK(vd) (vd->mask) argument 92 bool vdso_get_timestamp(const struct vdso_time_data *vd, const struct vdso_clock *vc, in vdso_get_timestamp() argument 101 cycles = __arch_get_hw_counter(vc->clock_mode, vd); in vdso_get_timestamp() 112 const struct vdso_time_data *__arch_get_vdso_u_timens_data(const struct vdso_time_data *vd) in __arch_get_vdso_u_timens_data() argument 114 return (void *)vd + PAGE_SIZE; in __arch_get_vdso_u_timens_data() 121 const struct vdso_time_data *vd = __arch_get_vdso_u_timens_data(vdns); in do_hres_timens() local 123 const struct vdso_clock *vc = vd->clock_data; in do_hres_timens() 136 if (!vdso_get_timestamp(vd, vc, clk, &sec, &ns)) in do_hres_timens() 150 bool do_hres(const struct vdso_time_data *vd, const struct vdso_clock *vc, in do_hres() argument [all …]
|
| /linux/drivers/dma/ |
| H A D | virt-dma.h | 57 struct virt_dma_desc *vd, unsigned long tx_flags) in vchan_tx_prep() argument 61 dma_async_tx_descriptor_init(&vd->tx, &vc->chan); in vchan_tx_prep() 62 vd->tx.flags = tx_flags; in vchan_tx_prep() 63 vd->tx.tx_submit = vchan_tx_submit; in vchan_tx_prep() 64 vd->tx.desc_free = vchan_tx_desc_free; in vchan_tx_prep() 66 vd->tx_result.result = DMA_TRANS_NOERROR; in vchan_tx_prep() 67 vd->tx_result.residue = 0; in vchan_tx_prep() 70 list_add_tail(&vd->node, &vc->desc_allocated); in vchan_tx_prep() 73 return &vd->tx; in vchan_tx_prep() 96 static inline void vchan_cookie_complete(struct virt_dma_desc *vd) in vchan_cookie_complete() argument [all …]
|
| H A D | virt-dma.c | 22 struct virt_dma_desc *vd = to_virt_desc(tx); in vchan_tx_submit() local 29 list_move_tail(&vd->node, &vc->desc_submitted); in vchan_tx_submit() 33 vc, vd, cookie); in vchan_tx_submit() 52 struct virt_dma_desc *vd = to_virt_desc(tx); in vchan_tx_desc_free() local 56 list_del(&vd->node); in vchan_tx_desc_free() 60 vc, vd, vd->tx.cookie); in vchan_tx_desc_free() 61 vc->desc_free(vd); in vchan_tx_desc_free() 69 struct virt_dma_desc *vd; in vchan_find_desc() local 71 list_for_each_entry(vd, &vc->desc_issued, node) in vchan_find_desc() 72 if (vd->tx.cookie == cookie) in vchan_find_desc() [all …]
|
| H A D | loongson1-apb-dma.c | 60 struct virt_dma_desc vd; member 90 container_of(d, struct ls1x_dma_desc, vd) 189 static void ls1x_dma_free_desc(struct virt_dma_desc *vd) in ls1x_dma_free_desc() argument 191 struct ls1x_dma_desc *desc = to_ls1x_dma_desc(vd); in ls1x_dma_free_desc() 192 struct ls1x_dma_chan *chan = to_ls1x_dma_chan(vd->tx.chan); in ls1x_dma_free_desc() 310 ls1x_dma_free_desc(&desc->vd); in ls1x_dma_prep_slave_sg() 314 return vchan_tx_prep(to_virt_chan(dchan), &desc->vd, flags); in ls1x_dma_prep_slave_sg() 354 ls1x_dma_free_desc(&desc->vd); in ls1x_dma_prep_dma_cyclic() 358 return vchan_tx_prep(to_virt_chan(dchan), &desc->vd, flags); in ls1x_dma_prep_dma_cyclic() 400 struct virt_dma_desc *vd; in ls1x_dma_terminate_all() local [all …]
|
| H A D | uniphier-mdmac.c | 52 struct virt_dma_desc vd; member 81 to_uniphier_mdmac_desc(struct virt_dma_desc *vd) in to_uniphier_mdmac_desc() argument 83 return container_of(vd, struct uniphier_mdmac_desc, vd); in to_uniphier_mdmac_desc() 90 struct virt_dma_desc *vd; in uniphier_mdmac_next_desc() local 92 vd = vchan_next_desc(&mc->vc); in uniphier_mdmac_next_desc() 93 if (!vd) { in uniphier_mdmac_next_desc() 98 list_del(&vd->node); in uniphier_mdmac_next_desc() 100 mc->md = to_uniphier_mdmac_desc(vd); in uniphier_mdmac_next_desc() 210 vchan_cookie_complete(&md->vd); in uniphier_mdmac_interrupt() 249 return vchan_tx_prep(vc, &md->vd, flags); in uniphier_mdmac_prep_slave_sg() [all …]
|
| H A D | pxa_dma.c | 84 struct virt_dma_desc vd; /* Virtual descriptor */ member 144 container_of((_vd), struct pxad_desc_sw, vd) 533 static bool is_desc_completed(struct virt_dma_desc *vd) in is_desc_completed() argument 535 struct pxad_desc_sw *sw_desc = to_pxad_sw_desc(vd); in is_desc_completed() 554 struct virt_dma_desc *vd) in pxad_try_hotchain() argument 570 to_pxad_sw_desc(vd)->misaligned) in pxad_try_hotchain() 575 pxad_desc_chain(vd_last_issued, vd); in pxad_try_hotchain() 576 if (is_chan_running(chan) || is_desc_completed(vd)) in pxad_try_hotchain() 606 struct virt_dma_desc *vd, *tmp; in pxad_chan_handler() local 618 list_for_each_entry_safe(vd, tmp, &chan->vc.desc_issued, node) { in pxad_chan_handler() [all …]
|
| H A D | milbeaut-xdmac.c | 62 struct virt_dma_desc vd; member 87 to_milbeaut_xdmac_desc(struct virt_dma_desc *vd) in to_milbeaut_xdmac_desc() argument 89 return container_of(vd, struct milbeaut_xdmac_desc, vd); in to_milbeaut_xdmac_desc() 96 struct virt_dma_desc *vd; in milbeaut_xdmac_next_desc() local 98 vd = vchan_next_desc(&mc->vc); in milbeaut_xdmac_next_desc() 99 if (!vd) { in milbeaut_xdmac_next_desc() 104 list_del(&vd->node); in milbeaut_xdmac_next_desc() 106 mc->md = to_milbeaut_xdmac_desc(vd); in milbeaut_xdmac_next_desc() 175 vchan_cookie_complete(&md->vd); in milbeaut_xdmac_interrupt() 203 return vchan_tx_prep(vc, &md->vd, flags); in milbeaut_xdmac_prep_memcpy() [all …]
|
| H A D | milbeaut-hdmac.c | 60 struct virt_dma_desc vd; member 90 to_milbeaut_hdmac_desc(struct virt_dma_desc *vd) in to_milbeaut_hdmac_desc() argument 92 return container_of(vd, struct milbeaut_hdmac_desc, vd); in to_milbeaut_hdmac_desc() 99 struct virt_dma_desc *vd; in milbeaut_hdmac_next_desc() local 101 vd = vchan_next_desc(&mc->vc); in milbeaut_hdmac_next_desc() 102 if (!vd) { in milbeaut_hdmac_next_desc() 107 list_del(&vd->node); in milbeaut_hdmac_next_desc() 109 mc->md = to_milbeaut_hdmac_desc(vd); in milbeaut_hdmac_next_desc() 194 vchan_cookie_complete(&md->vd); in milbeaut_hdmac_interrupt() 284 return vchan_tx_prep(vc, &md->vd, flags); in milbeaut_hdmac_prep_slave_sg() [all …]
|
| H A D | amba-pl08x.c | 199 struct virt_dma_desc vd; member 284 const struct vendor_data *vd; member 322 return container_of(tx, struct pl08x_txd, vd.tx); in to_pl08x_txd() 394 if (pl08x->vd->pl080s) in pl08x_write_lli() 514 if (pl08x->vd->pl080s) in pl08x_write_lli() 531 struct virt_dma_desc *vd = vchan_next_desc(&plchan->vc); in pl08x_start_next_txd() local 532 struct pl08x_txd *txd = to_pl08x_txd(&vd->tx); in pl08x_start_next_txd() 535 list_del(&txd->vd.node); in pl08x_start_next_txd() 825 for (i = 0; i < pl08x->vd->channels; i++) { in pl08x_get_phy_channel() 839 if (i == pl08x->vd->channels) { in pl08x_get_phy_channel() [all …]
|
| H A D | moxart-dma.c | 125 struct virt_dma_desc vd; member 170 return container_of(t, struct moxart_desc, vd.tx); in to_moxart_dma_desc() 173 static void moxart_dma_desc_free(struct virt_dma_desc *vd) in moxart_dma_desc_free() argument 175 kfree(container_of(vd, struct moxart_desc, vd)); in moxart_dma_desc_free() 190 moxart_dma_desc_free(&ch->desc->vd); in moxart_terminate_all() 320 return vchan_tx_prep(&ch->vc, &d->vd, tx_flags); in moxart_prep_slave_sg() 413 struct virt_dma_desc *vd; in moxart_dma_start_desc() local 415 vd = vchan_next_desc(&ch->vc); in moxart_dma_start_desc() 417 if (!vd) { in moxart_dma_start_desc() 422 list_del(&vd->node); in moxart_dma_start_desc() [all …]
|
| H A D | arm-dma350.c | 172 struct virt_dma_desc vd; member 205 static inline struct d350_desc *to_d350_desc(struct virt_dma_desc *vd) in to_d350_desc() argument 207 return container_of(vd, struct d350_desc, vd); in to_d350_desc() 210 static void d350_desc_free(struct virt_dma_desc *vd) in d350_desc_free() argument 212 kfree(to_d350_desc(vd)); in d350_desc_free() 250 return vchan_tx_prep(&dch->vc, &desc->vd, flags); in d350_prep_memcpy() 286 return vchan_tx_prep(&dch->vc, &desc->vd, flags); in d350_prep_memset() 347 vchan_terminate_vdesc(&dch->desc->vd); in d350_terminate_all() 374 struct virt_dma_desc *vd; in d350_tx_status() local 387 } else if ((vd = vchan_find_desc(&dch->vc, cookie))) { in d350_tx_status() [all …]
|
| /linux/arch/powerpc/kernel/vdso/ |
| H A D | vgettimeofday.c | 10 const struct vdso_time_data *vd) in __c_kernel_clock_gettime() argument 12 return __cvdso_clock_gettime_data(vd, clock, ts); in __c_kernel_clock_gettime() 16 const struct vdso_time_data *vd) in __c_kernel_clock_getres() argument 18 return __cvdso_clock_getres_data(vd, clock_id, res); in __c_kernel_clock_getres() 22 const struct vdso_time_data *vd) in __c_kernel_clock_gettime() argument 24 return __cvdso_clock_gettime32_data(vd, clock, ts); in __c_kernel_clock_gettime() 28 const struct vdso_time_data *vd) in __c_kernel_clock_gettime64() argument 30 return __cvdso_clock_gettime_data(vd, clock, ts); in __c_kernel_clock_gettime64() 34 const struct vdso_time_data *vd) in __c_kernel_clock_getres() argument 36 return __cvdso_clock_getres_time32_data(vd, clock_id, res); in __c_kernel_clock_getres() [all …]
|
| /linux/arch/powerpc/kernel/ |
| H A D | vecemu.c | 266 unsigned int va, vb, vc, vd; in emulate_altivec() local 275 vd = (word >> 21) & 0x1f; in emulate_altivec() 285 vaddfp(&vrs[vd], &vrs[va], &vrs[vb]); in emulate_altivec() 288 vsubfp(&vrs[vd], &vrs[va], &vrs[vb]); in emulate_altivec() 291 vrefp(&vrs[vd], &vrs[vb]); in emulate_altivec() 294 vrsqrtefp(&vrs[vd], &vrs[vb]); in emulate_altivec() 298 vrs[vd].u[i] = eexp2(vrs[vb].u[i]); in emulate_altivec() 302 vrs[vd].u[i] = elog2(vrs[vb].u[i]); in emulate_altivec() 306 vrs[vd].u[i] = rfin(vrs[vb].u[i]); in emulate_altivec() 310 vrs[vd].u[i] = rfiz(vrs[vb].u[i]); in emulate_altivec() [all …]
|
| /linux/arch/powerpc/include/asm/vdso/ |
| H A D | gettimeofday.h | 97 const struct vdso_time_data *vd) in __arch_get_hw_counter() argument 128 const struct vdso_time_data *vd); 130 const struct vdso_time_data *vd); 133 const struct vdso_time_data *vd); 135 const struct vdso_time_data *vd); 137 const struct vdso_time_data *vd); 139 const struct vdso_time_data *vd); 142 const struct vdso_time_data *vd); 144 const struct vdso_time_data *vd);
|
| /linux/drivers/dma/mediatek/ |
| H A D | mtk-hsdma.c | 131 struct virt_dma_desc vd; member 147 struct virt_dma_desc *vd; member 258 static struct mtk_hsdma_vdesc *to_hsdma_vdesc(struct virt_dma_desc *vd) in to_hsdma_vdesc() argument 260 return container_of(vd, struct mtk_hsdma_vdesc, vd); in to_hsdma_vdesc() 299 static void mtk_hsdma_vdesc_free(struct virt_dma_desc *vd) in mtk_hsdma_vdesc_free() argument 301 kfree(container_of(vd, struct mtk_hsdma_vdesc, vd)); in mtk_hsdma_vdesc_free() 459 ring->cb[ring->cur_tptr].vd = &hvd->vd; in mtk_hsdma_issue_pending_vdesc() 497 struct virt_dma_desc *vd, *vd2; in mtk_hsdma_issue_vchan_pending() local 502 list_for_each_entry_safe(vd, vd2, &hvc->vc.desc_issued, node) { in mtk_hsdma_issue_vchan_pending() 505 hvd = to_hsdma_vdesc(vd); in mtk_hsdma_issue_vchan_pending() [all …]
|
| H A D | mtk-cqdma.c | 79 struct virt_dma_desc vd; member 158 static struct mtk_cqdma_vdesc *to_cqdma_vdesc(struct virt_dma_desc *vd) in to_cqdma_vdesc() argument 160 return container_of(vd, struct mtk_cqdma_vdesc, vd); in to_cqdma_vdesc() 199 static void mtk_cqdma_vdesc_free(struct virt_dma_desc *vd) in mtk_cqdma_vdesc_free() argument 201 kfree(to_cqdma_vdesc(vd)); in mtk_cqdma_vdesc_free() 267 struct virt_dma_desc *vd, *vd2; in mtk_cqdma_issue_vchan_pending() local 275 list_for_each_entry_safe(vd, vd2, &cvc->vc.desc_issued, node) { in mtk_cqdma_issue_vchan_pending() 280 cvd = to_cqdma_vdesc(vd); in mtk_cqdma_issue_vchan_pending() 290 list_del(&vd->node); in mtk_cqdma_issue_vchan_pending() 338 vchan_cookie_complete(&cvd->parent->vd); in mtk_cqdma_consume_work_queue() [all …]
|
| /linux/arch/powerpc/platforms/pseries/ |
| H A D | mobility.c | 120 const char *name, u32 vd, char *value) in update_dt_property() argument 131 if (vd & 0x80000000) { in update_dt_property() 132 vd = ~vd + 1; in update_dt_property() 138 char *new_data = kzalloc(new_prop->length + vd, GFP_KERNEL); in update_dt_property() 143 memcpy(new_data + new_prop->length, value, vd); in update_dt_property() 147 new_prop->length += vd; in update_dt_property() 159 new_prop->length = vd; in update_dt_property() 167 memcpy(new_prop->value, value, vd); in update_dt_property() 189 u32 vd; in update_dt_node() local 218 vd = be32_to_cpu(*(__be32 *)prop_data); in update_dt_node() [all …]
|
| /linux/drivers/dma/amd/ptdma/ |
| H A D | ptdma-dmaengine.c | 70 static inline struct pt_dma_desc *to_pt_desc(struct virt_dma_desc *vd) in to_pt_desc() argument 72 return container_of(vd, struct pt_dma_desc, vd); in to_pt_desc() 89 static void pt_do_cleanup(struct virt_dma_desc *vd) in pt_do_cleanup() argument 91 struct pt_dma_desc *desc = to_pt_desc(vd); in pt_do_cleanup() 191 struct virt_dma_desc *vd = vchan_next_desc(&chan->vc); in pt_next_dma_desc() local 193 return vd ? to_pt_desc(vd) : NULL; in pt_next_dma_desc() 200 struct virt_dma_desc *vd; in pt_handle_active_desc() local 214 tx_desc = &desc->vd.tx; in pt_handle_active_desc() 215 vd = &desc->vd; in pt_handle_active_desc() 229 list_del(&desc->vd.node); in pt_handle_active_desc() [all …]
|
| /linux/arch/arm/vfp/ |
| H A D | vfpdouble.c | 54 static void vfp_double_normalise_denormal(struct vfp_double *vd) in vfp_double_normalise_denormal() argument 56 int bits = 31 - fls(vd->significand >> 32); in vfp_double_normalise_denormal() 58 bits = 63 - fls(vd->significand); in vfp_double_normalise_denormal() 60 vfp_double_dump("normalise_denormal: in", vd); in vfp_double_normalise_denormal() 63 vd->exponent -= bits - 1; in vfp_double_normalise_denormal() 64 vd->significand <<= bits; in vfp_double_normalise_denormal() 67 vfp_double_dump("normalise_denormal: out", vd); in vfp_double_normalise_denormal() 70 u32 vfp_double_normaliseround(int dd, struct vfp_double *vd, u32 fpscr, u32 exceptions, const char … in vfp_double_normaliseround() argument 76 vfp_double_dump("pack: in", vd); in vfp_double_normaliseround() 81 if (vd->exponent == 2047 && (vd->significand == 0 || exceptions)) in vfp_double_normaliseround() [all …]
|
| /linux/include/drm/ |
| H A D | drm_modes.h | 134 #define DRM_MODE(nm, t, c, hd, hss, hse, ht, hsk, vd, vss, vse, vt, vs, f) \ argument 137 .htotal = (ht), .hskew = (hsk), .vdisplay = (vd), \ 149 #define __DRM_MODE_INIT(pix, hd, vd, hd_mm, vd_mm) \ argument 152 .htotal = (hd), .vdisplay = (vd), .vsync_start = (vd), \ 153 .vsync_end = (vd), .vtotal = (vd), .width_mm = (hd_mm), \ 167 #define DRM_MODE_INIT(hz, hd, vd, hd_mm, vd_mm) \ argument 168 __DRM_MODE_INIT((hd) * (vd) * (hz) / 1000 /* kHz */, hd, vd, hd_mm, vd_mm) 180 #define DRM_SIMPLE_MODE(hd, vd, hd_mm, vd_mm) \ argument 181 __DRM_MODE_INIT(1 /* pass validation */, hd, vd, hd_mm, vd_mm)
|
| /linux/drivers/dma/dw-edma/ |
| H A D | dw-edma-core.c | 27 struct dw_edma_desc *vd2dw_edma_desc(struct virt_dma_desc *vd) in vd2dw_edma_desc() argument 29 return container_of(vd, struct dw_edma_desc, vd); in vd2dw_edma_desc() 179 struct virt_dma_desc *vd; in dw_edma_start_transfer() local 181 vd = vchan_next_desc(&chan->vc); in dw_edma_start_transfer() 182 if (!vd) in dw_edma_start_transfer() 185 desc = vd2dw_edma_desc(vd); in dw_edma_start_transfer() 320 struct virt_dma_desc *vd; in dw_edma_device_tx_status() local 336 vd = vchan_find_desc(&chan->vc, cookie); in dw_edma_device_tx_status() 337 if (vd) { in dw_edma_device_tx_status() 338 desc = vd2dw_edma_desc(vd); in dw_edma_device_tx_status() [all …]
|
| /linux/drivers/clk/versatile/ |
| H A D | icst.c | 76 unsigned int vd; in icst_hz_to_vco() local 81 vd = (f + fref_div / 2) / fref_div; in icst_hz_to_vco() 82 if (vd < p->vd_min || vd > p->vd_max) in icst_hz_to_vco() 85 f_pll = fref_div * vd; in icst_hz_to_vco() 91 vco.v = vd - 8; in icst_hz_to_vco()
|
| /linux/sound/core/ |
| H A D | control.c | 126 if (control->vd[idx].owner == ctl) in snd_ctl_release() 127 control->vd[idx].owner = NULL; in snd_ctl_release() 236 *kctl = kzalloc_flex(**kctl, vd, count); in snd_ctl_new() 242 (*kctl)->vd[idx].access = access; in snd_ctl_new() 243 (*kctl)->vd[idx].owner = file; in snd_ctl_new() 670 if (!(kctl->vd[0].access & SNDRV_CTL_ELEM_ACCESS_USER)) in snd_ctl_remove_user_ctl() 673 if (kctl->vd[idx].owner != NULL && kctl->vd[idx].owner != file) in snd_ctl_remove_user_ctl() 694 struct snd_kcontrol_volatile *vd; in snd_ctl_activate_id() local 705 vd = &kctl->vd[index_offset]; in snd_ctl_activate_id() 708 if (!(vd->access & SNDRV_CTL_ELEM_ACCESS_INACTIVE)) in snd_ctl_activate_id() [all …]
|
| /linux/drivers/dma/sh/ |
| H A D | usb-dmac.c | 52 struct virt_dma_desc vd; member 63 #define to_usb_dmac_desc(vd) container_of(vd, struct usb_dmac_desc, vd) argument 224 struct virt_dma_desc *vd; in usb_dmac_chan_start_desc() local 226 vd = vchan_next_desc(&chan->vc); in usb_dmac_chan_start_desc() 227 if (!vd) { in usb_dmac_chan_start_desc() 237 list_del(&vd->node); in usb_dmac_chan_start_desc() 239 chan->desc = to_usb_dmac_desc(vd); in usb_dmac_chan_start_desc() 442 return vchan_tx_prep(&uchan->vc, &desc->vd, dma_flags); in usb_dmac_prep_slave_sg() 507 struct virt_dma_desc *vd; in usb_dmac_chan_get_residue() local 512 vd = vchan_find_desc(&chan->vc, cookie); in usb_dmac_chan_get_residue() [all …]
|
| /linux/drivers/media/radio/ |
| H A D | tea575x.c | 522 tea->vd = tea575x_radio; in snd_tea575x_init() 523 video_set_drvdata(&tea->vd, tea); in snd_tea575x_init() 525 strscpy(tea->vd.name, tea->v4l2_dev->name, sizeof(tea->vd.name)); in snd_tea575x_init() 526 tea->vd.lock = &tea->mutex; in snd_tea575x_init() 527 tea->vd.v4l2_dev = tea->v4l2_dev; in snd_tea575x_init() 528 tea->vd.device_caps = V4L2_CAP_TUNER | V4L2_CAP_RADIO; in snd_tea575x_init() 530 tea->vd.device_caps |= V4L2_CAP_HW_FREQ_SEEK; in snd_tea575x_init() 533 tea->vd.fops = &tea->fops; in snd_tea575x_init() 536 v4l2_disable_ioctl(&tea->vd, VIDIOC_S_HW_FREQ_SEEK); in snd_tea575x_init() 539 tea->vd.ctrl_handler = &tea->ctrl_handler; in snd_tea575x_init() [all …]
|