Searched refs:sg_tmp (Results 1 – 6 of 6) sorted by relevance
| /linux/drivers/mmc/host/ ! |
| H A D | renesas_sdhi_sys_dmac.c | 160 struct scatterlist *sg = host->sg_ptr, *sg_tmp; in renesas_sdhi_sys_dmac_start_dma_rx() local 168 for_each_sg(sg, sg_tmp, host->sg_len, i) { in renesas_sdhi_sys_dmac_start_dma_rx() 169 if (sg_tmp->offset & align) in renesas_sdhi_sys_dmac_start_dma_rx() 171 if (sg_tmp->length & align) { in renesas_sdhi_sys_dmac_start_dma_rx() 232 struct scatterlist *sg = host->sg_ptr, *sg_tmp; in renesas_sdhi_sys_dmac_start_dma_tx() local 240 for_each_sg(sg, sg_tmp, host->sg_len, i) { in renesas_sdhi_sys_dmac_start_dma_tx() 241 if (sg_tmp->offset & align) in renesas_sdhi_sys_dmac_start_dma_tx() 243 if (sg_tmp->length & align) { in renesas_sdhi_sys_dmac_start_dma_tx()
|
| /linux/drivers/memstick/core/ ! |
| H A D | ms_block.c | 1609 struct scatterlist sg_tmp[10]; in msb_cache_write() local 1636 sg_init_table(sg_tmp, ARRAY_SIZE(sg_tmp)); in msb_cache_write() 1637 msb_sg_copy(sg, sg_tmp, ARRAY_SIZE(sg_tmp), offset, msb->page_size); in msb_cache_write() 1639 sg_copy_to_buffer(sg_tmp, sg_nents(sg_tmp), in msb_cache_write() 1650 struct scatterlist sg_tmp[10]; in msb_cache_read() local 1659 sg_init_table(sg_tmp, ARRAY_SIZE(sg_tmp)); in msb_cache_read() 1660 msb_sg_copy(sg, sg_tmp, ARRAY_SIZE(sg_tmp), in msb_cache_read() 1662 sg_copy_from_buffer(sg_tmp, sg_nents(sg_tmp), in msb_cache_read()
|
| /linux/drivers/crypto/intel/keembay/ ! |
| H A D | ocs-aes.c | 1425 struct scatterlist *sg_tmp; in ocs_create_linked_list_from_sg() local 1454 sg_tmp = sg; in ocs_create_linked_list_from_sg() 1457 if (!sg_tmp) in ocs_create_linked_list_from_sg() 1459 tmp += sg_dma_len(sg_tmp); in ocs_create_linked_list_from_sg() 1461 sg_tmp = sg_next(sg_tmp); in ocs_create_linked_list_from_sg()
|
| /linux/drivers/crypto/ ! |
| H A D | s5p-sss.c | 1125 struct scatterlist *sg_tmp = sg; in s5p_hash_prepare_sgs() local 1133 while (nbytes > 0 && sg_tmp) { in s5p_hash_prepare_sgs() 1135 if (skip >= sg_tmp->length) { in s5p_hash_prepare_sgs() 1136 skip -= sg_tmp->length; in s5p_hash_prepare_sgs() 1137 if (!sg_tmp->length) { in s5p_hash_prepare_sgs() 1142 if (!IS_ALIGNED(sg_tmp->length - skip, BUFLEN)) { in s5p_hash_prepare_sgs() 1147 if (nbytes < sg_tmp->length - skip) { in s5p_hash_prepare_sgs() 1152 nbytes -= sg_tmp->length - skip; in s5p_hash_prepare_sgs() 1156 sg_tmp = sg_next(sg_tmp); in s5p_hash_prepare_sgs()
|
| H A D | omap-sham.c | 732 struct scatterlist *sg_tmp = sg; in omap_sham_align_sgs() local 765 while (nbytes > 0 && sg_tmp) { in omap_sham_align_sgs() 782 if (page_zonenum(sg_page(sg_tmp)) != ZONE_DMA) { in omap_sham_align_sgs() 788 if (offset < sg_tmp->length) { in omap_sham_align_sgs() 789 if (!IS_ALIGNED(offset + sg_tmp->offset, 4)) { in omap_sham_align_sgs() 794 if (!IS_ALIGNED(sg_tmp->length - offset, bs)) { in omap_sham_align_sgs() 801 offset -= sg_tmp->length; in omap_sham_align_sgs() 807 nbytes -= sg_tmp->length; in omap_sham_align_sgs() 810 sg_tmp = sg_next(sg_tmp); in omap_sham_align_sgs()
|
| /linux/drivers/crypto/stm32/ ! |
| H A D | stm32-hash.c | 1140 struct scatterlist *sg_tmp = sg; in stm32_hash_align_sgs() local 1188 while (nbytes > 0 && sg_tmp) { in stm32_hash_align_sgs() 1204 if (offset < sg_tmp->length) { in stm32_hash_align_sgs() 1205 if (!IS_ALIGNED(offset + sg_tmp->offset, 4)) { in stm32_hash_align_sgs() 1210 if (!IS_ALIGNED(sg_tmp->length - offset, bs)) { in stm32_hash_align_sgs() 1217 offset -= sg_tmp->length; in stm32_hash_align_sgs() 1223 nbytes -= sg_tmp->length; in stm32_hash_align_sgs() 1226 sg_tmp = sg_next(sg_tmp); in stm32_hash_align_sgs()
|