Lines Matching refs:flaglen
344 __le32 flaglen; member
351 __le32 flaglen; member
952 return le32_to_cpu(prd->flaglen) in nv_descr_getlength()
958 return le32_to_cpu(prd->flaglen) & LEN_MASK_V2; in nv_descr_getlength_ex()
1833 np->put_rx.orig->flaglen = cpu_to_le32(np->rx_buf_sz | NV_RX_AVAIL); in nv_alloc_rx()
1875 np->put_rx.ex->flaglen = cpu_to_le32(np->rx_buf_sz | NV_RX2_AVAIL); in nv_alloc_rx_optimized()
1918 np->rx_ring.orig[i].flaglen = 0; in nv_init_rx()
1921 np->rx_ring.ex[i].flaglen = 0; in nv_init_rx()
1954 np->tx_ring.orig[i].flaglen = 0; in nv_init_tx()
1957 np->tx_ring.ex[i].flaglen = 0; in nv_init_tx()
2017 np->tx_ring.orig[i].flaglen = 0; in nv_drain_tx()
2020 np->tx_ring.ex[i].flaglen = 0; in nv_drain_tx()
2048 np->rx_ring.orig[i].flaglen = 0; in nv_drain_rx()
2051 np->rx_ring.ex[i].flaglen = 0; in nv_drain_rx()
2257 put_tx->flaglen = cpu_to_le32((bcnt-1) | tx_flags); in nv_start_xmit()
2307 put_tx->flaglen = cpu_to_le32((bcnt-1) | tx_flags); in nv_start_xmit()
2329 prev_tx->flaglen |= cpu_to_le32(tx_flags_extra); in nv_start_xmit()
2343 start_tx->flaglen |= cpu_to_le32(tx_flags | tx_flags_extra); in nv_start_xmit()
2435 put_tx->flaglen = cpu_to_le32((bcnt-1) | tx_flags); in nv_start_xmit_optimized()
2485 put_tx->flaglen = cpu_to_le32((bcnt-1) | tx_flags); in nv_start_xmit_optimized()
2507 prev_tx->flaglen |= cpu_to_le32(NV_TX2_LASTPACKET); in nv_start_xmit_optimized()
2548 start_tx->flaglen |= cpu_to_le32(tx_flags | tx_flags_extra); in nv_start_xmit_optimized()
2575 np->tx_change_owner->first_tx_desc->flaglen |= in nv_tx_flip_ownership()
2601 !((flags = le32_to_cpu(np->get_tx.orig->flaglen)) & NV_TX_VALID) && in nv_tx_done()
2671 !((flags = le32_to_cpu(np->get_tx.ex->flaglen)) & NV_TX2_VALID) && in nv_tx_done_optimized()
2761 le32_to_cpu(np->tx_ring.orig[i].flaglen), in nv_tx_timeout()
2763 le32_to_cpu(np->tx_ring.orig[i+1].flaglen), in nv_tx_timeout()
2765 le32_to_cpu(np->tx_ring.orig[i+2].flaglen), in nv_tx_timeout()
2767 le32_to_cpu(np->tx_ring.orig[i+3].flaglen)); in nv_tx_timeout()
2777 le32_to_cpu(np->tx_ring.ex[i].flaglen), in nv_tx_timeout()
2780 le32_to_cpu(np->tx_ring.ex[i+1].flaglen), in nv_tx_timeout()
2783 le32_to_cpu(np->tx_ring.ex[i+2].flaglen), in nv_tx_timeout()
2786 le32_to_cpu(np->tx_ring.ex[i+3].flaglen)); in nv_tx_timeout()
2886 !((flags = le32_to_cpu(np->get_rx.orig->flaglen)) & NV_RX_AVAIL) && in nv_rx_process()
2988 !((flags = le32_to_cpu(np->get_rx.ex->flaglen)) & NV_RX2_AVAIL) && in nv_rx_process_optimized()
5184 np->tx_ring.orig[0].flaglen = cpu_to_le32((pkt_len-1) | np->tx_flags | tx_flags_extra); in nv_loopback_test()
5188 np->tx_ring.ex[0].flaglen = cpu_to_le32((pkt_len-1) | np->tx_flags | tx_flags_extra); in nv_loopback_test()
5197 flags = le32_to_cpu(np->rx_ring.orig[0].flaglen); in nv_loopback_test()
5201 flags = le32_to_cpu(np->rx_ring.ex[0].flaglen); in nv_loopback_test()