Searched refs:softnet_data (Results 1 – 10 of 10) sorted by relevance
85 static u32 softnet_input_pkt_queue_len(struct softnet_data *sd) in softnet_input_pkt_queue_len()90 static u32 softnet_process_queue_len(struct softnet_data *sd) in softnet_process_queue_len()95 static struct softnet_data *softnet_get_online(loff_t *pos) in softnet_get_online()97 struct softnet_data *sd = NULL; in softnet_get_online()101 sd = &per_cpu(softnet_data, *pos); in softnet_get_online()125 struct softnet_data *sd = v; in softnet_seq_show()
231 static inline void backlog_lock_irq_save(struct softnet_data *sd, in backlog_lock_irq_save() 243 static inline void backlog_lock_irq_disable(struct softnet_data *sd) in backlog_lock_irq_disable() 251 static inline void backlog_unlock_irq_restore(struct softnet_data *sd, in backlog_unlock_irq_restore() 263 static inline void backlog_unlock_irq_enable(struct softnet_data *sd) in backlog_unlock_irq_enable() 462 DEFINE_PER_CPU_ALIGNED(struct softnet_data, softnet_data) = {465 EXPORT_PER_CPU_SYMBOL(softnet_data);3382 struct softnet_data *sd; in __netif_reschedule() 3386 sd = this_cpu_ptr(&softnet_data); in __netif_reschedule() 3457 skb->next = __this_cpu_read(softnet_data in dev_kfree_skb_irq_reason() [all...]
214 struct softnet_data *sd; in flow_limit_cpu_sysctl()229 sd = &per_cpu(softnet_data, i); in flow_limit_cpu_sysctl()253 sd = &per_cpu(softnet_data, i); in flow_limit_cpu_sysctl()
246 struct softnet_data *sd = &get_cpu_var(softnet_data); in zap_completion_queue()268 put_cpu_var(softnet_data); in zap_completion_queue()
180 static inline u32 rps_input_queue_tail_incr(struct softnet_data *sd) in rps_input_queue_tail_incr()196 static inline void rps_input_queue_head_add(struct softnet_data *sd, int val) in rps_input_queue_head_add()203 static inline void rps_input_queue_head_incr(struct softnet_data *sd) in rps_input_queue_head_incr()
2068 void xfrm_dev_backlog(struct softnet_data *sd);2133 static inline void xfrm_dev_backlog(struct softnet_data *sd) in xfrm_dev_backlog()
120 struct softnet_data *sd; in validate_xmit_xfrm()149 sd = this_cpu_ptr(&softnet_data); in validate_xmit_xfrm()472 struct softnet_data *sd; in xfrm_dev_resume()485 sd = this_cpu_ptr(&softnet_data); in xfrm_dev_resume()494 void xfrm_dev_backlog(struct softnet_data *sd) in xfrm_dev_backlog()
21 return this_cpu_ptr(&softnet_data.xmit.nf_dup_skb_recursion); in nf_get_nf_dup_skb_recursion()
3516 struct softnet_data { struct3525 struct softnet_data *rps_ipi_list; argument3551 struct softnet_data *rps_ipi_next; argument3572 DECLARE_PER_CPU_ALIGNED(struct softnet_data, softnet_data); argument3585 return this_cpu_read(softnet_data.xmit.recursion); in dev_recursion_level()3590 return unlikely(__this_cpu_read(softnet_data.xmit.recursion) > in dev_xmit_recursion()3596 __this_cpu_inc(softnet_data.xmit.recursion); in dev_xmit_recursion_inc()3601 __this_cpu_dec(softnet_data.xmit.recursion); in dev_xmit_recursion_dec()5301 __this_cpu_write(softnet_data.xmit.more, more); in netdev_xmit_set_more()5306 return __this_cpu_read(softnet_data.xmit.more); in netdev_xmit_more()
435 xmit = this_cpu_ptr(&softnet_data.xmit); in tcf_mirred_act()