| /linux/net/netfilter/ |
| H A D | nft_counter.c | 69 struct nft_counter __percpu *cpu_stats; in nft_counter_do_init() local 72 cpu_stats = alloc_percpu_gfp(struct nft_counter, GFP_KERNEL_ACCOUNT); in nft_counter_do_init() 73 if (cpu_stats == NULL) in nft_counter_do_init() 76 this_cpu = raw_cpu_ptr(cpu_stats); in nft_counter_do_init() 86 priv->counter = cpu_stats; in nft_counter_do_init() 254 struct nft_counter __percpu *cpu_stats; in nft_counter_clone() local 260 cpu_stats = alloc_percpu_gfp(struct nft_counter, gfp); in nft_counter_clone() 261 if (cpu_stats == NULL) in nft_counter_clone() 264 this_cpu = raw_cpu_ptr(cpu_stats); in nft_counter_clone() 268 priv_clone->counter = cpu_stats; in nft_counter_clone()
|
| H A D | nf_tables_api.c | 1948 struct nft_stats *cpu_stats, total; in nft_dump_stats() local 1959 cpu_stats = per_cpu_ptr(stats, cpu); in nft_dump_stats() 1961 seq = u64_stats_fetch_begin(&cpu_stats->syncp); in nft_dump_stats() 1962 pkts = cpu_stats->pkts; in nft_dump_stats() 1963 bytes = cpu_stats->bytes; in nft_dump_stats() 1964 } while (u64_stats_fetch_retry(&cpu_stats->syncp, seq)); in nft_dump_stats()
|
| /linux/net/core/ |
| H A D | drop_monitor.c | 1447 struct net_dm_stats *cpu_stats = &data->stats; in net_dm_stats_read() local 1452 start = u64_stats_fetch_begin(&cpu_stats->syncp); in net_dm_stats_read() 1453 dropped = u64_stats_read(&cpu_stats->dropped); in net_dm_stats_read() 1454 } while (u64_stats_fetch_retry(&cpu_stats->syncp, start)); in net_dm_stats_read() 1491 struct net_dm_stats *cpu_stats = &hw_data->stats; in net_dm_hw_stats_read() local 1496 start = u64_stats_fetch_begin(&cpu_stats->syncp); in net_dm_hw_stats_read() 1497 dropped = u64_stats_read(&cpu_stats->dropped); in net_dm_hw_stats_read() 1498 } while (u64_stats_fetch_retry(&cpu_stats->syncp, start)); in net_dm_hw_stats_read()
|
| /linux/net/devlink/ |
| H A D | trap.c | 163 struct devlink_stats *cpu_stats; in devlink_trap_stats_read() local 167 cpu_stats = per_cpu_ptr(trap_stats, i); in devlink_trap_stats_read() 169 start = u64_stats_fetch_begin(&cpu_stats->syncp); in devlink_trap_stats_read() 170 rx_packets = u64_stats_read(&cpu_stats->rx_packets); in devlink_trap_stats_read() 171 rx_bytes = u64_stats_read(&cpu_stats->rx_bytes); in devlink_trap_stats_read() 172 } while (u64_stats_fetch_retry(&cpu_stats->syncp, start)); in devlink_trap_stats_read()
|
| /linux/net/bridge/ |
| H A D | br_vlan.c | 1391 struct pcpu_sw_netstats *cpu_stats; in br_vlan_get_stats() local 1394 cpu_stats = per_cpu_ptr(v->stats, i); in br_vlan_get_stats() 1396 start = u64_stats_fetch_begin(&cpu_stats->syncp); in br_vlan_get_stats() 1397 rxpackets = u64_stats_read(&cpu_stats->rx_packets); in br_vlan_get_stats() 1398 rxbytes = u64_stats_read(&cpu_stats->rx_bytes); in br_vlan_get_stats() 1399 txbytes = u64_stats_read(&cpu_stats->tx_bytes); in br_vlan_get_stats() 1400 txpackets = u64_stats_read(&cpu_stats->tx_packets); in br_vlan_get_stats() 1401 } while (u64_stats_fetch_retry(&cpu_stats->syncp, start)); in br_vlan_get_stats()
|
| H A D | br_multicast.c | 5189 struct bridge_mcast_stats *cpu_stats = per_cpu_ptr(stats, i); in br_multicast_get_stats() local 5194 start = u64_stats_fetch_begin(&cpu_stats->syncp); in br_multicast_get_stats() 5195 u64_stats_copy(&temp, &cpu_stats->mstats, sizeof(temp)); in br_multicast_get_stats() 5196 } while (u64_stats_fetch_retry(&cpu_stats->syncp, start)); in br_multicast_get_stats()
|
| /linux/net/ipv4/ |
| H A D | nexthop.c | 674 struct nh_grp_entry_stats *cpu_stats; in nh_grp_entry_stats_inc() local 676 cpu_stats = get_cpu_ptr(nhge->stats); in nh_grp_entry_stats_inc() 677 u64_stats_update_begin(&cpu_stats->syncp); in nh_grp_entry_stats_inc() 678 u64_stats_inc(&cpu_stats->packets); in nh_grp_entry_stats_inc() 679 u64_stats_update_end(&cpu_stats->syncp); in nh_grp_entry_stats_inc() 680 put_cpu_ptr(cpu_stats); in nh_grp_entry_stats_inc() 691 struct nh_grp_entry_stats *cpu_stats; in nh_grp_entry_stats_read() local 695 cpu_stats = per_cpu_ptr(nhge->stats, i); in nh_grp_entry_stats_read() 697 start = u64_stats_fetch_begin(&cpu_stats->syncp); in nh_grp_entry_stats_read() 698 packets = u64_stats_read(&cpu_stats->packets); in nh_grp_entry_stats_read() [all …]
|
| /linux/drivers/net/ethernet/marvell/mvpp2/ |
| H A D | mvpp2_main.c | 2024 struct mvpp2_pcpu_stats *cpu_stats; in mvpp2_get_xdp_stats() local 2033 cpu_stats = per_cpu_ptr(port->stats, cpu); in mvpp2_get_xdp_stats() 2035 start = u64_stats_fetch_begin(&cpu_stats->syncp); in mvpp2_get_xdp_stats() 2036 xdp_redirect = cpu_stats->xdp_redirect; in mvpp2_get_xdp_stats() 2037 xdp_pass = cpu_stats->xdp_pass; in mvpp2_get_xdp_stats() 2038 xdp_drop = cpu_stats->xdp_drop; in mvpp2_get_xdp_stats() 2039 xdp_xmit = cpu_stats->xdp_xmit; in mvpp2_get_xdp_stats() 2040 xdp_xmit_err = cpu_stats->xdp_xmit_err; in mvpp2_get_xdp_stats() 2041 xdp_tx = cpu_stats->xdp_tx; in mvpp2_get_xdp_stats() 2042 xdp_tx_err = cpu_stats->xdp_tx_err; in mvpp2_get_xdp_stats() [all …]
|
| /linux/drivers/net/team/ |
| H A D | team_mode_loadbalance.c | 446 struct lb_stats *cpu_stats, in __lb_one_cpu_stats_add() argument 454 tmp.tx_bytes = cpu_stats->tx_bytes; in __lb_one_cpu_stats_add()
|
| /linux/drivers/net/ethernet/marvell/ |
| H A D | mvneta.c | 812 struct mvneta_pcpu_stats *cpu_stats; in mvneta_get_stats64() local 820 cpu_stats = per_cpu_ptr(pp->stats, cpu); in mvneta_get_stats64() 822 start = u64_stats_fetch_begin(&cpu_stats->syncp); in mvneta_get_stats64() 823 rx_packets = cpu_stats->es.ps.rx_packets; in mvneta_get_stats64() 824 rx_bytes = cpu_stats->es.ps.rx_bytes; in mvneta_get_stats64() 825 rx_dropped = cpu_stats->rx_dropped; in mvneta_get_stats64() 826 rx_errors = cpu_stats->rx_errors; in mvneta_get_stats64() 827 tx_packets = cpu_stats->es.ps.tx_packets; in mvneta_get_stats64() 828 tx_bytes = cpu_stats->es.ps.tx_bytes; in mvneta_get_stats64() 829 } while (u64_stats_fetch_retry(&cpu_stats->syncp, start)); in mvneta_get_stats64()
|
| /linux/mm/ |
| H A D | slab.h | 247 struct kmem_cache_stats __percpu *cpu_stats; member
|
| H A D | slub.c | 381 raw_cpu_inc(s->cpu_stats->stat[si]); in stat() 389 raw_cpu_add(s->cpu_stats->stat[si], v); in stat_add() 7452 s->cpu_stats = alloc_percpu(struct kmem_cache_stats); in alloc_kmem_cache_stats() 7454 if (!s->cpu_stats) in alloc_kmem_cache_stats() 7571 free_percpu(s->cpu_stats); in __kmem_cache_release() 9189 unsigned int x = per_cpu_ptr(s->cpu_stats, cpu)->stat[si]; in show_stat() 9215 per_cpu_ptr(s->cpu_stats, cpu)->stat[si] = 0; in clear_stat()
|
| /linux/kernel/locking/ |
| H A D | lockdep.c | 331 struct lock_class_stats *cpu_stats = in clear_lock_stats() local 334 memset(cpu_stats, 0, sizeof(struct lock_class_stats)); in clear_lock_stats()
|