Home
last modified time | relevance | path

Searched refs:tbl (Results 1 – 25 of 359) sorted by relevance

12345678910>>...15

/linux/fs/nfs/
H A Dnfs4session.c27 static void nfs4_init_slot_table(struct nfs4_slot_table *tbl, const char *queue) in nfs4_init_slot_table() argument
29 tbl->highest_used_slotid = NFS4_NO_SLOT; in nfs4_init_slot_table()
30 spin_lock_init(&tbl->slot_tbl_lock); in nfs4_init_slot_table()
31 rpc_init_priority_wait_queue(&tbl->slot_tbl_waitq, queue); in nfs4_init_slot_table()
32 init_waitqueue_head(&tbl->slot_waitq); in nfs4_init_slot_table()
33 init_completion(&tbl->complete); in nfs4_init_slot_table()
39 static void nfs4_shrink_slot_table(struct nfs4_slot_table *tbl, u32 newsize) in nfs4_shrink_slot_table() argument
42 if (newsize >= tbl->max_slots) in nfs4_shrink_slot_table()
45 p = &tbl->slots; in nfs4_shrink_slot_table()
53 tbl->max_slots--; in nfs4_shrink_slot_table()
[all …]
H A Dnfs4session.h83 extern int nfs4_setup_slot_table(struct nfs4_slot_table *tbl,
85 extern void nfs4_shutdown_slot_table(struct nfs4_slot_table *tbl);
86 extern struct nfs4_slot *nfs4_alloc_slot(struct nfs4_slot_table *tbl);
87 extern struct nfs4_slot *nfs4_lookup_slot(struct nfs4_slot_table *tbl, u32 slotid);
88 extern int nfs4_slot_wait_on_seqid(struct nfs4_slot_table *tbl,
91 extern bool nfs4_try_to_lock_slot(struct nfs4_slot_table *tbl, struct nfs4_slot *slot);
92 extern void nfs4_free_slot(struct nfs4_slot_table *tbl, struct nfs4_slot *slot);
93 extern void nfs4_slot_tbl_drain_complete(struct nfs4_slot_table *tbl);
94 bool nfs41_wake_and_assign_slot(struct nfs4_slot_table *tbl,
96 void nfs41_wake_slot_table(struct nfs4_slot_table *tbl);
[all …]
/linux/drivers/net/ethernet/mellanox/mlx5/core/steering/sws/
H A Ddr_table.c28 action->dest_tbl->tbl->rx.s_anchor->chunk : in dr_table_set_miss_action_nic()
29 action->dest_tbl->tbl->tx.s_anchor->chunk; in dr_table_set_miss_action_nic()
46 int mlx5dr_table_set_miss_action(struct mlx5dr_table *tbl, in mlx5dr_table_set_miss_action() argument
54 mlx5dr_domain_lock(tbl->dmn); in mlx5dr_table_set_miss_action()
56 if (tbl->dmn->type == MLX5DR_DOMAIN_TYPE_NIC_RX || in mlx5dr_table_set_miss_action()
57 tbl->dmn->type == MLX5DR_DOMAIN_TYPE_FDB) { in mlx5dr_table_set_miss_action()
58 ret = dr_table_set_miss_action_nic(tbl->dmn, &tbl->rx, action); in mlx5dr_table_set_miss_action()
63 if (tbl->dmn->type == MLX5DR_DOMAIN_TYPE_NIC_TX || in mlx5dr_table_set_miss_action()
64 tbl->dmn->type == MLX5DR_DOMAIN_TYPE_FDB) { in mlx5dr_table_set_miss_action()
65 ret = dr_table_set_miss_action_nic(tbl->dmn, &tbl->tx, action); in mlx5dr_table_set_miss_action()
[all …]
/linux/drivers/net/ethernet/mellanox/mlx5/core/steering/hws/
H A Dtable.c6 u32 mlx5hws_table_get_id(struct mlx5hws_table *tbl) in mlx5hws_table_get_id() argument
8 return tbl->ft_id; in mlx5hws_table_get_id()
11 static void hws_table_init_next_ft_attr(struct mlx5hws_table *tbl, in hws_table_init_next_ft_attr() argument
15 ft_attr->type = tbl->fw_ft_type; in hws_table_init_next_ft_attr()
16 if (tbl->type == MLX5HWS_TABLE_TYPE_FDB) in hws_table_init_next_ft_attr()
17 ft_attr->level = tbl->ctx->caps->fdb_ft.max_level - 1; in hws_table_init_next_ft_attr()
19 ft_attr->level = tbl->ctx->caps->nic_ft.max_level - 1; in hws_table_init_next_ft_attr()
25 static void hws_table_set_cap_attr(struct mlx5hws_table *tbl, in hws_table_set_cap_attr() argument
36 if (!MLX5_CAP_ESW_FLOWTABLE(tbl->ctx->mdev, fdb_dynamic_tunnel)) { in hws_table_set_cap_attr()
42 static int hws_table_up_default_fdb_miss_tbl(struct mlx5hws_table *tbl) in hws_table_up_default_fdb_miss_tbl() argument
[all …]
H A Dmatcher.c23 mlx5hws_table_destroy_default_ft(matcher->tbl, matcher->end_ft_id); in hws_matcher_destroy_end_ft()
26 int mlx5hws_matcher_update_end_ft_isolated(struct mlx5hws_table *tbl, in mlx5hws_matcher_update_end_ft_isolated() argument
31 if (list_empty(&tbl->matchers_list)) in mlx5hws_matcher_update_end_ft_isolated()
37 list_for_each_entry(tmp_matcher, &tbl->matchers_list, list_node) in mlx5hws_matcher_update_end_ft_isolated()
40 tmp_matcher = list_last_entry(&tbl->matchers_list, in mlx5hws_matcher_update_end_ft_isolated()
44 return mlx5hws_table_ft_set_next_ft(tbl->ctx, in mlx5hws_matcher_update_end_ft_isolated()
46 tbl->fw_ft_type, in mlx5hws_matcher_update_end_ft_isolated()
52 struct mlx5hws_table *tbl = matcher->tbl; in hws_matcher_connect_end_ft_isolated() local
57 ret = mlx5hws_table_ft_set_next_rtc(tbl->ctx, in hws_matcher_connect_end_ft_isolated()
59 matcher->tbl->fw_ft_type, in hws_matcher_connect_end_ft_isolated()
[all …]
/linux/arch/powerpc/kernel/
H A Diommu.c47 struct iommu_table *tbl = data; in iommu_debugfs_weight_get() local
48 *val = bitmap_weight(tbl->it_map, tbl->it_size); in iommu_debugfs_weight_get()
53 static void iommu_debugfs_add(struct iommu_table *tbl) in iommu_debugfs_add() argument
58 sprintf(name, "%08lx", tbl->it_index); in iommu_debugfs_add()
61 debugfs_create_file_unsafe("weight", 0400, liobn_entry, tbl, &iommu_debugfs_fops_weight); in iommu_debugfs_add()
62 debugfs_create_ulong("it_size", 0400, liobn_entry, &tbl->it_size); in iommu_debugfs_add()
63 debugfs_create_ulong("it_page_shift", 0400, liobn_entry, &tbl->it_page_shift); in iommu_debugfs_add()
64 debugfs_create_ulong("it_reserved_start", 0400, liobn_entry, &tbl->it_reserved_start); in iommu_debugfs_add()
65 debugfs_create_ulong("it_reserved_end", 0400, liobn_entry, &tbl->it_reserved_end); in iommu_debugfs_add()
66 debugfs_create_ulong("it_indirect_levels", 0400, liobn_entry, &tbl->it_indirect_levels); in iommu_debugfs_add()
[all …]
/linux/drivers/infiniband/hw/vmw_pvrdma/
H A Dpvrdma_doorbell.c56 struct pvrdma_id_table *tbl = &dev->uar_table.tbl; in pvrdma_uar_table_init() local
61 tbl->last = 0; in pvrdma_uar_table_init()
62 tbl->top = 0; in pvrdma_uar_table_init()
63 tbl->max = num; in pvrdma_uar_table_init()
64 tbl->mask = mask; in pvrdma_uar_table_init()
65 spin_lock_init(&tbl->lock); in pvrdma_uar_table_init()
66 tbl->table = bitmap_zalloc(num, GFP_KERNEL); in pvrdma_uar_table_init()
67 if (!tbl->table) in pvrdma_uar_table_init()
71 __set_bit(0, tbl->table); in pvrdma_uar_table_init()
78 struct pvrdma_id_table *tbl = &dev->uar_table.tbl; in pvrdma_uar_table_cleanup() local
[all …]
/linux/arch/powerpc/platforms/powernv/
H A Dpci-ioda-tce.c48 void pnv_pci_setup_iommu_table(struct iommu_table *tbl, in pnv_pci_setup_iommu_table() argument
52 tbl->it_blocksize = 16; in pnv_pci_setup_iommu_table()
53 tbl->it_base = (unsigned long)tce_mem; in pnv_pci_setup_iommu_table()
54 tbl->it_page_shift = page_shift; in pnv_pci_setup_iommu_table()
55 tbl->it_offset = dma_offset >> tbl->it_page_shift; in pnv_pci_setup_iommu_table()
56 tbl->it_index = 0; in pnv_pci_setup_iommu_table()
57 tbl->it_size = tce_size >> 3; in pnv_pci_setup_iommu_table()
58 tbl->it_busno = 0; in pnv_pci_setup_iommu_table()
59 tbl->it_type = TCE_PCI; in pnv_pci_setup_iommu_table()
83 static __be64 *pnv_tce(struct iommu_table *tbl, bool user, long idx, bool alloc) in pnv_tce() argument
[all …]
/linux/include/linux/
H A Drhashtable.h119 static inline unsigned int rht_bucket_index(const struct bucket_table *tbl, in rht_bucket_index() argument
122 return hash & (tbl->size - 1); in rht_bucket_index()
156 struct rhashtable *ht, const struct bucket_table *tbl, in rht_key_hashfn() argument
159 unsigned int hash = rht_key_get_hash(ht, key, params, tbl->hash_rnd); in rht_key_hashfn()
161 return rht_bucket_index(tbl, hash); in rht_key_hashfn()
165 struct rhashtable *ht, const struct bucket_table *tbl, in rht_head_hashfn() argument
171 rht_bucket_index(tbl, params.obj_hashfn(ptr, params.key_len ?: in rht_head_hashfn()
173 tbl->hash_rnd)) : in rht_head_hashfn()
174 rht_key_hashfn(ht, tbl, ptr + params.key_offset, params); in rht_head_hashfn()
180 * @tbl
183 rht_grow_above_75(const struct rhashtable * ht,const struct bucket_table * tbl) rht_grow_above_75() argument
196 rht_shrink_below_30(const struct rhashtable * ht,const struct bucket_table * tbl) rht_shrink_below_30() argument
209 rht_grow_above_100(const struct rhashtable * ht,const struct bucket_table * tbl) rht_grow_above_100() argument
221 rht_grow_above_max(const struct rhashtable * ht,const struct bucket_table * tbl) rht_grow_above_max() argument
235 lockdep_rht_bucket_is_held(const struct bucket_table * tbl,u32 hash) lockdep_rht_bucket_is_held() argument
277 rht_dereference_bucket(p,tbl,hash) global() argument
280 rht_dereference_bucket_rcu(p,tbl,hash) global() argument
287 rht_bucket(const struct bucket_table * tbl,unsigned int hash) rht_bucket() argument
294 rht_bucket_var(struct bucket_table * tbl,unsigned int hash) rht_bucket_var() argument
301 rht_bucket_insert(struct rhashtable * ht,struct bucket_table * tbl,unsigned int hash) rht_bucket_insert() argument
326 rht_lock(struct bucket_table * tbl,struct rhash_lock_head __rcu ** bkt) rht_lock() argument
337 rht_lock_nested(struct bucket_table * tbl,struct rhash_lock_head __rcu ** bucket,unsigned int subclass) rht_lock_nested() argument
349 rht_unlock(struct bucket_table * tbl,struct rhash_lock_head __rcu ** bkt,unsigned long flags) rht_unlock() argument
401 rht_ptr(struct rhash_lock_head __rcu * const * bkt,struct bucket_table * tbl,unsigned int hash) rht_ptr() argument
423 rht_assign_unlock(struct bucket_table * tbl,struct rhash_lock_head __rcu ** bkt,struct rhash_head * obj,unsigned long flags) rht_assign_unlock() argument
444 rht_for_each_from(pos,head,tbl,hash) global() argument
455 rht_for_each(pos,tbl,hash) global() argument
468 rht_for_each_entry_from(tpos,pos,head,tbl,hash,member) global() argument
481 rht_for_each_entry(tpos,pos,tbl,hash,member) global() argument
498 rht_for_each_entry_safe(tpos,pos,next,tbl,hash,member) global() argument
518 rht_for_each_rcu_from(pos,head,tbl,hash) global() argument
534 rht_for_each_rcu(pos,tbl,hash) global() argument
553 rht_for_each_entry_rcu_from(tpos,pos,head,tbl,hash,member) global() argument
571 rht_for_each_entry_rcu(tpos,pos,tbl,hash,member) global() argument
621 struct bucket_table *tbl; __rhashtable_lookup() local
761 struct bucket_table *tbl; __rhashtable_insert_fast() local
1042 __rhashtable_remove_fast_one(struct rhashtable * ht,struct bucket_table * tbl,struct rhash_head * obj,const struct rhashtable_params params,bool rhlist) __rhashtable_remove_fast_one() argument
1127 struct bucket_table *tbl; __rhashtable_remove_fast() local
1195 __rhashtable_replace_fast(struct rhashtable * ht,struct bucket_table * tbl,struct rhash_head * obj_old,struct rhash_head * obj_new,const struct rhashtable_params params) __rhashtable_replace_fast() argument
1262 struct bucket_table *tbl; rhashtable_replace_fast() local
[all...]
/linux/lib/
H A Drhashtable.c38 const struct bucket_table *tbl, in head_hashfn() argument
41 return rht_head_hashfn(ht, tbl, he, ht->p); in head_hashfn()
53 int lockdep_rht_bucket_is_held(const struct bucket_table *tbl, u32 hash) in lockdep_rht_bucket_is_held() argument
57 if (unlikely(tbl->nest)) in lockdep_rht_bucket_is_held()
59 return bit_spin_is_locked(0, (unsigned long *)&tbl->buckets[hash]); in lockdep_rht_bucket_is_held()
67 const struct bucket_table *tbl) in nested_table_top() argument
70 * because it's set at the same time as tbl->nest. in nested_table_top()
72 return (void *)rcu_dereference_protected(tbl->buckets[0], 1); in nested_table_top()
94 static void nested_bucket_table_free(const struct bucket_table *tbl) in nested_bucket_table_free() argument
96 unsigned int size = tbl in nested_bucket_table_free()
109 bucket_table_free(const struct bucket_table * tbl) bucket_table_free() argument
153 struct bucket_table *tbl; nested_bucket_table_alloc() local
181 struct bucket_table *tbl = NULL; bucket_table_alloc() local
216 rhashtable_last_table(struct rhashtable * ht,struct bucket_table * tbl) rhashtable_last_table() argument
417 struct bucket_table *tbl; rht_deferred_worker() local
447 rhashtable_insert_rehash(struct rhashtable * ht,struct bucket_table * tbl) rhashtable_insert_rehash() argument
496 rhashtable_lookup_one(struct rhashtable * ht,struct rhash_lock_head __rcu ** bkt,struct bucket_table * tbl,unsigned int hash,const void * key,struct rhash_head * obj) rhashtable_lookup_one() argument
547 rhashtable_insert_one(struct rhashtable * ht,struct rhash_lock_head __rcu ** bkt,struct bucket_table * tbl,unsigned int hash,struct rhash_head * obj,void * data) rhashtable_insert_one() argument
594 struct bucket_table *tbl; rhashtable_try_insert() local
801 struct bucket_table *tbl = iter->walker.tbl; __rhashtable_walk_find_next() local
946 struct bucket_table *tbl = iter->walker.tbl; rhashtable_walk_stop() local
1030 struct bucket_table *tbl; rhashtable_init_noprof() local
1148 struct bucket_table *tbl, *next_tbl; rhashtable_free_and_destroy() local
1189 __rht_bucket_nested(const struct bucket_table * tbl,unsigned int hash) __rht_bucket_nested() argument
1218 rht_bucket_nested(const struct bucket_table * tbl,unsigned int hash) rht_bucket_nested() argument
1229 rht_bucket_nested_insert(struct rhashtable * ht,struct bucket_table * tbl,unsigned int hash) rht_bucket_nested_insert() argument
[all...]
/linux/drivers/net/wireless/intel/iwlwifi/dvm/
H A Drs.c414 static s32 get_expected_tpt(struct iwl_scale_tbl_info *tbl, int rs_index) in get_expected_tpt() argument
416 if (tbl->expected_tpt) in get_expected_tpt()
417 return tbl->expected_tpt[rs_index]; in get_expected_tpt()
428 static int rs_collect_tx_data(struct iwl_scale_tbl_info *tbl, in rs_collect_tx_data() argument
439 window = &(tbl->win[scale_index]); in rs_collect_tx_data()
442 tpt = get_expected_tpt(tbl, scale_index); in rs_collect_tx_data()
507 struct iwl_scale_tbl_info *tbl, in rate_n_flags_from_tbl() argument
512 if (is_legacy(tbl->lq_type)) { in rate_n_flags_from_tbl()
517 } else if (is_Ht(tbl->lq_type)) { in rate_n_flags_from_tbl()
524 if (is_siso(tbl->lq_type)) in rate_n_flags_from_tbl()
[all …]
/linux/arch/powerpc/include/asm/
H A Diommu.h45 int (*set)(struct iommu_table *tbl,
56 int (*xchg_no_kill)(struct iommu_table *tbl,
61 void (*tce_kill)(struct iommu_table *tbl,
65 __be64 *(*useraddrptr)(struct iommu_table *tbl, long index, bool alloc);
67 void (*clear)(struct iommu_table *tbl,
70 unsigned long (*get)(struct iommu_table *tbl, long index);
71 void (*flush)(struct iommu_table *tbl);
72 void (*free)(struct iommu_table *tbl);
122 #define IOMMU_TABLE_USERSPACE_ENTRY_RO(tbl, entry) \ argument
123 ((tbl)->it_ops->useraddrptr((tbl), (entry), false))
[all …]
/linux/net/netfilter/ipvs/
H A Dip_vs_lblc.c166 ip_vs_lblc_hash(struct ip_vs_lblc_table *tbl, struct ip_vs_lblc_entry *en) in ip_vs_lblc_hash() argument
170 hlist_add_head_rcu(&en->list, &tbl->bucket[hash]); in ip_vs_lblc_hash()
171 atomic_inc(&tbl->entries); in ip_vs_lblc_hash()
177 ip_vs_lblc_get(int af, struct ip_vs_lblc_table *tbl, in ip_vs_lblc_get() argument
183 hlist_for_each_entry_rcu(en, &tbl->bucket[hash], list) in ip_vs_lblc_get()
196 ip_vs_lblc_new(struct ip_vs_lblc_table *tbl, const union nf_inet_addr *daddr, in ip_vs_lblc_new() argument
201 en = ip_vs_lblc_get(af, tbl, daddr); in ip_vs_lblc_new()
218 ip_vs_lblc_hash(tbl, en); in ip_vs_lblc_new()
229 struct ip_vs_lblc_table *tbl = svc->sched_data; in ip_vs_lblc_flush() local
235 tbl->dead = true; in ip_vs_lblc_flush()
[all …]
H A Dip_vs_lblcr.c329 ip_vs_lblcr_hash(struct ip_vs_lblcr_table *tbl, struct ip_vs_lblcr_entry *en) in ip_vs_lblcr_hash() argument
333 hlist_add_head_rcu(&en->list, &tbl->bucket[hash]); in ip_vs_lblcr_hash()
334 atomic_inc(&tbl->entries); in ip_vs_lblcr_hash()
340 ip_vs_lblcr_get(int af, struct ip_vs_lblcr_table *tbl, in ip_vs_lblcr_get() argument
346 hlist_for_each_entry_rcu(en, &tbl->bucket[hash], list) in ip_vs_lblcr_get()
359 ip_vs_lblcr_new(struct ip_vs_lblcr_table *tbl, const union nf_inet_addr *daddr, in ip_vs_lblcr_new() argument
364 en = ip_vs_lblcr_get(af, tbl, daddr); in ip_vs_lblcr_new()
380 ip_vs_lblcr_hash(tbl, en); in ip_vs_lblcr_new()
395 struct ip_vs_lblcr_table *tbl = svc->sched_data; in ip_vs_lblcr_flush() local
401 tbl->dead = true; in ip_vs_lblcr_flush()
[all …]
/linux/drivers/net/wireless/intel/iwlegacy/
H A D4965-rs.c383 il4965_get_expected_tpt(struct il_scale_tbl_info *tbl, int rs_idx) in il4965_get_expected_tpt() argument
385 if (tbl->expected_tpt) in il4965_get_expected_tpt()
386 return tbl->expected_tpt[rs_idx]; in il4965_get_expected_tpt()
398 il4965_rs_collect_tx_data(struct il_scale_tbl_info *tbl, int scale_idx, in il4965_rs_collect_tx_data() argument
409 win = &(tbl->win[scale_idx]); in il4965_rs_collect_tx_data()
412 tpt = il4965_get_expected_tpt(tbl, scale_idx); in il4965_rs_collect_tx_data()
476 il4965_rate_n_flags_from_tbl(struct il_priv *il, struct il_scale_tbl_info *tbl, in il4965_rate_n_flags_from_tbl() argument
481 if (is_legacy(tbl->lq_type)) { in il4965_rate_n_flags_from_tbl()
486 } else if (is_Ht(tbl->lq_type)) { in il4965_rate_n_flags_from_tbl()
493 if (is_siso(tbl->lq_type)) in il4965_rate_n_flags_from_tbl()
[all …]
/linux/drivers/net/wireless/marvell/mwifiex/
H A D11n_rxreorder.c98 struct mwifiex_rx_reorder_tbl *tbl, in mwifiex_11n_dispatch_pkt_until_start_win() argument
108 pkt_to_send = (start_win > tbl->start_win) ? in mwifiex_11n_dispatch_pkt_until_start_win()
109 min((start_win - tbl->start_win), tbl->win_size) : in mwifiex_11n_dispatch_pkt_until_start_win()
110 tbl->win_size; in mwifiex_11n_dispatch_pkt_until_start_win()
113 if (tbl->rx_reorder_ptr[i]) { in mwifiex_11n_dispatch_pkt_until_start_win()
114 skb = tbl->rx_reorder_ptr[i]; in mwifiex_11n_dispatch_pkt_until_start_win()
116 tbl->rx_reorder_ptr[i] = NULL; in mwifiex_11n_dispatch_pkt_until_start_win()
124 for (i = 0; i < tbl->win_size - pkt_to_send; ++i) { in mwifiex_11n_dispatch_pkt_until_start_win()
125 tbl->rx_reorder_ptr[i] = tbl->rx_reorder_ptr[pkt_to_send + i]; in mwifiex_11n_dispatch_pkt_until_start_win()
126 tbl->rx_reorder_ptr[pkt_to_send + i] = NULL; in mwifiex_11n_dispatch_pkt_until_start_win()
[all …]
/linux/drivers/firmware/efi/
H A Dmemattr.c24 efi_memory_attributes_table_t *tbl; in efi_memattr_init() local
30 tbl = early_memremap(efi_mem_attr_table, sizeof(*tbl)); in efi_memattr_init()
31 if (!tbl) { in efi_memattr_init()
37 if (tbl->version > 2) { in efi_memattr_init()
39 tbl->version); in efi_memattr_init()
51 size = tbl->num_entries * tbl->desc_size; in efi_memattr_init()
54 tbl->version, tbl->desc_size, tbl->num_entries); in efi_memattr_init()
58 tbl_size = sizeof(*tbl) + size; in efi_memattr_init()
63 early_memunmap(tbl, sizeof(*tbl)); in efi_memattr_init()
146 efi_memory_attributes_table_t *tbl; in efi_memattr_apply_permissions() local
[all …]
/linux/net/core/
H A Dneighbour.c56 static void pneigh_ifdown(struct neigh_table *tbl, struct net_device *dev,
141 atomic_dec(&n->tbl->gc_entries); in neigh_mark_dead()
151 spin_lock_bh(&n->tbl->lock); in neigh_update_gc_list()
166 atomic_dec(&n->tbl->gc_entries); in neigh_update_gc_list()
169 list_add_tail(&n->gc_list, &n->tbl->gc_list); in neigh_update_gc_list()
170 atomic_inc(&n->tbl->gc_entries); in neigh_update_gc_list()
174 spin_unlock_bh(&n->tbl->lock); in neigh_update_gc_list()
181 spin_lock_bh(&n->tbl->lock); in neigh_update_managed_list()
192 list_add_tail(&n->managed_list, &n->tbl->managed_list); in neigh_update_managed_list()
195 spin_unlock_bh(&n->tbl->lock); in neigh_update_managed_list()
[all …]
/linux/tools/perf/
H A Dcheck-headers.sh198 check_2 tools/perf/arch/x86/entry/syscalls/syscall_32.tbl arch/x86/entry/syscalls/syscall_32.tbl
199 check_2 tools/perf/arch/x86/entry/syscalls/syscall_64.tbl arch/x86/entry/syscalls/syscall_64.tbl
200 check_2 tools/perf/arch/powerpc/entry/syscalls/syscall.tbl arch/powerpc/kernel/syscalls/syscall.tbl
201 check_2 tools/perf/arch/s390/entry/syscalls/syscall.tbl arch/s390/kernel/syscalls/syscall.tbl
202 …eck_2 tools/perf/arch/mips/entry/syscalls/syscall_n64.tbl arch/mips/kernel/syscalls/syscall_n64.tbl
203 check_2 tools/perf/arch/arm/entry/syscalls/syscall.tbl arch/arm/tools/syscall.tbl
204 check_2 tools/perf/arch/sh/entry/syscalls/syscall.tbl arch/sh/kernel/syscalls/syscall.tbl
205 check_2 tools/perf/arch/sparc/entry/syscalls/syscall.tbl arch/sparc/kernel/syscalls/syscall.tbl
206 check_2 tools/perf/arch/xtensa/entry/syscalls/syscall.tbl arch/xtensa/kernel/syscalls/syscall.tbl
207 check_2 tools/perf/arch/alpha/entry/syscalls/syscall.tbl arch/alpha/entry/syscalls/syscall.tbl
[all …]
/linux/drivers/vfio/
H A Dvfio_iommu_spapr_tce.c209 struct iommu_table *tbl = container->tables[i]; in tce_iommu_find_table() local
211 if (tbl) { in tce_iommu_find_table()
212 unsigned long entry = ioba >> tbl->it_page_shift; in tce_iommu_find_table()
213 unsigned long start = tbl->it_offset; in tce_iommu_find_table()
214 unsigned long end = start + tbl->it_size; in tce_iommu_find_table()
217 *ptbl = tbl; in tce_iommu_find_table()
339 struct iommu_table *tbl,
342 struct iommu_table *tbl);
362 struct iommu_table *tbl = container->tables[i]; in tce_iommu_release() local
364 if (!tbl) in tce_iommu_release()
[all …]
/linux/ipc/
H A Dipc_sysctl.c247 struct ctl_table *tbl; in setup_ipc_sysctls() local
251 tbl = kmemdup(ipc_sysctls, sizeof(ipc_sysctls), GFP_KERNEL); in setup_ipc_sysctls()
252 if (tbl) { in setup_ipc_sysctls()
256 if (tbl[i].data == &init_ipc_ns.shm_ctlmax) in setup_ipc_sysctls()
257 tbl[i].data = &ns->shm_ctlmax; in setup_ipc_sysctls()
259 else if (tbl[i].data == &init_ipc_ns.shm_ctlall) in setup_ipc_sysctls()
260 tbl[i].data = &ns->shm_ctlall; in setup_ipc_sysctls()
262 else if (tbl[i].data == &init_ipc_ns.shm_ctlmni) in setup_ipc_sysctls()
263 tbl[i].data = &ns->shm_ctlmni; in setup_ipc_sysctls()
265 else if (tbl[i].data == &init_ipc_ns.shm_rmid_forced) in setup_ipc_sysctls()
[all …]
H A Dmq_sysctl.c119 struct ctl_table *tbl; in setup_mq_sysctls() local
123 tbl = kmemdup(mq_sysctls, sizeof(mq_sysctls), GFP_KERNEL); in setup_mq_sysctls()
124 if (tbl) { in setup_mq_sysctls()
128 if (tbl[i].data == &init_ipc_ns.mq_queues_max) in setup_mq_sysctls()
129 tbl[i].data = &ns->mq_queues_max; in setup_mq_sysctls()
131 else if (tbl[i].data == &init_ipc_ns.mq_msg_max) in setup_mq_sysctls()
132 tbl[i].data = &ns->mq_msg_max; in setup_mq_sysctls()
134 else if (tbl[i].data == &init_ipc_ns.mq_msgsize_max) in setup_mq_sysctls()
135 tbl[i].data = &ns->mq_msgsize_max; in setup_mq_sysctls()
137 else if (tbl[i].data == &init_ipc_ns.mq_msg_default) in setup_mq_sysctls()
[all …]
/linux/net/mac80211/
H A Dmesh_pathtbl.c20 static void mesh_path_free_rcu(struct mesh_table *tbl, struct mesh_path *mpath);
82 struct mesh_table *tbl = tblptr; in mesh_path_rht_free() local
84 mesh_path_free_rcu(tbl, mpath); in mesh_path_rht_free()
87 static void mesh_table_init(struct mesh_table *tbl) in mesh_table_init() argument
89 INIT_HLIST_HEAD(&tbl->known_gates); in mesh_table_init()
90 INIT_HLIST_HEAD(&tbl->walk_head); in mesh_table_init()
91 atomic_set(&tbl->entries, 0); in mesh_table_init()
92 spin_lock_init(&tbl->gates_lock); in mesh_table_init()
93 spin_lock_init(&tbl->walk_lock); in mesh_table_init()
98 WARN_ON(rhashtable_init(&tbl->rhead, &mesh_rht_params)); in mesh_table_init()
[all …]
/linux/arch/powerpc/kvm/
H A Dbook3s_64_vio.c66 iommu_tce_table_put(stit->tbl); in kvm_spapr_tce_iommu_table_free()
98 if (table_group->tables[i] != stit->tbl) in kvm_spapr_tce_release_iommu_group()
114 struct iommu_table *tbl = NULL; in kvm_spapr_tce_attach_iommu_group() local
154 tbl = iommu_tce_table_get(tbltmp); in kvm_spapr_tce_attach_iommu_group()
158 if (!tbl) in kvm_spapr_tce_attach_iommu_group()
163 if (tbl != stit->tbl) in kvm_spapr_tce_attach_iommu_group()
168 iommu_tce_table_put(tbl); in kvm_spapr_tce_attach_iommu_group()
183 iommu_tce_table_put(tbl); in kvm_spapr_tce_attach_iommu_group()
187 stit->tbl = tbl; in kvm_spapr_tce_attach_iommu_group()
390 long shift = stit->tbl->it_page_shift; in kvmppc_tce_validate()
[all …]
/linux/net/netfilter/
H A Dxt_repldata.h24 } *tbl; \
26 size_t term_offset = (offsetof(typeof(*tbl), entries[nhooks]) + \
28 tbl = kzalloc(term_offset + sizeof(*term), GFP_KERNEL); \
29 if (tbl == NULL) \
31 term = (struct type##_error *)&(((char *)tbl)[term_offset]); \
32 strscpy(tbl->repl.name, info->name); \
34 tbl->repl.valid_hooks = hook_mask; \
35 tbl->repl.num_entries = nhooks + 1; \
36 tbl->repl.size = nhooks * sizeof(struct type##_standard) + \
41 tbl->repl.hook_entry[hooknum] = bytes; \
[all …]

12345678910>>...15