/linux/fs/nfs/ |
H A D | nfs4session.c | 27 static void nfs4_init_slot_table(struct nfs4_slot_table *tbl, const char *queue) in nfs4_init_slot_table() argument 29 tbl->highest_used_slotid = NFS4_NO_SLOT; in nfs4_init_slot_table() 30 spin_lock_init(&tbl->slot_tbl_lock); in nfs4_init_slot_table() 31 rpc_init_priority_wait_queue(&tbl->slot_tbl_waitq, queue); in nfs4_init_slot_table() 32 init_waitqueue_head(&tbl->slot_waitq); in nfs4_init_slot_table() 33 init_completion(&tbl->complete); in nfs4_init_slot_table() 39 static void nfs4_shrink_slot_table(struct nfs4_slot_table *tbl, u32 newsize) in nfs4_shrink_slot_table() argument 42 if (newsize >= tbl->max_slots) in nfs4_shrink_slot_table() 45 p = &tbl->slots; in nfs4_shrink_slot_table() 53 tbl in nfs4_shrink_slot_table() 62 nfs4_slot_tbl_drain_complete(struct nfs4_slot_table * tbl) nfs4_slot_tbl_drain_complete() argument 83 nfs4_free_slot(struct nfs4_slot_table * tbl,struct nfs4_slot * slot) nfs4_free_slot() argument 104 nfs4_new_slot(struct nfs4_slot_table * tbl,u32 slotid,u32 seq_init,gfp_t gfp_mask) nfs4_new_slot() argument 120 nfs4_find_or_create_slot(struct nfs4_slot_table * tbl,u32 slotid,u32 seq_init,gfp_t gfp_mask) nfs4_find_or_create_slot() argument 142 nfs4_lock_slot(struct nfs4_slot_table * tbl,struct nfs4_slot * slot) nfs4_lock_slot() argument 159 nfs4_try_to_lock_slot(struct nfs4_slot_table * tbl,struct nfs4_slot * slot) nfs4_try_to_lock_slot() argument 172 nfs4_lookup_slot(struct nfs4_slot_table * tbl,u32 slotid) nfs4_lookup_slot() argument 179 nfs4_slot_get_seqid(struct nfs4_slot_table * tbl,u32 slotid,u32 * seq_nr) nfs4_slot_get_seqid() argument 201 nfs4_slot_seqid_in_use(struct nfs4_slot_table * tbl,u32 slotid,u32 seq_nr) nfs4_slot_seqid_in_use() argument 222 nfs4_slot_wait_on_seqid(struct nfs4_slot_table * tbl,u32 slotid,u32 seq_nr,unsigned long timeout) nfs4_slot_wait_on_seqid() argument 242 nfs4_alloc_slot(struct nfs4_slot_table * tbl) nfs4_alloc_slot() argument 262 nfs4_grow_slot_table(struct nfs4_slot_table * tbl,u32 max_reqs,u32 ivalue) nfs4_grow_slot_table() argument 272 nfs4_reset_slot_table(struct nfs4_slot_table * tbl,u32 server_highest_slotid,u32 ivalue) nfs4_reset_slot_table() argument 297 nfs4_realloc_slot_table(struct nfs4_slot_table * tbl,u32 max_reqs,u32 ivalue) nfs4_realloc_slot_table() argument 326 nfs4_release_slot_table(struct nfs4_slot_table * tbl) nfs4_release_slot_table() argument 336 nfs4_shutdown_slot_table(struct nfs4_slot_table * tbl) nfs4_shutdown_slot_table() argument 350 nfs4_setup_slot_table(struct nfs4_slot_table * tbl,unsigned int max_reqs,const char * queue) nfs4_setup_slot_table() argument 362 struct nfs4_slot_table *tbl = slot->table; nfs41_assign_slot() local 375 __nfs41_wake_and_assign_slot(struct nfs4_slot_table * tbl,struct nfs4_slot * slot) __nfs41_wake_and_assign_slot() argument 383 nfs41_wake_and_assign_slot(struct nfs4_slot_table * tbl,struct nfs4_slot * slot) nfs41_wake_and_assign_slot() argument 391 nfs41_try_wake_next_slot_table_entry(struct nfs4_slot_table * tbl) nfs41_try_wake_next_slot_table_entry() argument 403 nfs41_wake_slot_table(struct nfs4_slot_table * tbl) nfs41_wake_slot_table() argument 413 nfs41_set_max_slotid_locked(struct nfs4_slot_table * tbl,u32 target_highest_slotid) nfs41_set_max_slotid_locked() argument 428 nfs41_set_target_slotid_locked(struct nfs4_slot_table * tbl,u32 target_highest_slotid) nfs41_set_target_slotid_locked() argument 437 nfs41_set_target_slotid(struct nfs4_slot_table * tbl,u32 target_highest_slotid) nfs41_set_target_slotid() argument 448 nfs41_set_server_slotid_locked(struct nfs4_slot_table * tbl,u32 highest_slotid) nfs41_set_server_slotid_locked() argument 489 nfs41_is_outlier_target_slotid(struct nfs4_slot_table * tbl,u32 new_target) nfs41_is_outlier_target_slotid() argument 510 nfs41_update_target_slotid(struct nfs4_slot_table * tbl,struct nfs4_slot * slot,struct nfs4_sequence_res * res) nfs41_update_target_slotid() argument 538 struct nfs4_slot_table *tbl; nfs4_setup_session_slot_tables() local [all...] |
H A D | nfs4session.h | 83 extern int nfs4_setup_slot_table(struct nfs4_slot_table *tbl, 85 extern void nfs4_shutdown_slot_table(struct nfs4_slot_table *tbl); 86 extern struct nfs4_slot *nfs4_alloc_slot(struct nfs4_slot_table *tbl); 87 extern struct nfs4_slot *nfs4_lookup_slot(struct nfs4_slot_table *tbl, u32 slotid); 88 extern int nfs4_slot_wait_on_seqid(struct nfs4_slot_table *tbl, 91 extern bool nfs4_try_to_lock_slot(struct nfs4_slot_table *tbl, struct nfs4_slot *slot); 92 extern void nfs4_free_slot(struct nfs4_slot_table *tbl, struct nfs4_slot *slot); 93 extern void nfs4_slot_tbl_drain_complete(struct nfs4_slot_table *tbl); 94 bool nfs41_wake_and_assign_slot(struct nfs4_slot_table *tbl, 96 void nfs41_wake_slot_table(struct nfs4_slot_table *tbl); 98 nfs4_slot_tbl_draining(struct nfs4_slot_table * tbl) nfs4_slot_tbl_draining() argument 103 nfs4_test_locked_slot(const struct nfs4_slot_table * tbl,u32 slotid) nfs4_test_locked_slot() argument [all...] |
/linux/drivers/net/ethernet/mellanox/mlx5/core/steering/sws/ |
H A D | dr_table.c | 28 action->dest_tbl->tbl->rx.s_anchor->chunk : in dr_table_set_miss_action_nic() 29 action->dest_tbl->tbl->tx.s_anchor->chunk; in dr_table_set_miss_action_nic() 46 int mlx5dr_table_set_miss_action(struct mlx5dr_table *tbl, in mlx5dr_table_set_miss_action() argument 54 mlx5dr_domain_lock(tbl->dmn); in mlx5dr_table_set_miss_action() 56 if (tbl->dmn->type == MLX5DR_DOMAIN_TYPE_NIC_RX || in mlx5dr_table_set_miss_action() 57 tbl->dmn->type == MLX5DR_DOMAIN_TYPE_FDB) { in mlx5dr_table_set_miss_action() 58 ret = dr_table_set_miss_action_nic(tbl->dmn, &tbl->rx, action); in mlx5dr_table_set_miss_action() 63 if (tbl->dmn->type == MLX5DR_DOMAIN_TYPE_NIC_TX || in mlx5dr_table_set_miss_action() 64 tbl in mlx5dr_table_set_miss_action() 92 dr_table_uninit_fdb(struct mlx5dr_table * tbl) dr_table_uninit_fdb() argument 98 dr_table_uninit(struct mlx5dr_table * tbl) dr_table_uninit() argument 159 dr_table_init_fdb(struct mlx5dr_table * tbl) dr_table_init_fdb() argument 178 dr_table_init(struct mlx5dr_table * tbl) dr_table_init() argument 213 dr_table_destroy_sw_owned_tbl(struct mlx5dr_table * tbl) dr_table_destroy_sw_owned_tbl() argument 220 dr_table_create_sw_owned_tbl(struct mlx5dr_table * tbl,u16 uid) dr_table_create_sw_owned_tbl() argument 253 struct mlx5dr_table *tbl; mlx5dr_table_create() local 288 mlx5dr_table_destroy(struct mlx5dr_table * tbl) mlx5dr_table_destroy() argument 311 mlx5dr_table_get_id(struct mlx5dr_table * tbl) mlx5dr_table_get_id() argument [all...] |
/linux/drivers/net/ethernet/mellanox/mlx5/core/steering/hws/ |
H A D | table.c | 6 u32 mlx5hws_table_get_id(struct mlx5hws_table *tbl) in mlx5hws_table_get_id() argument 8 return tbl->ft_id; in mlx5hws_table_get_id() 11 static void hws_table_init_next_ft_attr(struct mlx5hws_table *tbl, in hws_table_init_next_ft_attr() argument 14 ft_attr->type = tbl->fw_ft_type; in hws_table_init_next_ft_attr() 15 if (tbl->type == MLX5HWS_TABLE_TYPE_FDB) in hws_table_init_next_ft_attr() 16 ft_attr->level = tbl->ctx->caps->fdb_ft.max_level - 1; in hws_table_init_next_ft_attr() 18 ft_attr->level = tbl->ctx->caps->nic_ft.max_level - 1; in hws_table_init_next_ft_attr() 22 static void hws_table_set_cap_attr(struct mlx5hws_table *tbl, in hws_table_set_cap_attr() argument 33 if (!MLX5_CAP_ESW_FLOWTABLE(tbl->ctx->mdev, fdb_dynamic_tunnel)) { in hws_table_set_cap_attr() 39 static int hws_table_up_default_fdb_miss_tbl(struct mlx5hws_table *tbl) in hws_table_up_default_fdb_miss_tbl() argument 81 hws_table_down_default_fdb_miss_tbl(struct mlx5hws_table * tbl) hws_table_down_default_fdb_miss_tbl() argument 98 hws_table_connect_to_default_miss_tbl(struct mlx5hws_table * tbl,u32 ft_id) hws_table_connect_to_default_miss_tbl() argument 121 mlx5hws_table_create_default_ft(struct mlx5_core_dev * mdev,struct mlx5hws_table * tbl,u32 * ft_id) mlx5hws_table_create_default_ft() argument 159 mlx5hws_table_destroy_default_ft(struct mlx5hws_table * tbl,u32 ft_id) mlx5hws_table_destroy_default_ft() argument 167 hws_table_init_check_hws_support(struct mlx5hws_context * ctx,struct mlx5hws_table * tbl) hws_table_init_check_hws_support() argument 177 hws_table_init(struct mlx5hws_table * tbl) hws_table_init() argument 216 hws_table_uninit(struct mlx5hws_table * tbl) hws_table_uninit() argument 227 struct mlx5hws_table *tbl; mlx5hws_table_create() local 260 mlx5hws_table_destroy(struct mlx5hws_table * tbl) mlx5hws_table_destroy() argument 291 hws_table_get_last_ft(struct mlx5hws_table * tbl) hws_table_get_last_ft() argument 302 mlx5hws_table_ft_set_default_next_ft(struct mlx5hws_table * tbl,u32 ft_id) mlx5hws_table_ft_set_default_next_ft() argument 444 hws_table_set_default_miss_not_valid(struct mlx5hws_table * tbl,struct mlx5hws_table * miss_tbl) hws_table_set_default_miss_not_valid() argument 460 mlx5hws_table_set_default_miss(struct mlx5hws_table * tbl,struct mlx5hws_table * miss_tbl) mlx5hws_table_set_default_miss() argument [all...] |
H A D | matcher.c | 23 mlx5hws_table_destroy_default_ft(matcher->tbl, matcher->end_ft_id); in hws_matcher_destroy_end_ft() 26 int mlx5hws_matcher_update_end_ft_isolated(struct mlx5hws_table *tbl, in mlx5hws_matcher_update_end_ft_isolated() argument 31 if (list_empty(&tbl->matchers_list)) in mlx5hws_matcher_update_end_ft_isolated() 37 list_for_each_entry(tmp_matcher, &tbl->matchers_list, list_node) in mlx5hws_matcher_update_end_ft_isolated() 40 tmp_matcher = list_last_entry(&tbl->matchers_list, in mlx5hws_matcher_update_end_ft_isolated() 44 return mlx5hws_table_ft_set_next_ft(tbl->ctx, in mlx5hws_matcher_update_end_ft_isolated() 46 tbl->fw_ft_type, in mlx5hws_matcher_update_end_ft_isolated() 52 struct mlx5hws_table *tbl = matcher->tbl; in hws_matcher_connect_end_ft_isolated() local 57 ret = mlx5hws_table_ft_set_next_rtc(tbl in hws_matcher_connect_end_ft_isolated() 83 struct mlx5hws_table *tbl = matcher->tbl; hws_matcher_create_end_ft_isolated() local 109 struct mlx5hws_table *tbl = matcher->tbl; hws_matcher_create_end_ft() local 128 struct mlx5hws_table *tbl = matcher->tbl; hws_matcher_connect_isolated_first() local 162 struct mlx5hws_table *tbl = matcher->tbl; hws_matcher_connect_isolated_last() local 221 struct mlx5hws_table *tbl = matcher->tbl; hws_matcher_connect() local 310 struct mlx5hws_table *tbl = matcher->tbl; hws_matcher_disconnect_isolated() local 396 struct mlx5hws_table *tbl = matcher->tbl; hws_matcher_disconnect() local 472 struct mlx5hws_table *tbl = matcher->tbl; hws_matcher_create_rtc() local 1122 mlx5hws_matcher_create(struct mlx5hws_table * tbl,struct mlx5hws_match_template * mt[],u8 num_of_mt,struct mlx5hws_action_template * at[],u8 num_of_at,struct mlx5hws_matcher_attr * attr) mlx5hws_matcher_create() argument [all...] |
/linux/drivers/infiniband/hw/vmw_pvrdma/ |
H A D | pvrdma_doorbell.c | 56 struct pvrdma_id_table *tbl = &dev->uar_table.tbl; in pvrdma_uar_table_init() local 61 tbl->last = 0; in pvrdma_uar_table_init() 62 tbl->top = 0; in pvrdma_uar_table_init() 63 tbl->max = num; in pvrdma_uar_table_init() 64 tbl->mask = mask; in pvrdma_uar_table_init() 65 spin_lock_init(&tbl->lock); in pvrdma_uar_table_init() 66 tbl->table = bitmap_zalloc(num, GFP_KERNEL); in pvrdma_uar_table_init() 67 if (!tbl->table) in pvrdma_uar_table_init() 71 __set_bit(0, tbl in pvrdma_uar_table_init() 78 struct pvrdma_id_table *tbl = &dev->uar_table.tbl; pvrdma_uar_table_cleanup() local 85 struct pvrdma_id_table *tbl; pvrdma_uar_alloc() local 117 struct pvrdma_id_table *tbl = &dev->uar_table.tbl; pvrdma_uar_free() local [all...] |
/linux/arch/powerpc/kernel/ |
H A D | iommu.c | 47 struct iommu_table *tbl = data; in iommu_debugfs_weight_get() local 48 *val = bitmap_weight(tbl->it_map, tbl->it_size); in iommu_debugfs_weight_get() 53 static void iommu_debugfs_add(struct iommu_table *tbl) in iommu_debugfs_add() argument 58 sprintf(name, "%08lx", tbl->it_index); in iommu_debugfs_add() 61 debugfs_create_file_unsafe("weight", 0400, liobn_entry, tbl, &iommu_debugfs_fops_weight); in iommu_debugfs_add() 62 debugfs_create_ulong("it_size", 0400, liobn_entry, &tbl->it_size); in iommu_debugfs_add() 63 debugfs_create_ulong("it_page_shift", 0400, liobn_entry, &tbl->it_page_shift); in iommu_debugfs_add() 64 debugfs_create_ulong("it_reserved_start", 0400, liobn_entry, &tbl->it_reserved_start); in iommu_debugfs_add() 65 debugfs_create_ulong("it_reserved_end", 0400, liobn_entry, &tbl in iommu_debugfs_add() 70 iommu_debugfs_del(struct iommu_table * tbl) iommu_debugfs_del() argument 78 iommu_debugfs_add(struct iommu_table * tbl) iommu_debugfs_add() argument 79 iommu_debugfs_del(struct iommu_table * tbl) iommu_debugfs_del() argument 216 iommu_range_alloc(struct device * dev,struct iommu_table * tbl,unsigned long npages,unsigned long * handle,unsigned long mask,unsigned int align_order) iommu_range_alloc() argument 347 iommu_alloc(struct device * dev,struct iommu_table * tbl,void * page,unsigned int npages,enum dma_data_direction direction,unsigned long mask,unsigned int align_order,unsigned long attrs) iommu_alloc() argument 390 iommu_free_check(struct iommu_table * tbl,dma_addr_t dma_addr,unsigned int npages) iommu_free_check() argument 418 get_pool(struct iommu_table * tbl,unsigned long entry) get_pool() argument 437 __iommu_free(struct iommu_table * tbl,dma_addr_t dma_addr,unsigned int npages) __iommu_free() argument 459 iommu_free(struct iommu_table * tbl,dma_addr_t dma_addr,unsigned int npages) iommu_free() argument 472 ppc_iommu_map_sg(struct device * dev,struct iommu_table * tbl,struct scatterlist * sglist,int nelems,unsigned long mask,enum dma_data_direction direction,unsigned long attrs) ppc_iommu_map_sg() argument 615 ppc_iommu_unmap_sg(struct iommu_table * tbl,struct scatterlist * sglist,int nelems,enum dma_data_direction direction,unsigned long attrs) ppc_iommu_unmap_sg() argument 647 iommu_table_clear(struct iommu_table * tbl) iommu_table_clear() argument 688 iommu_table_reserve_pages(struct iommu_table * tbl,unsigned long res_start,unsigned long res_end) iommu_table_reserve_pages() argument 726 iommu_init_table(struct iommu_table * tbl,int nid,unsigned long res_start,unsigned long res_end) iommu_init_table() argument 783 iommu_table_in_use(struct iommu_table * tbl) iommu_table_in_use() argument 806 struct iommu_table *tbl; iommu_table_free() local 831 iommu_tce_table_get(struct iommu_table * tbl) iommu_tce_table_get() argument 840 iommu_tce_table_put(struct iommu_table * tbl) iommu_tce_table_put() argument 854 iommu_map_page(struct device * dev,struct iommu_table * tbl,struct page * page,unsigned long offset,size_t size,unsigned long mask,enum dma_data_direction direction,unsigned long attrs) iommu_map_page() argument 893 iommu_unmap_page(struct iommu_table * tbl,dma_addr_t dma_handle,size_t size,enum dma_data_direction direction,unsigned long attrs) iommu_unmap_page() argument 912 iommu_alloc_coherent(struct device * dev,struct iommu_table * tbl,size_t size,dma_addr_t * dma_handle,unsigned long mask,gfp_t flag,int node) iommu_alloc_coherent() argument 962 iommu_free_coherent(struct iommu_table * tbl,size_t size,void * vaddr,dma_addr_t dma_handle) iommu_free_coherent() argument 1054 iommu_flush_tce(struct iommu_table * tbl) iommu_flush_tce() argument 1097 iommu_tce_xchg_no_kill(struct mm_struct * mm,struct iommu_table * tbl,unsigned long entry,unsigned long * hpa,enum dma_data_direction * direction) iommu_tce_xchg_no_kill() argument 1115 iommu_tce_kill(struct iommu_table * tbl,unsigned long entry,unsigned long pages) iommu_tce_kill() argument [all...] |
/linux/arch/powerpc/platforms/powernv/ |
H A D | pci-ioda-tce.c | 48 void pnv_pci_setup_iommu_table(struct iommu_table *tbl, in pnv_pci_setup_iommu_table() argument 52 tbl->it_blocksize = 16; in pnv_pci_setup_iommu_table() 53 tbl->it_base = (unsigned long)tce_mem; in pnv_pci_setup_iommu_table() 54 tbl->it_page_shift = page_shift; in pnv_pci_setup_iommu_table() 55 tbl->it_offset = dma_offset >> tbl->it_page_shift; in pnv_pci_setup_iommu_table() 56 tbl->it_index = 0; in pnv_pci_setup_iommu_table() 57 tbl->it_size = tce_size >> 3; in pnv_pci_setup_iommu_table() 58 tbl->it_busno = 0; in pnv_pci_setup_iommu_table() 59 tbl in pnv_pci_setup_iommu_table() 83 pnv_tce(struct iommu_table * tbl,bool user,long idx,bool alloc) pnv_tce() argument 124 pnv_tce_build(struct iommu_table * tbl,long index,long npages,unsigned long uaddr,enum dma_data_direction direction,unsigned long attrs) pnv_tce_build() argument 147 pnv_tce_xchg(struct iommu_table * tbl,long index,unsigned long * hpa,enum dma_data_direction * direction) pnv_tce_xchg() argument 181 pnv_tce_useraddrptr(struct iommu_table * tbl,long index,bool alloc) pnv_tce_useraddrptr() argument 190 pnv_tce_free(struct iommu_table * tbl,long index,long npages) pnv_tce_free() argument 206 pnv_tce_get(struct iommu_table * tbl,long index) pnv_tce_get() argument 240 pnv_pci_ioda2_table_free_pages(struct iommu_table * tbl) pnv_pci_ioda2_table_free_pages() argument 292 pnv_pci_ioda2_table_alloc_pages(int nid,__u64 bus_offset,__u32 page_shift,__u64 window_size,__u32 levels,bool alloc_userspace_copy,struct iommu_table * tbl) pnv_pci_ioda2_table_alloc_pages() argument 370 pnv_pci_unlink_table_and_group(struct iommu_table * tbl,struct iommu_table_group * table_group) pnv_pci_unlink_table_and_group() argument 411 pnv_pci_link_table_and_group(int node,int num,struct iommu_table * tbl,struct iommu_table_group * table_group) pnv_pci_link_table_and_group() argument [all...] |
/linux/lib/ |
H A D | rhashtable.c | 38 const struct bucket_table *tbl, in head_hashfn() argument 41 return rht_head_hashfn(ht, tbl, he, ht->p); in head_hashfn() 53 int lockdep_rht_bucket_is_held(const struct bucket_table *tbl, u32 hash) in lockdep_rht_bucket_is_held() argument 57 if (unlikely(tbl->nest)) in lockdep_rht_bucket_is_held() 59 return bit_spin_is_locked(0, (unsigned long *)&tbl->buckets[hash]); in lockdep_rht_bucket_is_held() 67 const struct bucket_table *tbl) in nested_table_top() argument 70 * because it's set at the same time as tbl->nest. in nested_table_top() 72 return (void *)rcu_dereference_protected(tbl->buckets[0], 1); in nested_table_top() 94 static void nested_bucket_table_free(const struct bucket_table *tbl) in nested_bucket_table_free() argument 96 unsigned int size = tbl in nested_bucket_table_free() 109 bucket_table_free(const struct bucket_table * tbl) bucket_table_free() argument 153 struct bucket_table *tbl; nested_bucket_table_alloc() local 181 struct bucket_table *tbl = NULL; bucket_table_alloc() local 216 rhashtable_last_table(struct rhashtable * ht,struct bucket_table * tbl) rhashtable_last_table() argument 417 struct bucket_table *tbl; rht_deferred_worker() local 447 rhashtable_insert_rehash(struct rhashtable * ht,struct bucket_table * tbl) rhashtable_insert_rehash() argument 496 rhashtable_lookup_one(struct rhashtable * ht,struct rhash_lock_head __rcu ** bkt,struct bucket_table * tbl,unsigned int hash,const void * key,struct rhash_head * obj) rhashtable_lookup_one() argument 547 rhashtable_insert_one(struct rhashtable * ht,struct rhash_lock_head __rcu ** bkt,struct bucket_table * tbl,unsigned int hash,struct rhash_head * obj,void * data) rhashtable_insert_one() argument 594 struct bucket_table *tbl; rhashtable_try_insert() local 801 struct bucket_table *tbl = iter->walker.tbl; __rhashtable_walk_find_next() local 946 struct bucket_table *tbl = iter->walker.tbl; rhashtable_walk_stop() local 1030 struct bucket_table *tbl; rhashtable_init_noprof() local 1148 struct bucket_table *tbl, *next_tbl; rhashtable_free_and_destroy() local 1189 __rht_bucket_nested(const struct bucket_table * tbl,unsigned int hash) __rht_bucket_nested() argument 1218 rht_bucket_nested(const struct bucket_table * tbl,unsigned int hash) rht_bucket_nested() argument 1229 rht_bucket_nested_insert(struct rhashtable * ht,struct bucket_table * tbl,unsigned int hash) rht_bucket_nested_insert() argument [all...] |
/linux/arch/powerpc/include/asm/ |
H A D | iommu.h | 45 int (*set)(struct iommu_table *tbl, 56 int (*xchg_no_kill)(struct iommu_table *tbl, 61 void (*tce_kill)(struct iommu_table *tbl, 65 __be64 *(*useraddrptr)(struct iommu_table *tbl, long index, bool alloc); 67 void (*clear)(struct iommu_table *tbl, 70 unsigned long (*get)(struct iommu_table *tbl, long index); 71 void (*flush)(struct iommu_table *tbl); 72 void (*free)(struct iommu_table *tbl); 122 #define IOMMU_TABLE_USERSPACE_ENTRY_RO(tbl, entry) \ argument 123 ((tbl) 124 IOMMU_TABLE_USERSPACE_ENTRY(tbl,entry) global() argument 129 get_iommu_order(unsigned long size,struct iommu_table * tbl) get_iommu_order() argument 305 iommu_tce_clear_param_check(tbl,ioba,tce_value,npages) global() argument 309 iommu_tce_put_param_check(tbl,ioba,gpa) global() argument [all...] |
/linux/drivers/net/wireless/intel/iwlwifi/dvm/ |
H A D | rs.c | 414 static s32 get_expected_tpt(struct iwl_scale_tbl_info *tbl, int rs_index) in get_expected_tpt() argument 416 if (tbl->expected_tpt) in get_expected_tpt() 417 return tbl->expected_tpt[rs_index]; in get_expected_tpt() 428 static int rs_collect_tx_data(struct iwl_scale_tbl_info *tbl, in rs_collect_tx_data() argument 439 window = &(tbl->win[scale_index]); in rs_collect_tx_data() 442 tpt = get_expected_tpt(tbl, scale_index); in rs_collect_tx_data() 507 struct iwl_scale_tbl_info *tbl, in rate_n_flags_from_tbl() argument 512 if (is_legacy(tbl->lq_type)) { in rate_n_flags_from_tbl() 517 } else if (is_Ht(tbl->lq_type)) { in rate_n_flags_from_tbl() 524 if (is_siso(tbl in rate_n_flags_from_tbl() 564 rs_get_tbl_info_from_mcs(const u32 rate_n_flags,enum nl80211_band band,struct iwl_scale_tbl_info * tbl,int * rate_idx) rs_get_tbl_info_from_mcs() argument 629 rs_toggle_antenna(u32 valid_ant,u32 * rate_n_flags,struct iwl_scale_tbl_info * tbl) rs_toggle_antenna() argument 753 rs_get_lower_rate(struct iwl_lq_sta * lq_sta,struct iwl_scale_tbl_info * tbl,u8 scale_index,u8 ht_possible) rs_get_lower_rate() argument 825 struct iwl_scale_tbl_info *tbl; rs_bt_update_lq() local 1065 rs_set_expected_tpt_table(struct iwl_lq_sta * lq_sta,struct iwl_scale_tbl_info * tbl) rs_set_expected_tpt_table() argument 1122 rs_get_best_rate(struct iwl_priv * priv,struct iwl_lq_sta * lq_sta,struct iwl_scale_tbl_info * tbl,u16 rate_mask,s8 index) rs_get_best_rate() argument 1217 rs_switch_to_mimo2(struct iwl_priv * priv,struct iwl_lq_sta * lq_sta,struct ieee80211_conf * conf,struct ieee80211_sta * sta,struct iwl_scale_tbl_info * tbl,int index) rs_switch_to_mimo2() argument 1272 rs_switch_to_mimo3(struct iwl_priv * priv,struct iwl_lq_sta * lq_sta,struct ieee80211_conf * conf,struct ieee80211_sta * sta,struct iwl_scale_tbl_info * tbl,int index) rs_switch_to_mimo3() argument 1328 rs_switch_to_siso(struct iwl_priv * priv,struct iwl_lq_sta * lq_sta,struct ieee80211_conf * conf,struct ieee80211_sta * sta,struct iwl_scale_tbl_info * tbl,int index) rs_switch_to_siso() argument 1379 struct iwl_scale_tbl_info *tbl = &(lq_sta->lq_info[lq_sta->active_tbl]); rs_move_legacy_other() local 1552 struct iwl_scale_tbl_info *tbl = &(lq_sta->lq_info[lq_sta->active_tbl]); rs_move_siso_to_other() local 1722 struct iwl_scale_tbl_info *tbl = &(lq_sta->lq_info[lq_sta->active_tbl]); rs_move_mimo2_to_other() local 1890 struct iwl_scale_tbl_info *tbl = &(lq_sta->lq_info[lq_sta->active_tbl]); rs_move_mimo3_to_other() local 2063 struct iwl_scale_tbl_info *tbl; rs_stay_in_table() local 2143 rs_update_rate_tbl(struct iwl_priv * priv,struct iwl_rxon_context * ctx,struct iwl_lq_sta * lq_sta,struct iwl_scale_tbl_info * tbl,int index,u8 is_green) rs_update_rate_tbl() argument 2178 struct iwl_scale_tbl_info *tbl, *tbl1; rs_rate_scale_perform() local 2611 struct iwl_scale_tbl_info *tbl; rs_initialize_lq() local 3089 struct iwl_scale_tbl_info *tbl = &(lq_sta->lq_info[lq_sta->active_tbl]); rs_sta_dbgfs_scale_table_read() local 3217 struct iwl_scale_tbl_info *tbl = &lq_sta->lq_info[lq_sta->active_tbl]; rs_sta_dbgfs_rate_scale_data_read() local [all...] |
H A D | rs.h | 256 #define is_legacy(tbl) (((tbl) == LQ_G) || ((tbl) == LQ_A)) argument 257 #define is_siso(tbl) ((tbl) == LQ_SISO) argument 258 #define is_mimo2(tbl) ((tbl) == LQ_MIMO2) argument 259 #define is_mimo3(tbl) ((tbl) == LQ_MIMO3) argument 260 #define is_mimo(tbl) (is_mimo argument 261 is_Ht(tbl) global() argument 262 is_a_band(tbl) global() argument 263 is_g_and(tbl) global() argument [all...] |
/linux/net/netfilter/ipvs/ |
H A D | ip_vs_lblc.c | 9 * *lock(tbl->lock) ==> *lock(&tbl->lock) 10 * Wensong Zhang : fixed the uninitialized tbl->lock bug 167 ip_vs_lblc_hash(struct ip_vs_lblc_table *tbl, struct ip_vs_lblc_entry *en) in ip_vs_lblc_hash() argument 171 hlist_add_head_rcu(&en->list, &tbl->bucket[hash]); in ip_vs_lblc_hash() 172 atomic_inc(&tbl->entries); in ip_vs_lblc_hash() 178 ip_vs_lblc_get(int af, struct ip_vs_lblc_table *tbl, in ip_vs_lblc_get() argument 184 hlist_for_each_entry_rcu(en, &tbl->bucket[hash], list) in ip_vs_lblc_get() 197 ip_vs_lblc_new(struct ip_vs_lblc_table *tbl, const union nf_inet_addr *daddr, in ip_vs_lblc_new() argument 202 en = ip_vs_lblc_get(af, tbl, dadd in ip_vs_lblc_new() 230 struct ip_vs_lblc_table *tbl = svc->sched_data; ip_vs_lblc_flush() local 257 struct ip_vs_lblc_table *tbl = svc->sched_data; ip_vs_lblc_full_check() local 295 struct ip_vs_lblc_table *tbl = timer_container_of(tbl, t, ip_vs_lblc_check_expire() local 346 struct ip_vs_lblc_table *tbl; ip_vs_lblc_init_svc() local 384 struct ip_vs_lblc_table *tbl = svc->sched_data; ip_vs_lblc_done_svc() local 484 struct ip_vs_lblc_table *tbl = svc->sched_data; ip_vs_lblc_schedule() local [all...] |
H A D | ip_vs_lblcr.c | 330 ip_vs_lblcr_hash(struct ip_vs_lblcr_table *tbl, struct ip_vs_lblcr_entry *en) in ip_vs_lblcr_hash() argument 334 hlist_add_head_rcu(&en->list, &tbl->bucket[hash]); in ip_vs_lblcr_hash() 335 atomic_inc(&tbl->entries); in ip_vs_lblcr_hash() 341 ip_vs_lblcr_get(int af, struct ip_vs_lblcr_table *tbl, in ip_vs_lblcr_get() argument 347 hlist_for_each_entry_rcu(en, &tbl->bucket[hash], list) in ip_vs_lblcr_get() 360 ip_vs_lblcr_new(struct ip_vs_lblcr_table *tbl, const union nf_inet_addr *daddr, in ip_vs_lblcr_new() argument 365 en = ip_vs_lblcr_get(af, tbl, daddr); in ip_vs_lblcr_new() 381 ip_vs_lblcr_hash(tbl, en); in ip_vs_lblcr_new() 396 struct ip_vs_lblcr_table *tbl = svc->sched_data; in ip_vs_lblcr_flush() local 402 tbl in ip_vs_lblcr_flush() 422 struct ip_vs_lblcr_table *tbl = svc->sched_data; ip_vs_lblcr_full_check() local 459 struct ip_vs_lblcr_table *tbl = timer_container_of(tbl, t, ip_vs_lblcr_check_expire() local 509 struct ip_vs_lblcr_table *tbl; ip_vs_lblcr_init_svc() local 547 struct ip_vs_lblcr_table *tbl = svc->sched_data; ip_vs_lblcr_done_svc() local 648 struct ip_vs_lblcr_table *tbl = svc->sched_data; ip_vs_lblcr_schedule() local [all...] |
/linux/drivers/firmware/efi/ |
H A D | memattr.c | 24 efi_memory_attributes_table_t *tbl; in efi_memattr_init() local 30 tbl = early_memremap(efi_mem_attr_table, sizeof(*tbl)); in efi_memattr_init() 31 if (!tbl) { in efi_memattr_init() 37 if (tbl->version > 2) { in efi_memattr_init() 39 tbl->version); in efi_memattr_init() 51 size = tbl->num_entries * tbl->desc_size; in efi_memattr_init() 54 tbl->version, tbl in efi_memattr_init() 147 efi_memory_attributes_table_t *tbl; efi_memattr_apply_permissions() local [all...] |
/linux/drivers/net/wireless/marvell/mwifiex/ |
H A D | 11n_rxreorder.c | 98 struct mwifiex_rx_reorder_tbl *tbl, in mwifiex_11n_dispatch_pkt_until_start_win() argument 108 pkt_to_send = (start_win > tbl->start_win) ? in mwifiex_11n_dispatch_pkt_until_start_win() 109 min((start_win - tbl->start_win), tbl->win_size) : in mwifiex_11n_dispatch_pkt_until_start_win() 110 tbl->win_size; in mwifiex_11n_dispatch_pkt_until_start_win() 113 if (tbl->rx_reorder_ptr[i]) { in mwifiex_11n_dispatch_pkt_until_start_win() 114 skb = tbl->rx_reorder_ptr[i]; in mwifiex_11n_dispatch_pkt_until_start_win() 116 tbl->rx_reorder_ptr[i] = NULL; in mwifiex_11n_dispatch_pkt_until_start_win() 124 for (i = 0; i < tbl->win_size - pkt_to_send; ++i) { in mwifiex_11n_dispatch_pkt_until_start_win() 125 tbl in mwifiex_11n_dispatch_pkt_until_start_win() 146 mwifiex_11n_scan_and_dispatch(struct mwifiex_private * priv,struct mwifiex_rx_reorder_tbl * tbl) mwifiex_11n_scan_and_dispatch() argument 190 mwifiex_del_rx_reorder_entry(struct mwifiex_private * priv,struct mwifiex_rx_reorder_tbl * tbl) mwifiex_del_rx_reorder_entry() argument 232 struct mwifiex_rx_reorder_tbl *tbl; mwifiex_11n_get_rx_reorder_tbl() local 251 struct mwifiex_rx_reorder_tbl *tbl, *tmp; mwifiex_11n_del_rx_reorder_tbl_by_ta() local 333 struct mwifiex_rx_reorder_tbl *tbl, *new_node; mwifiex_11n_create_rx_reorder_tbl() local 410 mwifiex_11n_rxreorder_timer_restart(struct mwifiex_rx_reorder_tbl * tbl) mwifiex_11n_rxreorder_timer_restart() argument 551 struct mwifiex_rx_reorder_tbl *tbl; mwifiex_11n_rx_reorder_pkt() local 663 struct mwifiex_rx_reorder_tbl *tbl; mwifiex_del_ba_tbl() local 717 struct mwifiex_rx_reorder_tbl *tbl; mwifiex_ret_11n_addba_resp() local 808 struct mwifiex_rx_reorder_tbl *tbl; mwifiex_update_rxreor_flags() local [all...] |
/linux/net/core/ |
H A D | neighbour.c | 57 static void pneigh_ifdown(struct neigh_table *tbl, struct net_device *dev, 84 Neighbour hash table buckets are protected with rwlock tbl->lock. 142 atomic_dec(&n->tbl->gc_entries); in neigh_mark_dead() 152 write_lock_bh(&n->tbl->lock); in neigh_update_gc_list() 167 atomic_dec(&n->tbl->gc_entries); in neigh_update_gc_list() 170 list_add_tail(&n->gc_list, &n->tbl->gc_list); in neigh_update_gc_list() 171 atomic_inc(&n->tbl->gc_entries); in neigh_update_gc_list() 175 write_unlock_bh(&n->tbl->lock); in neigh_update_gc_list() 182 write_lock_bh(&n->tbl->lock); in neigh_update_managed_list() 193 list_add_tail(&n->managed_list, &n->tbl in neigh_update_managed_list() 254 neigh_forced_gc(struct neigh_table * tbl) neigh_forced_gc() argument 419 neigh_flush_dev(struct neigh_table * tbl,struct net_device * dev,bool skip_perm) neigh_flush_dev() argument 438 neigh_flush_table(struct neigh_table * tbl) neigh_flush_table() argument 455 neigh_changeaddr(struct neigh_table * tbl,struct net_device * dev) neigh_changeaddr() argument 463 __neigh_ifdown(struct neigh_table * tbl,struct net_device * dev,bool skip_perm) __neigh_ifdown() argument 483 neigh_carrier_down(struct neigh_table * tbl,struct net_device * dev) neigh_carrier_down() argument 490 neigh_ifdown(struct neigh_table * tbl,struct net_device * dev) neigh_ifdown() argument 497 neigh_alloc(struct neigh_table * tbl,struct net_device * dev,u32 flags,bool exempt_from_gc) neigh_alloc() argument 592 neigh_hash_grow(struct neigh_table * tbl,unsigned long new_shift) neigh_hash_grow() argument 626 neigh_lookup(struct neigh_table * tbl,const void * pkey,struct net_device * dev) neigh_lookup() argument 647 ___neigh_create(struct neigh_table * tbl,const void * pkey,struct net_device * dev,u32 flags,bool exempt_from_gc,bool want_ref) ___neigh_create() argument 739 __neigh_create(struct neigh_table * tbl,const void * pkey,struct net_device * dev,bool want_ref) __neigh_create() argument 758 pneigh_lookup(struct neigh_table * tbl,struct net * net,const void * pkey,struct net_device * dev) pneigh_lookup() argument 784 pneigh_create(struct neigh_table * tbl,struct net * net,const void * pkey,struct net_device * dev,u32 flags,u8 protocol,bool permanent) pneigh_create() argument 838 pneigh_delete(struct neigh_table * tbl,struct net * net,const void * pkey,struct net_device * dev) pneigh_delete() argument 871 pneigh_ifdown(struct neigh_table * tbl,struct net_device * dev,bool skip_perm) pneigh_ifdown() argument 977 struct neigh_table *tbl = container_of(work, struct neigh_table, gc_work.work); neigh_periodic_work() local 1542 neigh_event_ns(struct neigh_table * tbl,u8 * lladdr,void * saddr,struct net_device * dev) neigh_event_ns() argument 1641 struct neigh_table *tbl = container_of(work, struct neigh_table, neigh_managed_work() local 1655 struct neigh_table *tbl = timer_container_of(tbl, t, proxy_timer); neigh_proxy_process() local 1700 pneigh_enqueue(struct neigh_table * tbl,struct neigh_parms * p,struct sk_buff * skb) pneigh_enqueue() argument 1727 lookup_neigh_parms(struct neigh_table * tbl,struct net * net,int ifindex) lookup_neigh_parms() argument 1742 neigh_parms_alloc(struct net_device * dev,struct neigh_table * tbl) neigh_parms_alloc() argument 1784 neigh_parms_release(struct neigh_table * tbl,struct neigh_parms * parms) neigh_parms_release() argument 1801 neigh_table_init(int index,struct neigh_table * tbl) neigh_table_init() argument 1865 neigh_table_clear(int index,struct neigh_table * tbl) neigh_table_clear() argument 1897 struct neigh_table *tbl = NULL; neigh_find_table() local 1934 struct neigh_table *tbl; neigh_delete() local 2001 struct neigh_table *tbl; neigh_add() local 2223 neightbl_fill_info(struct sk_buff * skb,struct neigh_table * tbl,u32 pid,u32 seq,int type,int flags) neightbl_fill_info() argument 2314 neightbl_fill_param_info(struct sk_buff * skb,struct neigh_table * tbl,struct neigh_parms * parms,u32 pid,u32 seq,int type,unsigned int flags) neightbl_fill_param_info() argument 2378 struct neigh_table *tbl; neightbl_set() local 2571 struct neigh_table *tbl; neightbl_dump_info() local 2689 pneigh_fill_info(struct sk_buff * skb,struct pneigh_entry * pn,u32 pid,u32 seq,int type,unsigned int flags,struct neigh_table * tbl) pneigh_fill_info() argument 2770 neigh_dump_table(struct neigh_table * tbl,struct sk_buff * skb,struct netlink_callback * cb,struct neigh_dump_filter * filter) neigh_dump_table() argument 2811 pneigh_dump_table(struct neigh_table * tbl,struct sk_buff * skb,struct netlink_callback * cb,struct neigh_dump_filter * filter) pneigh_dump_table() argument 2916 struct neigh_table *tbl; neigh_dump_info() local 3042 struct neigh_table *tbl; neigh_get() local 3122 neigh_for_each(struct neigh_table * tbl,void (* cb)(struct neighbour *,void *),void * cookie) neigh_for_each() argument 3143 __neigh_for_each_release(struct neigh_table * tbl,int (* cb)(struct neighbour *)) __neigh_for_each_release() argument 3179 struct neigh_table *tbl; neigh_xmit() local 3319 struct neigh_table *tbl = state->tbl; pneigh_get_first() local 3343 struct neigh_table *tbl = state->tbl; pneigh_get_next() local 3395 neigh_seq_start(struct seq_file * seq,loff_t * pos,struct neigh_table * tbl,unsigned int neigh_seq_flags) neigh_seq_start() argument 3445 struct neigh_table *tbl = state->tbl; neigh_seq_stop() local 3456 struct neigh_table *tbl = pde_data(file_inode(seq->file)); neigh_stat_seq_start() local 3473 struct neigh_table *tbl = pde_data(file_inode(seq->file)); neigh_stat_seq_next() local 3493 struct neigh_table *tbl = pde_data(file_inode(seq->file)); neigh_stat_seq_show() local 3842 struct neigh_table *tbl = p->tbl; neigh_sysctl_register() local [all...] |
/linux/drivers/net/wireless/intel/iwlegacy/ |
H A D | 4965-rs.c | 383 il4965_get_expected_tpt(struct il_scale_tbl_info *tbl, int rs_idx) in il4965_get_expected_tpt() argument 385 if (tbl->expected_tpt) in il4965_get_expected_tpt() 386 return tbl->expected_tpt[rs_idx]; in il4965_get_expected_tpt() 398 il4965_rs_collect_tx_data(struct il_scale_tbl_info *tbl, int scale_idx, in il4965_rs_collect_tx_data() argument 409 win = &(tbl->win[scale_idx]); in il4965_rs_collect_tx_data() 412 tpt = il4965_get_expected_tpt(tbl, scale_idx); in il4965_rs_collect_tx_data() 476 il4965_rate_n_flags_from_tbl(struct il_priv *il, struct il_scale_tbl_info *tbl, in il4965_rate_n_flags_from_tbl() argument 481 if (is_legacy(tbl->lq_type)) { in il4965_rate_n_flags_from_tbl() 486 } else if (is_Ht(tbl->lq_type)) { in il4965_rate_n_flags_from_tbl() 493 if (is_siso(tbl in il4965_rate_n_flags_from_tbl() 532 il4965_rs_get_tbl_info_from_mcs(const u32 rate_n_flags,enum nl80211_band band,struct il_scale_tbl_info * tbl,int * rate_idx) il4965_rs_get_tbl_info_from_mcs() argument 592 il4965_rs_toggle_antenna(u32 valid_ant,u32 * rate_n_flags,struct il_scale_tbl_info * tbl) il4965_rs_toggle_antenna() argument 709 il4965_rs_get_lower_rate(struct il_lq_sta * lq_sta,struct il_scale_tbl_info * tbl,u8 scale_idx,u8 ht_possible) il4965_rs_get_lower_rate() argument 1006 il4965_rs_set_expected_tpt_table(struct il_lq_sta * lq_sta,struct il_scale_tbl_info * tbl) il4965_rs_set_expected_tpt_table() argument 1059 il4965_rs_get_best_rate(struct il_priv * il,struct il_lq_sta * lq_sta,struct il_scale_tbl_info * tbl,u16 rate_mask,s8 idx) il4965_rs_get_best_rate() argument 1155 il4965_rs_switch_to_mimo2(struct il_priv * il,struct il_lq_sta * lq_sta,struct ieee80211_conf * conf,struct ieee80211_sta * sta,struct il_scale_tbl_info * tbl,int idx) il4965_rs_switch_to_mimo2() argument 1208 il4965_rs_switch_to_siso(struct il_priv * il,struct il_lq_sta * lq_sta,struct ieee80211_conf * conf,struct ieee80211_sta * sta,struct il_scale_tbl_info * tbl,int idx) il4965_rs_switch_to_siso() argument 1257 struct il_scale_tbl_info *tbl = &(lq_sta->lq_info[lq_sta->active_tbl]); il4965_rs_move_legacy_other() local 1377 struct il_scale_tbl_info *tbl = &(lq_sta->lq_info[lq_sta->active_tbl]); il4965_rs_move_siso_to_other() local 1500 struct il_scale_tbl_info *tbl = &(lq_sta->lq_info[lq_sta->active_tbl]); il4965_rs_move_mimo2_to_other() local 1626 struct il_scale_tbl_info *tbl; il4965_rs_stay_in_table() local 1705 il4965_rs_update_rate_tbl(struct il_priv * il,struct il_lq_sta * lq_sta,struct il_scale_tbl_info * tbl,int idx,u8 is_green) il4965_rs_update_rate_tbl() argument 1739 struct il_scale_tbl_info *tbl, *tbl1; il4965_rs_rate_scale_perform() local 2129 struct il_scale_tbl_info *tbl; il4965_rs_initialize_lq() local 2573 struct il_scale_tbl_info *tbl = &(lq_sta->lq_info[lq_sta->active_tbl]); il4965_rs_sta_dbgfs_scale_table_read() local 2722 struct il_scale_tbl_info *tbl = &lq_sta->lq_info[lq_sta->active_tbl]; il4965_rs_sta_dbgfs_rate_scale_data_read() local [all...] |
/linux/tools/perf/ |
H A D | check-headers.sh | 72 "scripts/syscall.tbl" 196 check_2 tools/perf/arch/x86/entry/syscalls/syscall_32.tbl arch/x86/entry/syscalls/syscall_32.tbl 197 check_2 tools/perf/arch/x86/entry/syscalls/syscall_64.tbl arch/x86/entry/syscalls/syscall_64.tbl 198 check_2 tools/perf/arch/powerpc/entry/syscalls/syscall.tbl arch/powerpc/kernel/syscalls/syscall.tbl 199 check_2 tools/perf/arch/s390/entry/syscalls/syscall.tbl arch/s390/kernel/syscalls/syscall.tbl 200 check_2 tools/perf/arch/mips/entry/syscalls/syscall_n64.tbl arc [all...] |
/linux/ipc/ |
H A D | ipc_sysctl.c | 247 struct ctl_table *tbl; in setup_ipc_sysctls() local 251 tbl = kmemdup(ipc_sysctls, sizeof(ipc_sysctls), GFP_KERNEL); in setup_ipc_sysctls() 252 if (tbl) { in setup_ipc_sysctls() 256 if (tbl[i].data == &init_ipc_ns.shm_ctlmax) in setup_ipc_sysctls() 257 tbl[i].data = &ns->shm_ctlmax; in setup_ipc_sysctls() 259 else if (tbl[i].data == &init_ipc_ns.shm_ctlall) in setup_ipc_sysctls() 260 tbl[i].data = &ns->shm_ctlall; in setup_ipc_sysctls() 262 else if (tbl[i].data == &init_ipc_ns.shm_ctlmni) in setup_ipc_sysctls() 263 tbl[i].data = &ns->shm_ctlmni; in setup_ipc_sysctls() 265 else if (tbl[ in setup_ipc_sysctls() 307 const struct ctl_table *tbl; retire_ipc_sysctls() local [all...] |
/linux/net/netfilter/ |
H A D | xt_repldata.h | 9 /* tbl has the following structure equivalent, but is C99 compliant: 14 * } *tbl; 24 } *tbl; \ 26 size_t term_offset = (offsetof(typeof(*tbl), entries[nhooks]) + \ 28 tbl = kzalloc(term_offset + sizeof(*term), GFP_KERNEL); \ 29 if (tbl == NULL) \ 31 term = (struct type##_error *)&(((char *)tbl)[term_offset]); \ 32 strscpy(tbl->repl.name, info->name); \ 34 tbl->repl.valid_hooks = hook_mask; \ 35 tbl [all...] |
/linux/drivers/vfio/ |
H A D | vfio_iommu_spapr_tce.c | 209 struct iommu_table *tbl = container->tables[i]; in tce_iommu_find_table() local 211 if (tbl) { in tce_iommu_find_table() 212 unsigned long entry = ioba >> tbl->it_page_shift; in tce_iommu_find_table() 213 unsigned long start = tbl->it_offset; in tce_iommu_find_table() 214 unsigned long end = start + tbl->it_size; in tce_iommu_find_table() 217 *ptbl = tbl; in tce_iommu_find_table() 339 struct iommu_table *tbl, 342 struct iommu_table *tbl); 362 struct iommu_table *tbl = container->tables[i]; in tce_iommu_release() local 364 if (!tbl) in tce_iommu_release() 410 tce_iommu_unuse_page_v2(struct tce_container * container,struct iommu_table * tbl,unsigned long entry) tce_iommu_unuse_page_v2() argument 432 tce_iommu_clear(struct tce_container * container,struct iommu_table * tbl,unsigned long entry,unsigned long pages) tce_iommu_clear() argument 500 tce_iommu_build(struct tce_container * container,struct iommu_table * tbl,unsigned long entry,unsigned long tce,unsigned long pages,enum dma_data_direction direction) tce_iommu_build() argument 548 tce_iommu_build_v2(struct tce_container * container,struct iommu_table * tbl,unsigned long entry,unsigned long tce,unsigned long pages,enum dma_data_direction direction) tce_iommu_build_v2() argument 635 tce_iommu_free_table(struct tce_container * container,struct iommu_table * tbl) tce_iommu_free_table() argument 649 struct iommu_table *tbl = NULL; tce_iommu_create_window() local 712 struct iommu_table *tbl; tce_iommu_remove_window() local 907 struct iommu_table *tbl = NULL; tce_iommu_ioctl() local 974 struct iommu_table *tbl = NULL; tce_iommu_ioctl() local 1217 struct iommu_table *tbl = container->tables[i]; tce_iommu_take_ownership() local [all...] |
/linux/net/mac80211/ |
H A D | mesh_pathtbl.c | 20 static void mesh_path_free_rcu(struct mesh_table *tbl, struct mesh_path *mpath); 82 struct mesh_table *tbl = tblptr; in mesh_path_rht_free() local 84 mesh_path_free_rcu(tbl, mpath); in mesh_path_rht_free() 87 static void mesh_table_init(struct mesh_table *tbl) in mesh_table_init() argument 89 INIT_HLIST_HEAD(&tbl->known_gates); in mesh_table_init() 90 INIT_HLIST_HEAD(&tbl->walk_head); in mesh_table_init() 91 atomic_set(&tbl->entries, 0); in mesh_table_init() 92 spin_lock_init(&tbl->gates_lock); in mesh_table_init() 93 spin_lock_init(&tbl->walk_lock); in mesh_table_init() 98 WARN_ON(rhashtable_init(&tbl in mesh_table_init() 101 mesh_table_free(struct mesh_table * tbl) mesh_table_free() argument 243 mpath_lookup(struct mesh_table * tbl,const u8 * dst,struct ieee80211_sub_if_data * sdata) mpath_lookup() argument 280 __mesh_path_lookup_by_idx(struct mesh_table * tbl,int idx) __mesh_path_lookup_by_idx() argument 339 struct mesh_table *tbl; mesh_path_add_gate() local 374 mesh_gate_del(struct mesh_table * tbl,struct mesh_path * mpath) mesh_gate_del() argument 683 struct mesh_table *tbl; mesh_path_add() local 725 struct mesh_table *tbl; mpp_path_add() local 773 struct mesh_table *tbl = &sdata->u.mesh.mesh_paths; mesh_plink_broken() local 795 mesh_path_free_rcu(struct mesh_table * tbl,struct mesh_path * mpath) mesh_path_free_rcu() argument 811 __mesh_path_del(struct mesh_table * tbl,struct mesh_path * mpath) __mesh_path_del() argument 836 struct mesh_table *tbl = &sdata->u.mesh.mesh_paths; mesh_path_flush_by_nexthop() local 851 struct mesh_table *tbl = &sdata->u.mesh.mpp_paths; mpp_flush_by_proxy() local 863 table_flush_by_iface(struct mesh_table * tbl) table_flush_by_iface() argument 897 table_path_del(struct mesh_table * tbl,struct ieee80211_sub_if_data * sdata,const u8 * addr) table_path_del() argument 966 struct mesh_table *tbl; mesh_path_send_to_gates() local 1074 mesh_path_tbl_expire(struct ieee80211_sub_if_data * sdata,struct mesh_table * tbl) mesh_path_tbl_expire() argument [all...] |
/linux/arch/powerpc/kvm/ |
H A D | book3s_64_vio.c | 66 iommu_tce_table_put(stit->tbl); in kvm_spapr_tce_iommu_table_free() 98 if (table_group->tables[i] != stit->tbl) in kvm_spapr_tce_release_iommu_group() 114 struct iommu_table *tbl = NULL; in kvm_spapr_tce_attach_iommu_group() local 154 tbl = iommu_tce_table_get(tbltmp); in kvm_spapr_tce_attach_iommu_group() 158 if (!tbl) in kvm_spapr_tce_attach_iommu_group() 163 if (tbl != stit->tbl) in kvm_spapr_tce_attach_iommu_group() 168 iommu_tce_table_put(tbl); in kvm_spapr_tce_attach_iommu_group() 183 iommu_tce_table_put(tbl); in kvm_spapr_tce_attach_iommu_group() 187 stit->tbl in kvm_spapr_tce_attach_iommu_group() 412 u64 *tbl; kvmppc_tce_put() local 434 kvmppc_clear_tce(struct mm_struct * mm,struct kvmppc_spapr_tce_table * stt,struct iommu_table * tbl,unsigned long entry) kvmppc_clear_tce() argument 449 kvmppc_tce_iommu_mapped_dec(struct kvm * kvm,struct iommu_table * tbl,unsigned long entry) kvmppc_tce_iommu_mapped_dec() argument 470 kvmppc_tce_iommu_do_unmap(struct kvm * kvm,struct iommu_table * tbl,unsigned long entry) kvmppc_tce_iommu_do_unmap() argument 491 kvmppc_tce_iommu_unmap(struct kvm * kvm,struct kvmppc_spapr_tce_table * stt,struct iommu_table * tbl,unsigned long entry) kvmppc_tce_iommu_unmap() argument 509 kvmppc_tce_iommu_do_map(struct kvm * kvm,struct iommu_table * tbl,unsigned long entry,unsigned long ua,enum dma_data_direction dir) kvmppc_tce_iommu_do_map() argument 548 kvmppc_tce_iommu_map(struct kvm * kvm,struct kvmppc_spapr_tce_table * stt,struct iommu_table * tbl,unsigned long entry,unsigned long ua,enum dma_data_direction dir) kvmppc_tce_iommu_map() argument 774 u64 *tbl; kvmppc_h_get_tce() local [all...] |
/linux/drivers/net/ethernet/mellanox/mlx5/core/en/ |
H A D | mod_hdr.c | 42 void mlx5e_mod_hdr_tbl_init(struct mod_hdr_tbl *tbl) in mlx5e_mod_hdr_tbl_init() argument 44 mutex_init(&tbl->lock); in mlx5e_mod_hdr_tbl_init() 45 hash_init(tbl->hlist); in mlx5e_mod_hdr_tbl_init() 48 void mlx5e_mod_hdr_tbl_destroy(struct mod_hdr_tbl *tbl) in mlx5e_mod_hdr_tbl_destroy() argument 50 WARN_ON(!hash_empty(tbl->hlist)); in mlx5e_mod_hdr_tbl_destroy() 51 mutex_destroy(&tbl->lock); in mlx5e_mod_hdr_tbl_destroy() 54 static struct mlx5e_mod_hdr_handle *mod_hdr_get(struct mod_hdr_tbl *tbl, in mod_hdr_get() argument 60 hash_for_each_possible(tbl->hlist, mh, mod_hdr_hlist, hash_key) { in mod_hdr_get() 73 struct mod_hdr_tbl *tbl, in mlx5e_mod_hdr_attach() argument 90 mutex_lock(&tbl in mlx5e_mod_hdr_attach() 140 mlx5e_mod_hdr_detach(struct mlx5_core_dev * mdev,struct mod_hdr_tbl * tbl,struct mlx5e_mod_hdr_handle * mh) mlx5e_mod_hdr_detach() argument [all...] |