| /linux/drivers/net/ethernet/chelsio/cxgb4/ |
| H A D | l2t.c | 65 static inline unsigned int vlan_prio(const struct l2t_entry *e) in vlan_prio() argument 67 return e->vlan >> VLAN_PRIO_SHIFT; in vlan_prio() 70 static inline void l2t_hold(struct l2t_data *d, struct l2t_entry *e) in l2t_hold() argument 72 if (atomic_add_return(1, &e->refcnt) == 1) /* 0 -> 1 transition */ in l2t_hold() 118 static int addreq(const struct l2t_entry *e, const u32 *addr) in addreq() argument 120 if (e->v6) in addreq() 121 return (e->addr[0] ^ addr[0]) | (e->addr[1] ^ addr[1]) | in addreq() 122 (e->addr[2] ^ addr[2]) | (e->addr[3] ^ addr[3]); in addreq() 123 return e->addr[0] ^ addr[0]; in addreq() 126 static void neigh_replace(struct l2t_entry *e, struct neighbour *n) in neigh_replace() argument [all …]
|
| H A D | smt.c | 68 struct smt_entry *e, *end; in find_or_alloc_smte() local 70 for (e = &s->smtab[0], end = &s->smtab[s->smt_size]; e != end; ++e) { in find_or_alloc_smte() 71 if (e->refcnt == 0) { in find_or_alloc_smte() 73 first_free = e; in find_or_alloc_smte() 75 if (e->state == SMT_STATE_SWITCHING) { in find_or_alloc_smte() 79 if (memcmp(e->src_mac, smac, ETH_ALEN) == 0) in find_or_alloc_smte() 86 e = first_free; in find_or_alloc_smte() 92 e->state = SMT_STATE_UNUSED; in find_or_alloc_smte() 95 return e; in find_or_alloc_smte() 98 static void t4_smte_free(struct smt_entry *e) in t4_smte_free() argument [all …]
|
| H A D | sched.c | 47 struct sched_class *e; in t4_sched_class_fw_cmd() local 50 e = &s->tab[p->u.params.class]; in t4_sched_class_fw_cmd() 58 p->u.params.channel, e->idx, in t4_sched_class_fw_cmd() 125 struct sched_class *e, *end; in t4_sched_entry_lookup() local 130 for (e = &s->tab[0]; e != end; ++e) { in t4_sched_entry_lookup() 131 if (e->state == SCHED_STATE_UNUSED || in t4_sched_entry_lookup() 132 e->bind_type != type) in t4_sched_entry_lookup() 139 list_for_each_entry(qe, &e->entry_list, list) { in t4_sched_entry_lookup() 150 list_for_each_entry(fe, &e->entry_list, list) { in t4_sched_entry_lookup() 190 struct sched_class *e; in t4_sched_queue_unbind() local [all …]
|
| /linux/drivers/net/ethernet/chelsio/cxgb3/ |
| H A D | l2t.c | 63 static inline unsigned int vlan_prio(const struct l2t_entry *e) in vlan_prio() argument 65 return e->vlan >> 13; in vlan_prio() 74 static inline void neigh_replace(struct l2t_entry *e, struct neighbour *n) in neigh_replace() argument 77 if (e->neigh) in neigh_replace() 78 neigh_release(e->neigh); in neigh_replace() 79 e->neigh = n; in neigh_replace() 88 struct l2t_entry *e) in setup_l2e_send_pending() argument 101 OPCODE_TID(req) = htonl(MK_OPCODE_TID(CPL_L2T_WRITE_REQ, e->idx)); in setup_l2e_send_pending() 102 req->params = htonl(V_L2T_W_IDX(e->idx) | V_L2T_W_IFF(e->smt_idx) | in setup_l2e_send_pending() 103 V_L2T_W_VLAN(e->vlan & VLAN_VID_MASK) | in setup_l2e_send_pending() [all …]
|
| /linux/tools/testing/selftests/powerpc/pmu/ |
| H A D | event.c | 24 static void __event_init_opts(struct event *e, u64 config, in __event_init_opts() argument 27 memset(e, 0, sizeof(*e)); in __event_init_opts() 29 e->name = name; in __event_init_opts() 31 e->attr.type = type; in __event_init_opts() 32 e->attr.config = config; in __event_init_opts() 33 e->attr.size = sizeof(e->attr); in __event_init_opts() 35 e->attr.read_format = PERF_FORMAT_TOTAL_TIME_ENABLED | \ in __event_init_opts() 38 e->attr.sample_period = 1000; in __event_init_opts() 39 e->attr.sample_type = PERF_SAMPLE_REGS_INTR; in __event_init_opts() 40 e->attr.disabled = 1; in __event_init_opts() [all …]
|
| /linux/drivers/media/test-drivers/vidtv/ |
| H A D | vidtv_s302m.c | 165 static void vidtv_s302m_access_unit_destroy(struct vidtv_encoder *e) in vidtv_s302m_access_unit_destroy() argument 167 struct vidtv_access_unit *head = e->access_units; in vidtv_s302m_access_unit_destroy() 176 e->access_units = NULL; in vidtv_s302m_access_unit_destroy() 179 static void vidtv_s302m_alloc_au(struct vidtv_encoder *e) in vidtv_s302m_alloc_au() argument 184 if (e->sync && e->sync->is_video_encoder) { in vidtv_s302m_alloc_au() 185 sync_au = e->sync->access_units; in vidtv_s302m_alloc_au() 188 temp = vidtv_s302m_access_unit_init(e->access_units); in vidtv_s302m_alloc_au() 189 if (!e->access_units) in vidtv_s302m_alloc_au() 190 e->access_units = temp; in vidtv_s302m_alloc_au() 198 e->access_units = vidtv_s302m_access_unit_init(NULL); in vidtv_s302m_alloc_au() [all …]
|
| /linux/fs/ |
| H A D | binfmt_misc.c | 94 Node *e; in search_binfmt_handler() local 97 list_for_each_entry(e, &misc->entries, list) { in search_binfmt_handler() 102 if (!test_bit(Enabled, &e->flags)) in search_binfmt_handler() 106 if (!test_bit(Magic, &e->flags)) { in search_binfmt_handler() 107 if (p && !strcmp(e->magic, p + 1)) in search_binfmt_handler() 108 return e; in search_binfmt_handler() 113 s = bprm->buf + e->offset; in search_binfmt_handler() 114 if (e->mask) { in search_binfmt_handler() 115 for (j = 0; j < e->size; j++) in search_binfmt_handler() 116 if ((*s++ ^ e->magic[j]) & e->mask[j]) in search_binfmt_handler() [all …]
|
| /linux/security/apparmor/ |
| H A D | policy_unpack_test.c | 52 struct aa_ext *e; member 60 struct aa_ext *e; in build_aa_ext_struct() local 65 e = kunit_kmalloc(test, sizeof(*e), GFP_USER); in build_aa_ext_struct() 66 KUNIT_EXPECT_NOT_ERR_OR_NULL(test, e); in build_aa_ext_struct() 68 e->start = buf; in build_aa_ext_struct() 69 e->end = e->start + buf_size; in build_aa_ext_struct() 70 e->pos = e->start; in build_aa_ext_struct() 74 strscpy(buf + 3, TEST_STRING_NAME, e->end - (void *)(buf + 3)); in build_aa_ext_struct() 76 buf = e->start + TEST_STRING_BUF_OFFSET; in build_aa_ext_struct() 79 strscpy(buf + 3, TEST_STRING_DATA, e->end - (void *)(buf + 3)); in build_aa_ext_struct() [all …]
|
| H A D | policy_unpack.c | 64 const char *name, const char *info, struct aa_ext *e, in audit_iface() argument 69 if (e) in audit_iface() 70 ad.iface.pos = e->pos - e->start; in audit_iface() 163 VISIBLE_IF_KUNIT bool aa_inbounds(struct aa_ext *e, size_t size) in aa_inbounds() argument 165 return (size <= e->end - e->pos); in aa_inbounds() 176 VISIBLE_IF_KUNIT size_t aa_unpack_u16_chunk(struct aa_ext *e, char **chunk) in aa_unpack_u16_chunk() argument 179 void *pos = e->pos; in aa_unpack_u16_chunk() 181 if (!aa_inbounds(e, sizeof(u16))) in aa_unpack_u16_chunk() 183 size = le16_to_cpu(get_unaligned((__le16 *) e->pos)); in aa_unpack_u16_chunk() 184 e->pos += sizeof(__le16); in aa_unpack_u16_chunk() [all …]
|
| /linux/scripts/kconfig/ |
| H A D | expr.c | 21 static struct expr *expr_eliminate_yn(struct expr *e); 35 struct expr *e; in expr_lookup() local 40 hash_for_each_possible(expr_hashtable, e, node, hash) { in expr_lookup() 41 if (e->type == type && e->left._initdata == l && in expr_lookup() 42 e->right._initdata == r) in expr_lookup() 43 return e; in expr_lookup() 46 e = xmalloc(sizeof(*e)); in expr_lookup() 47 e->type = type; in expr_lookup() 48 e->left._initdata = l; in expr_lookup() 49 e->right._initdata = r; in expr_lookup() [all …]
|
| /linux/drivers/md/ |
| H A D | dm-cache-policy-smq.c | 88 struct entry *e; in __get_entry() local 90 e = es->begin + block; in __get_entry() 91 BUG_ON(e >= es->end); in __get_entry() 93 return e; in __get_entry() 96 static unsigned int to_index(struct entry_space *es, struct entry *e) in to_index() argument 98 BUG_ON(e < es->begin || e >= es->end); in to_index() 99 return e - es->begin; in to_index() 133 static struct entry *l_next(struct entry_space *es, struct entry *e) in l_next() argument 135 return to_entry(es, e->next); in l_next() 138 static struct entry *l_prev(struct entry_space *es, struct entry *e) in l_prev() argument [all …]
|
| /linux/tools/testing/selftests/powerpc/pmu/ebb/ |
| H A D | trace.c | 78 struct trace_entry *e; in trace_alloc_entry() local 80 e = trace_alloc(tb, sizeof(*e) + payload_size); in trace_alloc_entry() 81 if (e) in trace_alloc_entry() 82 e->length = payload_size; in trace_alloc_entry() 84 return e; in trace_alloc_entry() 89 struct trace_entry *e; in trace_log_reg() local 92 e = trace_alloc_entry(tb, sizeof(reg) + sizeof(value)); in trace_log_reg() 93 if (!e) in trace_log_reg() 96 e->type = TRACE_TYPE_REG; in trace_log_reg() 97 p = (u64 *)e->data; in trace_log_reg() [all …]
|
| /linux/lib/ |
| H A D | lru_cache.c | 39 #define PARANOIA_LC_ELEMENT(lc, e) do { \ argument 41 struct lc_element *e_ = (e); \ 84 struct lc_element *e; in lc_create() local 128 e = p + e_off; in lc_create() 129 e->lc_index = i; in lc_create() 130 e->lc_number = LC_FREE; in lc_create() 131 e->lc_new_number = LC_FREE; in lc_create() 132 list_add(&e->list, &lc->free); in lc_create() 133 element[i] = e; in lc_create() 202 struct lc_element *e = lc->lc_element[i]; in lc_reset() local [all …]
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/esw/ |
| H A D | indir_table.c | 97 struct mlx5_esw_indir_table_entry *e) in mlx5_esw_indir_table_rule_get() argument 109 if (e->recirc_rule) { in mlx5_esw_indir_table_rule_get() 110 refcount_inc(&e->recirc_rule->refcnt); in mlx5_esw_indir_table_rule_get() 139 flow_act.fg = e->recirc_grp; in mlx5_esw_indir_table_rule_get() 146 handle = mlx5_add_flow_rules(e->ft, NULL, &flow_act, &dest, 1); in mlx5_esw_indir_table_rule_get() 156 e->recirc_rule = rule; in mlx5_esw_indir_table_rule_get() 172 struct mlx5_esw_indir_table_entry *e) in mlx5_esw_indir_table_rule_put() argument 174 struct mlx5_esw_indir_table_rule *rule = e->recirc_rule; in mlx5_esw_indir_table_rule_put() 187 e->recirc_rule = NULL; in mlx5_esw_indir_table_rule_put() 190 static int mlx5_create_indir_recirc_group(struct mlx5_esw_indir_table_entry *e) in mlx5_create_indir_recirc_group() argument [all …]
|
| /linux/drivers/mtd/ubi/ |
| H A D | wl.c | 127 struct ubi_wl_entry *e, struct rb_root *root); 129 struct ubi_wl_entry *e); 139 static void wl_tree_add(struct ubi_wl_entry *e, struct rb_root *root) in wl_tree_add() argument 150 if (e->ec < e1->ec) in wl_tree_add() 152 else if (e->ec > e1->ec) in wl_tree_add() 155 ubi_assert(e->pnum != e1->pnum); in wl_tree_add() 156 if (e->pnum < e1->pnum) in wl_tree_add() 163 rb_link_node(&e->u.rb, parent, p); in wl_tree_add() 164 rb_insert_color(&e->u.rb, root); in wl_tree_add() 175 static void wl_entry_destroy(struct ubi_device *ubi, struct ubi_wl_entry *e) in wl_entry_destroy() argument [all …]
|
| /linux/block/ |
| H A D | elevator.c | 76 struct elevator_queue *e = q->elevator; in elv_iosched_allow_bio_merge() local 78 if (e->type->ops.allow_merge) in elv_iosched_allow_bio_merge() 79 return e->type->ops.allow_merge(q, rq, bio); in elv_iosched_allow_bio_merge() 106 static bool elevator_match(const struct elevator_type *e, const char *name) in elevator_match() argument 108 return !strcmp(e->elevator_name, name) || in elevator_match() 109 (e->elevator_alias && !strcmp(e->elevator_alias, name)); in elevator_match() 114 struct elevator_type *e; in __elevator_find() local 116 list_for_each_entry(e, &elv_list, list) in __elevator_find() 117 if (elevator_match(e, name)) in __elevator_find() 118 return e; in __elevator_find() [all …]
|
| /linux/arch/sparc/vdso/ |
| H A D | vma.c | 68 static void *one_section64(struct vdso_elfinfo64 *e, const char *name, in one_section64() argument 75 shdrs = (void *)e->hdr + e->hdr->e_shoff; in one_section64() 76 snames = (void *)e->hdr + shdrs[e->hdr->e_shstrndx].sh_offset; in one_section64() 77 for (i = 1; i < e->hdr->e_shnum; i++) { in one_section64() 81 return (void *)e->hdr + shdrs[i].sh_offset; in one_section64() 89 struct vdso_elfinfo64 *e = &_e->u.elf64; in find_sections64() local 91 e->hdr = image->data; in find_sections64() 92 e->dynsym = one_section64(e, ".dynsym", &e->dynsymsize); in find_sections64() 93 e->dynstr = one_section64(e, ".dynstr", NULL); in find_sections64() 95 if (!e->dynsym || !e->dynstr) { in find_sections64() [all …]
|
| /linux/lib/crypto/ |
| H A D | sm3.c | 41 #define R(i, a, b, c, d, e, f, g, h, t, w1, w2) \ argument 43 ss1 = rol32((rol32((a), 12) + (e) + (t)), 7); \ 46 h += GG ## i(e, f, g) + ss1 + (w1); \ 52 #define R1(a, b, c, d, e, f, g, h, t, w1, w2) \ argument 53 R(1, a, b, c, d, e, f, g, h, t, w1, w2) 54 #define R2(a, b, c, d, e, f, g, h, t, w1, w2) \ argument 55 R(2, a, b, c, d, e, f, g, h, t, w1, w2) 77 u32 a, b, c, d, e, f, g, h, ss1, ss2; in sm3_transform() local 83 e = sctx->state[4]; in sm3_transform() 88 R1(a, b, c, d, e, f, g, h, K[0], I(0), I(4)); in sm3_transform() [all …]
|
| /linux/drivers/edac/ |
| H A D | edac_mc.c | 56 static struct mem_ctl_info *error_desc_to_mci(struct edac_raw_error_desc *e) in error_desc_to_mci() argument 58 return container_of(e, struct mem_ctl_info, error_desc); in error_desc_to_mci() 804 static void edac_inc_ce_error(struct edac_raw_error_desc *e) in edac_inc_ce_error() argument 806 int pos[EDAC_MAX_LAYERS] = { e->top_layer, e->mid_layer, e->low_layer }; in edac_inc_ce_error() 807 struct mem_ctl_info *mci = error_desc_to_mci(e); in edac_inc_ce_error() 810 mci->ce_mc += e->error_count; in edac_inc_ce_error() 813 dimm->ce_count += e->error_count; in edac_inc_ce_error() 815 mci->ce_noinfo_count += e->error_count; in edac_inc_ce_error() 818 static void edac_inc_ue_error(struct edac_raw_error_desc *e) in edac_inc_ue_error() argument 820 int pos[EDAC_MAX_LAYERS] = { e->top_layer, e->mid_layer, e->low_layer }; in edac_inc_ue_error() [all …]
|
| /linux/net/netfilter/ipset/ |
| H A D | ip_set_hash_netnet.c | 136 hash_netnet4_init(struct hash_netnet4_elem *e) in hash_netnet4_init() argument 138 e->cidr[0] = HOST_MASK; in hash_netnet4_init() 139 e->cidr[1] = HOST_MASK; in hash_netnet4_init() 149 struct hash_netnet4_elem e = { }; in hash_netnet4_kadt() local 152 e.cidr[0] = INIT_CIDR(h->nets[0].cidr[0], HOST_MASK); in hash_netnet4_kadt() 153 e.cidr[1] = INIT_CIDR(h->nets[0].cidr[1], HOST_MASK); in hash_netnet4_kadt() 155 e.ccmp = (HOST_MASK << (sizeof(e.cidr[0]) * 8)) | HOST_MASK; in hash_netnet4_kadt() 157 ip4addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip[0]); in hash_netnet4_kadt() 158 ip4addrptr(skb, opt->flags & IPSET_DIM_TWO_SRC, &e.ip[1]); in hash_netnet4_kadt() 159 e.ip[0] &= (ip_set_netmask(e.cidr[0]) & h->bitmask.ip); in hash_netnet4_kadt() [all …]
|
| H A D | ip_set_hash_netportnet.c | 144 hash_netportnet4_init(struct hash_netportnet4_elem *e) in hash_netportnet4_init() argument 146 e->cidr[0] = HOST_MASK; in hash_netportnet4_init() 147 e->cidr[1] = HOST_MASK; in hash_netportnet4_init() 157 struct hash_netportnet4_elem e = { }; in hash_netportnet4_kadt() local 160 e.cidr[0] = INIT_CIDR(h->nets[0].cidr[0], HOST_MASK); in hash_netportnet4_kadt() 161 e.cidr[1] = INIT_CIDR(h->nets[0].cidr[1], HOST_MASK); in hash_netportnet4_kadt() 163 e.ccmp = (HOST_MASK << (sizeof(e.cidr[0]) * 8)) | HOST_MASK; in hash_netportnet4_kadt() 166 &e.port, &e.proto)) in hash_netportnet4_kadt() 169 ip4addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip[0]); in hash_netportnet4_kadt() 170 ip4addrptr(skb, opt->flags & IPSET_DIM_THREE_SRC, &e.ip[1]); in hash_netportnet4_kadt() [all …]
|
| H A D | ip_set_list_set.c | 55 struct set_elem *e; in list_set_ktest() local 63 list_for_each_entry_rcu(e, &map->members, list) { in list_set_ktest() 64 ret = ip_set_test(e->id, skb, par, opt); in list_set_ktest() 67 if (ip_set_match_extensions(set, ext, mext, flags, e)) in list_set_ktest() 79 struct set_elem *e; in list_set_kadd() local 82 list_for_each_entry_rcu(e, &map->members, list) { in list_set_kadd() 84 ip_set_timeout_expired(ext_timeout(e, set))) in list_set_kadd() 86 ret = ip_set_add(e->id, skb, par, opt); in list_set_kadd() 99 struct set_elem *e; in list_set_kdel() local 102 list_for_each_entry_rcu(e, &map->members, list) { in list_set_kdel() [all …]
|
| /linux/drivers/gpu/drm/amd/amdgpu/ |
| H A D | amdgpu_ring_mux.c | 78 struct amdgpu_mux_entry *e = NULL; in amdgpu_mux_resubmit_chunks() local 89 e = &mux->ring_entry[i]; in amdgpu_mux_resubmit_chunks() 94 if (!e) { in amdgpu_mux_resubmit_chunks() 99 last_seq = atomic_read(&e->ring->fence_drv.last_seq); in amdgpu_mux_resubmit_chunks() 103 list_for_each_entry(chunk, &e->list, entry) { in amdgpu_mux_resubmit_chunks() 105 amdgpu_fence_update_start_timestamp(e->ring, in amdgpu_mux_resubmit_chunks() 109 le32_to_cpu(*(e->ring->fence_drv.cpu_addr + 2))) { in amdgpu_mux_resubmit_chunks() 110 if (chunk->cntl_offset <= e->ring->buf_mask) in amdgpu_mux_resubmit_chunks() 111 amdgpu_ring_patch_cntl(e->ring, in amdgpu_mux_resubmit_chunks() 113 if (chunk->ce_offset <= e->ring->buf_mask) in amdgpu_mux_resubmit_chunks() [all …]
|
| H A D | amdgpu_sync.c | 135 struct amdgpu_sync_entry *e; in amdgpu_sync_add_later() local 137 hash_for_each_possible(sync->fences, e, node, f->context) { in amdgpu_sync_add_later() 138 if (dma_fence_is_signaled(e->fence)) { in amdgpu_sync_add_later() 139 dma_fence_put(e->fence); in amdgpu_sync_add_later() 140 e->fence = dma_fence_get(f); in amdgpu_sync_add_later() 144 if (likely(e->fence->context == f->context)) { in amdgpu_sync_add_later() 145 amdgpu_sync_keep_later(&e->fence, f); in amdgpu_sync_add_later() 164 struct amdgpu_sync_entry *e; in amdgpu_sync_fence() local 172 e = kmem_cache_alloc(amdgpu_sync_slab, flags); in amdgpu_sync_fence() 173 if (!e) in amdgpu_sync_fence() [all …]
|
| /linux/tools/testing/selftests/powerpc/pmu/event_code_tests/ |
| H A D | event_alternatives_tests_p10.c | 27 struct event *e, events[5]; in event_alternatives_tests_p10() local 48 e = &events[0]; in event_alternatives_tests_p10() 49 event_init(e, 0x0001e); in event_alternatives_tests_p10() 51 e = &events[1]; in event_alternatives_tests_p10() 52 event_init(e, EventCode_1); in event_alternatives_tests_p10() 54 e = &events[2]; in event_alternatives_tests_p10() 55 event_init(e, EventCode_2); in event_alternatives_tests_p10() 57 e = &events[3]; in event_alternatives_tests_p10() 58 event_init(e, EventCode_3); in event_alternatives_tests_p10() 60 e = &events[4]; in event_alternatives_tests_p10() [all …]
|