| /linux/drivers/net/ethernet/chelsio/cxgb4/ |
| A D | l2t.c | 159 memcpy(e->dmac, e->neigh->ha, sizeof(e->dmac)); in write_l2e() 272 for (e = d->rover, end = &d->l2tab[d->l2t_size]; e != end; ++e) in alloc_l2e() 276 for (e = d->l2tab; atomic_read(&e->refcnt); ++e) in alloc_l2e() 304 for (e = &d->l2tab[0], end = &d->l2tab[d->l2t_size]; e != end; ++e) { in find_or_alloc_l2e() 440 for (e = d->l2tab[hash].first; e; e = e->next) in cxgb4_l2t_get() 442 e->vlan == vlan && e->lport == lport) { in cxgb4_l2t_get() 520 for (e = d->l2tab[hash].first; e; e = e->next) in t4_l2t_update() 727 e->idx + d->l2t_start, ip, e->dmac, in l2t_seq_show() 728 e->vlan & VLAN_VID_MASK, vlan_prio(e), e->lport, in l2t_seq_show() 729 l2e_state(e), atomic_read(&e->refcnt), in l2t_seq_show() [all …]
|
| A D | smt.c | 70 for (e = &s->smtab[0], end = &s->smtab[s->smt_size]; e != end; ++e) { in find_or_alloc_smte() 73 first_free = e; in find_or_alloc_smte() 86 e = first_free; in find_or_alloc_smte() 95 return e; in find_or_alloc_smte() 115 t4_smte_free(e); in cxgb4_smt_release() 196 row = e->idx; in write_smt_entry() 217 if (e) { in t4_smt_alloc_switching() 220 e->refcnt = 1; in t4_smt_alloc_switching() 222 e->pfvf = pfvf; in t4_smt_alloc_switching() 226 ++e->refcnt; in t4_smt_alloc_switching() [all …]
|
| A D | sched.c | 130 for (e = &s->tab[0]; e != end; ++e) { in t4_sched_entry_lookup() 328 if (!e) in t4_sched_class_unbind_all() 489 for (e = &s->tab[0]; e != end; ++e) { in t4_sched_class_lookup() 491 found = e; in t4_sched_class_lookup() 505 for (e = &s->tab[0]; e != end; ++e) { in t4_sched_class_lookup() 550 if (!e) { in t4_sched_class_alloc() 555 if (!e) in t4_sched_class_alloc() 564 memcpy(&e->info, &np, sizeof(e->info)); in t4_sched_class_alloc() 569 return e; in t4_sched_class_alloc() 639 memset(&e->info, 0, sizeof(e->info)); in cxgb4_sched_class_free() [all …]
|
| /linux/drivers/net/ethernet/chelsio/cxgb3/ |
| A D | l2t.c | 105 memcpy(e->dmac, e->neigh->ha, sizeof(e->dmac)); in setup_l2e_send_pending() 227 for (e = d->rover, end = &d->l2tab[d->nentries]; e != end; ++e) in alloc_l2e() 231 for (e = &d->l2tab[1]; atomic_read(&e->refcnt); ++e) ; in alloc_l2e() 241 int hash = arp_hash(e->addr, e->ifindex, d); in alloc_l2e() 250 return e; in alloc_l2e() 334 for (e = d->l2tab[hash].first; e; e = e->next) in t3_l2t_get() 335 if (e->addr == addr && e->ifindex == ifidx && in t3_l2t_get() 345 if (e) { in t3_l2t_get() 367 return e; in t3_l2t_get() 409 for (e = d->l2tab[hash].first; e; e = e->next) in t3_l2t_update() [all …]
|
| /linux/tools/testing/selftests/powerpc/pmu/ |
| A D | event.c | 27 memset(e, 0, sizeof(*e)); in __event_init_opts() 29 e->name = name; in __event_init_opts() 31 e->attr.type = type; in __event_init_opts() 32 e->attr.config = config; in __event_init_opts() 33 e->attr.size = sizeof(e->attr); in __event_init_opts() 40 e->attr.disabled = 1; in __event_init_opts() 71 e->fd = perf_event_open(&e->attr, pid, cpu, group_fd, 0); in event_open_with_options() 72 if (e->fd == -1) { in event_open_with_options() 102 close(e->fd); in event_close() 124 rc = read(e->fd, &e->result, sizeof(e->result)); in event_read() [all …]
|
| /linux/drivers/media/test-drivers/vidtv/ |
| A D | vidtv_s302m.c | 184 if (e->sync && e->sync->is_video_encoder) { in vidtv_s302m_alloc_au() 269 if (e->src_buf_offset > e->src_buf_sz) { in vidtv_s302m_get_sample() 277 if (e->src_buf_offset >= e->src_buf_sz) { in vidtv_s302m_get_sample() 280 e->last_sample_cb(e->sample_count); in vidtv_s302m_get_sample() 285 sample = *(u16 *)(e->src_buf + e->src_buf_offset); in vidtv_s302m_get_sample() 410 if (e->sync && e->sync->is_video_encoder) { in vidtv_s302m_encode() 448 e = kzalloc(sizeof(*e), GFP_KERNEL); in vidtv_s302m_encoder_init() 449 if (!e) in vidtv_s302m_encoder_init() 501 return e; in vidtv_s302m_encoder_init() 508 kfree(e); in vidtv_s302m_encoder_init() [all …]
|
| /linux/fs/ |
| A D | binfmt_misc.c | 94 Node *e; in search_binfmt_handler() local 116 if ((*s++ ^ e->magic[j]) & e->mask[j]) in search_binfmt_handler() 147 if (e) in get_binfmt_handler() 451 DUMP_PREFIX_NONE, e->mask, p - e->mask); in create_entry() 464 BINPRM_BUF_SIZE - e->size < e->offset) in create_entry() 470 DUMP_PREFIX_NONE, e->magic, e->size); in create_entry() 478 DUMP_PREFIX_NONE, e->mask, e->size); in create_entry() 482 masked[i] = e->magic[i] & e->mask[i]; in create_entry() 506 if (!e->magic[0] || strchr(e->magic, '/')) in create_entry() 605 dp = bin2hex(dp, e->magic, e->size); in entry_status() [all …]
|
| /linux/security/apparmor/ |
| A D | policy_unpack_test.c | 63 e = kunit_kmalloc(test, sizeof(*e), GFP_USER); in build_aa_ext_struct() 66 e->start = buf; in build_aa_ext_struct() 67 e->end = e->start + buf_size; in build_aa_ext_struct() 68 e->pos = e->start; in build_aa_ext_struct() 108 return e; in build_aa_ext_struct() 220 puf->e->end = puf->e->start + TEST_BLOB_BUF_OFFSET in policy_unpack_test_unpack_blob_out_of_bounds() 261 puf->e->end = puf->e->pos + TEST_STRING_BUF_OFFSET in policy_unpack_test_unpack_str_out_of_bounds() 308 puf->e->end = puf->e->pos + TEST_STRING_BUF_OFFSET in policy_unpack_test_unpack_strdup_out_of_bounds() 405 puf->e->pos = puf->e->end - 1; in policy_unpack_test_unpack_u16_chunk_out_of_bounds_1() 427 puf->e->end = puf->e->pos + TEST_U16_DATA - 1; in policy_unpack_test_unpack_u16_chunk_out_of_bounds_2() [all …]
|
| A D | policy_unpack.c | 68 if (e) in audit_iface() 69 ad.iface.pos = e->pos - e->start; in audit_iface() 164 return (size <= e->end - e->pos); in aa_inbounds() 191 e->pos = pos; in aa_unpack_u16_chunk() 203 e->pos++; in aa_unpack_X() 250 e->pos = pos; in aa_unpack_nameX() 269 e->pos = pos; in unpack_u8() 438 ((e->pos - e->start) & 7); in unpack_dfa() 1136 if (!aa_unpack_u32(e, &e->version, "version")) { in verify_header() 1424 while (e.pos < e.end) { in aa_unpack() [all …]
|
| /linux/scripts/kconfig/ |
| A D | expr.c | 46 e = xmalloc(sizeof(*e)); in expr_lookup() 271 if (e) switch (e->type) { in expr_eliminate_yn() 557 e = expr_eliminate_yn(e); in expr_eliminate_dups() 607 e = expr_alloc_two(e->type, in expr_transform() 657 e = e->left.expr->left.expr; in expr_transform() 684 e = expr_transform(e); in expr_transform() 690 e = expr_transform(e); in expr_transform() 791 e = expr_alloc_one(E_NOT, e); in expr_trans_compare() 803 e = expr_alloc_one(E_NOT, e); in expr_trans_compare() 813 e = expr_alloc_one(E_NOT, e); in expr_trans_compare() [all …]
|
| /linux/block/ |
| A D | elevator.c | 96 (e->elevator_alias && !strcmp(e->elevator_alias, name)); in elevator_match() 105 return e; in __elevator_find() 115 if (e && (!elevator_tryget(e))) in elevator_find_get() 116 e = NULL; in elevator_find_get() 118 return e; in elevator_find_get() 133 eq->type = e; in elevator_alloc() 148 kfree(e); in elevator_release() 422 error = e->type ? entry->show(e, page) : -ENOENT; in elv_attr_show() 509 e->icq_cache = kmem_cache_create(e->icq_cache_name, e->icq_size, in elv_register() 592 if (!e) in elevator_init_mq() [all …]
|
| /linux/drivers/md/ |
| A D | dm-cache-policy-smq.c | 98 BUG_ON(e < es->begin || e >= es->end); in to_index() 221 for (e = l_head(es, l); e; e = l_next(es, e)) in l_pop_head() 234 for (e = l_tail(es, l); e; e = l_prev(es, e)) in l_pop_tail() 303 l_add_tail(q->es, q->qs + e->level, e); in q_push() 313 l_add_head(q->es, q->qs + e->level, e); in q_push_front() 328 l_del(q->es, q->qs + e->level, e); in q_del() 344 for (e = l_head(q->es, q->qs + level); e; e = l_next(q->es, e)) { in q_peek() 378 for (e = l_head(q->es, q->qs + level); e; e = l_next(q->es, e)) in __redist_pop_from() 380 l_del(q->es, q->qs + e->level, e); in __redist_pop_from() 637 for (e = h_head(ht, h); e; e = h_next(ht, e)) { in __h_lookup() [all …]
|
| /linux/tools/testing/selftests/powerpc/pmu/ebb/ |
| A D | trace.c | 80 e = trace_alloc(tb, sizeof(*e) + payload_size); in trace_alloc_entry() 81 if (e) in trace_alloc_entry() 84 return e; in trace_alloc_entry() 93 if (!e) in trace_log_reg() 110 if (!e) in trace_log_counter() 130 if (!e) in trace_log_string() 147 if (!e) in trace_log_indent() 160 if (!e) in trace_log_outdent() 264 printf("entry @ %p type %d\n", e, e->type); in trace_print_entry() 288 e = p; in trace_buffer_print() [all …]
|
| /linux/lib/ |
| A D | lru_cache.c | 288 return e && e->refcnt; in lc_is_used() 305 e->lc_number = e->lc_new_number = LC_FREE; in lc_del() 332 return e; in lc_prepare_for_change() 366 if (e) { in __lc_get() 367 if (e->lc_new_number != e->lc_number) { in __lc_get() 427 RETURN(e); in __lc_get() 532 e->lc_number = e->lc_new_number; in lc_committed() 554 BUG_ON(e->lc_number != e->lc_new_number); in lc_put() 596 if (e->lc_number != e->lc_new_number) in lc_seq_dump_details() 598 i, e->lc_number, e->lc_new_number, e->refcnt); in lc_seq_dump_details() [all …]
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/esw/ |
| A D | indir_table.c | 201 e->recirc_grp = mlx5_create_flow_group(e->ft, in); in mlx5_create_indir_recirc_group() 224 e->fwd_grp = mlx5_create_flow_group(e->ft, in); in mlx5_create_indir_fwd_group() 261 e = kzalloc(sizeof(*e), GFP_KERNEL); in mlx5_esw_indir_table_entry_create() 262 if (!e) in mlx5_esw_indir_table_entry_create() 275 e->ft = ft; in mlx5_esw_indir_table_entry_create() 296 return e; in mlx5_esw_indir_table_entry_create() 306 kfree(e); in mlx5_esw_indir_table_entry_create() 332 if (e) { in mlx5_esw_indir_table_get() 363 if (!e) in mlx5_esw_indir_table_put() 371 if (e->fwd_ref || e->recirc_rule) in mlx5_esw_indir_table_put() [all …]
|
| /linux/drivers/mtd/ubi/ |
| A D | wl.c | 524 e->pnum, e->ec); in serve_prot_queue() 603 e->pnum, e->ec, torture); in schedule_erase() 610 wl_wrk->e = e; in schedule_erase() 639 wl_wrk.e = e; in do_sync_erase() 1761 ubi->lookuptbl[e->pnum] = e; 1845 ubi->lookuptbl[e->pnum] = e; 1862 ubi->lookuptbl[e->pnum] = e; 1866 e->pnum, e->ec); 1870 e->pnum, e->ec); 2054 e->pnum, e->ec, root); [all …]
|
| /linux/tools/testing/selftests/powerpc/pmu/event_code_tests/ |
| A D | event_alternatives_tests_p10.c | 47 e = &events[0]; in event_alternatives_tests_p10() 48 event_init(e, 0x0001e); in event_alternatives_tests_p10() 50 e = &events[1]; in event_alternatives_tests_p10() 53 e = &events[2]; in event_alternatives_tests_p10() 56 e = &events[3]; in event_alternatives_tests_p10() 59 e = &events[4]; in event_alternatives_tests_p10() 75 e = &events[0]; in event_alternatives_tests_p10() 78 e = &events[1]; in event_alternatives_tests_p10() 81 e = &events[2]; in event_alternatives_tests_p10() 84 e = &events[3]; in event_alternatives_tests_p10() [all …]
|
| /linux/arch/sparc/vdso/ |
| A D | vma.c | 75 shdrs = (void *)e->hdr + e->hdr->e_shoff; in one_section64() 76 snames = (void *)e->hdr + shdrs[e->hdr->e_shstrndx].sh_offset; in one_section64() 91 e->hdr = image->data; in find_sections64() 92 e->dynsym = one_section64(e, ".dynsym", &e->dynsymsize); in find_sections64() 93 e->dynstr = one_section64(e, ".dynstr", NULL); in find_sections64() 95 if (!e->dynsym || !e->dynstr) { in find_sections64() 143 shdrs = (void *)e->hdr + e->hdr->e_shoff; in one_section32() 144 snames = (void *)e->hdr + shdrs[e->hdr->e_shstrndx].sh_offset; in one_section32() 160 e->dynsym = one_section32(e, ".dynsym", &e->dynsymsize); in find_sections32() 161 e->dynstr = one_section32(e, ".dynstr", NULL); in find_sections32() [all …]
|
| /linux/drivers/edac/ |
| A D | edac_mc.c | 806 int pos[EDAC_MAX_LAYERS] = { e->top_layer, e->mid_layer, e->low_layer }; in edac_inc_ce_error() 820 int pos[EDAC_MAX_LAYERS] = { e->top_layer, e->mid_layer, e->low_layer }; in edac_inc_ue_error() 840 e->error_count, e->msg, in edac_ce_error() 842 e->label, e->location, e->page_frame_number, e->offset_in_page, in edac_ce_error() 843 e->grain, e->syndrome, in edac_ce_error() 877 e->error_count, e->msg, in edac_ue_error() 879 e->label, e->location, e->page_frame_number, e->offset_in_page, in edac_ue_error() 889 e->msg, in edac_ue_error() 891 e->label, e->location, e->page_frame_number, e->offset_in_page, in edac_ue_error() 931 trace_mc_event(e->type, e->msg, e->label, e->error_count, in edac_raw_mc_handle_error() [all …]
|
| /linux/net/netfilter/ipset/ |
| A D | ip_set_hash_netnet.c | 155 e.ccmp = (HOST_MASK << (sizeof(e.cidr[0]) * 8)) | HOST_MASK; in hash_netnet4_kadt() 159 e.ip[0] &= (ip_set_netmask(e.cidr[0]) & h->bitmask.ip); in hash_netnet4_kadt() 160 e.ip[1] &= (ip_set_netmask(e.cidr[1]) & h->bitmask.ip); in hash_netnet4_kadt() 199 if (!e.cidr[0] || e.cidr[0] > HOST_MASK) in hash_netnet4_uadt() 205 if (!e.cidr[1] || e.cidr[1] > HOST_MASK) in hash_netnet4_uadt() 398 ip6_netmask(&e.ip[0], e.cidr[0]); in hash_netnet6_kadt() 399 ip6_netmask(&e.ip[1], e.cidr[1]); in hash_netnet6_kadt() 443 if (!e.cidr[0] || e.cidr[0] > HOST_MASK) in hash_netnet6_uadt() 449 if (!e.cidr[1] || e.cidr[1] > HOST_MASK) in hash_netnet6_uadt() 453 ip6_netmask(&e.ip[0], e.cidr[0]); in hash_netnet6_uadt() [all …]
|
| A D | ip_set_hash_netportnet.c | 166 &e.port, &e.proto)) in hash_netportnet4_kadt() 171 e.ip[0] &= ip_set_netmask(e.cidr[0]); in hash_netportnet4_kadt() 172 e.ip[1] &= ip_set_netmask(e.cidr[1]); in hash_netportnet4_kadt() 247 e.port = 0; in hash_netportnet4_uadt() 259 e.ip[0] = htonl(ip & ip_set_hostmask(e.cidr[0])); in hash_netportnet4_uadt() 260 e.ip[1] = htonl(ip2_from & ip_set_hostmask(e.cidr[1])); in hash_netportnet4_uadt() 461 &e.port, &e.proto)) in hash_netportnet6_kadt() 466 ip6_netmask(&e.ip[0], e.cidr[0]); in hash_netportnet6_kadt() 467 ip6_netmask(&e.ip[1], e.cidr[1]); in hash_netportnet6_kadt() 520 ip6_netmask(&e.ip[0], e.cidr[0]); in hash_netportnet6_uadt() [all …]
|
| A D | ip_set_list_set.c | 55 struct set_elem *e; in list_set_ktest() local 149 kfree(e); in __list_set_del_rcu() 199 prev = e; in list_set_utest() 251 n = e; in list_set_uadd() 255 next = e; in list_set_uadd() 257 prev = e; in list_set_uadd() 298 if (!e) in list_set_uadd() 300 e->id = d->id; in list_set_uadd() 301 e->set = set; in list_set_uadd() 330 prev = e; in list_set_udel() [all …]
|
| /linux/drivers/gpu/drm/amd/amdgpu/ |
| A D | amdgpu_ring_mux.c | 94 if (!e) { in amdgpu_mux_resubmit_chunks() 222 if (!e) { in amdgpu_ring_mux_set_wptr() 234 e->sw_cptr = e->sw_wptr; in amdgpu_ring_mux_set_wptr() 257 if (!e) { in amdgpu_ring_mux_get_wptr() 287 if (!e) { in amdgpu_ring_mux_get_rptr() 306 e->sw_rptr = e->sw_cptr; in amdgpu_ring_mux_get_rptr() 309 e->sw_rptr = e->sw_wptr; in amdgpu_ring_mux_get_rptr() 442 if (!e) { in amdgpu_ring_mux_start_ib() 468 if (!e) { in scan_and_remove_signaled_chunk() 491 if (!e) { in amdgpu_ring_mux_ib_mark_offset() [all …]
|
| A D | amdgpu_sync.c | 135 struct amdgpu_sync_entry *e; in amdgpu_sync_add_later() local 157 struct amdgpu_sync_entry *e; in amdgpu_sync_fence() local 166 if (!e) in amdgpu_sync_fence() 170 e->fence = dma_fence_get(f); in amdgpu_sync_fence() 296 hash_del(&e->node); in amdgpu_sync_entry_free() 297 dma_fence_put(e->fence); in amdgpu_sync_entry_free() 359 f = e->fence; in amdgpu_sync_get_fence() 361 hash_del(&e->node); in amdgpu_sync_get_fence() 389 f = e->fence; in amdgpu_sync_clone() 417 f = e->fence; in amdgpu_sync_push_to_job() [all …]
|
| /linux/arch/arm64/kvm/vgic/ |
| A D | vgic-irqfd.c | 46 e->set = vgic_irqfd_set_irq; in kvm_set_routing_entry() 54 e->set = kvm_set_msi; in kvm_set_routing_entry() 57 e->msi.data = ue->u.msi.data; in kvm_set_routing_entry() 58 e->msi.flags = ue->flags; in kvm_set_routing_entry() 59 e->msi.devid = ue->u.msi.devid; in kvm_set_routing_entry() 74 msi->data = e->msi.data; in kvm_populate_msi() 75 msi->flags = e->msi.flags; in kvm_populate_msi() 76 msi->devid = e->msi.devid; in kvm_populate_msi() 98 kvm_populate_msi(e, &msi); in kvm_set_msi() 112 switch (e->type) { in kvm_arch_set_irq_inatomic() [all …]
|