| /linux/drivers/gpu/drm/xe/ |
| A D | xe_guc_ct.c | 67 return container_of(ct, struct xe_guc, ct); in ct_to_guc() 73 return container_of(ct, struct xe_gt, uc.guc.ct); in ct_to_gt() 202 ct->bo = bo; in xe_guc_ct_init() 335 xe_gt_assert(ct_to_gt(ct), ct->g2h_outstanding == 0 || in xe_guc_ct_set_state() 527 xe_gt_assert(ct_to_gt(ct), ct->g2h_outstanding); in __g2h_release_space() 766 (desc_read(ct_to_xe(ct), (&ct->ctbs.g2h), tail) != ct->ctbs.g2h.info.head) in guc_ct_send_locked() 1323 len = g2h_read(ct, ct->fast_msg, true); in xe_guc_ct_fast_path() 1325 g2h_fast_path(ct, ct->fast_msg, len); in xe_guc_ct_fast_path() 1342 len = g2h_read(ct, ct->msg, false); in dequeue_one_g2h() 1347 ret = parse_g2h_msg(ct, ct->msg, len); in dequeue_one_g2h() [all …]
|
| A D | xe_guc_ct.h | 13 int xe_guc_ct_init(struct xe_guc_ct *ct); 14 int xe_guc_ct_enable(struct xe_guc_ct *ct); 15 void xe_guc_ct_disable(struct xe_guc_ct *ct); 16 void xe_guc_ct_stop(struct xe_guc_ct *ct); 17 void xe_guc_ct_fast_path(struct xe_guc_ct *ct); 28 return ct->state == XE_GUC_CT_STATE_ENABLED; in xe_guc_ct_enabled() 33 if (!xe_guc_ct_enabled(ct)) in xe_guc_ct_irq_handler() 36 wake_up_all(&ct->wq); in xe_guc_ct_irq_handler() 37 queue_work(ct->g2h_wq, &ct->g2h_worker); in xe_guc_ct_irq_handler() 38 xe_guc_ct_fast_path(ct); in xe_guc_ct_irq_handler() [all …]
|
| /linux/drivers/gpu/drm/i915/gt/uc/ |
| A D | intel_guc_ct.c | 42 return container_of(ct, struct intel_guc, ct); in ct_to_guc() 321 memset(ct, 0, sizeof(*ct)); in intel_guc_ct_fini() 374 ct->enabled = true; in intel_guc_ct_enable() 385 CT_DEAD(ct, SETUP); in intel_guc_ct_enable() 583 CT_ERROR(ct, "Head: %u (Dwords)\n", ct->ctbs.send.desc->head); in ct_deadlocked() 584 CT_ERROR(ct, "Tail: %u (Dwords)\n", ct->ctbs.send.desc->tail); in ct_deadlocked() 587 CT_ERROR(ct, "Head: %u\n (Dwords)", ct->ctbs.recv.desc->head); in ct_deadlocked() 588 CT_ERROR(ct, "Tail: %u\n (Dwords)", ct->ctbs.recv.desc->tail); in ct_deadlocked() 998 CT_DEAD(ct, READ); in ct_read() 1339 struct intel_guc_ct *ct = from_tasklet(ct, t, receive_tasklet); in ct_receive_tasklet_func() local [all …]
|
| /linux/tools/testing/selftests/bpf/progs/ |
| A D | test_bpf_nf_fail.c | 36 if (!ct) in alloc_release() 50 if (!ct) in insert_insert() 52 ct = bpf_ct_insert_entry(ct); in insert_insert() 53 if (!ct) in insert_insert() 55 ct = bpf_ct_insert_entry(ct); in insert_insert() 67 if (!ct) in lookup_insert() 81 if (!ct) in write_not_allowlisted_field() 95 if (!ct) in set_timeout_after_insert() 97 ct = bpf_ct_insert_entry(ct); in set_timeout_after_insert() 98 if (!ct) in set_timeout_after_insert() [all …]
|
| A D | test_bpf_nf.c | 103 if (ct) in nf_ct_test() 113 if (ct) in nf_ct_test() 122 if (ct) in nf_ct_test() 129 if (ct) in nf_ct_test() 138 if (ct) in nf_ct_test() 147 if (ct) in nf_ct_test() 154 if (ct) in nf_ct_test() 161 if (ct) in nf_ct_test() 173 if (ct) { in nf_ct_test() 232 if (ct) { in nf_ct_test() [all …]
|
| /linux/drivers/macintosh/ |
| A D | windfarm.h | 30 s32 (*get_min)(struct wf_control *ct); 31 s32 (*get_max)(struct wf_control *ct); 32 void (*release)(struct wf_control *ct); 63 s32 vmax = ct->ops->get_max(ct); in wf_control_set_max() 64 return ct->ops->set_value(ct, vmax); in wf_control_set_max() 69 s32 vmin = ct->ops->get_min(ct); in wf_control_set_min() 70 return ct->ops->set_value(ct, vmin); in wf_control_set_min() 75 return ct->ops->set_value(ct, val); in wf_control_set() 80 return ct->ops->get_value(ct, val); in wf_control_get() 85 return ct->ops->get_min(ct); in wf_control_get_min() [all …]
|
| /linux/drivers/video/fbdev/aty/ |
| A D | mach64_ct.c | 266 …ret = par->ref_clk_per * pll->ct.pll_ref_div * pll->ct.vclk_post_div_real / pll->ct.vclk_fb_div / … in aty_pll_to_var_ct() 268 if(pll->ct.xres > 0) { in aty_pll_to_var_ct() 270 ret /= pll->ct.xres; in aty_pll_to_var_ct() 293 pll->ct.pll_ext_cntl, pll->ct.pll_gen_cntl, pll->ct.pll_vclk_cntl); in aty_set_pll_ct() 298 pll->ct.pll_ref_div, pll->ct.vclk_post_div, pll->ct.vclk_post_div_real); in aty_set_pll_ct() 411 pll->ct.xclk_post_div = pll->ct.pll_ext_cntl & 0x07; in aty_init_pll_ct() 434 __func__, pll->ct.mclk_fb_mult, pll->ct.xclk_post_div); in aty_init_pll_ct() 487 if (pll->ct.xclkmaxrasdelay <= pll->ct.xclkpagefaultdelay) in aty_init_pll_ct() 488 pll->ct.xclkmaxrasdelay = pll->ct.xclkpagefaultdelay + 1; in aty_init_pll_ct() 540 pll->ct.mclk_fb_div = q * pll->ct.xclk_post_div_real / 8; in aty_init_pll_ct() [all …]
|
| /linux/net/netfilter/ |
| A D | nf_conntrack_proto_tcp.c | 857 memset(&ct->proto.tcp, 0, sizeof(ct->proto.tcp)); in tcp_new() 873 memset(&ct->proto.tcp, 0, sizeof(ct->proto.tcp)); in tcp_new() 926 spin_lock_bh(&ct->lock); in nf_conntrack_tcp_set_closing() 986 if (!nf_ct_is_confirmed(ct) && !tcp_new(ct, skb, dataoff, th, state)) in nf_conntrack_tcp_packet() 1024 if (nf_ct_kill(ct)) in nf_conntrack_tcp_packet() 1055 ct->proto.tcp.seen[ct->proto.tcp.last_dir].td_end = in nf_conntrack_tcp_packet() 1057 ct->proto.tcp.seen[ct->proto.tcp.last_dir].td_maxend = in nf_conntrack_tcp_packet() 1059 ct->proto.tcp.seen[ct->proto.tcp.last_dir].td_maxwin = in nf_conntrack_tcp_packet() 1062 ct->proto.tcp.seen[ct->proto.tcp.last_dir].td_scale = in nf_conntrack_tcp_packet() 1065 ct->proto.tcp.seen[ct->proto.tcp.last_dir].flags = in nf_conntrack_tcp_packet() [all …]
|
| A D | nf_conntrack_h323_main.c | 183 if (nf_ct_l3num(ct) != AF_INET) in get_h245_addr() 189 if (nf_ct_l3num(ct) != AF_INET6) in get_h245_addr() 254 ct->status & IPS_NAT_MASK) { in expect_rtp_rtcp() 280 struct nf_conn *ct, in expect_t120() argument 314 ct->status & IPS_NAT_MASK) { in expect_t120() 332 struct nf_conn *ct, in process_h245_channel() argument 735 struct nf_conn *ct, in expect_callforwarding() argument 758 nf_ct_l3num(ct))) { in expect_callforwarding() 866 struct nf_conn *ct, in process_callproceeding() argument 1198 if (exp && exp->master == ct) in find_expect() [all …]
|
| A D | nf_conntrack_core.c | 595 if (ct->master) in nf_ct_destroy() 680 nf_ct_put(ct); in nf_ct_delete() 724 nf_ct_kill(ct); in nf_ct_gc_expired() 726 nf_ct_put(ct); in nf_ct_gc_expired() 1481 return !nf_ct_is_confirmed(ct) || nf_ct_is_dying(ct); in gc_worker_skip_ct() 1698 return ct; in __nf_conntrack_alloc() 2050 if (!ct) { in nf_conntrack_in() 2260 if (!ct) in nf_conntrack_update() 2276 if (ct) { in nf_conntrack_get_tuple_skb() 2296 nf_ct_put(ct); in nf_conntrack_get_tuple_skb() [all …]
|
| A D | nf_nat_sip.c | 45 if (nf_ct_protonum(ct) == IPPROTO_TCP) { in mangle_packet() 73 if (nf_ct_l3num(ct) == NFPROTO_IPV4) in sip_sprintf_addr() 86 if (nf_ct_l3num(ct) == NFPROTO_IPV4) in sip_sprintf_addr_port() 173 if (nf_ct_protonum(ct) == IPPROTO_TCP) in nf_nat_sip() 189 &ct->tuplehash[dir].tuple.src.u3) || in nf_nat_sip() 194 &ct->tuplehash[dir].tuple.dst.u3) || in nf_nat_sip() 215 buflen = sip_sprintf_addr(ct, buffer, in nf_nat_sip() 216 &ct->tuplehash[!dir].tuple.dst.u3, in nf_nat_sip() 232 buflen = sip_sprintf_addr(ct, buffer, in nf_nat_sip() 233 &ct->tuplehash[!dir].tuple.src.u3, in nf_nat_sip() [all …]
|
| A D | nf_conntrack_proto_sctp.c | 253 memset(&ct->proto.sctp, 0, sizeof(ct->proto.sctp)); in sctp_new() 386 spin_lock_bh(&ct->lock); in nf_conntrack_sctp_packet() 474 if (ct->proto.sctp.init[dir] && ct->proto.sctp.init[!dir]) in nf_conntrack_sctp_packet() 501 if (ct->proto.sctp.init[dir] && ct->proto.sctp.init[!dir] && in nf_conntrack_sctp_packet() 517 spin_unlock_bh(&ct->lock); in nf_conntrack_sctp_packet() 532 spin_unlock_bh(&ct->lock); in nf_conntrack_sctp_packet() 561 spin_lock_bh(&ct->lock); in sctp_to_nlattr() 579 spin_unlock_bh(&ct->lock); in sctp_to_nlattr() 585 spin_unlock_bh(&ct->lock); in sctp_to_nlattr() 620 spin_lock_bh(&ct->lock); in nlattr_to_sctp() [all …]
|
| A D | xt_CT.c | 26 if (ct) { in xt_ct_target() 40 struct nf_conn *ct = info->ct; in xt_ct_target_v0() local 49 struct nf_conn *ct = info->ct; in xt_ct_target_v1() local 159 struct nf_conn *ct; in xt_ct_tg_check() local 163 ct = NULL; in xt_ct_tg_check() 185 if (!ct) { in xt_ct_tg_check() 220 info->ct = ct; in xt_ct_tg_check() 254 info->ct = info_v1.ct; in xt_ct_tg_check_v0() 282 struct nf_conn *ct = info->ct; in xt_ct_tg_destroy() local 285 if (ct) { in xt_ct_tg_destroy() [all …]
|
| A D | nf_conntrack_proto_dccp.c | 539 if (!nf_ct_is_confirmed(ct) && !dccp_new(ct, skb, dh, state)) in nf_conntrack_dccp_packet() 549 spin_lock_bh(&ct->lock); in nf_conntrack_dccp_packet() 593 spin_unlock_bh(&ct->lock); in nf_conntrack_dccp_packet() 597 spin_unlock_bh(&ct->lock); in nf_conntrack_dccp_packet() 602 ct->proto.dccp.last_dir = dir; in nf_conntrack_dccp_packet() 605 spin_unlock_bh(&ct->lock); in nf_conntrack_dccp_packet() 645 spin_lock_bh(&ct->lock); in dccp_to_nlattr() 663 spin_unlock_bh(&ct->lock); in dccp_to_nlattr() 668 spin_unlock_bh(&ct->lock); in dccp_to_nlattr() 706 spin_lock_bh(&ct->lock); in nlattr_to_dccp() [all …]
|
| A D | nf_nat_core.c | 133 if (ct == NULL) in __nf_nat_decode_session() 254 struct nf_conn *ct; in nf_nat_used_tuple_new() local 315 nf_ct_put(ct); in nf_nat_used_tuple_new() 355 struct nf_conn *ct; in nf_nat_used_tuple_harder() local 394 if (nf_ct_kill(ct)) in nf_nat_used_tuple_harder() 397 nf_ct_put(ct); in nf_nat_used_tuple_harder() 608 if (!ct->master) in nf_nat_l4proto_unique_tuple() 823 if (nfct_help(ct) && !nfct_seqadj(ct)) in nf_nat_setup_info() 919 struct nf_conn *ct; in nf_nat_inet_fn() local 968 ct, ct->status); in nf_nat_inet_fn() [all …]
|
| A D | nf_conntrack_netlink.c | 730 struct nf_conn *ct = item->ct; in ctnetlink_conntrack_event() local 1640 nf_ct_put(ct); in ctnetlink_del_conntrack() 1700 nf_ct_put(ct); in ctnetlink_get_conntrack() 2330 memset(&ct->proto, 0, sizeof(ct->proto)); in ctnetlink_create_conntrack() 2378 return ct; in ctnetlink_create_conntrack() 2484 nf_ct_put(ct); in ctnetlink_new_conntrack() 2723 if (ct->secmark && ctnetlink_dump_secctx(skb, ct) < 0) in __ctnetlink_glue_build() 2726 if (ct->master && ctnetlink_dump_master(skb, ct) < 0) in __ctnetlink_glue_build() 3287 c.data = ct; in ctnetlink_dump_exp_ct() 3290 nf_ct_put(ct); in ctnetlink_dump_exp_ct() [all …]
|
| A D | nf_conntrack_sip.c | 150 if (!ct) in sip_parse_addr() 154 switch (nf_ct_l3num(ct)) { in sip_parse_addr() 641 *proto = nf_ct_protonum(ct); in ct_sip_parse_transport() 654 switch (nf_ct_l3num(ct)) { in sdp_parse_addr() 878 switch (nf_ct_l3num(ct)) { in set_expected_rtp_rtcp() 1106 nf_ct_helper_log(skb, ct, in process_sdp() 1148 flush_expectations(ct, true); in process_invite_response() 1165 flush_expectations(ct, true); in process_update_response() 1196 flush_expectations(ct, true); in process_invite_request() 1211 flush_expectations(ct, true); in process_bye_request() [all …]
|
| A D | nf_conntrack_ecache.c | 98 nf_ct_put(ct); in ecache_work_evict_list() 132 struct net *net = nf_ct_net(item->ct); in __nf_conntrack_eventmask_report() 173 if (!nf_ct_is_confirmed(ct)) in nf_conntrack_eventmask_report() 176 e = nf_ct_ecache_find(ct); in nf_conntrack_eventmask_report() 182 item.ct = ct; in nf_conntrack_eventmask_report() 210 if (!nf_ct_is_confirmed(ct) || nf_ct_is_dying(ct)) in nf_ct_deliver_cached_events() 213 e = nf_ct_ecache_find(ct); in nf_ct_deliver_cached_events() 219 item.ct = ct; in nf_ct_deliver_cached_events() 291 net->ct.ecache_dwork_pending = true; in nf_conntrack_ecache_work() 302 struct net *net = nf_ct_net(ct); in nf_ct_ecache_ext_add() [all …]
|
| A D | nf_nat_ovs.c | 29 if (!nf_nat_icmp_reply_translation(skb, ct, ctinfo, in nf_ct_nat_execute() 41 if (!nf_nat_icmpv6_reply_translation(skb, ct, in nf_ct_nat_execute() 55 if (!nf_nat_initialized(ct, maniptype)) { in nf_ct_nat_execute() 61 ? nf_nat_setup_info(ct, range, maniptype) in nf_ct_nat_execute() 62 : nf_nat_alloc_null_binding(ct, hooknum); in nf_ct_nat_execute() 77 err = nf_nat_packet(ct, ctinfo, hooknum, skb); in nf_ct_nat_execute() 95 if (!nf_ct_is_confirmed(ct) && !nf_ct_nat_ext_add(ct)) in nf_ct_nat() 106 maniptype = ct->status & IPS_SRC_NAT in nf_ct_nat() 109 maniptype = ct->status & IPS_SRC_NAT in nf_ct_nat() 120 if (err == NF_ACCEPT && ct->status & IPS_DST_NAT) { in nf_ct_nat() [all …]
|
| /linux/include/net/netfilter/ |
| A D | nf_conntrack.h | 151 #define nf_ct_tuple(ct, dir) (&(ct)->tuplehash[dir].tuple) argument 160 return read_pnet(&ct->ct_net); in nf_ct_net() 185 if (ct && refcount_dec_and_test(&ct->ct_general.use)) in nf_ct_put() 186 nf_ct_destroy(&ct->ct_general); in nf_ct_put() 234 return nf_ct_delete(ct, 0, 0); in nf_ct_kill() 254 void nf_conntrack_free(struct nf_conn *ct); 287 if (WARN_ON(nf_ct_is_confirmed(ct))) in nf_conntrack_alter_reply() 311 return nf_ct_is_expired(ct) && nf_ct_is_confirmed(ct) && in nf_ct_should_gc() 312 !nf_ct_is_dying(ct); in nf_ct_should_gc() 323 if (nf_ct_expires(ct) < NF_CT_DAY / 2) in nf_ct_offload_timeout() [all …]
|
| A D | nf_conntrack_ecache.h | 30 nf_ct_ecache_find(const struct nf_conn *ct) in nf_ct_ecache_find() argument 33 return nf_ct_ext_find(ct, NF_CT_EXT_ECACHE); in nf_ct_ecache_find() 42 return nf_ct_ext_exist(ct, NF_CT_EXT_ECACHE); in nf_ct_ecache_exist() 52 struct nf_conn *ct; member 72 void nf_ct_deliver_cached_events(struct nf_conn *ct); 84 struct nf_conn *ct, in nf_conntrack_eventmask_report() argument 101 struct net *net = nf_ct_net(ct); in nf_conntrack_event_cache() 107 e = nf_ct_ecache_find(ct); in nf_conntrack_event_cache() 120 if (nf_ct_ecache_exist(ct)) in nf_conntrack_event_report() 130 if (nf_ct_ecache_exist(ct)) in nf_conntrack_event() [all …]
|
| A D | nf_conntrack_l4proto.h | 35 struct nf_conn *ct, bool destroy); 164 const struct nf_conn *ct, 180 const struct nf_conn *ct, in nf_ct_l4proto_log_invalid() argument 188 return &net->ct.nf_ct_proto.generic; in nf_generic_pernet() 193 return &net->ct.nf_ct_proto.tcp; in nf_tcp_pernet() 198 return &net->ct.nf_ct_proto.udp; in nf_udp_pernet() 203 return &net->ct.nf_ct_proto.icmp; in nf_icmp_pernet() 208 return &net->ct.nf_ct_proto.icmpv6; in nf_icmpv6_pernet() 229 return &net->ct.nf_ct_proto.dccp; in nf_dccp_pernet() 236 return &net->ct.nf_ct_proto.sctp; in nf_sctp_pernet() [all …]
|
| /linux/net/netfilter/ipvs/ |
| A D | ip_vs_nfct.c | 77 if (ct == NULL || nf_ct_is_confirmed(ct) || in ip_vs_update_conntrack() 78 nf_ct_is_dying(ct)) in ip_vs_update_conntrack() 95 !nfct_seqadj(ct) && !nfct_seqadj_ext_add(ct)) in ip_vs_update_conntrack() 121 __func__, ct, ct->status, ctinfo, in ip_vs_update_conntrack() 125 __func__, ct, ct->status, ctinfo, in ip_vs_update_conntrack() 166 __func__, ct, ct->status, ARG_CONN(cp)); in ip_vs_nfct_expect_callback() 182 __func__, ct, ct->status, ARG_CONN(cp)); in ip_vs_nfct_expect_callback() 194 __func__, ct, ct->status, ARG_TUPLE(orig)); in ip_vs_nfct_expect_callback() 217 if (ct == NULL) in ip_vs_nfct_expect_related() 245 struct nf_conn *ct; in ip_vs_conn_drop_conntrack() local [all …]
|
| /linux/kernel/ |
| A D | context_tracking.c | 129 WRITE_ONCE(ct->nmi_nesting, 0); in ct_kernel_exit() 134 ct->nesting--; in ct_kernel_exit() 145 instrument_atomic_write(&ct->state, sizeof(ct->state)); in ct_kernel_exit() 173 ct->nesting++; in ct_kernel_enter() 183 instrument_atomic_write(&ct->state, sizeof(ct->state)); in ct_kernel_enter() 187 WRITE_ONCE(ct->nesting, 1); in ct_kernel_enter() 235 instrument_atomic_write(&ct->state, sizeof(ct->state)); in ct_nmi_exit() 285 instrument_atomic_read(&ct->state, sizeof(ct->state)); in ct_nmi_enter() 287 instrument_atomic_write(&ct->state, sizeof(ct->state)); in ct_nmi_enter() 480 if (ct->active) { in __ct_user_enter() [all …]
|
| /linux/net/openvswitch/ |
| A D | conntrack.c | 157 return ct ? READ_ONCE(ct->mark) : 0; in ovs_ct_get_mark() 173 if (ct) { in ovs_ct_get_labels() 174 if (ct->master && !nf_ct_is_confirmed(ct)) in ovs_ct_get_labels() 175 ct = ct->master; in ovs_ct_get_labels() 204 key->ct.mark = ovs_ct_get_mark(ct); in __ovs_ct_update_key() 205 ovs_ct_get_labels(ct, &key->ct.labels); in __ovs_ct_update_key() 207 if (ct) { in __ovs_ct_update_key() 212 ct = ct->master; in __ovs_ct_update_key() 253 if (ct) { in ovs_ct_update_key() 574 if (ct) in skb_nfct_cached() [all …]
|