| /linux/drivers/base/ |
| A D | class.c | 186 if (cls->ns_type && !cls->namespace) { in class_register() 191 if (!cls->ns_type && cls->namespace) { in class_register() 251 kfree(cls); in class_create_release() 271 cls = kzalloc(sizeof(*cls), GFP_KERNEL); in class_create() 272 if (!cls) { in class_create() 284 return cls; in class_create() 287 kfree(cls); in class_create() 570 if (!cls) in class_compat_register() 574 kfree(cls); in class_compat_register() 577 return cls; in class_compat_register() [all …]
|
| /linux/scripts/ |
| A D | rust_is_available_test.py | 27 def generate_executable(cls, content): argument 37 def generate_clang(cls, stdout): argument 47 def generate_rustc(cls, stdout): argument 84 return cls.generate_bindgen(cls.bindgen_default_bindgen_version_stdout, None) 88 return cls.generate_bindgen(cls.bindgen_default_bindgen_version_stdout, stderr) 91 def setUpClass(cls): argument 94 cls.missing = pathlib.Path(cls.tempdir.name) / "missing" 96 cls.nonexecutable = pathlib.Path(cls.tempdir.name) / "nonexecutable" 111 cls.default_rustc = cls.generate_rustc(f"rustc {cls.rustc_default_version}") 112 …cls.default_bindgen = cls.generate_bindgen(cls.bindgen_default_bindgen_version_stdout, cls.bindge… [all …]
|
| /linux/drivers/net/ethernet/chelsio/cxgb4/ |
| A D | cxgb4_tc_u32.c | 46 struct tc_cls_u32_offload *cls, in fill_match_fields() argument 56 off = cls->knode.sel->keys[i].off; in fill_match_fields() 57 val = cls->knode.sel->keys[i].val; in fill_match_fields() 58 mask = cls->knode.sel->keys[i].mask; in fill_match_fields() 62 if (!cls->knode.sel->keys[i].offmask) in fill_match_fields() 66 if (cls->knode.sel->keys[i].offmask) in fill_match_fields() 99 exts = cls->knode.exts; in fill_action_fields() 177 TC_U32_NODE(cls->knode.handle)); in cxgb4_config_knode() 202 fs.tc_prio = cls->common.prio; in cxgb4_config_knode() 203 fs.tc_cookie = cls->knode.handle; in cxgb4_config_knode() [all …]
|
| A D | cxgb4_tc_matchall.c | 52 struct tc_cls_matchall_offload *cls) in cxgb4_matchall_egress_validate() argument 55 struct flow_action *actions = &cls->rule->action; in cxgb4_matchall_egress_validate() 165 struct tc_cls_matchall_offload *cls) in cxgb4_matchall_alloc_tc() argument 215 tc_port_matchall->egress.cookie = cls->cookie; in cxgb4_matchall_alloc_tc() 240 struct tc_cls_matchall_offload *cls) in cxgb4_matchall_mirror_alloc() argument 300 struct tc_cls_matchall_offload *cls, in cxgb4_matchall_add_filter() argument 315 false, cls->common.prio); in cxgb4_matchall_add_filter() 328 fs->tc_prio = cls->common.prio; in cxgb4_matchall_add_filter() 329 fs->tc_cookie = cls->cookie; in cxgb4_matchall_add_filter() 348 struct tc_cls_matchall_offload *cls) in cxgb4_matchall_alloc_filter() argument [all …]
|
| A D | cxgb4_tc_flower.h | 135 struct flow_cls_offload *cls); 137 struct flow_cls_offload *cls); 139 struct flow_cls_offload *cls);
|
| A D | cxgb4_tc_u32.h | 47 int cxgb4_config_knode(struct net_device *dev, struct tc_cls_u32_offload *cls); 48 int cxgb4_delete_knode(struct net_device *dev, struct tc_cls_u32_offload *cls);
|
| A D | cxgb4_tc_flower.c | 926 struct flow_cls_offload *cls) in cxgb4_tc_flower_replace() argument 928 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in cxgb4_tc_flower_replace() 929 struct netlink_ext_ack *extack = cls->common.extack; in cxgb4_tc_flower_replace() 943 fs->tc_cookie = cls->cookie; in cxgb4_tc_flower_replace() 950 ch_flower->tc_flower_cookie = cls->cookie; in cxgb4_tc_flower_replace() 960 cxgb4_tc_flower_hash_prio_del(adap, cls->common.prio); in cxgb4_tc_flower_replace() 989 struct flow_cls_offload *cls) in cxgb4_tc_flower_destroy() argument 995 ch_flower = ch_flower_lookup(adap, cls->cookie); in cxgb4_tc_flower_destroy() 1060 struct flow_cls_offload *cls) in cxgb4_tc_flower_stats() argument 1069 ch_flower = ch_flower_lookup(adap, cls->cookie); in cxgb4_tc_flower_stats() [all …]
|
| /linux/drivers/net/ethernet/stmicro/stmmac/ |
| A D | stmmac_tc.c | 33 u32 loc = cls->knode.handle; in tc_find_entry() 66 exts = cls->knode.exts; in tc_fill_actions() 193 ret = tc_fill_entry(priv, cls); in tc_config_knode() 205 tc_unfill_entry(priv, cls); in tc_config_knode() 213 tc_unfill_entry(priv, cls); in tc_delete_knode() 222 switch (cls->command) { in tc_setup_cls_u32() 224 tc_unfill_entry(priv, cls); in tc_setup_cls_u32() 593 cls->common.extack); in tc_add_flow() 606 entry->cookie = cls->cookie; in tc_add_flow() 695 entry->cookie = cls->cookie; in tc_add_vlan_flow() [all …]
|
| A D | stmmac_selftests.c | 1361 cls = kzalloc(sizeof(*cls), GFP_KERNEL); in __stmmac_test_l3filt() 1362 if (!cls) { in __stmmac_test_l3filt() 1367 cls->common.chain_index = 0; in __stmmac_test_l3filt() 1369 cls->cookie = dummy_cookie; in __stmmac_test_l3filt() 1386 cls->rule = rule; in __stmmac_test_l3filt() 1414 kfree(cls); in __stmmac_test_l3filt() 1489 cls = kzalloc(sizeof(*cls), GFP_KERNEL); in __stmmac_test_l4filt() 1490 if (!cls) { in __stmmac_test_l4filt() 1497 cls->cookie = dummy_cookie; in __stmmac_test_l4filt() 1515 cls->rule = rule; in __stmmac_test_l4filt() [all …]
|
| /linux/drivers/net/ethernet/freescale/dpaa2/ |
| A D | dpaa2-switch-flower.c | 494 struct flow_cls_offload *cls) in dpaa2_switch_cls_flower_replace_acl() argument 522 acl_entry->prio = cls->common.prio; in dpaa2_switch_cls_flower_replace_acl() 523 acl_entry->cookie = cls->cookie; in dpaa2_switch_cls_flower_replace_acl() 642 mirror_entry->cookie = cls->cookie; in dpaa2_switch_cls_flower_replace_mirror() 682 cls->cookie); in dpaa2_switch_cls_flower_destroy() 688 cls->cookie); in dpaa2_switch_cls_flower_destroy() 721 acl_entry->prio = cls->common.prio; in dpaa2_switch_cls_matchall_replace_acl() 722 acl_entry->cookie = cls->cookie; in dpaa2_switch_cls_matchall_replace_acl() 788 mirror_entry->cookie = cls->cookie; in dpaa2_switch_cls_matchall_replace_mirror() 878 cls->cookie); in dpaa2_switch_cls_matchall_destroy() [all …]
|
| A D | dpaa2-switch.h | 258 struct flow_cls_offload *cls); 261 struct flow_cls_offload *cls); 264 struct tc_cls_matchall_offload *cls); 267 struct tc_cls_matchall_offload *cls);
|
| /linux/include/linux/ |
| A D | dynamic_debug.h | 163 #define DEFINE_DYNAMIC_DEBUG_METADATA_CLS(name, cls, fmt) \ argument 172 .class_id = cls, \ 175 BUILD_BUG_ON_MSG(cls > _DPRINTK_CLASS_DFLT, \ 222 DEFINE_DYNAMIC_DEBUG_METADATA_CLS(id, cls, fmt); \ 231 DEFINE_DYNAMIC_DEBUG_METADATA_CLS(id, cls, fmt); \ 247 #define _dynamic_func_call_cls(cls, fmt, func, ...) \ argument 264 #define dynamic_pr_debug_cls(cls, fmt, ...) \ argument 265 _dynamic_func_call_cls(cls, fmt, __dynamic_pr_debug, \ 292 #define __pr_debug_cls(cls, fmt, ...) do { \ argument 293 BUILD_BUG_ON_MSG(!__builtin_constant_p(cls), \ [all …]
|
| A D | transport_class.h | 27 #define DECLARE_TRANSPORT_CLASS(cls, nm, su, rm, cfg) \ argument 28 struct transport_class cls = { \ 43 #define DECLARE_ANON_TRANSPORT_CLASS(cls, mtch, cfg) \ argument 44 struct anon_transport_class cls = { \
|
| /linux/scripts/gdb/linux/ |
| A D | device.py | 78 def class_for_each_device(cls): argument 79 for kn in klist_for_each(cls['klist_devices']): 125 for cls in for_each_class(): 126 gdb.write("class {}:\t{}\n".format(cls['class']['name'].string(), cls)) 127 for dev in class_for_each_device(cls): 130 cls = get_class_by_name(arg) 131 for dev in class_for_each_device(cls): 170 def invoke(self, cls, name): argument 172 cls = get_class_by_name(cls.string()) 173 for dev in class_for_each_device(cls):
|
| /linux/drivers/net/dsa/microchip/ |
| A D | ksz9477_tc_flower.c | 159 struct flow_cls_offload *cls, in ksz9477_flower_parse_action() argument 162 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in ksz9477_flower_parse_action() 170 if (TC_H_MIN(cls->classid)) { in ksz9477_flower_parse_action() 216 struct flow_cls_offload *cls, bool ingress) in ksz9477_cls_flower_add() argument 218 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in ksz9477_cls_flower_add() 219 struct netlink_ext_ack *extack = cls->common.extack; in ksz9477_cls_flower_add() 237 ret = ksz9477_flower_parse_key(dev, port, extack, rule, cls->cookie, in ksz9477_cls_flower_add() 238 cls->common.prio); in ksz9477_cls_flower_add() 242 ret = ksz9477_flower_parse_action(dev, port, extack, cls, in ksz9477_cls_flower_add() 270 struct flow_cls_offload *cls, bool ingress) in ksz9477_cls_flower_del() argument [all …]
|
| /linux/tools/perf/scripts/python/ |
| A D | compaction-times.py | 114 def add_filter(cls, filter): argument 115 cls.fobj = filter 121 head = cls.heads[pid] 124 if cls.fobj != None: 133 head = cls.heads[pid] 142 head = cls.heads[pid] 150 def gen(cls): argument 152 for i in cls.heads: 153 yield cls.heads[i] 156 def str(cls): argument [all …]
|
| /linux/tools/testing/selftests/hid/tests/ |
| A D | base_device.py | 136 def _init_pyudev(cls: Type["UdevHIDIsReady"]) -> None: 137 if cls._pyudev_context is None: 138 cls._pyudev_context = pyudev.Context() 139 cls._pyudev_monitor = pyudev.Monitor.from_netlink(cls._pyudev_context) 140 cls._pyudev_monitor.filter_by("hid") 141 cls._pyudev_monitor.start() 144 cls._pyudev_monitor.fileno(), cls._cls_udev_event_callback 148 def _cls_udev_event_callback(cls: Type["UdevHIDIsReady"]) -> None: 149 if cls._pyudev_monitor is None: 160 device_ready, count = cls._uhid_devices.get(id, (False, 0)) [all …]
|
| /linux/drivers/acpi/acpica/ |
| A D | utids.c | 341 struct acpi_pnp_device_id *cls; in acpi_ut_execute_CLS() local 377 cls = in acpi_ut_execute_CLS() 380 if (!cls) { in acpi_ut_execute_CLS() 387 cls->string = in acpi_ut_execute_CLS() 388 ACPI_ADD_PTR(char, cls, sizeof(struct acpi_pnp_device_id)); in acpi_ut_execute_CLS() 392 acpi_ex_pci_cls_to_string(cls->string, class_code); in acpi_ut_execute_CLS() 393 cls->length = length; in acpi_ut_execute_CLS() 394 *return_id = cls; in acpi_ut_execute_CLS()
|
| A D | nsxfname.c | 234 struct acpi_pnp_device_id *cls = NULL; in acpi_get_object_info() local 317 status = acpi_ut_execute_CLS(node, &cls); in acpi_get_object_info() 319 info_size += cls->length; in acpi_get_object_info() 427 if (cls) { in acpi_get_object_info() 429 cls, next_id_string); in acpi_get_object_info() 453 if (cls) { in acpi_get_object_info() 454 ACPI_FREE(cls); in acpi_get_object_info()
|
| /linux/drivers/net/dsa/sja1105/ |
| A D | sja1105_flower.c | 197 struct flow_cls_offload *cls, in sja1105_flower_parse_key() argument 200 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in sja1105_flower_parse_key() 347 struct flow_cls_offload *cls, bool ingress) in sja1105_cls_flower_add() argument 349 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in sja1105_cls_flower_add() 350 struct netlink_ext_ack *extack = cls->common.extack; in sja1105_cls_flower_add() 353 unsigned long cookie = cls->cookie; in sja1105_cls_flower_add() 360 rc = sja1105_flower_parse_key(priv, extack, cls, &key); in sja1105_cls_flower_add() 463 struct flow_cls_offload *cls, bool ingress) in sja1105_cls_flower_del() argument 503 struct flow_cls_offload *cls, bool ingress) in sja1105_cls_flower_stats() argument 515 rc = sja1105_vl_stats(priv, port, rule, &cls->stats, in sja1105_cls_flower_stats() [all …]
|
| /linux/drivers/net/ethernet/meta/fbnic/ |
| A D | fbnic_mac.c | 12 unsigned int cls, unsigned int readrq) in fbnic_init_readrq() argument 22 FIELD_PREP(FBNIC_QM_TNI_TDF_CTL_CLS, cls); in fbnic_init_readrq() 28 unsigned int cls, unsigned int mps) in fbnic_init_mps() argument 36 FIELD_PREP(FBNIC_QM_TNI_TCM_CTL_CLS, cls); in fbnic_init_mps() 44 int readrq, mps, cls; in fbnic_mac_init_axi() local 60 cls = ilog2(L1_CACHE_BYTES) - 6; in fbnic_mac_init_axi() 61 cls = clamp(cls, 0, 3); in fbnic_mac_init_axi() 65 fbnic_init_mps(fbd, FBNIC_QM_TNI_TCM_CTL, cls, mps); in fbnic_mac_init_axi() 77 FIELD_PREP(FBNIC_QM_TNI_TDE_CTL_CLS, cls)); in fbnic_mac_init_axi() 80 fbnic_init_mps(fbd, FBNIC_QM_RNI_RDE_CTL, cls, mps); in fbnic_mac_init_axi() [all …]
|
| /linux/tools/testing/selftests/bpf/progs/ |
| A D | test_spin_lock.c | 56 struct cls_elem *cls; in bpf_spin_lock_test() local 96 cls = bpf_get_local_storage(&cls_map, 0); in bpf_spin_lock_test() 97 bpf_spin_lock(&cls->lock); in bpf_spin_lock_test() 98 cls->cnt++; in bpf_spin_lock_test() 99 bpf_spin_unlock(&cls->lock); in bpf_spin_lock_test()
|
| /linux/arch/arm64/mm/ |
| A D | dma-mapping.c | 43 int cls = cache_line_size_of_cpu(); in arch_setup_dma_ops() local 45 WARN_TAINT(!coherent && cls > ARCH_DMA_MINALIGN, in arch_setup_dma_ops() 49 ARCH_DMA_MINALIGN, cls); in arch_setup_dma_ops()
|
| /linux/include/linux/device/ |
| A D | class.h | 84 void class_compat_unregister(struct class_compat *cls); 85 int class_compat_create_link(struct class_compat *cls, struct device *dev, 87 void class_compat_remove_link(struct class_compat *cls, struct device *dev, 229 void class_destroy(const struct class *cls);
|
| /linux/net/dsa/ |
| A D | user.c | 1384 cls->common.extack)) in dsa_user_add_cls_matchall_mirred() 1387 act = &cls->rule->action.entries[0]; in dsa_user_add_cls_matchall_mirred() 1407 mall_tc_entry->cookie = cls->cookie; in dsa_user_add_cls_matchall_mirred() 1451 cls->common.extack)) in dsa_user_add_cls_matchall_police() 1462 act = &cls->rule->action.entries[0]; in dsa_user_add_cls_matchall_police() 1536 if (cls->common.chain_index) in dsa_user_setup_tc_cls_matchall() 1539 switch (cls->command) { in dsa_user_setup_tc_cls_matchall() 1551 struct flow_cls_offload *cls, in dsa_user_add_cls_flower() argument 1565 struct flow_cls_offload *cls, in dsa_user_del_cls_flower() argument 1593 struct flow_cls_offload *cls, in dsa_user_setup_tc_cls_flower() argument [all …]
|