| /linux/drivers/net/ethernet/chelsio/cxgb4/ |
| H A D | cxgb4_tc_u32.c | 46 struct tc_cls_u32_offload *cls, in fill_match_fields() argument 55 for (i = 0; i < cls->knode.sel->nkeys; i++) { in fill_match_fields() 56 off = cls->knode.sel->keys[i].off; in fill_match_fields() 57 val = cls->knode.sel->keys[i].val; in fill_match_fields() 58 mask = cls->knode.sel->keys[i].mask; in fill_match_fields() 62 if (!cls->knode.sel->keys[i].offmask) in fill_match_fields() 66 if (cls->knode.sel->keys[i].offmask) in fill_match_fields() 92 struct tc_cls_u32_offload *cls) in fill_action_fields() argument 99 exts = cls->knode.exts; in fill_action_fields() 149 int cxgb4_config_knode(struct net_device *dev, struct tc_cls_u32_offload *cls) in cxgb4_config_knode() argument [all …]
|
| H A D | cxgb4_tc_flower.h | 135 struct flow_cls_offload *cls); 137 struct flow_cls_offload *cls); 139 struct flow_cls_offload *cls);
|
| H A D | cxgb4_tc_u32.h | 47 int cxgb4_config_knode(struct net_device *dev, struct tc_cls_u32_offload *cls); 48 int cxgb4_delete_knode(struct net_device *dev, struct tc_cls_u32_offload *cls);
|
| /linux/drivers/net/ethernet/stmicro/stmmac/ |
| H A D | stmmac_tc.c | 29 struct tc_cls_u32_offload *cls, in tc_find_entry() argument 33 u32 loc = cls->knode.handle; in tc_find_entry() 59 struct tc_cls_u32_offload *cls) in tc_fill_actions() argument 66 exts = cls->knode.exts; in tc_fill_actions() 92 struct tc_cls_u32_offload *cls) in tc_fill_entry() argument 95 struct tc_u32_sel *sel = cls->knode.sel; in tc_fill_entry() 97 u32 prio = cls->common.prio << 16; in tc_fill_entry() 108 switch (ntohs(cls->common.protocol)) { in tc_fill_entry() 124 entry = tc_find_entry(priv, cls, true); in tc_fill_entry() 129 frag = tc_find_entry(priv, cls, true); in tc_fill_entry() [all …]
|
| /linux/drivers/net/ethernet/freescale/dpaa2/ |
| H A D | dpaa2-switch-flower.c | 11 static int dpaa2_switch_flower_parse_key(struct flow_cls_offload *cls, in dpaa2_switch_flower_parse_key() argument 14 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in dpaa2_switch_flower_parse_key() 16 struct netlink_ext_ack *extack = cls->common.extack; in dpaa2_switch_flower_parse_key() 494 struct flow_cls_offload *cls) in dpaa2_switch_cls_flower_replace_acl() argument 496 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in dpaa2_switch_cls_flower_replace_acl() 497 struct netlink_ext_ack *extack = cls->common.extack; in dpaa2_switch_cls_flower_replace_acl() 512 err = dpaa2_switch_flower_parse_key(cls, &acl_entry->key); in dpaa2_switch_cls_flower_replace_acl() 522 acl_entry->prio = cls->common.prio; in dpaa2_switch_cls_flower_replace_acl() 523 acl_entry->cookie = cls->cookie; in dpaa2_switch_cls_flower_replace_acl() 537 static int dpaa2_switch_flower_parse_mirror_key(struct flow_cls_offload *cls, in dpaa2_switch_flower_parse_mirror_key() argument [all …]
|
| H A D | dpaa2-switch.h | 258 struct flow_cls_offload *cls); 261 struct flow_cls_offload *cls); 264 struct tc_cls_matchall_offload *cls); 267 struct tc_cls_matchall_offload *cls);
|
| /linux/include/linux/ |
| H A D | dynamic_debug.h | 170 #define DEFINE_DYNAMIC_DEBUG_METADATA_CLS(name, cls, fmt) \ argument 179 .class_id = cls, \ 182 BUILD_BUG_ON_MSG(cls > _DPRINTK_CLASS_DFLT, \ 228 #define __dynamic_func_call_cls(id, cls, fmt, func, ...) do { \ argument 229 DEFINE_DYNAMIC_DEBUG_METADATA_CLS(id, cls, fmt); \ 239 #define __dynamic_func_call_cls_no_desc(id, cls, fmt, func, ...) do { \ argument 240 DEFINE_DYNAMIC_DEBUG_METADATA_CLS(id, cls, fmt); \ 258 #define _dynamic_func_call_cls(cls, fmt, func, ...) \ argument 259 __dynamic_func_call_cls(__UNIQUE_ID(ddebug), cls, fmt, func, ##__VA_ARGS__) 268 #define _dynamic_func_call_cls_no_desc(cls, fmt, func, ...) \ argument [all …]
|
| H A D | transport_class.h | 27 #define DECLARE_TRANSPORT_CLASS(cls, nm, su, rm, cfg) \ argument 28 struct transport_class cls = { \ 43 #define DECLARE_ANON_TRANSPORT_CLASS(cls, mtch, cfg) \ argument 44 struct anon_transport_class cls = { \
|
| /linux/scripts/gdb/linux/ |
| H A D | device.py | 78 def class_for_each_device(cls): argument 79 for kn in klist_for_each(cls['klist_devices']): 125 for cls in for_each_class(): 126 gdb.write("class {}:\t{}\n".format(cls['class']['name'].string(), cls)) 127 for dev in class_for_each_device(cls): 130 cls = get_class_by_name(arg) 131 for dev in class_for_each_device(cls): 170 def invoke(self, cls, name): argument 172 cls = get_class_by_name(cls.string()) 173 for dev in class_for_each_device(cls):
|
| /linux/drivers/net/dsa/microchip/ |
| H A D | ksz9477_tc_flower.c | 159 struct flow_cls_offload *cls, in ksz9477_flower_parse_action() argument 162 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in ksz9477_flower_parse_action() 170 if (TC_H_MIN(cls->classid)) { in ksz9477_flower_parse_action() 216 struct flow_cls_offload *cls, bool ingress) in ksz9477_cls_flower_add() argument 218 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in ksz9477_cls_flower_add() 219 struct netlink_ext_ack *extack = cls->common.extack; in ksz9477_cls_flower_add() 237 ret = ksz9477_flower_parse_key(dev, port, extack, rule, cls->cookie, in ksz9477_cls_flower_add() 238 cls->common.prio); in ksz9477_cls_flower_add() 242 ret = ksz9477_flower_parse_action(dev, port, extack, cls, in ksz9477_cls_flower_add() 270 struct flow_cls_offload *cls, bool ingress) in ksz9477_cls_flower_del() argument [all …]
|
| /linux/tools/perf/scripts/python/ |
| H A D | compaction-times.py | 114 def add_filter(cls, filter): argument 115 cls.fobj = filter 118 def create_pending(cls, pid, comm, start_secs, start_nsecs): argument 121 head = cls.heads[pid] 124 if cls.fobj != None: 125 filtered = cls.fobj.filter(pid, comm) 126 head = cls.heads[pid] = chead(comm, pid, filtered) 132 def increment_pending(cls, pid, migrated, fscan, mscan): argument 133 head = cls.heads[pid] 141 def complete_pending(cls, pid, secs, nsecs): argument [all …]
|
| /linux/tools/testing/selftests/hid/tests/ |
| H A D | base_device.py | 144 def _init_pyudev(cls: Type["UdevHIDIsReady"]) -> None: 145 if cls._pyudev_context is None: 146 cls._pyudev_context = pyudev.Context() 147 cls._pyudev_monitor = pyudev.Monitor.from_netlink(cls._pyudev_context) 148 cls._pyudev_monitor.filter_by("hid") 149 cls._pyudev_monitor.start() 152 cls._pyudev_monitor.fileno(), cls._cls_udev_event_callback 156 def _cls_udev_event_callback(cls: Type["UdevHIDIsReady"]) -> None: 157 if cls._pyudev_monitor is None: 160 for event in iter(functools.partial(cls._pyudev_monitor.poll, 0.02), None): [all …]
|
| H A D | base.py | 198 def debug_reports(cls, reports, uhdev=None, events=None): argument 429 def instance(cls): argument 430 if not cls._instance: 431 cls._instance = HIDTestUdevRule() 432 return cls._instance
|
| /linux/drivers/acpi/acpica/ |
| H A D | utids.c | 341 struct acpi_pnp_device_id *cls; in acpi_ut_execute_CLS() local 377 cls = in acpi_ut_execute_CLS() 380 if (!cls) { in acpi_ut_execute_CLS() 387 cls->string = in acpi_ut_execute_CLS() 388 ACPI_ADD_PTR(char, cls, sizeof(struct acpi_pnp_device_id)); in acpi_ut_execute_CLS() 392 acpi_ex_pci_cls_to_string(cls->string, class_code); in acpi_ut_execute_CLS() 393 cls->length = length; in acpi_ut_execute_CLS() 394 *return_id = cls; in acpi_ut_execute_CLS()
|
| H A D | nsxfname.c | 234 struct acpi_pnp_device_id *cls = NULL; in acpi_get_object_info() local 317 status = acpi_ut_execute_CLS(node, &cls); in acpi_get_object_info() 319 info_size += cls->length; in acpi_get_object_info() 427 if (cls) { in acpi_get_object_info() 429 cls, next_id_string); in acpi_get_object_info() 453 if (cls) { in acpi_get_object_info() 454 ACPI_FREE(cls); in acpi_get_object_info()
|
| /linux/drivers/net/dsa/sja1105/ |
| H A D | sja1105_flower.c | 197 struct flow_cls_offload *cls, in sja1105_flower_parse_key() argument 200 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in sja1105_flower_parse_key() 347 struct flow_cls_offload *cls, bool ingress) in sja1105_cls_flower_add() argument 349 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in sja1105_cls_flower_add() 350 struct netlink_ext_ack *extack = cls->common.extack; in sja1105_cls_flower_add() 353 unsigned long cookie = cls->cookie; in sja1105_cls_flower_add() 360 rc = sja1105_flower_parse_key(priv, extack, cls, &key); in sja1105_cls_flower_add() 463 struct flow_cls_offload *cls, bool ingress) in sja1105_cls_flower_del() argument 466 struct sja1105_rule *rule = sja1105_rule_find(priv, cls->cookie); in sja1105_cls_flower_del() 474 return sja1105_vl_delete(priv, port, rule, cls->common.extack); in sja1105_cls_flower_del() [all …]
|
| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | test_spin_lock.c | 56 struct cls_elem *cls; in bpf_spin_lock_test() local 96 cls = bpf_get_local_storage(&cls_map, 0); in bpf_spin_lock_test() 97 bpf_spin_lock(&cls->lock); in bpf_spin_lock_test() 98 cls->cnt++; in bpf_spin_lock_test() 99 bpf_spin_unlock(&cls->lock); in bpf_spin_lock_test()
|
| H A D | test_tc_bpf.c | 11 int cls(struct __sk_buff *skb) in cls() function
|
| /linux/arch/arm64/mm/ |
| H A D | dma-mapping.c | 43 int cls = cache_line_size_of_cpu(); in arch_setup_dma_ops() local 45 WARN_TAINT(!coherent && cls > ARCH_DMA_MINALIGN, in arch_setup_dma_ops() 49 ARCH_DMA_MINALIGN, cls); in arch_setup_dma_ops()
|
| /linux/drivers/hid/ |
| H A D | hid-multitouch.c | 645 struct mt_class *cls = &td->mtclass; in mt_allocate_report_data() local 673 (cls->quirks & MT_QUIRK_APPLE_TOUCHBAR && in mt_allocate_report_data() 756 struct mt_class *cls = &td->mtclass; in mt_touch_input_mapping() local 789 set_abs(hi->input, code, field, cls->sn_move); in mt_touch_input_mapping() 814 set_abs(hi->input, code, field, cls->sn_move); in mt_touch_input_mapping() 830 if ((cls->name == MT_CLS_WIN_8 || in mt_touch_input_mapping() 831 cls->name == MT_CLS_WIN_8_FORCE_MULTI_INPUT || in mt_touch_input_mapping() 832 cls->name == MT_CLS_WIN_8_FORCE_MULTI_INPUT_NSMU || in mt_touch_input_mapping() 833 cls->name == MT_CLS_WIN_8_DISABLE_WAKEUP) && in mt_touch_input_mapping() 854 if (!(cls->quirks & MT_QUIRK_APPLE_TOUCHBAR)) in mt_touch_input_mapping() [all …]
|
| /linux/net/dsa/ |
| H A D | user.c | 1369 struct tc_cls_matchall_offload *cls, in dsa_user_add_cls_matchall_mirred() argument 1372 struct netlink_ext_ack *extack = cls->common.extack; in dsa_user_add_cls_matchall_mirred() 1382 if (cls->common.protocol != htons(ETH_P_ALL)) { in dsa_user_add_cls_matchall_mirred() 1394 if (!flow_action_basic_hw_stats_check(&cls->rule->action, extack)) in dsa_user_add_cls_matchall_mirred() 1397 act = &cls->rule->action.entries[0]; in dsa_user_add_cls_matchall_mirred() 1405 if (cls->common.skip_sw) { in dsa_user_add_cls_matchall_mirred() 1419 if (cls->common.skip_sw) { in dsa_user_add_cls_matchall_mirred() 1437 mall_tc_entry->cookie = cls->cookie; in dsa_user_add_cls_matchall_mirred() 1456 struct tc_cls_matchall_offload *cls, in dsa_user_add_cls_matchall_police() argument 1459 struct netlink_ext_ack *extack = cls->common.extack; in dsa_user_add_cls_matchall_police() [all …]
|
| /linux/drivers/ata/ |
| H A D | sata_sil.c | 654 u8 cls; in sil_init_controller() local 659 cls = sil_get_device_cache_line(pdev); in sil_init_controller() 660 if (cls) { in sil_init_controller() 661 cls >>= 3; in sil_init_controller() 662 cls++; /* cls = (line_size/8)+1 */ in sil_init_controller() 664 writew(cls << 8 | cls, in sil_init_controller()
|
| /linux/drivers/net/ethernet/mediatek/ |
| H A D | mtk_ppe_offload.c | 548 int mtk_flow_offload_cmd(struct mtk_eth *eth, struct flow_cls_offload *cls, in mtk_flow_offload_cmd() argument 554 switch (cls->command) { in mtk_flow_offload_cmd() 556 err = mtk_flow_offload_replace(eth, cls, ppe_index); in mtk_flow_offload_cmd() 559 err = mtk_flow_offload_destroy(eth, cls); in mtk_flow_offload_cmd() 562 err = mtk_flow_offload_stats(eth, cls); in mtk_flow_offload_cmd() 576 struct flow_cls_offload *cls = type_data; in mtk_eth_setup_tc_block_cb() local 590 return mtk_flow_offload_cmd(eth, cls, 0); in mtk_eth_setup_tc_block_cb()
|
| /linux/drivers/net/ethernet/ti/ |
| H A D | am65-cpsw-qos.c | 1025 struct flow_cls_offload *cls, in am65_cpsw_qos_clsflower_add_policer() argument 1028 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in am65_cpsw_qos_clsflower_add_policer() 1066 qos->ale_bc_ratelimit.cookie = cls->cookie; in am65_cpsw_qos_clsflower_add_policer() 1074 qos->ale_mc_ratelimit.cookie = cls->cookie; in am65_cpsw_qos_clsflower_add_policer() 1119 struct flow_cls_offload *cls) in am65_cpsw_qos_configure_clsflower() argument 1121 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in am65_cpsw_qos_configure_clsflower() 1122 struct netlink_ext_ack *extack = cls->common.extack; in am65_cpsw_qos_configure_clsflower() 1133 return am65_cpsw_qos_clsflower_add_policer(port, extack, cls, in am65_cpsw_qos_configure_clsflower() 1144 static int am65_cpsw_qos_delete_clsflower(struct am65_cpsw_port *port, struct flow_cls_offload *cls) in am65_cpsw_qos_delete_clsflower() argument 1148 if (cls->cookie == qos->ale_bc_ratelimit.cookie) { in am65_cpsw_qos_delete_clsflower() [all …]
|
| /linux/drivers/net/ethernet/marvell/octeontx2/nic/ |
| H A D | otx2_tc.c | 234 struct tc_cls_matchall_offload *cls) in otx2_tc_egress_matchall_install() argument 236 struct netlink_ext_ack *extack = cls->common.extack; in otx2_tc_egress_matchall_install() 237 struct flow_action *actions = &cls->rule->action; in otx2_tc_egress_matchall_install() 251 entry = &cls->rule->action.entries[0]; in otx2_tc_egress_matchall_install() 254 err = otx2_policer_validate(&cls->rule->action, entry, extack); in otx2_tc_egress_matchall_install() 278 struct tc_cls_matchall_offload *cls) in otx2_tc_egress_matchall_delete() argument 280 struct netlink_ext_ack *extack = cls->common.extack; in otx2_tc_egress_matchall_delete() 1427 struct tc_cls_matchall_offload *cls) in otx2_tc_ingress_matchall_install() argument 1429 struct netlink_ext_ack *extack = cls->common.extack; in otx2_tc_ingress_matchall_install() 1430 struct flow_action *actions = &cls->rule->action; in otx2_tc_ingress_matchall_install() [all …]
|