| /linux/drivers/media/platform/amphion/ |
| H A D | vpu_rpc.h | 120 struct vpu_iface_ops *ops = vpu_core_get_iface(core); in vpu_iface_check_codec() local 122 if (ops && ops->check_codec) in vpu_iface_check_codec() 123 return ops->check_codec(core->type); in vpu_iface_check_codec() 130 struct vpu_iface_ops *ops = vpu_inst_get_iface(inst); in vpu_iface_check_format() local 132 if (ops && ops->check_fmt) in vpu_iface_check_format() 133 return ops->check_fmt(inst->type, pixelfmt); in vpu_iface_check_format() 140 struct vpu_iface_ops *ops = vpu_core_get_iface(core); in vpu_iface_boot_core() local 142 if (ops && ops->boot_core) in vpu_iface_boot_core() 143 return ops->boot_core(core); in vpu_iface_boot_core() 149 struct vpu_iface_ops *ops = vpu_core_get_iface(core); in vpu_iface_get_power_state() local [all …]
|
| /linux/drivers/net/wireless/ti/wlcore/ |
| H A D | hw_ops.h | 17 if (!wl->ops->calc_tx_blocks) in wlcore_hw_calc_tx_blocks() 20 return wl->ops->calc_tx_blocks(wl, len, spare_blks); in wlcore_hw_calc_tx_blocks() 27 if (!wl->ops->set_tx_desc_blocks) in wlcore_hw_set_tx_desc_blocks() 30 return wl->ops->set_tx_desc_blocks(wl, desc, blks, spare_blks); in wlcore_hw_set_tx_desc_blocks() 38 if (!wl->ops->set_tx_desc_data_len) in wlcore_hw_set_tx_desc_data_len() 41 wl->ops->set_tx_desc_data_len(wl, desc, skb); in wlcore_hw_set_tx_desc_data_len() 48 if (!wl->ops->get_rx_buf_align) in wlcore_hw_get_rx_buf_align() 51 return wl->ops->get_rx_buf_align(wl, rx_desc); in wlcore_hw_get_rx_buf_align() 57 if (wl->ops->prepare_read) in wlcore_hw_prepare_read() 58 return wl->ops->prepare_read(wl, rx_desc, len); in wlcore_hw_prepare_read() [all …]
|
| /linux/drivers/gpu/drm/amd/display/amdgpu_dm/ |
| H A D | amdgpu_dm_colorop.c | 60 struct drm_colorop *ops[MAX_COLOR_PIPELINE_OPS]; in amdgpu_dm_initialize_default_pipeline() local 67 memset(ops, 0, sizeof(ops)); in amdgpu_dm_initialize_default_pipeline() 70 ops[i] = kzalloc_obj(*ops[0]); in amdgpu_dm_initialize_default_pipeline() 71 if (!ops[i]) { in amdgpu_dm_initialize_default_pipeline() 76 ret = drm_plane_colorop_curve_1d_init(dev, ops[i], plane, in amdgpu_dm_initialize_default_pipeline() 82 list->type = ops[i]->base.id; in amdgpu_dm_initialize_default_pipeline() 87 ops[i] = kzalloc_obj(struct drm_colorop); in amdgpu_dm_initialize_default_pipeline() 88 if (!ops[i]) { in amdgpu_dm_initialize_default_pipeline() 93 ret = drm_plane_colorop_mult_init(dev, ops[i], plane, DRM_COLOROP_FLAG_ALLOW_BYPASS); in amdgpu_dm_initialize_default_pipeline() 97 drm_colorop_set_next_property(ops[i-1], ops[i]); in amdgpu_dm_initialize_default_pipeline() [all …]
|
| /linux/drivers/mfd/ |
| H A D | abx500-core.c | 19 struct abx500_ops ops; member 23 static void lookup_ops(struct device *dev, struct abx500_ops **ops) in lookup_ops() argument 27 *ops = NULL; in lookup_ops() 30 *ops = &dev_entry->ops; in lookup_ops() 36 int abx500_register_ops(struct device *dev, struct abx500_ops *ops) in abx500_register_ops() argument 45 memcpy(&dev_entry->ops, ops, sizeof(*ops)); in abx500_register_ops() 65 struct abx500_ops *ops; in abx500_set_register_interruptible() local 67 lookup_ops(dev->parent, &ops); in abx500_set_register_interruptible() 68 if (ops && ops->set_register) in abx500_set_register_interruptible() 69 return ops->set_register(dev, bank, reg, value); in abx500_set_register_interruptible() [all …]
|
| /linux/include/linux/ |
| H A D | ntb.h | 178 static inline int ntb_client_ops_is_valid(const struct ntb_client_ops *ops) in ntb_client_ops_is_valid() argument 182 ops->probe && in ntb_client_ops_is_valid() 183 ops->remove && in ntb_client_ops_is_valid() 199 static inline int ntb_ctx_ops_is_valid(const struct ntb_ctx_ops *ops) in ntb_ctx_ops_is_valid() argument 334 static inline int ntb_dev_ops_is_valid(const struct ntb_dev_ops *ops) in ntb_dev_ops_is_valid() argument 339 !ops->peer_port_count == !ops->port_number && in ntb_dev_ops_is_valid() 340 !ops->peer_port_number == !ops->port_number && in ntb_dev_ops_is_valid() 341 !ops->peer_port_idx == !ops->port_number && in ntb_dev_ops_is_valid() 344 ops->link_is_up && in ntb_dev_ops_is_valid() 345 ops->link_enable && in ntb_dev_ops_is_valid() [all …]
|
| /linux/net/netfilter/ |
| H A D | nf_sockopt.c | 27 struct nf_sockopt_ops *ops; in nf_register_sockopt() local 31 list_for_each_entry(ops, &nf_sockopts, list) { in nf_register_sockopt() 32 if (ops->pf == reg->pf in nf_register_sockopt() 33 && (overlap(ops->set_optmin, ops->set_optmax, in nf_register_sockopt() 35 || overlap(ops->get_optmin, ops->get_optmax, in nf_register_sockopt() 38 ops->set_optmin, ops->set_optmax, in nf_register_sockopt() 39 ops->get_optmin, ops->get_optmax, in nf_register_sockopt() 65 struct nf_sockopt_ops *ops; in nf_sockopt_find() local 68 list_for_each_entry(ops, &nf_sockopts, list) { in nf_sockopt_find() 69 if (ops->pf == pf) { in nf_sockopt_find() [all …]
|
| /linux/tools/testing/cxl/test/ |
| H A D | mock.c | 16 void register_cxl_mock_ops(struct cxl_mock_ops *ops) 18 list_add_rcu(&ops->list, &mock); 24 void unregister_cxl_mock_ops(struct cxl_mock_ops *ops) in register_cxl_mock_ops() 26 list_del_rcu(&ops->list); in register_cxl_mock_ops() 49 struct cxl_mock_ops *ops = get_cxl_mock_ops(&index); 52 if (ops) in put_cxl_mock_ops() 53 retval = ops->is_mock_adev(adev); in put_cxl_mock_ops() 68 struct cxl_mock_ops *ops = get_cxl_mock_ops(&index); in __wrap_is_acpi_device_node() 70 if (ops) in __wrap_is_acpi_device_node() 71 rc = ops in __wrap_is_acpi_device_node() 22 register_cxl_mock_ops(struct cxl_mock_ops * ops) register_cxl_mock_ops() argument 33 unregister_cxl_mock_ops(struct cxl_mock_ops * ops) unregister_cxl_mock_ops() argument 61 struct cxl_mock_ops *ops = get_cxl_mock_ops(&index); __wrap_is_acpi_device_node() local 80 struct cxl_mock_ops *ops = get_cxl_mock_ops(&index); __wrap_acpi_table_parse_cedt() local 99 struct cxl_mock_ops *ops = get_cxl_mock_ops(&index); __wrap_acpi_evaluate_integer() local 119 struct cxl_mock_ops *ops = get_cxl_mock_ops(&index); __wrap_hmat_get_extended_linear_cache_size() local 138 struct cxl_mock_ops *ops = get_cxl_mock_ops(&index); __wrap_acpi_pci_find_root() local 156 struct cxl_mock_ops *ops = get_cxl_mock_ops(&index); __wrap_nvdimm_bus_register() local 169 struct cxl_mock_ops *ops = get_cxl_mock_ops(&index); redirect_devm_cxl_switch_port_decoders_setup() local 183 struct cxl_mock_ops *ops = get_cxl_mock_ops(&index); __wrap_devm_cxl_endpoint_decoders_setup() local 198 struct cxl_mock_ops *ops = get_cxl_mock_ops(&index); __wrap_cxl_await_media_ready() local 217 struct cxl_mock_ops *ops = get_cxl_mock_ops(&index); __wrap_devm_cxl_add_rch_dport() local 237 struct cxl_mock_ops *ops = get_cxl_mock_ops(&index); __wrap_cxl_endpoint_parse_cdat() local 251 struct cxl_mock_ops *ops = get_cxl_mock_ops(&index); __wrap_cxl_dport_init_ras_reporting() local 264 struct cxl_mock_ops *ops = get_cxl_mock_ops(&index); redirect_devm_cxl_add_dport_by_dev() local [all...] |
| /linux/drivers/gpu/drm/vkms/ |
| H A D | vkms_colorop.c | 19 struct drm_colorop *ops[MAX_COLOR_PIPELINE_OPS]; in vkms_initialize_color_pipeline() local 24 memset(ops, 0, sizeof(ops)); in vkms_initialize_color_pipeline() 27 ops[i] = kzalloc_obj(*ops[i]); in vkms_initialize_color_pipeline() 28 if (!ops[i]) { in vkms_initialize_color_pipeline() 34 ret = drm_plane_colorop_curve_1d_init(dev, ops[i], plane, supported_tfs, in vkms_initialize_color_pipeline() 39 list->type = ops[i]->base.id; in vkms_initialize_color_pipeline() 44 ops[i] = kzalloc_obj(*ops[i]); in vkms_initialize_color_pipeline() 45 if (!ops[i]) { in vkms_initialize_color_pipeline() 51 ret = drm_plane_colorop_ctm_3x4_init(dev, ops[i], plane, DRM_COLOROP_FLAG_ALLOW_BYPASS); in vkms_initialize_color_pipeline() 55 drm_colorop_set_next_property(ops[i - 1], ops[i]); in vkms_initialize_color_pipeline() [all …]
|
| /linux/mm/ |
| H A D | pagewalk.c | 33 const struct mm_walk_ops *ops = walk->ops; in walk_pte_range_inner() local 37 if (ops->install_pte && pte_none(ptep_get(pte))) { in walk_pte_range_inner() 40 err = ops->install_pte(addr, addr + PAGE_SIZE, &new_pte, in walk_pte_range_inner() 50 err = ops->pte_entry(pte, addr, addr + PAGE_SIZE, walk); in walk_pte_range_inner() 103 const struct mm_walk_ops *ops = walk->ops; in walk_pmd_range() 104 bool has_handler = ops->pte_entry; in walk_pmd_range() 105 bool has_install = ops->install_pte; in walk_pmd_range() 134 else if (ops in walk_pmd_range() 102 const struct mm_walk_ops *ops = walk->ops; walk_pmd_range() local 172 const struct mm_walk_ops *ops = walk->ops; walk_pud_range() local 237 const struct mm_walk_ops *ops = walk->ops; walk_p4d_range() local 275 const struct mm_walk_ops *ops = walk->ops; walk_pgd_range() local 328 const struct mm_walk_ops *ops = walk->ops; walk_hugetlb_range() local 366 const struct mm_walk_ops *ops = walk->ops; walk_page_test() local 393 const struct mm_walk_ops *ops = walk->ops; __walk_page_range() local 456 walk_page_range_mm_unsafe(struct mm_struct * mm,unsigned long start,unsigned long end,const struct mm_walk_ops * ops,void * private) walk_page_range_mm_unsafe() argument 524 check_ops_safe(const struct mm_walk_ops * ops) check_ops_safe() argument 579 walk_page_range(struct mm_struct * mm,unsigned long start,unsigned long end,const struct mm_walk_ops * ops,void * private) walk_page_range() argument 607 walk_kernel_page_table_range(unsigned long start,unsigned long end,const struct mm_walk_ops * ops,pgd_t * pgd,void * private) walk_kernel_page_table_range() argument 630 walk_kernel_page_table_range_lockless(unsigned long start,unsigned long end,const struct mm_walk_ops * ops,pgd_t * pgd,void * private) walk_kernel_page_table_range_lockless() argument 664 walk_page_range_debug(struct mm_struct * mm,unsigned long start,unsigned long end,const struct mm_walk_ops * ops,pgd_t * pgd,void * private) walk_page_range_debug() argument 698 walk_page_range_vma_unsafe(struct vm_area_struct * vma,unsigned long start,unsigned long end,const struct mm_walk_ops * ops,void * private) walk_page_range_vma_unsafe() argument 718 walk_page_range_vma(struct vm_area_struct * vma,unsigned long start,unsigned long end,const struct mm_walk_ops * ops,void * private) walk_page_range_vma() argument 727 walk_page_vma(struct vm_area_struct * vma,const struct mm_walk_ops * ops,void * private) walk_page_vma() argument 778 walk_page_mapping(struct address_space * mapping,pgoff_t first_index,pgoff_t nr,const struct mm_walk_ops * ops,void * private) walk_page_mapping() argument [all...] |
| /linux/arch/s390/appldata/ |
| H A D | appldata_base.c | 111 struct appldata_ops *ops; in appldata_work_fn() local 115 ops = list_entry(lh, struct appldata_ops, list); in appldata_work_fn() 116 if (ops->active == 1) { in appldata_work_fn() 117 ops->callback(ops->data); in appldata_work_fn() 268 struct appldata_ops *ops = NULL, *tmp_ops; in appldata_generic_handler() local 291 ops = ctl->data; in appldata_generic_handler() 292 if (!try_module_get(ops->owner)) { // protect this function in appldata_generic_handler() 298 active = ops->active; in appldata_generic_handler() 301 module_put(ops->owner); in appldata_generic_handler() 306 if (active && (ops->active == 0)) { in appldata_generic_handler() [all …]
|
| /linux/drivers/media/v4l2-core/ |
| H A D | v4l2-dev.c | 531 #define SET_VALID_IOCTL(ops, cmd, op) \ argument 532 do { if ((ops)->op) __set_bit(_IOC_NR(cmd), valid_ioctls); } while (0) 557 const struct v4l2_ioctl_ops *ops = vdev->ioctl_ops; in determine_valid_ioctls() local 576 SET_VALID_IOCTL(ops, VIDIOC_QUERYCAP, vidioc_querycap); in determine_valid_ioctls() 584 if (vdev->ctrl_handler || ops->vidioc_query_ext_ctrl) in determine_valid_ioctls() 586 if (vdev->ctrl_handler || ops->vidioc_query_ext_ctrl) in determine_valid_ioctls() 588 if (vdev->ctrl_handler || ops->vidioc_g_ext_ctrls) in determine_valid_ioctls() 590 if (vdev->ctrl_handler || ops->vidioc_s_ext_ctrls) in determine_valid_ioctls() 592 if (vdev->ctrl_handler || ops->vidioc_g_ext_ctrls) in determine_valid_ioctls() 594 if (vdev->ctrl_handler || ops->vidioc_s_ext_ctrls) in determine_valid_ioctls() [all …]
|
| /linux/tools/perf/util/ |
| H A D | disasm.c | 89 int arch__associate_ins_ops(struct arch *arch, const char *name, const struct ins_ops *ops) in arch__associate_ins_ops() argument 102 ins->ops = ops; in arch__associate_ins_ops() 205 static void ins_ops__delete(struct ins_operands *ops) in ins_ops__delete() argument 207 if (ops == NULL) in ins_ops__delete() 209 zfree(&ops->source.raw); in ins_ops__delete() 210 zfree(&ops->source.name); in ins_ops__delete() 211 zfree(&ops->target.raw); in ins_ops__delete() 212 zfree(&ops->target.name); in ins_ops__delete() 216 struct ins_operands *ops, int max_ins_name) in ins__raw_scnprintf() argument 218 return scnprintf(bf, size, "%-*s %s", max_ins_name, ins->name, ops->raw); in ins__raw_scnprintf() [all …]
|
| /linux/net/netlabel/ |
| H A D | netlabel_calipso.c | 68 netlbl_calipso_ops_register(const struct netlbl_calipso_ops *ops) in netlbl_calipso_ops_register() argument 70 return xchg(&calipso_ops, ops); in netlbl_calipso_ops_register() 124 const struct netlbl_calipso_ops *ops = netlbl_calipso_ops_get(); in netlbl_calipso_add() local 130 if (!ops) in netlbl_calipso_add() 408 const struct netlbl_calipso_ops *ops = netlbl_calipso_ops_get(); in calipso_doi_add() local 410 if (ops) in calipso_doi_add() 411 ret_val = ops->doi_add(doi_def, audit_info); in calipso_doi_add() 425 const struct netlbl_calipso_ops *ops = netlbl_calipso_ops_get(); in calipso_doi_free() local 427 if (ops) in calipso_doi_free() 428 ops->doi_free(doi_def); in calipso_doi_free() [all …]
|
| /linux/net/core/ |
| H A D | lwtunnel.c | 82 int lwtunnel_encap_add_ops(const struct lwtunnel_encap_ops *ops, in lwtunnel_encap_add_ops() argument 90 NULL, ops) ? 0 : -1; in lwtunnel_encap_add_ops() 94 int lwtunnel_encap_del_ops(const struct lwtunnel_encap_ops *ops, in lwtunnel_encap_del_ops() argument 105 ops, NULL) == ops) ? 0 : -1; in lwtunnel_encap_del_ops() 118 const struct lwtunnel_encap_ops *ops; in lwtunnel_build_state() local 131 ops = rcu_dereference(lwtun_encaps[encap_type]); in lwtunnel_build_state() 132 if (likely(ops && ops->build_state && try_module_get(ops->owner))) in lwtunnel_build_state() 137 ret = ops->build_state(net, encap, family, cfg, lws, extack); in lwtunnel_build_state() 139 module_put(ops->owner); in lwtunnel_build_state() 154 const struct lwtunnel_encap_ops *ops; in lwtunnel_valid_encap_type() local [all …]
|
| H A D | fib_rules.c | 56 int fib_default_rule_add(struct fib_rules_ops *ops, in fib_default_rule_add() argument 61 r = kzalloc(ops->rule_size, GFP_KERNEL_ACCOUNT); in fib_default_rule_add() 70 r->fr_net = ops->fro_net; in fib_default_rule_add() 78 list_add_tail(&r->list, &ops->rules_list); in fib_default_rule_add() 83 static u32 fib_default_rule_pref(struct fib_rules_ops *ops) in fib_default_rule_pref() argument 88 if (!list_empty(&ops->rules_list)) { in fib_default_rule_pref() 89 pos = ops->rules_list.next; in fib_default_rule_pref() 90 if (pos->next != &ops->rules_list) { in fib_default_rule_pref() 101 struct fib_rules_ops *ops, struct nlmsghdr *nlh, 107 struct fib_rules_ops *ops; in lookup_rules_ops() local [all …]
|
| /linux/tools/perf/util/annotate-arch/ |
| H A D | annotate-loongarch.c | 17 static int loongarch_call__parse(const struct arch *arch, struct ins_operands *ops, in loongarch_call__parse() argument 25 c = strchr(ops->raw, '#'); in loongarch_call__parse() 29 ops->target.addr = strtoull(c, &endptr, 16); in loongarch_call__parse() 43 ops->target.name = strdup(name); in loongarch_call__parse() 46 if (ops->target.name == NULL) in loongarch_call__parse() 51 .addr = map__objdump_2mem(map, ops->target.addr), in loongarch_call__parse() 55 map__rip_2objdump(target.ms.map, map__map_ip(target.ms.map, target.addr)) == ops->target.addr) in loongarch_call__parse() 56 ops->target.sym = target.ms.sym; in loongarch_call__parse() 68 static int loongarch_jump__parse(const struct arch *arch, struct ins_operands *ops, in loongarch_jump__parse() argument 78 const char *c = strchr(ops->raw, '#'); in loongarch_jump__parse() [all …]
|
| H A D | annotate-x86.c | 23 { .name = "adc", .ops = &mov_ops, }, 24 { .name = "add", .ops = &mov_ops, }, 25 { .name = "addsd", .ops = &mov_ops, }, 26 { .name = "and", .ops = &mov_ops, }, 27 { .name = "andpd", .ops = &mov_ops, }, 28 { .name = "andps", .ops = &mov_ops, }, 29 { .name = "bsr", .ops = &mov_ops, }, 30 { .name = "bt", .ops = &mov_ops, }, 31 { .name = "btr", .ops = &mov_ops, }, 32 { .name = "bts", .ops = &mov_ops, }, [all …]
|
| /linux/tools/testing/selftests/mm/ |
| H A D | khugepaged.c | 58 struct mem_ops *ops, bool expect); 304 static void *alloc_hpage(struct mem_ops *ops) in alloc_hpage() argument 306 void *p = ops->setup_area(1); in alloc_hpage() 308 ops->fault(p, 0, hpage_pmd_size); in alloc_hpage() 322 if (!ops->check_huge(p, 1)) { in alloc_hpage() 497 struct mem_ops *ops, bool expect) in __madvise_collapse() argument 517 else if (!ops->check_huge(p, expect ? nr_hpages : 0)) in __madvise_collapse() 526 struct mem_ops *ops, bool expect) in madvise_collapse() argument 529 if (!ops->check_huge(p, 0)) { in madvise_collapse() 533 __madvise_collapse(msg, p, nr_hpages, ops, expect); in madvise_collapse() [all …]
|
| /linux/rust/kernel/ |
| H A D | num.rs | 5 use core::ops; 25 + ops::Add<Output = Self> 26 + ops::AddAssign 27 + ops::Sub<Output = Self> 28 + ops::SubAssign 29 + ops::Mul<Output = Self> 30 + ops::MulAssign 31 + ops::Div<Output = Self> 32 + ops::DivAssign 33 + ops::Rem<Output = Self> [all …]
|
| /linux/drivers/misc/c2port/ |
| H A D | core.c | 63 struct c2port_ops *ops = dev->ops; in c2port_reset() local 69 ops->c2ck_set(dev, 0); in c2port_reset() 71 ops->c2ck_set(dev, 1); in c2port_reset() 79 struct c2port_ops *ops = dev->ops; in c2port_strobe_ck() local 87 ops->c2ck_set(dev, 0); in c2port_strobe_ck() 89 ops->c2ck_set(dev, 1); in c2port_strobe_ck() 101 struct c2port_ops *ops = dev->ops; in c2port_write_ar() local 108 ops->c2d_dir(dev, 0); in c2port_write_ar() 109 ops->c2d_set(dev, 1); in c2port_write_ar() 111 ops->c2d_set(dev, 1); in c2port_write_ar() [all …]
|
| /linux/kernel/livepatch/ |
| H A D | patch.c | 27 struct klp_ops *ops; in klp_find_ops() local 30 list_for_each_entry(ops, &klp_ops, node) { in klp_find_ops() 31 func = list_first_entry(&ops->func_stack, struct klp_func, in klp_find_ops() 34 return ops; in klp_find_ops() 45 struct klp_ops *ops; in klp_ftrace_handler() local 50 ops = container_of(fops, struct klp_ops, fops); in klp_ftrace_handler() 62 func = list_first_or_null_rcu(&ops->func_stack, struct klp_func, in klp_ftrace_handler() 109 if (&func->stack_node == &ops->func_stack) in klp_ftrace_handler() 129 struct klp_ops *ops; in klp_unpatch_func() local 136 ops = klp_find_ops(func->old_func); in klp_unpatch_func() [all …]
|
| /linux/net/dsa/ |
| H A D | conduit.c | 23 const struct ethtool_ops *ops = cpu_dp->orig_ethtool_ops; in dsa_conduit_get_regs_len() local 29 if (ops && ops->get_regs_len) { in dsa_conduit_get_regs_len() 31 len = ops->get_regs_len(dev); in dsa_conduit_get_regs_len() 41 if (ds->ops->get_regs_len) { in dsa_conduit_get_regs_len() 42 len = ds->ops->get_regs_len(ds, port); in dsa_conduit_get_regs_len() 55 const struct ethtool_ops *ops = cpu_dp->orig_ethtool_ops; in dsa_conduit_get_regs() local 62 if (ops && ops->get_regs_len && ops in dsa_conduit_get_regs() 95 const struct ethtool_ops *ops = cpu_dp->orig_ethtool_ops; dsa_conduit_get_ethtool_stats() local 116 const struct ethtool_ops *ops = cpu_dp->orig_ethtool_ops; dsa_conduit_get_ethtool_phy_stats() local 142 const struct ethtool_ops *ops = cpu_dp->orig_ethtool_ops; dsa_conduit_get_sset_count() local 167 const struct ethtool_ops *ops = cpu_dp->orig_ethtool_ops; dsa_conduit_get_strings() local 242 struct ethtool_ops *ops; dsa_conduit_ethtool_setup() local 285 const struct dsa_device_ops *ops = dev->dsa_ptr->tag_ops; dsa_conduit_set_promiscuity() local [all...] |
| /linux/drivers/net/ethernet/intel/igb/ |
| H A D | e1000_phy.c | 62 phy->ops.write_reg(hw, I347AT4_PAGE_SELECT, 0); in igb_get_phy_id() 64 ret_val = phy->ops.read_reg(hw, PHY_ID1, &phy_id); in igb_get_phy_id() 70 ret_val = phy->ops.read_reg(hw, PHY_ID2, &phy_id); in igb_get_phy_id() 91 if (!(hw->phy.ops.write_reg)) in igb_phy_reset_dsp() 94 ret_val = hw->phy.ops.write_reg(hw, M88E1000_PHY_GEN_CONTROL, 0xC1); in igb_phy_reset_dsp() 98 ret_val = hw->phy.ops.write_reg(hw, M88E1000_PHY_GEN_CONTROL, 0); in igb_phy_reset_dsp() 384 if (!(hw->phy.ops.acquire)) in igb_read_phy_reg_igp() 387 ret_val = hw->phy.ops.acquire(hw); in igb_read_phy_reg_igp() 396 hw->phy.ops.release(hw); in igb_read_phy_reg_igp() 404 hw->phy.ops.release(hw); in igb_read_phy_reg_igp() [all …]
|
| /linux/net/mac80211/ |
| H A D | driver-ops.h | 38 local->ops->tx(&local->hw, control, skb); in drv_tx() 47 if (local->ops->sync_rx_queues) { in drv_sync_rx_queues() 49 local->ops->sync_rx_queues(&local->hw); in drv_sync_rx_queues() 58 if (local->ops->get_et_strings) { in drv_get_et_strings() 60 local->ops->get_et_strings(&local->hw, &sdata->vif, sset, data); in drv_get_et_strings() 70 if (local->ops->get_et_stats) { in drv_get_et_stats() 72 local->ops->get_et_stats(&local->hw, &sdata->vif, stats, data); in drv_get_et_stats() 82 if (local->ops->get_et_sset_count) { in drv_get_et_sset_count() 84 rv = local->ops->get_et_sset_count(&local->hw, &sdata->vif, in drv_get_et_sset_count() 104 ret = local->ops in drv_suspend() [all...] |
| /linux/sound/soc/sof/ |
| H A D | ipc.c | 19 #include "ops.h" 80 return ipc->ops->tx_msg(ipc->sdev, msg_data, msg_bytes, reply_data, in sof_ipc_tx_message() 89 return ipc->ops->set_get_data(ipc->sdev, msg_data, msg_bytes, set); in sof_ipc_set_get_data() 105 return ipc->ops->tx_msg(ipc->sdev, msg_data, msg_bytes, reply_data, in sof_ipc_tx_message_no_pm() 123 sdev->msg->reply_error = sdev->ipc->ops->get_reply(sdev); in snd_sof_ipc_get_reply() 149 const struct sof_ipc_ops *ops; in snd_sof_ipc_init() 167 ops = &ipc3_ops; in snd_sof_ipc_init() 172 ops = &ipc4_ops; in snd_sof_ipc_init() 181 /* check for mandatory ops */ in snd_sof_ipc_init() 182 if (!ops in snd_sof_ipc_init() 151 const struct sof_ipc_ops *ops; snd_sof_ipc_init() local [all...] |