Home
last modified time | relevance | path

Searched full:hw_data (Results 1 – 25 of 49) sorted by relevance

12

/linux/drivers/video/backlight/
H A Dapple_bl.c31 struct hw_data { struct
40 static const struct hw_data *hw_data; argument
81 static const struct hw_data intel_chipset_data = {
126 static const struct hw_data nvidia_chipset_data = {
151 hw_data = &intel_chipset_data; in apple_bl_add()
153 hw_data = &nvidia_chipset_data; in apple_bl_add()
157 if (!hw_data) { in apple_bl_add()
164 intensity = hw_data->backlight_ops.get_brightness(NULL); in apple_bl_add()
167 hw_data->set_brightness(1); in apple_bl_add()
168 if (!hw_data->backlight_ops.get_brightness(NULL)) in apple_bl_add()
[all …]
/linux/drivers/crypto/intel/qat/qat_common/
H A Dadf_isr.c24 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_enable_msix() local
25 u32 msix_num_entries = hw_data->num_banks + 1; in adf_enable_msix()
28 if (hw_data->set_msix_rttable) in adf_enable_msix()
29 hw_data->set_msix_rttable(accel_dev); in adf_enable_msix()
126 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_handle_pm_int() local
128 if (hw_data->handle_pm_interrupt && in adf_handle_pm_int()
129 hw_data->handle_pm_interrupt(accel_dev)) in adf_handle_pm_int()
180 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_free_irqs() local
183 int clust_irq = hw_data->num_banks; in adf_free_irqs()
187 for (i = 0; i < hw_data->num_banks; i++) { in adf_free_irqs()
[all …]
H A Dadf_gen4_vf_mig.c96 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_gen4_vfmig_suspend_device() local
106 for (i = 0; i < hw_data->num_banks_per_vf; i++) { in adf_gen4_vfmig_suspend_device()
107 u32 pf_bank_nr = i + vf_nr * hw_data->num_banks_per_vf; in adf_gen4_vfmig_suspend_device()
129 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_gen4_vfmig_resume_device() local
138 for (i = 0; i < hw_data->num_banks_per_vf; i++) { in adf_gen4_vfmig_resume_device()
139 u32 pf_bank_nr = i + vf_nr * hw_data->num_banks_per_vf; in adf_gen4_vfmig_resume_device()
316 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_gen4_vfmig_get_slas() local
327 rp_index = vf_nr * hw_data->num_banks_per_vf; in adf_gen4_vfmig_get_slas()
356 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_gen4_vfmig_load_etr_regs() local
360 pf_bank_nr = vf_bank_info->bank_nr + vf_bank_info->vf_nr * hw_data->num_banks_per_vf; in adf_gen4_vfmig_load_etr_regs()
[all …]
H A Dadf_transport.c164 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_init_ring() local
187 if (hw_data->tx_rings_mask & (1 << ring->ring_number)) in adf_init_ring()
388 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_init_bank() local
389 u8 num_rings_per_bank = hw_data->num_rings_per_bank; in adf_init_bank()
390 struct adf_hw_csr_ops *csr_ops = &hw_data->csr_ops; in adf_init_bank()
426 if (hw_data->tx_rings_mask & (1 << i)) { in adf_init_bank()
434 if (i < hw_data->tx_rx_gap) { in adf_init_bank()
439 tx_ring = &bank->rings[i - hw_data->tx_rx_gap]; in adf_init_bank()
454 ring_mask = hw_data->tx_rings_mask; in adf_init_bank()
525 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in cleanup_bank() local
[all …]
H A Dadf_fw_counters.c52 struct adf_hw_device_data *hw_data = GET_HW_DATA(accel_dev); in adf_fw_counters_load_from_device() local
58 ae_mask = hw_data->ae_mask & ~hw_data->admin_ae_mask; in adf_fw_counters_load_from_device()
109 struct adf_hw_device_data *hw_data = GET_HW_DATA(accel_dev); in adf_fw_counters_get() local
120 ae_count = hweight_long(hw_data->ae_mask & ~hw_data->admin_ae_mask); in adf_fw_counters_get()
H A Dadf_sriov.c56 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_enable_sriov() local
74 if (hw_data->configure_iov_threads) in adf_enable_sriov()
75 hw_data->configure_iov_threads(accel_dev, true); in adf_enable_sriov()
239 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_disable_sriov() local
255 if (hw_data->configure_iov_threads) in adf_disable_sriov()
256 hw_data->configure_iov_threads(accel_dev, false); in adf_disable_sriov()
H A Dadf_accel_engine.c128 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_ae_start() local
131 if (!hw_data->fw_name) in adf_ae_start()
144 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_ae_stop() local
147 if (!hw_data->fw_name) in adf_ae_stop()
151 if (hw_data->ae_mask & (1 << ae)) { in adf_ae_stop()
H A Dadf_heartbeat_inject.c13 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_hb_set_timer_to_max() local
17 if (hw_data->stop_timer) in adf_hb_set_timer_to_max()
18 hw_data->stop_timer(accel_dev); in adf_hb_set_timer_to_max()
H A Dadf_pfvf_vf_msg.c112 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_vf2pf_get_capabilities() local
133 hw_data->clock_frequency = cap_msg.frequency; in adf_vf2pf_get_capabilities()
139 hw_data->accel_capabilities_mask = cap_msg.capabilities; in adf_vf2pf_get_capabilities()
145 hw_data->extended_dc_capabilities = cap_msg.ext_dc_caps; in adf_vf2pf_get_capabilities()
/linux/net/core/
H A Ddrop_monitor.c304 net_dm_hw_reset_per_cpu_data(struct per_cpu_dm_data *hw_data) in net_dm_hw_reset_per_cpu_data() argument
315 mod_timer(&hw_data->send_timer, jiffies + HZ / 10); in net_dm_hw_reset_per_cpu_data()
318 raw_spin_lock_irqsave(&hw_data->lock, flags); in net_dm_hw_reset_per_cpu_data()
319 swap(hw_data->hw_entries, hw_entries); in net_dm_hw_reset_per_cpu_data()
320 raw_spin_unlock_irqrestore(&hw_data->lock, flags); in net_dm_hw_reset_per_cpu_data()
411 struct per_cpu_dm_data *hw_data; in net_dm_hw_summary_work() local
415 hw_data = container_of(work, struct per_cpu_dm_data, dm_alert_work); in net_dm_hw_summary_work()
417 hw_entries = net_dm_hw_reset_per_cpu_data(hw_data); in net_dm_hw_summary_work()
444 struct per_cpu_dm_data *hw_data; in net_dm_hw_trap_summary_probe() local
451 hw_data = this_cpu_ptr(&dm_hw_cpu_data); in net_dm_hw_trap_summary_probe()
[all …]
/linux/drivers/crypto/intel/qat/qat_dh895xccvf/
H A Dadf_drv.c79 struct adf_hw_device_data *hw_data; in adf_probe() local
112 hw_data = kzalloc_node(sizeof(*hw_data), GFP_KERNEL, in adf_probe()
114 if (!hw_data) { in adf_probe()
118 accel_dev->hw_device = hw_data; in adf_probe()
122 hw_data->accel_mask = hw_data->get_accel_mask(hw_data); in adf_probe()
123 hw_data->ae_mask = hw_data->get_ae_mask(hw_data); in adf_probe()
124 accel_pci_dev->sku = hw_data->get_sku(hw_data); in adf_probe()
/linux/drivers/crypto/intel/qat/qat_c62xvf/
H A Dadf_drv.c79 struct adf_hw_device_data *hw_data; in adf_probe() local
112 hw_data = kzalloc_node(sizeof(*hw_data), GFP_KERNEL, in adf_probe()
114 if (!hw_data) { in adf_probe()
118 accel_dev->hw_device = hw_data; in adf_probe()
122 hw_data->accel_mask = hw_data->get_accel_mask(hw_data); in adf_probe()
123 hw_data->ae_mask = hw_data->get_ae_mask(hw_data); in adf_probe()
124 accel_pci_dev->sku = hw_data->get_sku(hw_data); in adf_probe()
/linux/drivers/crypto/intel/qat/qat_c3xxxvf/
H A Dadf_drv.c79 struct adf_hw_device_data *hw_data; in adf_probe() local
112 hw_data = kzalloc_node(sizeof(*hw_data), GFP_KERNEL, in adf_probe()
114 if (!hw_data) { in adf_probe()
118 accel_dev->hw_device = hw_data; in adf_probe()
122 hw_data->accel_mask = hw_data->get_accel_mask(hw_data); in adf_probe()
123 hw_data->ae_mask = hw_data->get_ae_mask(hw_data); in adf_probe()
124 accel_pci_dev->sku = hw_data->get_sku(hw_data); in adf_probe()
/linux/drivers/clk/
H A Dclk-lan966x.c226 struct clk_hw_onecell_data *hw_data, in lan966x_gate_clk_register() argument
235 hw_data->hws[i] = in lan966x_gate_clk_register()
241 if (IS_ERR(hw_data->hws[i])) in lan966x_gate_clk_register()
242 return dev_err_probe(dev, PTR_ERR(hw_data->hws[i]), in lan966x_gate_clk_register()
253 struct clk_hw_onecell_data *hw_data; in lan966x_clk_probe() local
263 hw_data = devm_kzalloc(dev, struct_size(hw_data, hws, data->num_total_clks), in lan966x_clk_probe()
265 if (!hw_data) in lan966x_clk_probe()
274 hw_data->num = data->num_generic_clks; in lan966x_clk_probe()
278 hw_data->hws[i] = lan966x_gck_clk_register(dev, i); in lan966x_clk_probe()
279 if (IS_ERR(hw_data->hws[i])) { in lan966x_clk_probe()
[all …]
H A Dclk-twl.c145 const struct twl_clks_data *hw_data; in twl_clks_probe() local
152 hw_data = twl6032_clks; in twl_clks_probe()
153 for (count = 0; hw_data[count].init.name; count++) in twl_clks_probe()
168 cinfo[i].base = hw_data[i].base; in twl_clks_probe()
171 cinfo[i].hw.init = &hw_data[i].init; in twl_clks_probe()
176 hw_data[i].init.name); in twl_clks_probe()
H A Dclk-bm1880.c65 struct clk_hw_onecell_data hw_data; member
543 data->hw_data.hws[clks[i].pll.id] = hw; in bm1880_clk_register_plls()
550 clk_hw_unregister(data->hw_data.hws[clks[i].pll.id]); in bm1880_clk_register_plls()
577 data->hw_data.hws[clks[i].id] = hw; in bm1880_clk_register_mux()
584 clk_hw_unregister_mux(data->hw_data.hws[clks[i].id]); in bm1880_clk_register_mux()
707 data->hw_data.hws[id] = hw; in bm1880_clk_register_divs()
714 clk_hw_unregister(data->hw_data.hws[clks[i].div.id]); in bm1880_clk_register_divs()
740 data->hw_data.hws[clks[i].id] = hw; in bm1880_clk_register_gate()
747 clk_hw_unregister_gate(data->hw_data.hws[clks[i].id]); in bm1880_clk_register_gate()
860 data->hw_data.hws[clks[i].id] = hw; in bm1880_clk_register_composites()
[all …]
/linux/include/sound/
H A Dpcm-indirect.h17 unsigned int hw_data; /* Offset to next dst (or src) in hw ring buffer */ member
53 unsigned int hw_to_end = rec->hw_buffer_size - rec->hw_data; in snd_pcm_indirect_playback_transfer()
65 rec->hw_data += bytes; in snd_pcm_indirect_playback_transfer()
66 if (rec->hw_data == rec->hw_buffer_size) in snd_pcm_indirect_playback_transfer()
67 rec->hw_data = 0; in snd_pcm_indirect_playback_transfer()
126 size_t hw_to_end = rec->hw_buffer_size - rec->hw_data; in snd_pcm_indirect_capture_transfer()
138 rec->hw_data += bytes; in snd_pcm_indirect_capture_transfer()
139 if ((int)rec->hw_data == rec->hw_buffer_size) in snd_pcm_indirect_capture_transfer()
140 rec->hw_data = 0; in snd_pcm_indirect_capture_transfer()
/linux/drivers/clk/keystone/
H A Dsyscon-clk.c117 struct clk_hw_onecell_data *hw_data; in ti_syscon_gate_clk_probe() local
148 hw_data = devm_kzalloc(dev, struct_size(hw_data, hws, num_clks), in ti_syscon_gate_clk_probe()
150 if (!hw_data) in ti_syscon_gate_clk_probe()
153 hw_data->num = num_clks; in ti_syscon_gate_clk_probe()
157 hw_data->hws[i] = ti_syscon_gate_clk_register(dev, regmap, in ti_syscon_gate_clk_probe()
160 if (IS_ERR(hw_data->hws[i])) in ti_syscon_gate_clk_probe()
167 hw_data->hws[0]); in ti_syscon_gate_clk_probe()
168 return devm_of_clk_add_hw_provider(dev, of_clk_hw_onecell_get, hw_data); in ti_syscon_gate_clk_probe()
/linux/drivers/clk/uniphier/
H A Dclk-uniphier-core.c44 struct clk_hw_onecell_data *hw_data; in uniphier_clk_probe() local
66 hw_data = devm_kzalloc(dev, struct_size(hw_data, hws, clk_num), in uniphier_clk_probe()
68 if (!hw_data) in uniphier_clk_probe()
71 hw_data->num = clk_num; in uniphier_clk_probe()
75 hw_data->hws[clk_num] = ERR_PTR(-EINVAL); in uniphier_clk_probe()
86 hw_data->hws[p->idx] = hw; in uniphier_clk_probe()
90 hw_data); in uniphier_clk_probe()
/linux/drivers/clk/qcom/
H A Dipq-cmn-pll.c327 struct clk_hw_onecell_data *hw_data; in ipq_cmn_pll_register_clks() local
342 hw_data = devm_kzalloc(dev, struct_size(hw_data, hws, num_clks + 1), in ipq_cmn_pll_register_clks()
344 if (!hw_data) in ipq_cmn_pll_register_clks()
365 hw_data->hws[fixed_clk[i].id] = hw; in ipq_cmn_pll_register_clks()
372 hw_data->hws[CMN_PLL_CLK] = cmn_pll_hw; in ipq_cmn_pll_register_clks()
373 hw_data->num = num_clks + 1; in ipq_cmn_pll_register_clks()
375 ret = devm_of_clk_add_hw_provider(dev, of_clk_hw_onecell_get, hw_data); in ipq_cmn_pll_register_clks()
379 platform_set_drvdata(pdev, hw_data); in ipq_cmn_pll_register_clks()
385 clk_hw_unregister(hw_data->hws[fixed_clk[--i].id]); in ipq_cmn_pll_register_clks()
431 struct clk_hw_onecell_data *hw_data = platform_get_drvdata(pdev); in ipq_cmn_pll_clk_remove() local
[all …]
/linux/drivers/gpu/drm/amd/pm/powerplay/hwmgr/
H A Dsmu8_hwmgr.c798 struct smu8_hwmgr *hw_data = hwmgr->backend; in smu8_nbdpm_pstate_enable_disable() local
800 if (hw_data->is_nb_dpm_enabled) { in smu8_nbdpm_pstate_enable_disable()
869 struct smu8_hwmgr *hw_data = hwmgr->backend; in smu8_update_low_mem_pstate() local
873 if (hw_data->sys_info.nb_dpm_enable) { in smu8_update_low_mem_pstate()
874 disable_switch = hw_data->cc6_settings.nb_pstate_switch_disable ? true : false; in smu8_update_low_mem_pstate()
875 enable_low_mem_state = hw_data->cc6_settings.nb_pstate_switch_disable ? false : true; in smu8_update_low_mem_pstate()
931 struct smu8_hwmgr *hw_data = hwmgr->backend; in smu8_power_up_display_clock_sys_pll() local
933 hw_data->disp_clk_bypass_pending = false; in smu8_power_up_display_clock_sys_pll()
934 hw_data->disp_clk_bypass = false; in smu8_power_up_display_clock_sys_pll()
939 struct smu8_hwmgr *hw_data = hwmgr->backend; in smu8_clear_nb_dpm_flag() local
[all …]
/linux/drivers/clk/microchip/
H A Dclk-mpfs-ccc.c39 struct clk_hw_onecell_data hw_data; member
181 data->hw_data.hws[out_hw->id] = &out_hw->divider.hw; in mpfs_ccc_register_outputs()
219 data->hw_data.hws[pll_hw->id] = &pll_hw->hw; in mpfs_ccc_register_plls()
240 clk_data = devm_kzalloc(&pdev->dev, struct_size(clk_data, hw_data.hws, num_clks), in mpfs_ccc_probe()
254 clk_data->hw_data.num = num_clks; in mpfs_ccc_probe()
263 &clk_data->hw_data); in mpfs_ccc_probe()
H A Dclk-mpfs.c56 struct clk_hw_onecell_data hw_data; member
195 data->hw_data.hws[msspll_hw->id] = &msspll_hw->hw; in mpfs_clk_register_mssplls()
243 data->hw_data.hws[msspll_out_hw->id] = &msspll_out_hw->output.hw; in mpfs_clk_register_msspll_outs()
349 data->hw_data.hws[id] = &cfg_hw->hw; in mpfs_clk_register_cfgs()
466 data->hw_data.hws[id] = &periph_hw->hw; in mpfs_clk_register_periphs()
519 clk_data = devm_kzalloc(dev, struct_size(clk_data, hw_data.hws, num_clks), GFP_KERNEL); in mpfs_clk_probe()
530 clk_data->hw_data.num = num_clks; in mpfs_clk_probe()
554 return devm_of_clk_add_hw_provider(dev, of_clk_hw_onecell_get, &clk_data->hw_data); in mpfs_clk_probe()
/linux/drivers/clk/mvebu/
H A Darmada-37xx-periph.c58 struct clk_hw_onecell_data *hw_data; member
750 driver_data->hw_data = devm_kzalloc(dev, in armada_3700_periph_clock_probe()
751 struct_size(driver_data->hw_data, in armada_3700_periph_clock_probe()
754 if (!driver_data->hw_data) in armada_3700_periph_clock_probe()
756 driver_data->hw_data->num = num_periph; in armada_3700_periph_clock_probe()
765 struct clk_hw **hw = &driver_data->hw_data->hws[i]; in armada_3700_periph_clock_probe()
773 driver_data->hw_data); in armada_3700_periph_clock_probe()
776 clk_hw_unregister(driver_data->hw_data->hws[i]); in armada_3700_periph_clock_probe()
787 struct clk_hw_onecell_data *hw_data = data->hw_data; in armada_3700_periph_clock_remove() local
792 for (i = 0; i < hw_data->num; i++) in armada_3700_periph_clock_remove()
[all …]
/linux/drivers/gpu/drm/amd/pm/powerplay/smumgr/
H A Dvegam_smumgr.c866 struct smu7_hwmgr *hw_data = (struct smu7_hwmgr *)(hwmgr->backend); in vegam_populate_all_graphic_levels() local
868 struct smu7_dpm_table *dpm_table = &hw_data->dpm_table; in vegam_populate_all_graphic_levels()
872 uint8_t pcie_entry_cnt = (uint8_t) hw_data->dpm_table.pcie_speed_table.count; in vegam_populate_all_graphic_levels()
910 hw_data->dpm_level_enable_mask.sclk_dpm_enable_mask = in vegam_populate_all_graphic_levels()
915 (hw_data->dpm_level_enable_mask.sclk_dpm_enable_mask >> i) & 0x1; in vegam_populate_all_graphic_levels()
926 while (hw_data->dpm_level_enable_mask.pcie_dpm_enable_mask && in vegam_populate_all_graphic_levels()
927 ((hw_data->dpm_level_enable_mask.pcie_dpm_enable_mask & in vegam_populate_all_graphic_levels()
931 while (hw_data->dpm_level_enable_mask.pcie_dpm_enable_mask && in vegam_populate_all_graphic_levels()
932 ((hw_data->dpm_level_enable_mask.pcie_dpm_enable_mask & in vegam_populate_all_graphic_levels()
937 ((hw_data->dpm_level_enable_mask.pcie_dpm_enable_mask & in vegam_populate_all_graphic_levels()
[all …]

12