| /linux/drivers/soundwire/ |
| H A D | qcom.c | 220 int (*reg_read)(struct qcom_swrm_ctrl *ctrl, int reg, u32 *val); 221 int (*reg_write)(struct qcom_swrm_ctrl *ctrl, int reg, int val); 331 static int qcom_swrm_ahb_reg_read(struct qcom_swrm_ctrl *ctrl, int reg, in qcom_swrm_ahb_reg_read() argument 334 struct regmap *wcd_regmap = ctrl->regmap; in qcom_swrm_ahb_reg_read() 351 static int qcom_swrm_ahb_reg_write(struct qcom_swrm_ctrl *ctrl, in qcom_swrm_ahb_reg_write() argument 354 struct regmap *wcd_regmap = ctrl->regmap; in qcom_swrm_ahb_reg_write() 371 static int qcom_swrm_cpu_reg_read(struct qcom_swrm_ctrl *ctrl, int reg, in qcom_swrm_cpu_reg_read() argument 374 *val = readl(ctrl->mmio + reg); in qcom_swrm_cpu_reg_read() 378 static int qcom_swrm_cpu_reg_write(struct qcom_swrm_ctrl *ctrl, int reg, in qcom_swrm_cpu_reg_write() argument 381 writel(val, ctrl->mmio + reg); in qcom_swrm_cpu_reg_write() [all …]
|
| /linux/drivers/nvme/target/ |
| H A D | auth.c | 64 int nvmet_setup_dhgroup(struct nvmet_ctrl *ctrl, u8 dhgroup_id) in nvmet_setup_dhgroup() argument 69 pr_debug("%s: ctrl %d selecting dhgroup %d\n", in nvmet_setup_dhgroup() 70 __func__, ctrl->cntlid, dhgroup_id); in nvmet_setup_dhgroup() 72 if (ctrl->dh_tfm) { in nvmet_setup_dhgroup() 73 if (ctrl->dh_gid == dhgroup_id) { in nvmet_setup_dhgroup() 74 pr_debug("%s: ctrl %d reuse existing DH group %d\n", in nvmet_setup_dhgroup() 75 __func__, ctrl->cntlid, dhgroup_id); in nvmet_setup_dhgroup() 78 crypto_free_kpp(ctrl->dh_tfm); in nvmet_setup_dhgroup() 79 ctrl->dh_tfm = NULL; in nvmet_setup_dhgroup() 80 ctrl->dh_gid = 0; in nvmet_setup_dhgroup() [all …]
|
| H A D | core.c | 137 static void nvmet_async_events_failall(struct nvmet_ctrl *ctrl) in nvmet_async_events_failall() argument 141 mutex_lock(&ctrl->lock); in nvmet_async_events_failall() 142 while (ctrl->nr_async_event_cmds) { in nvmet_async_events_failall() 143 req = ctrl->async_event_cmds[--ctrl->nr_async_event_cmds]; in nvmet_async_events_failall() 144 mutex_unlock(&ctrl->lock); in nvmet_async_events_failall() 146 mutex_lock(&ctrl->lock); in nvmet_async_events_failall() 148 mutex_unlock(&ctrl->lock); in nvmet_async_events_failall() 151 static void nvmet_async_events_process(struct nvmet_ctrl *ctrl) in nvmet_async_events_process() argument 156 mutex_lock(&ctrl->lock); in nvmet_async_events_process() 157 while (ctrl->nr_async_event_cmds && !list_empty(&ctrl->async_events)) { in nvmet_async_events_process() [all …]
|
| H A D | pci-epf.c | 84 struct nvmet_pci_epf_ctrl *ctrl; member 129 struct nvmet_pci_epf_ctrl *ctrl; member 209 struct nvmet_pci_epf_ctrl ctrl; member 226 static inline u32 nvmet_pci_epf_bar_read32(struct nvmet_pci_epf_ctrl *ctrl, in nvmet_pci_epf_bar_read32() argument 229 __le32 *bar_reg = ctrl->bar + off; in nvmet_pci_epf_bar_read32() 234 static inline void nvmet_pci_epf_bar_write32(struct nvmet_pci_epf_ctrl *ctrl, in nvmet_pci_epf_bar_write32() argument 237 __le32 *bar_reg = ctrl->bar + off; in nvmet_pci_epf_bar_write32() 242 static inline u64 nvmet_pci_epf_bar_read64(struct nvmet_pci_epf_ctrl *ctrl, in nvmet_pci_epf_bar_read64() argument 245 return (u64)nvmet_pci_epf_bar_read32(ctrl, off) | in nvmet_pci_epf_bar_read64() 246 ((u64)nvmet_pci_epf_bar_read32(ctrl, off + 4) << 32); in nvmet_pci_epf_bar_read64() [all …]
|
| H A D | fabrics-cmd-auth.c | 19 pr_debug("%s: ctrl %d qid %d transaction %u expired, resetting\n", in nvmet_auth_expired_work() 20 __func__, sq->ctrl->cntlid, sq->qid, sq->dhchap_tid); in nvmet_auth_expired_work() 35 struct nvmet_ctrl *ctrl = req->sq->ctrl; in nvmet_auth_negotiate() local 39 pr_debug("%s: ctrl %d qid %d: data sc_d %d napd %d authid %d halen %d dhlen %d\n", in nvmet_auth_negotiate() 40 __func__, ctrl->cntlid, req->sq->qid, in nvmet_auth_negotiate() 64 ctrl->concat = true; in nvmet_auth_negotiate() 79 if (ctrl->shash_id != host_hmac_id) in nvmet_auth_negotiate() 81 hash_id = ctrl->shash_id; in nvmet_auth_negotiate() 86 pr_debug("%s: ctrl %d qid %d: no usable hash found\n", in nvmet_auth_negotiate() 87 __func__, ctrl->cntlid, req->sq->qid); in nvmet_auth_negotiate() [all …]
|
| /linux/drivers/slimbus/ |
| H A D | qcom-ngd-ctrl.c | 135 struct qcom_slim_ngd_ctrl *ctrl; member 150 struct slim_controller ctrl; member 347 static int qcom_slim_qmi_send_select_inst_req(struct qcom_slim_ngd_ctrl *ctrl, in qcom_slim_qmi_send_select_inst_req() argument 354 rc = qmi_txn_init(ctrl->qmi.handle, &txn, in qcom_slim_qmi_send_select_inst_req() 357 dev_err(ctrl->dev, "QMI TXN init fail: %d\n", rc); in qcom_slim_qmi_send_select_inst_req() 361 rc = qmi_send_request(ctrl->qmi.handle, NULL, &txn, in qcom_slim_qmi_send_select_inst_req() 366 dev_err(ctrl->dev, "QMI send req fail %d\n", rc); in qcom_slim_qmi_send_select_inst_req() 373 dev_err(ctrl->dev, "QMI TXN wait fail: %d\n", rc); in qcom_slim_qmi_send_select_inst_req() 378 dev_err(ctrl->dev, "QMI request failed 0x%x\n", in qcom_slim_qmi_send_select_inst_req() 400 static int qcom_slim_qmi_send_power_request(struct qcom_slim_ngd_ctrl *ctrl, in qcom_slim_qmi_send_power_request() argument [all …]
|
| H A D | core.c | 153 static int slim_add_device(struct slim_controller *ctrl, in slim_add_device() argument 158 sbdev->dev.parent = ctrl->dev; in slim_add_device() 161 sbdev->ctrl = ctrl; in slim_add_device() 176 static struct slim_device *slim_alloc_device(struct slim_controller *ctrl, in slim_alloc_device() argument 188 ret = slim_add_device(ctrl, sbdev, node); in slim_alloc_device() 197 static void of_register_slim_devices(struct slim_controller *ctrl) in of_register_slim_devices() argument 199 struct device *dev = ctrl->dev; in of_register_slim_devices() 202 if (!ctrl->dev->of_node) in of_register_slim_devices() 205 for_each_child_of_node(ctrl->dev->of_node, node) { in of_register_slim_devices() 235 sbdev = slim_alloc_device(ctrl, &e_addr, node); in of_register_slim_devices() [all …]
|
| /linux/drivers/nvme/host/ |
| H A D | rdma.c | 89 struct nvme_rdma_ctrl *ctrl; member 125 struct nvme_ctrl ctrl; member 130 static inline struct nvme_rdma_ctrl *to_rdma_ctrl(struct nvme_ctrl *ctrl) in to_rdma_ctrl() argument 132 return container_of(ctrl, struct nvme_rdma_ctrl, ctrl); in to_rdma_ctrl() 161 return queue - queue->ctrl->queues; in nvme_rdma_queue_idx() 167 queue->ctrl->io_queues[HCTX_TYPE_DEFAULT] + in nvme_rdma_poll_queue() 168 queue->ctrl->io_queues[HCTX_TYPE_READ]; in nvme_rdma_poll_queue() 297 struct nvme_rdma_ctrl *ctrl = to_rdma_ctrl(set->driver_data); in nvme_rdma_init_request() local 299 int queue_idx = (set == &ctrl->tag_set) ? hctx_idx + 1 : 0; in nvme_rdma_init_request() 300 struct nvme_rdma_queue *queue = &ctrl->queues[queue_idx]; in nvme_rdma_init_request() [all …]
|
| H A D | fc.c | 35 struct nvme_fc_ctrl *ctrl; member 99 struct nvme_fc_ctrl *ctrl; member 179 struct nvme_ctrl ctrl; member 183 to_fc_ctrl(struct nvme_ctrl *ctrl) in to_fc_ctrl() argument 185 return container_of(ctrl, struct nvme_fc_ctrl, ctrl); in to_fc_ctrl() 552 nvme_fc_resume_controller(struct nvme_fc_ctrl *ctrl) in nvme_fc_resume_controller() argument 554 switch (nvme_ctrl_state(&ctrl->ctrl)) { in nvme_fc_resume_controller() 561 dev_info(ctrl->ctrl.device, in nvme_fc_resume_controller() 563 "Attempting reconnect\n", ctrl->cnum); in nvme_fc_resume_controller() 565 queue_delayed_work(nvme_wq, &ctrl->connect_work, 0); in nvme_fc_resume_controller() [all …]
|
| H A D | core.c | 151 static void nvme_remove_invalid_namespaces(struct nvme_ctrl *ctrl, 153 static void nvme_update_keep_alive(struct nvme_ctrl *ctrl, 155 static int nvme_get_log_lsi(struct nvme_ctrl *ctrl, u32 nsid, u8 log_page, 158 void nvme_queue_scan(struct nvme_ctrl *ctrl) in nvme_queue_scan() argument 163 if (nvme_ctrl_state(ctrl) == NVME_CTRL_LIVE && ctrl->tagset) in nvme_queue_scan() 164 queue_work(nvme_wq, &ctrl->scan_work); in nvme_queue_scan() 173 int nvme_try_sched_reset(struct nvme_ctrl *ctrl) in nvme_try_sched_reset() argument 175 if (nvme_ctrl_state(ctrl) != NVME_CTRL_RESETTING) in nvme_try_sched_reset() 177 if (!queue_work(nvme_reset_wq, &ctrl->reset_work)) in nvme_try_sched_reset() 185 struct nvme_ctrl *ctrl = container_of(to_delayed_work(work), in nvme_failfast_work() local [all …]
|
| H A D | tcp.c | 165 struct nvme_tcp_ctrl *ctrl; member 195 struct nvme_ctrl ctrl; member 210 static inline struct nvme_tcp_ctrl *to_tcp_ctrl(struct nvme_ctrl *ctrl) in to_tcp_ctrl() argument 212 return container_of(ctrl, struct nvme_tcp_ctrl, ctrl); in to_tcp_ctrl() 217 return queue - queue->ctrl->queues; in nvme_tcp_queue_id() 247 static inline bool nvme_tcp_tls_configured(struct nvme_ctrl *ctrl) in nvme_tcp_tls_configured() argument 252 return ctrl->opts->tls || ctrl->opts->concat; in nvme_tcp_tls_configured() 260 return queue->ctrl->admin_tag_set.tags[queue_idx]; in nvme_tcp_tagset() 261 return queue->ctrl->tag_set.tags[queue_idx - 1]; in nvme_tcp_tagset() 295 return req == &req->queue->ctrl->async_req; in nvme_tcp_async_req() [all …]
|
| H A D | auth.c | 23 struct nvme_ctrl *ctrl; member 52 static inline int ctrl_max_dhchaps(struct nvme_ctrl *ctrl) in ctrl_max_dhchaps() argument 54 return ctrl->opts->nr_io_queues + ctrl->opts->nr_write_queues + in ctrl_max_dhchaps() 55 ctrl->opts->nr_poll_queues + 1; in ctrl_max_dhchaps() 58 static int nvme_auth_submit(struct nvme_ctrl *ctrl, int qid, in nvme_auth_submit() argument 63 struct request_queue *q = ctrl->fabrics_q; in nvme_auth_submit() 68 q = ctrl->connect_q; in nvme_auth_submit() 86 dev_warn(ctrl->device, in nvme_auth_submit() 89 dev_err(ctrl->device, in nvme_auth_submit() 94 static int nvme_auth_receive_validate(struct nvme_ctrl *ctrl, int qid, in nvme_auth_receive_validate() argument [all …]
|
| H A D | nvme.h | 251 struct nvme_ctrl *ctrl; member 471 static inline enum nvme_ctrl_state nvme_ctrl_state(struct nvme_ctrl *ctrl) in nvme_ctrl_state() argument 473 return READ_ONCE(ctrl->state); in nvme_ctrl_state() 586 struct nvme_ctrl *ctrl; member 616 static inline unsigned long nvme_get_virt_boundary(struct nvme_ctrl *ctrl, in nvme_get_virt_boundary() argument 631 int (*reg_read32)(struct nvme_ctrl *ctrl, u32 off, u32 *val); 632 int (*reg_write32)(struct nvme_ctrl *ctrl, u32 off, u32 val); 633 int (*reg_read64)(struct nvme_ctrl *ctrl, u32 off, u64 *val); 634 void (*free_ctrl)(struct nvme_ctrl *ctrl); 635 void (*submit_async_event)(struct nvme_ctrl *ctrl); [all …]
|
| /linux/drivers/pci/hotplug/ |
| H A D | pciehp_hpc.c | 49 static inline struct pci_dev *ctrl_dev(struct controller *ctrl) in ctrl_dev() argument 51 return ctrl->pcie->port; in ctrl_dev() 58 static inline int pciehp_request_irq(struct controller *ctrl) in pciehp_request_irq() argument 60 int retval, irq = ctrl->pcie->irq; in pciehp_request_irq() 63 ctrl->poll_thread = kthread_run(&pciehp_poll, ctrl, in pciehp_request_irq() 65 slot_name(ctrl)); in pciehp_request_irq() 66 return PTR_ERR_OR_ZERO(ctrl->poll_thread); in pciehp_request_irq() 71 IRQF_SHARED, "pciehp", ctrl); in pciehp_request_irq() 73 ctrl_err(ctrl, "Cannot get irq %d for the hotplug controller\n", in pciehp_request_irq() 78 static inline void pciehp_free_irq(struct controller *ctrl) in pciehp_free_irq() argument [all …]
|
| H A D | shpchp_hpc.c | 169 static void start_int_poll_timer(struct controller *ctrl, int sec); 171 static inline u8 shpc_readb(struct controller *ctrl, int reg) in shpc_readb() argument 173 return readb(ctrl->creg + reg); in shpc_readb() 176 static inline u16 shpc_readw(struct controller *ctrl, int reg) in shpc_readw() argument 178 return readw(ctrl->creg + reg); in shpc_readw() 181 static inline void shpc_writew(struct controller *ctrl, int reg, u16 val) in shpc_writew() argument 183 writew(val, ctrl->creg + reg); in shpc_writew() 186 static inline u32 shpc_readl(struct controller *ctrl, int reg) in shpc_readl() argument 188 return readl(ctrl->creg + reg); in shpc_readl() 191 static inline void shpc_writel(struct controller *ctrl, int reg, u32 val) in shpc_writel() argument [all …]
|
| H A D | cpqphp_core.c | 116 * @ctrl: controller to use 120 static int init_SERR(struct controller *ctrl) in init_SERR() argument 125 if (!ctrl) in init_SERR() 128 tempdword = ctrl->first_slot; in init_SERR() 130 number_of_slots = readb(ctrl->hpc_reg + SLOT_MASK) & 0x0F; in init_SERR() 133 writeb(0, ctrl->hpc_reg + SLOT_SERR); in init_SERR() 266 static int ctrl_slot_cleanup(struct controller *ctrl) in ctrl_slot_cleanup() argument 270 old_slot = ctrl->slot; in ctrl_slot_cleanup() 271 ctrl->slot = NULL; in ctrl_slot_cleanup() 280 cpqhp_remove_debugfs_files(ctrl); in ctrl_slot_cleanup() [all …]
|
| H A D | shpchp_ctrl.c | 45 u8 shpchp_handle_attention_button(u8 hp_slot, struct controller *ctrl) in shpchp_handle_attention_button() argument 51 ctrl_dbg(ctrl, "Attention button interrupt received\n"); in shpchp_handle_attention_button() 53 p_slot = shpchp_find_slot(ctrl, hp_slot + ctrl->slot_device_offset); in shpchp_handle_attention_button() 59 ctrl_info(ctrl, "Button pressed on Slot(%s)\n", slot_name(p_slot)); in shpchp_handle_attention_button() 68 u8 shpchp_handle_switch_change(u8 hp_slot, struct controller *ctrl) in shpchp_handle_switch_change() argument 75 ctrl_dbg(ctrl, "Switch interrupt received\n"); in shpchp_handle_switch_change() 77 p_slot = shpchp_find_slot(ctrl, hp_slot + ctrl->slot_device_offset); in shpchp_handle_switch_change() 80 ctrl_dbg(ctrl, "Card present %x Power status %x\n", in shpchp_handle_switch_change() 87 ctrl_info(ctrl, "Latch open on Slot(%s)\n", slot_name(p_slot)); in shpchp_handle_switch_change() 91 ctrl_err(ctrl, "Surprise Removal of card\n"); in shpchp_handle_switch_change() [all …]
|
| H A D | cpqphp_ctrl.c | 28 static u32 configure_new_device(struct controller *ctrl, struct pci_func *func, 30 static int configure_new_function(struct controller *ctrl, struct pci_func *func, 32 static void interrupt_event_handler(struct controller *ctrl); 53 static u8 handle_switch_change(u8 change, struct controller *ctrl) in handle_switch_change() argument 72 func = cpqhp_slot_find(ctrl->bus, in handle_switch_change() 73 (hp_slot + ctrl->slot_device_offset), 0); in handle_switch_change() 78 taskInfo = &(ctrl->event_queue[ctrl->next_event]); in handle_switch_change() 79 ctrl->next_event = (ctrl->next_event + 1) % 10; in handle_switch_change() 84 temp_word = ctrl->ctrl_int_comp >> 16; in handle_switch_change() 88 if (ctrl->ctrl_int_comp & (0x1L << hp_slot)) { in handle_switch_change() [all …]
|
| /linux/drivers/media/v4l2-core/ |
| H A D | v4l2-ctrls-api.c | 41 struct v4l2_ctrl *ctrl, in ptr_to_user() argument 46 if (ctrl->is_ptr && !ctrl->is_string) in ptr_to_user() 50 switch (ctrl->type) { in ptr_to_user() 54 c->size = ctrl->elem_size; in ptr_to_user() 70 static int cur_to_user(struct v4l2_ext_control *c, struct v4l2_ctrl *ctrl) in cur_to_user() argument 72 return ptr_to_user(c, ctrl, ctrl->p_cur); in cur_to_user() 77 struct v4l2_ctrl *ctrl) in new_to_user() argument 79 return ptr_to_user(c, ctrl, ctrl->p_new); in new_to_user() 86 return ptr_to_user(c, ref->ctrl, ref->p_req); in req_to_user() 90 static int def_to_user(struct v4l2_ext_control *c, struct v4l2_ctrl *ctrl) in def_to_user() argument [all …]
|
| H A D | v4l2-ctrls-core.c | 19 static void fill_event(struct v4l2_event *ev, struct v4l2_ctrl *ctrl, in fill_event() argument 24 ev->id = ctrl->id; in fill_event() 25 ev->u.ctrl.changes = changes; in fill_event() 26 ev->u.ctrl.type = ctrl->type; in fill_event() 27 ev->u.ctrl.flags = user_flags(ctrl); in fill_event() 28 if (ctrl->is_ptr) in fill_event() 29 ev->u.ctrl.value64 = 0; in fill_event() 31 ev->u.ctrl.value64 = *ctrl->p_cur.p_s64; in fill_event() 32 ev->u.ctrl.minimum = ctrl->minimum; in fill_event() 33 ev->u.ctrl.maximum = ctrl->maximum; in fill_event() [all …]
|
| /linux/drivers/mtd/nand/raw/brcmnand/ |
| H A D | brcmnand.c | 346 struct brcmnand_controller *ctrl; member 666 static inline bool brcmnand_non_mmio_ops(struct brcmnand_controller *ctrl) in brcmnand_non_mmio_ops() argument 675 static inline u32 nand_readreg(struct brcmnand_controller *ctrl, u32 offs) in nand_readreg() argument 677 if (brcmnand_non_mmio_ops(ctrl)) in nand_readreg() 678 return brcmnand_soc_read(ctrl->soc, offs); in nand_readreg() 679 return brcmnand_readl(ctrl->nand_base + offs); in nand_readreg() 682 static inline void nand_writereg(struct brcmnand_controller *ctrl, u32 offs, in nand_writereg() argument 685 if (brcmnand_non_mmio_ops(ctrl)) in nand_writereg() 686 brcmnand_soc_write(ctrl->soc, val, offs); in nand_writereg() 688 brcmnand_writel(val, ctrl->nand_base + offs); in nand_writereg() [all …]
|
| /linux/drivers/net/mdio/ |
| H A D | mdio-bitbang.c | 44 static void mdiobb_send_bit(struct mdiobb_ctrl *ctrl, int val) in mdiobb_send_bit() argument 46 const struct mdiobb_ops *ops = ctrl->ops; in mdiobb_send_bit() 48 ops->set_mdio_data(ctrl, val); in mdiobb_send_bit() 50 ops->set_mdc(ctrl, 1); in mdiobb_send_bit() 52 ops->set_mdc(ctrl, 0); in mdiobb_send_bit() 56 static int mdiobb_get_bit(struct mdiobb_ctrl *ctrl) in mdiobb_get_bit() argument 58 const struct mdiobb_ops *ops = ctrl->ops; in mdiobb_get_bit() 61 ops->set_mdc(ctrl, 1); in mdiobb_get_bit() 63 ops->set_mdc(ctrl, 0); in mdiobb_get_bit() 65 return ops->get_mdio_data(ctrl); in mdiobb_get_bit() [all …]
|
| /linux/drivers/clk/bcm/ |
| H A D | clk-iproc-pll.c | 68 const struct iproc_pll_ctrl *ctrl; member 76 const struct iproc_clk_ctrl *ctrl; member 150 const struct iproc_pll_ctrl *ctrl = pll->ctrl; in pll_wait_for_lock() local 153 u32 val = readl(pll->status_base + ctrl->status.offset); in pll_wait_for_lock() 155 if (val & (1 << ctrl->status.shift)) in pll_wait_for_lock() 166 const struct iproc_pll_ctrl *ctrl = pll->ctrl; in iproc_pll_write() local 170 if (unlikely(ctrl->flags & IPROC_CLK_NEEDS_READ_BACK && in iproc_pll_write() 177 const struct iproc_pll_ctrl *ctrl = pll->ctrl; in __pll_disable() local 180 if (ctrl->flags & IPROC_CLK_PLL_ASIU) { in __pll_disable() 181 val = readl(pll->asiu_base + ctrl->asiu.offset); in __pll_disable() [all …]
|
| /linux/drivers/watchdog/ |
| H A D | realtek_otto_wdt.c | 76 struct otto_wdt_ctrl *ctrl = watchdog_get_drvdata(wdev); in otto_wdt_start() local 79 v = ioread32(ctrl->base + OTTO_WDT_REG_CTRL); in otto_wdt_start() 81 iowrite32(v, ctrl->base + OTTO_WDT_REG_CTRL); in otto_wdt_start() 88 struct otto_wdt_ctrl *ctrl = watchdog_get_drvdata(wdev); in otto_wdt_stop() local 91 v = ioread32(ctrl->base + OTTO_WDT_REG_CTRL); in otto_wdt_stop() 93 iowrite32(v, ctrl->base + OTTO_WDT_REG_CTRL); in otto_wdt_stop() 100 struct otto_wdt_ctrl *ctrl = watchdog_get_drvdata(wdev); in otto_wdt_ping() local 102 iowrite32(OTTO_WDT_CNTR_PING, ctrl->base + OTTO_WDT_REG_CNTR); in otto_wdt_ping() 107 static int otto_wdt_tick_ms(struct otto_wdt_ctrl *ctrl, int prescale) in otto_wdt_tick_ms() argument 109 return DIV_ROUND_CLOSEST(1 << (25 + prescale), ctrl->clk_rate_khz); in otto_wdt_tick_ms() [all …]
|
| /linux/sound/soc/codecs/ |
| H A D | wcd-clsh-v2.c | 119 static inline void wcd_enable_clsh_block(struct wcd_clsh_ctrl *ctrl, in wcd_enable_clsh_block() argument 122 struct snd_soc_component *comp = ctrl->comp; in wcd_enable_clsh_block() 124 if ((enable && ++ctrl->clsh_users == 1) || in wcd_enable_clsh_block() 125 (!enable && --ctrl->clsh_users == 0)) in wcd_enable_clsh_block() 129 if (ctrl->clsh_users < 0) in wcd_enable_clsh_block() 130 ctrl->clsh_users = 0; in wcd_enable_clsh_block() 175 static void wcd_clsh_buck_ctrl(struct wcd_clsh_ctrl *ctrl, in wcd_clsh_buck_ctrl() argument 179 struct snd_soc_component *comp = ctrl->comp; in wcd_clsh_buck_ctrl() 182 if ((enable && (++ctrl->buck_users == 1)) || in wcd_clsh_buck_ctrl() 183 (!enable && (--ctrl in wcd_clsh_buck_ctrl() 195 wcd_clsh_v3_buck_ctrl(struct snd_soc_component * component,struct wcd_clsh_ctrl * ctrl,int mode,bool enable) wcd_clsh_v3_buck_ctrl() argument 224 wcd_clsh_flyback_ctrl(struct wcd_clsh_ctrl * ctrl,int mode,bool enable) wcd_clsh_flyback_ctrl() argument 246 wcd_clsh_set_gain_path(struct wcd_clsh_ctrl * ctrl,int mode) wcd_clsh_set_gain_path() argument 348 wcd_clsh_set_hph_mode(struct wcd_clsh_ctrl * ctrl,int mode) wcd_clsh_set_hph_mode() argument 453 wcd_clsh_v3_flyback_ctrl(struct snd_soc_component * component,struct wcd_clsh_ctrl * ctrl,int mode,bool enable) wcd_clsh_v3_flyback_ctrl() argument 490 wcd_clsh_v3_state_aux(struct wcd_clsh_ctrl * ctrl,int req_state,bool is_enable,int mode) wcd_clsh_v3_state_aux() argument 509 wcd_clsh_state_lo(struct wcd_clsh_ctrl * ctrl,int req_state,bool is_enable,int mode) wcd_clsh_state_lo() argument 536 wcd_clsh_v3_state_hph_r(struct wcd_clsh_ctrl * ctrl,int req_state,bool is_enable,int mode) wcd_clsh_v3_state_hph_r() argument 568 wcd_clsh_state_hph_r(struct wcd_clsh_ctrl * ctrl,int req_state,bool is_enable,int mode) wcd_clsh_state_hph_r() argument 626 wcd_clsh_v3_state_hph_l(struct wcd_clsh_ctrl * ctrl,int req_state,bool is_enable,int mode) wcd_clsh_v3_state_hph_l() argument 658 wcd_clsh_state_hph_l(struct wcd_clsh_ctrl * ctrl,int req_state,bool is_enable,int mode) wcd_clsh_state_hph_l() argument 716 wcd_clsh_v3_state_ear(struct wcd_clsh_ctrl * ctrl,int req_state,bool is_enable,int mode) wcd_clsh_v3_state_ear() argument 742 wcd_clsh_state_ear(struct wcd_clsh_ctrl * ctrl,int req_state,bool is_enable,int mode) wcd_clsh_state_ear() argument 777 _wcd_clsh_ctrl_set_state(struct wcd_clsh_ctrl * ctrl,int req_state,bool is_enable,int mode) _wcd_clsh_ctrl_set_state() argument 844 wcd_clsh_ctrl_set_state(struct wcd_clsh_ctrl * ctrl,enum wcd_clsh_event clsh_event,int nstate,enum wcd_clsh_mode mode) wcd_clsh_ctrl_set_state() argument 875 wcd_clsh_ctrl_get_state(struct wcd_clsh_ctrl * ctrl) wcd_clsh_ctrl_get_state() argument 884 struct wcd_clsh_ctrl *ctrl; wcd_clsh_ctrl_alloc() local 898 wcd_clsh_ctrl_free(struct wcd_clsh_ctrl * ctrl) wcd_clsh_ctrl_free() argument [all...] |