| /linux/drivers/nvme/target/ |
| H A D | loop.c | 36 struct nvme_ctrl ctrl; member 44 static inline struct nvme_loop_ctrl *to_loop_ctrl(struct nvme_ctrl *ctrl) in to_loop_ctrl() argument 46 return container_of(ctrl, struct nvme_loop_ctrl, ctrl); in to_loop_ctrl() 56 struct nvme_loop_ctrl *ctrl; member 67 static void nvme_loop_delete_ctrl(struct nvmet_ctrl *ctrl); 73 return queue - queue->ctrl->queues; in nvme_loop_queue_idx() 89 return queue->ctrl->admin_tag_set.tags[queue_idx]; in nvme_loop_tagset() 90 return queue->ctrl->tag_set.tags[queue_idx - 1]; in nvme_loop_tagset() 107 nvme_complete_async_event(&queue->ctrl in nvme_loop_queue_response() 176 struct nvme_loop_ctrl *ctrl = to_loop_ctrl(arg); nvme_loop_submit_async_event() local 193 nvme_loop_init_iod(struct nvme_loop_ctrl * ctrl,struct nvme_loop_iod * iod,unsigned int queue_idx) nvme_loop_init_iod() argument 207 struct nvme_loop_ctrl *ctrl = to_loop_ctrl(set->driver_data); nvme_loop_init_request() local 221 struct nvme_loop_ctrl *ctrl = to_loop_ctrl(data); nvme_loop_init_hctx() local 241 struct nvme_loop_ctrl *ctrl = to_loop_ctrl(data); nvme_loop_init_admin_hctx() local 264 nvme_loop_destroy_admin_queue(struct nvme_loop_ctrl * ctrl) nvme_loop_destroy_admin_queue() argument 282 struct nvme_loop_ctrl *ctrl = to_loop_ctrl(nctrl); nvme_loop_free_ctrl() local 299 nvme_loop_destroy_io_queues(struct nvme_loop_ctrl * ctrl) nvme_loop_destroy_io_queues() argument 317 nvme_loop_init_io_queues(struct nvme_loop_ctrl * ctrl) nvme_loop_init_io_queues() argument 350 nvme_loop_connect_io_queues(struct nvme_loop_ctrl * ctrl) nvme_loop_connect_io_queues() argument 364 nvme_loop_configure_admin_queue(struct nvme_loop_ctrl * ctrl) nvme_loop_configure_admin_queue() argument 418 nvme_loop_shutdown_ctrl(struct nvme_loop_ctrl * ctrl) nvme_loop_shutdown_ctrl() argument 434 nvme_loop_delete_ctrl_host(struct nvme_ctrl * ctrl) nvme_loop_delete_ctrl_host() argument 441 struct nvme_loop_ctrl *ctrl; nvme_loop_delete_ctrl() local 453 struct nvme_loop_ctrl *ctrl = nvme_loop_reset_ctrl_work() local 517 nvme_loop_create_io_queues(struct nvme_loop_ctrl * ctrl) nvme_loop_create_io_queues() argument 545 nvme_loop_find_port(struct nvme_ctrl * ctrl) nvme_loop_find_port() argument 565 struct nvme_loop_ctrl *ctrl; nvme_loop_create_ctrl() local 706 struct nvme_loop_ctrl *ctrl, *next; nvme_loop_cleanup_module() local [all...] |
| H A D | auth.c | 64 int nvmet_setup_dhgroup(struct nvmet_ctrl *ctrl, u8 dhgroup_id) in nvmet_auth_set_key() 69 pr_debug("%s: ctrl %d selecting dhgroup %d\n", in nvmet_auth_set_key() 70 __func__, ctrl->cntlid, dhgroup_id); in nvmet_auth_set_key() 72 if (ctrl->dh_tfm) { in nvmet_auth_set_key() 73 if (ctrl->dh_gid == dhgroup_id) { 74 pr_debug("%s: ctrl %d reuse existing DH group %d\n", in nvmet_setup_dhgroup() argument 75 __func__, ctrl->cntlid, dhgroup_id); in nvmet_setup_dhgroup() 78 crypto_free_kpp(ctrl->dh_tfm); in nvmet_setup_dhgroup() 79 ctrl->dh_tfm = NULL; in nvmet_setup_dhgroup() 80 ctrl in nvmet_setup_dhgroup() 143 nvmet_setup_auth(struct nvmet_ctrl * ctrl,struct nvmet_sq * sq) nvmet_setup_auth() argument 253 nvmet_destroy_auth(struct nvmet_ctrl * ctrl) nvmet_destroy_auth() argument 297 struct nvmet_ctrl *ctrl = req->sq->ctrl; nvmet_auth_host_hash() local 404 struct nvmet_ctrl *ctrl = req->sq->ctrl; nvmet_auth_ctrl_hash() local 513 struct nvmet_ctrl *ctrl = req->sq->ctrl; nvmet_auth_ctrl_exponential() local 536 struct nvmet_ctrl *ctrl = req->sq->ctrl; nvmet_auth_ctrl_sesskey() local [all...] |
| H A D | core.c | 137 static void nvmet_async_events_failall(struct nvmet_ctrl *ctrl) in nvmet_async_events_failall() argument 141 mutex_lock(&ctrl->lock); in nvmet_async_events_failall() 142 while (ctrl->nr_async_event_cmds) { in nvmet_async_events_failall() 143 req = ctrl->async_event_cmds[--ctrl->nr_async_event_cmds]; in nvmet_async_events_failall() 144 mutex_unlock(&ctrl->lock); in nvmet_async_events_failall() 146 mutex_lock(&ctrl->lock); in nvmet_async_events_failall() 148 mutex_unlock(&ctrl->lock); in nvmet_async_events_failall() 151 static void nvmet_async_events_process(struct nvmet_ctrl *ctrl) in nvmet_async_events_process() argument 156 mutex_lock(&ctrl in nvmet_async_events_process() 174 nvmet_async_events_free(struct nvmet_ctrl * ctrl) nvmet_async_events_free() argument 188 struct nvmet_ctrl *ctrl = nvmet_async_event_work() local 194 nvmet_add_async_event(struct nvmet_ctrl * ctrl,u8 event_type,u8 event_info,u8 log_page) nvmet_add_async_event() argument 214 nvmet_add_to_changed_ns_log(struct nvmet_ctrl * ctrl,__le32 nsid) nvmet_add_to_changed_ns_log() argument 240 struct nvmet_ctrl *ctrl; nvmet_ns_changed() local 257 struct nvmet_ctrl *ctrl; nvmet_send_ana_event() local 306 struct nvmet_ctrl *ctrl; nvmet_port_del_ctrls() local 398 struct nvmet_ctrl *ctrl = container_of(to_delayed_work(work), nvmet_keep_alive_timer() local 416 nvmet_start_keep_alive_timer(struct nvmet_ctrl * ctrl) nvmet_start_keep_alive_timer() argument 427 nvmet_stop_keep_alive_timer(struct nvmet_ctrl * ctrl) nvmet_stop_keep_alive_timer() argument 519 nvmet_p2pmem_ns_add_p2p(struct nvmet_ctrl * ctrl,struct nvmet_ns * ns) nvmet_p2pmem_ns_add_p2p() argument 572 struct nvmet_ctrl *ctrl; nvmet_ns_enable() local 629 struct nvmet_ctrl *ctrl; nvmet_ns_disable() local 750 struct nvmet_ctrl *ctrl = req->sq->ctrl; nvmet_set_error() local 828 nvmet_cq_setup(struct nvmet_ctrl * ctrl,struct nvmet_cq * cq,u16 qid,u16 size) nvmet_cq_setup() argument 839 struct nvmet_ctrl *ctrl = cq->ctrl; nvmet_cq_destroy() local 848 nvmet_sq_setup(struct nvmet_ctrl * ctrl,struct nvmet_sq * sq,u16 qid,u16 size) nvmet_sq_setup() argument 865 nvmet_check_cqid(struct nvmet_ctrl * ctrl,u16 cqid,bool create) nvmet_check_cqid() argument 879 nvmet_check_io_cqid(struct nvmet_ctrl * ctrl,u16 cqid,bool create) nvmet_check_io_cqid() argument 892 nvmet_cq_create(struct nvmet_ctrl * ctrl,struct nvmet_cq * cq,u16 qid,u16 size) nvmet_cq_create() argument 912 nvmet_check_sqid(struct nvmet_ctrl * ctrl,u16 sqid,bool create) nvmet_check_sqid() argument 928 nvmet_sq_create(struct nvmet_ctrl * ctrl,struct nvmet_sq * sq,struct nvmet_cq * cq,u16 sqid,u16 size) nvmet_sq_create() argument 960 struct nvmet_ctrl *ctrl = sq->ctrl; nvmet_sq_destroy() local 1375 nvmet_start_ctrl(struct nvmet_ctrl * ctrl) nvmet_start_ctrl() argument 1411 nvmet_clear_ctrl(struct nvmet_ctrl * ctrl) nvmet_clear_ctrl() argument 1420 nvmet_update_cc(struct nvmet_ctrl * ctrl,u32 new) nvmet_update_cc() argument 1442 nvmet_init_cap(struct nvmet_ctrl * ctrl) nvmet_init_cap() argument 1465 struct nvmet_ctrl *ctrl = NULL; nvmet_ctrl_find_get() local 1544 nvmet_setup_p2p_ns_map(struct nvmet_ctrl * ctrl,struct device * p2p_client) nvmet_setup_p2p_ns_map() argument 1561 nvmet_release_p2p_ns_map(struct nvmet_ctrl * ctrl) nvmet_release_p2p_ns_map() argument 1576 struct nvmet_ctrl *ctrl = nvmet_fatal_error_handler() local 1586 struct nvmet_ctrl *ctrl; nvmet_alloc_ctrl() local 1735 struct nvmet_ctrl *ctrl = container_of(ref, struct nvmet_ctrl, ref); nvmet_ctrl_free() local 1764 nvmet_ctrl_put(struct nvmet_ctrl * ctrl) nvmet_ctrl_put() argument 1770 nvmet_ctrl_fatal_error(struct nvmet_ctrl * ctrl) nvmet_ctrl_fatal_error() argument 1781 nvmet_ctrl_host_traddr(struct nvmet_ctrl * ctrl,char * traddr,size_t traddr_len) nvmet_ctrl_host_traddr() argument 1922 struct nvmet_ctrl *ctrl; nvmet_subsys_del_ctrls() local [all...] |
| H A D | pci-epf.c | 84 struct nvmet_pci_epf_ctrl *ctrl; member 129 struct nvmet_pci_epf_ctrl *ctrl; member 209 struct nvmet_pci_epf_ctrl ctrl; member 226 static inline u32 nvmet_pci_epf_bar_read32(struct nvmet_pci_epf_ctrl *ctrl, in nvmet_pci_epf_bar_read32() argument 229 __le32 *bar_reg = ctrl->bar + off; in nvmet_pci_epf_bar_read32() 234 static inline void nvmet_pci_epf_bar_write32(struct nvmet_pci_epf_ctrl *ctrl, in nvmet_pci_epf_bar_write32() argument 237 __le32 *bar_reg = ctrl->bar + off; in nvmet_pci_epf_bar_write32() 242 static inline u64 nvmet_pci_epf_bar_read64(struct nvmet_pci_epf_ctrl *ctrl, in nvmet_pci_epf_bar_read64() argument 245 return (u64)nvmet_pci_epf_bar_read32(ctrl, off) | in nvmet_pci_epf_bar_read64() 246 ((u64)nvmet_pci_epf_bar_read32(ctrl, off + 4) << 32); in nvmet_pci_epf_bar_read64() [all …]
|
| H A D | fabrics-cmd-auth.c | 19 pr_debug("%s: ctrl %d qid %d transaction %u expired, resetting\n", in nvmet_auth_expired_work() 20 __func__, sq->ctrl->cntlid, sq->qid, sq->dhchap_tid); in nvmet_auth_expired_work() 35 struct nvmet_ctrl *ctrl = req->sq->ctrl; in nvmet_auth_negotiate() 39 pr_debug("%s: ctrl %d qid %d: data sc_d %d napd %d authid %d halen %d dhlen %d\n", in nvmet_auth_negotiate() 40 __func__, ctrl->cntlid, req->sq->qid, in nvmet_auth_negotiate() 64 ctrl->concat = true; in nvmet_auth_negotiate() 79 if (ctrl->shash_id != host_hmac_id) in nvmet_auth_negotiate() 81 hash_id = ctrl->shash_id; in nvmet_auth_negotiate() 86 pr_debug("%s: ctrl in nvmet_auth_negotiate() 36 struct nvmet_ctrl *ctrl = req->sq->ctrl; nvmet_auth_negotiate() local 139 struct nvmet_ctrl *ctrl = req->sq->ctrl; nvmet_auth_reply() local 236 struct nvmet_ctrl *ctrl = req->sq->ctrl; nvmet_execute_auth_send() local 410 struct nvmet_ctrl *ctrl = req->sq->ctrl; nvmet_auth_challenge() local 450 struct nvmet_ctrl *ctrl = req->sq->ctrl; nvmet_auth_success1() local 493 struct nvmet_ctrl *ctrl = req->sq->ctrl; nvmet_execute_auth_receive() local [all...] |
| /linux/drivers/soundwire/ |
| H A D | qcom.c | 220 int (*reg_read)(struct qcom_swrm_ctrl *ctrl, int reg, u32 *val); 221 int (*reg_write)(struct qcom_swrm_ctrl *ctrl, int reg, int val); 331 static int qcom_swrm_ahb_reg_read(struct qcom_swrm_ctrl *ctrl, int reg, in qcom_swrm_ahb_reg_read() argument 334 struct regmap *wcd_regmap = ctrl->regmap; in qcom_swrm_ahb_reg_read() 351 static int qcom_swrm_ahb_reg_write(struct qcom_swrm_ctrl *ctrl, in qcom_swrm_ahb_reg_write() argument 354 struct regmap *wcd_regmap = ctrl->regmap; in qcom_swrm_ahb_reg_write() 371 static int qcom_swrm_cpu_reg_read(struct qcom_swrm_ctrl *ctrl, int reg, in qcom_swrm_cpu_reg_read() argument 374 *val = readl(ctrl->mmio + reg); in qcom_swrm_cpu_reg_read() 378 static int qcom_swrm_cpu_reg_write(struct qcom_swrm_ctrl *ctrl, int reg, in qcom_swrm_cpu_reg_write() argument 381 writel(val, ctrl->mmio + reg); in qcom_swrm_cpu_reg_write() [all …]
|
| /linux/drivers/slimbus/ |
| H A D | qcom-ngd-ctrl.c | 135 struct qcom_slim_ngd_ctrl *ctrl; member 150 struct slim_controller ctrl; member 347 static int qcom_slim_qmi_send_select_inst_req(struct qcom_slim_ngd_ctrl *ctrl, in qcom_slim_qmi_send_select_inst_req() argument 354 rc = qmi_txn_init(ctrl->qmi.handle, &txn, in qcom_slim_qmi_send_select_inst_req() 357 dev_err(ctrl->dev, "QMI TXN init fail: %d\n", rc); in qcom_slim_qmi_send_select_inst_req() 361 rc = qmi_send_request(ctrl->qmi.handle, NULL, &txn, in qcom_slim_qmi_send_select_inst_req() 366 dev_err(ctrl->dev, "QMI send req fail %d\n", rc); in qcom_slim_qmi_send_select_inst_req() 373 dev_err(ctrl->dev, "QMI TXN wait fail: %d\n", rc); in qcom_slim_qmi_send_select_inst_req() 378 dev_err(ctrl->dev, "QMI request failed 0x%x\n", in qcom_slim_qmi_send_select_inst_req() 400 static int qcom_slim_qmi_send_power_request(struct qcom_slim_ngd_ctrl *ctrl, in qcom_slim_qmi_send_power_request() argument [all …]
|
| /linux/drivers/nvme/host/ |
| H A D | rdma.c | 89 struct nvme_rdma_ctrl *ctrl; member 125 struct nvme_ctrl ctrl; member 130 static inline struct nvme_rdma_ctrl *to_rdma_ctrl(struct nvme_ctrl *ctrl) in to_rdma_ctrl() argument 132 return container_of(ctrl, struct nvme_rdma_ctrl, ctrl); in to_rdma_ctrl() 161 return queue - queue->ctrl->queues; in nvme_rdma_queue_idx() 167 queue->ctrl->io_queues[HCTX_TYPE_DEFAULT] + in nvme_rdma_poll_queue() 168 queue->ctrl->io_queues[HCTX_TYPE_READ]; in nvme_rdma_poll_queue() 297 struct nvme_rdma_ctrl *ctrl = to_rdma_ctrl(set->driver_data); in nvme_rdma_init_request() local 299 int queue_idx = (set == &ctrl->tag_set) ? hctx_idx + 1 : 0; in nvme_rdma_init_request() 300 struct nvme_rdma_queue *queue = &ctrl->queues[queue_idx]; in nvme_rdma_init_request() [all …]
|
| H A D | fc.c | 35 struct nvme_fc_ctrl *ctrl; member 99 struct nvme_fc_ctrl *ctrl; member 179 struct nvme_ctrl ctrl; member 183 to_fc_ctrl(struct nvme_ctrl *ctrl) in to_fc_ctrl() argument 185 return container_of(ctrl, struct nvme_fc_ctrl, ctrl); in to_fc_ctrl() 552 nvme_fc_resume_controller(struct nvme_fc_ctrl *ctrl) in nvme_fc_resume_controller() argument 554 switch (nvme_ctrl_state(&ctrl->ctrl)) { in nvme_fc_resume_controller() 561 dev_info(ctrl->ctrl.device, in nvme_fc_resume_controller() 563 "Attempting reconnect\n", ctrl->cnum); in nvme_fc_resume_controller() 565 queue_delayed_work(nvme_wq, &ctrl->connect_work, 0); in nvme_fc_resume_controller() [all …]
|
| H A D | core.c | 151 static void nvme_remove_invalid_namespaces(struct nvme_ctrl *ctrl, 153 static void nvme_update_keep_alive(struct nvme_ctrl *ctrl, 155 static int nvme_get_log_lsi(struct nvme_ctrl *ctrl, u32 nsid, u8 log_page, 158 void nvme_queue_scan(struct nvme_ctrl *ctrl) in nvme_queue_scan() argument 163 if (nvme_ctrl_state(ctrl) == NVME_CTRL_LIVE && ctrl->tagset) in nvme_queue_scan() 164 queue_work(nvme_wq, &ctrl->scan_work); in nvme_queue_scan() 173 int nvme_try_sched_reset(struct nvme_ctrl *ctrl) in nvme_try_sched_reset() argument 175 if (nvme_ctrl_state(ctrl) != NVME_CTRL_RESETTING) in nvme_try_sched_reset() 177 if (!queue_work(nvme_reset_wq, &ctrl in nvme_try_sched_reset() 185 struct nvme_ctrl *ctrl = container_of(to_delayed_work(work), nvme_failfast_work() local 196 nvme_start_failfast_work(struct nvme_ctrl * ctrl) nvme_start_failfast_work() argument 205 nvme_stop_failfast_work(struct nvme_ctrl * ctrl) nvme_stop_failfast_work() argument 215 nvme_reset_ctrl(struct nvme_ctrl * ctrl) nvme_reset_ctrl() argument 225 nvme_reset_ctrl_sync(struct nvme_ctrl * ctrl) nvme_reset_ctrl_sync() argument 239 nvme_do_delete_ctrl(struct nvme_ctrl * ctrl) nvme_do_delete_ctrl() argument 253 struct nvme_ctrl *ctrl = nvme_delete_ctrl_work() local 259 nvme_delete_ctrl(struct nvme_ctrl * ctrl) nvme_delete_ctrl() argument 269 nvme_delete_ctrl_sync(struct nvme_ctrl * ctrl) nvme_delete_ctrl_sync() argument 459 struct nvme_ctrl *ctrl = nvme_req(req)->ctrl; nvme_complete_rq() local 537 nvme_cancel_tagset(struct nvme_ctrl * ctrl) nvme_cancel_tagset() argument 547 nvme_cancel_admin_tagset(struct nvme_ctrl * ctrl) nvme_cancel_admin_tagset() argument 557 nvme_change_ctrl_state(struct nvme_ctrl * ctrl,enum nvme_ctrl_state new_state) nvme_change_ctrl_state() argument 656 nvme_wait_reset(struct nvme_ctrl * ctrl) nvme_wait_reset() argument 756 nvme_fail_nonready_command(struct nvme_ctrl * ctrl,struct request * rq) nvme_fail_nonready_command() argument 775 __nvme_check_ready(struct nvme_ctrl * ctrl,struct request * rq,bool queue_live,enum nvme_ctrl_state state) __nvme_check_ready() argument 1070 struct nvme_ctrl *ctrl = nvme_req(req)->ctrl; nvme_cleanup_cmd() local 1206 nvme_command_effects(struct nvme_ctrl * ctrl,struct nvme_ns * ns,u8 opcode) nvme_command_effects() argument 1235 nvme_passthru_start(struct nvme_ctrl * ctrl,struct nvme_ns * ns,u8 opcode) nvme_passthru_start() argument 1255 nvme_passthru_end(struct nvme_ctrl * ctrl,struct nvme_ns * ns,u32 effects,struct nvme_command * cmd,int status) nvme_passthru_end() argument 1306 nvme_keep_alive_work_period(struct nvme_ctrl * ctrl) nvme_keep_alive_work_period() argument 1321 nvme_queue_keep_alive_work(struct nvme_ctrl * ctrl) nvme_queue_keep_alive_work() argument 1339 struct nvme_ctrl *ctrl = rq->end_io_data; nvme_keep_alive_end_io() local 1374 struct nvme_ctrl *ctrl = container_of(to_delayed_work(work), nvme_keep_alive_work() local 1405 nvme_start_keep_alive(struct nvme_ctrl * ctrl) nvme_start_keep_alive() argument 1413 nvme_stop_keep_alive(struct nvme_ctrl * ctrl) nvme_stop_keep_alive() argument 1422 nvme_update_keep_alive(struct nvme_ctrl * ctrl,struct nvme_command * cmd) nvme_update_keep_alive() argument 1437 nvme_id_cns_ok(struct nvme_ctrl * ctrl,u8 cns) nvme_id_cns_ok() argument 1485 nvme_process_ns_desc(struct nvme_ctrl * ctrl,struct nvme_ns_ids * ids,struct nvme_ns_id_desc * cur,bool * csi_seen) nvme_process_ns_desc() argument 1537 nvme_identify_ns_descs(struct nvme_ctrl * ctrl,struct nvme_ns_info * info) nvme_identify_ns_descs() argument 1591 nvme_identify_ns(struct nvme_ctrl * ctrl,unsigned nsid,struct nvme_id_ns ** id) nvme_identify_ns() argument 1615 nvme_ns_info_from_identify(struct nvme_ctrl * ctrl,struct nvme_ns_info * info) nvme_ns_info_from_identify() argument 1655 nvme_ns_info_from_id_cs_indep(struct nvme_ctrl * ctrl,struct nvme_ns_info * info) nvme_ns_info_from_id_cs_indep() argument 1720 nvme_set_queue_count(struct nvme_ctrl * ctrl,int * count) nvme_set_queue_count() argument 1758 nvme_enable_aen(struct nvme_ctrl * ctrl) nvme_enable_aen() argument 1888 struct nvme_ctrl *ctrl = ns->ctrl; nvme_config_discard() local 1914 nvme_identify_ns_nvm(struct nvme_ctrl * ctrl,unsigned int nsid,struct nvme_id_ns_nvm ** nvmp) nvme_identify_ns_nvm() argument 1966 nvme_configure_metadata(struct nvme_ctrl * ctrl,struct nvme_ns_head * head,struct nvme_id_ns * id,struct nvme_id_ns_nvm * nvm,struct nvme_ns_info * info) nvme_configure_metadata() argument 2063 nvme_max_drv_segments(struct nvme_ctrl * ctrl) nvme_max_drv_segments() argument 2068 nvme_set_ctrl_limits(struct nvme_ctrl * ctrl,struct queue_limits * lim,bool is_admin) nvme_set_ctrl_limits() argument 2140 struct nvme_ctrl *ctrl = ns->ctrl; nvme_set_chunk_sectors() local 2190 nvme_query_fdp_granularity(struct nvme_ctrl * ctrl,struct nvme_ns_info * info,u8 fdp_idx) nvme_query_fdp_granularity() argument 2266 struct nvme_ctrl *ctrl = ns->ctrl; nvme_query_fdp_info() local 2564 struct nvme_ctrl *ctrl = data; nvme_sec_submit() local 2579 nvme_configure_opal(struct nvme_ctrl * ctrl,bool was_suspended) nvme_configure_opal() argument 2592 nvme_configure_opal(struct nvme_ctrl * ctrl,bool was_suspended) nvme_configure_opal() argument 2619 nvme_wait_ready(struct nvme_ctrl * ctrl,u32 mask,u32 val,u32 timeout,const char * op) nvme_wait_ready() argument 2646 nvme_disable_ctrl(struct nvme_ctrl * ctrl,bool shutdown) nvme_disable_ctrl() argument 2672 nvme_enable_ctrl(struct nvme_ctrl * ctrl) nvme_enable_ctrl() argument 2751 nvme_configure_timestamp(struct nvme_ctrl * ctrl) nvme_configure_timestamp() argument 2768 nvme_configure_host_options(struct nvme_ctrl * ctrl) nvme_configure_host_options() argument 2849 nvme_configure_apst(struct nvme_ctrl * ctrl) nvme_configure_apst() argument 2956 struct nvme_ctrl *ctrl = dev_get_drvdata(dev); nvme_set_latency_tolerance() local 3067 nvme_init_subnqn(struct nvme_subsystem * subsys,struct nvme_ctrl * ctrl,struct nvme_id_ctrl * id) nvme_init_subnqn() argument 3156 nvme_discovery_ctrl(struct nvme_ctrl * ctrl) nvme_discovery_ctrl() argument 3161 nvme_admin_ctrl(struct nvme_ctrl * ctrl) nvme_admin_ctrl() argument 3166 nvme_is_io_ctrl(struct nvme_ctrl * ctrl) nvme_is_io_ctrl() argument 3172 nvme_validate_cntlid(struct nvme_subsystem * subsys,struct nvme_ctrl * ctrl,struct nvme_id_ctrl * id) nvme_validate_cntlid() argument 3202 nvme_init_subsystem(struct nvme_ctrl * ctrl,struct nvme_id_ctrl * id) nvme_init_subsystem() argument 3288 nvme_get_log_lsi(struct nvme_ctrl * ctrl,u32 nsid,u8 log_page,u8 lsp,u8 csi,void * log,size_t size,u64 offset,u16 lsi) nvme_get_log_lsi() argument 3308 nvme_get_log(struct nvme_ctrl * ctrl,u32 nsid,u8 log_page,u8 lsp,u8 csi,void * log,size_t size,u64 offset) nvme_get_log() argument 3315 nvme_get_effects_log(struct nvme_ctrl * ctrl,u8 csi,struct nvme_effects_log ** log) nvme_get_effects_log() argument 3345 nvme_mps_to_sectors(struct nvme_ctrl * ctrl,u32 units) nvme_mps_to_sectors() argument 3354 nvme_init_non_mdts_limits(struct nvme_ctrl * ctrl) nvme_init_non_mdts_limits() argument 3401 nvme_init_effects_log(struct nvme_ctrl * ctrl,u8 csi,struct nvme_effects_log ** log) nvme_init_effects_log() argument 3420 nvme_init_known_nvm_effects(struct nvme_ctrl * ctrl) nvme_init_known_nvm_effects() argument 3452 nvme_init_effects(struct nvme_ctrl * ctrl,struct nvme_id_ctrl * id) nvme_init_effects() argument 3475 nvme_check_ctrl_fabric_info(struct nvme_ctrl * ctrl,struct nvme_id_ctrl * id) nvme_check_ctrl_fabric_info() argument 3517 nvme_init_identify(struct nvme_ctrl * ctrl) nvme_init_identify() argument 3664 nvme_init_ctrl_finish(struct nvme_ctrl * ctrl,bool was_suspended) nvme_init_ctrl_finish() argument 3729 struct nvme_ctrl *ctrl = nvme_dev_open() local 3751 struct nvme_ctrl *ctrl = nvme_dev_release() local 3768 nvme_find_ns_head(struct nvme_ctrl * ctrl,unsigned nsid) nvme_find_ns_head() argument 3881 nvme_alloc_ns_head(struct nvme_ctrl * ctrl,struct nvme_ns_info * info) nvme_alloc_ns_head() argument 3968 struct nvme_ctrl *ctrl = ns->ctrl; nvme_init_ns_head() local 4066 nvme_find_get_ns(struct nvme_ctrl * ctrl,unsigned nsid) nvme_find_get_ns() argument 4104 nvme_alloc_ns(struct nvme_ctrl * ctrl,struct nvme_ns_info * info) nvme_alloc_ns() argument 4269 nvme_ns_remove_by_nsid(struct nvme_ctrl * ctrl,u32 nsid) nvme_ns_remove_by_nsid() argument 4301 nvme_scan_ns(struct nvme_ctrl * ctrl,unsigned nsid) nvme_scan_ns() argument 4359 struct nvme_ctrl *ctrl; global() member 4376 nvme_remove_invalid_namespaces(struct nvme_ctrl * ctrl,unsigned nsid) nvme_remove_invalid_namespaces() argument 4396 nvme_scan_ns_list(struct nvme_ctrl * ctrl) nvme_scan_ns_list() argument 4447 nvme_scan_ns_sequential(struct nvme_ctrl * ctrl) nvme_scan_ns_sequential() argument 4463 nvme_clear_changed_ns_log(struct nvme_ctrl * ctrl) nvme_clear_changed_ns_log() argument 4490 struct nvme_ctrl *ctrl = nvme_scan_work() local 4547 nvme_remove_namespaces(struct nvme_ctrl * ctrl) nvme_remove_namespaces() argument 4592 const struct nvme_ctrl *ctrl = nvme_class_uevent() local 4622 nvme_change_uevent(struct nvme_ctrl * ctrl,char * envdata) nvme_change_uevent() argument 4629 nvme_aen_uevent(struct nvme_ctrl * ctrl) nvme_aen_uevent() argument 4647 struct nvme_ctrl *ctrl = nvme_async_event_work() local 4661 nvme_ctrl_pp_status(struct nvme_ctrl * ctrl) nvme_ctrl_pp_status() argument 4675 nvme_get_fw_slot_info(struct nvme_ctrl * ctrl) nvme_get_fw_slot_info() argument 4707 struct nvme_ctrl *ctrl = container_of(work, nvme_fw_act_work() local 4750 nvme_handle_aen_notice(struct nvme_ctrl * ctrl,u32 result) nvme_handle_aen_notice() argument 4787 nvme_handle_aer_persistent_error(struct nvme_ctrl * ctrl) nvme_handle_aer_persistent_error() argument 4794 nvme_complete_async_event(struct nvme_ctrl * ctrl,__le16 status,volatile union nvme_result * res) nvme_complete_async_event() argument 4834 nvme_alloc_admin_tag_set(struct nvme_ctrl * ctrl,struct blk_mq_tag_set * set,const struct blk_mq_ops * ops,unsigned int cmd_size) nvme_alloc_admin_tag_set() argument 4891 nvme_remove_admin_tag_set(struct nvme_ctrl * ctrl) nvme_remove_admin_tag_set() argument 4907 nvme_alloc_io_tag_set(struct nvme_ctrl * ctrl,struct blk_mq_tag_set * set,const struct blk_mq_ops * ops,unsigned int nr_maps,unsigned int cmd_size) nvme_alloc_io_tag_set() argument 4959 nvme_remove_io_tag_set(struct nvme_ctrl * ctrl) nvme_remove_io_tag_set() argument 4969 nvme_stop_ctrl(struct nvme_ctrl * ctrl) nvme_stop_ctrl() argument 4981 nvme_start_ctrl(struct nvme_ctrl * ctrl) nvme_start_ctrl() argument 5012 nvme_uninit_ctrl(struct nvme_ctrl * ctrl) nvme_uninit_ctrl() argument 5023 nvme_free_cels(struct nvme_ctrl * ctrl) nvme_free_cels() argument 5038 struct nvme_ctrl *ctrl = nvme_free_ctrl() local 5075 nvme_init_ctrl(struct nvme_ctrl * ctrl,struct device * dev,const struct nvme_ctrl_ops * ops,unsigned long quirks) nvme_init_ctrl() argument 5157 nvme_add_ctrl(struct nvme_ctrl * ctrl) nvme_add_ctrl() argument 5187 nvme_mark_namespaces_dead(struct nvme_ctrl * ctrl) nvme_mark_namespaces_dead() argument 5200 nvme_unfreeze(struct nvme_ctrl * ctrl) nvme_unfreeze() argument 5214 nvme_wait_freeze_timeout(struct nvme_ctrl * ctrl,long timeout) nvme_wait_freeze_timeout() argument 5231 nvme_wait_freeze(struct nvme_ctrl * ctrl) nvme_wait_freeze() argument 5244 nvme_start_freeze(struct nvme_ctrl * ctrl) nvme_start_freeze() argument 5263 nvme_quiesce_io_queues(struct nvme_ctrl * ctrl) nvme_quiesce_io_queues() argument 5274 nvme_unquiesce_io_queues(struct nvme_ctrl * ctrl) nvme_unquiesce_io_queues() argument 5283 nvme_quiesce_admin_queue(struct nvme_ctrl * ctrl) nvme_quiesce_admin_queue() argument 5292 nvme_unquiesce_admin_queue(struct nvme_ctrl * ctrl) nvme_unquiesce_admin_queue() argument 5299 nvme_sync_io_queues(struct nvme_ctrl * ctrl) nvme_sync_io_queues() argument 5312 nvme_sync_queues(struct nvme_ctrl * ctrl) nvme_sync_queues() argument [all...] |
| H A D | tcp.c | 165 struct nvme_tcp_ctrl *ctrl; member 195 struct nvme_ctrl ctrl; member 210 static inline struct nvme_tcp_ctrl *to_tcp_ctrl(struct nvme_ctrl *ctrl) in to_tcp_ctrl() argument 212 return container_of(ctrl, struct nvme_tcp_ctrl, ctrl); in to_tcp_ctrl() 217 return queue - queue->ctrl->queues; in nvme_tcp_queue_id() 247 static inline bool nvme_tcp_tls_configured(struct nvme_ctrl *ctrl) in nvme_tcp_tls_configured() argument 252 return ctrl->opts->tls || ctrl->opts->concat; in nvme_tcp_tls_configured() 260 return queue->ctrl->admin_tag_set.tags[queue_idx]; in nvme_tcp_tagset() 261 return queue->ctrl->tag_set.tags[queue_idx - 1]; in nvme_tcp_tagset() 295 return req == &req->queue->ctrl->async_req; in nvme_tcp_async_req() [all …]
|
| H A D | auth.c | 23 struct nvme_ctrl *ctrl; 52 static inline int ctrl_max_dhchaps(struct nvme_ctrl *ctrl) 54 return ctrl->opts->nr_io_queues + ctrl->opts->nr_write_queues + in ctrl_max_dhchaps() argument 55 ctrl->opts->nr_poll_queues + 1; in ctrl_max_dhchaps() 58 static int nvme_auth_submit(struct nvme_ctrl *ctrl, int qid, in ctrl_max_dhchaps() 63 struct request_queue *q = ctrl->fabrics_q; in nvme_auth_submit() 68 q = ctrl->connect_q; in nvme_auth_submit() 86 dev_warn(ctrl->device, in nvme_auth_submit() 89 dev_err(ctrl in nvme_auth_submit() 24 struct nvme_ctrl *ctrl; global() member 60 nvme_auth_submit(struct nvme_ctrl * ctrl,int qid,void * data,size_t data_len,bool auth_send) nvme_auth_submit() argument 96 nvme_auth_receive_validate(struct nvme_ctrl * ctrl,int qid,struct nvmf_auth_dhchap_failure_data * data,u16 transaction,u8 expected_msg) nvme_auth_receive_validate() argument 123 nvme_auth_set_dhchap_negotiate_data(struct nvme_ctrl * ctrl,struct nvme_dhchap_queue_context * chap) nvme_auth_set_dhchap_negotiate_data() argument 163 nvme_auth_process_dhchap_challenge(struct nvme_ctrl * ctrl,struct nvme_dhchap_queue_context * chap) nvme_auth_process_dhchap_challenge() argument 302 nvme_auth_set_dhchap_reply_data(struct nvme_ctrl * ctrl,struct nvme_dhchap_queue_context * chap) nvme_auth_set_dhchap_reply_data() argument 354 nvme_auth_process_dhchap_success1(struct nvme_ctrl * ctrl,struct nvme_dhchap_queue_context * chap) nvme_auth_process_dhchap_success1() argument 404 nvme_auth_set_dhchap_success2_data(struct nvme_ctrl * ctrl,struct nvme_dhchap_queue_context * chap) nvme_auth_set_dhchap_success2_data() argument 418 nvme_auth_set_dhchap_failure2_data(struct nvme_ctrl * ctrl,struct nvme_dhchap_queue_context * chap) nvme_auth_set_dhchap_failure2_data() argument 434 nvme_auth_dhchap_setup_host_response(struct nvme_ctrl * ctrl,struct nvme_dhchap_queue_context * chap) nvme_auth_dhchap_setup_host_response() argument 521 nvme_auth_dhchap_setup_ctrl_response(struct nvme_ctrl * ctrl,struct nvme_dhchap_queue_context * chap) nvme_auth_dhchap_setup_ctrl_response() argument 607 nvme_auth_dhchap_exponential(struct nvme_ctrl * ctrl,struct nvme_dhchap_queue_context * chap) nvme_auth_dhchap_exponential() argument 698 nvme_auth_revoke_tls_key(struct nvme_ctrl * ctrl) nvme_auth_revoke_tls_key() argument 708 nvme_auth_secure_concat(struct nvme_ctrl * ctrl,struct nvme_dhchap_queue_context * chap) nvme_auth_secure_concat() argument 788 struct nvme_ctrl *ctrl = chap->ctrl; nvme_queue_auth_work() local 967 nvme_auth_negotiate(struct nvme_ctrl * ctrl,int qid) nvme_auth_negotiate() argument 988 nvme_auth_wait(struct nvme_ctrl * ctrl,int qid) nvme_auth_wait() argument 1004 struct nvme_ctrl *ctrl = nvme_ctrl_auth_work() local 1065 nvme_auth_init_ctrl(struct nvme_ctrl * ctrl) nvme_auth_init_ctrl() argument 1111 nvme_auth_stop(struct nvme_ctrl * ctrl) nvme_auth_stop() argument 1117 nvme_auth_free(struct nvme_ctrl * ctrl) nvme_auth_free() argument [all...] |
| /linux/drivers/gpu/drm/msm/dp/ |
| H A D | dp_ctrl.c | 145 static inline u32 msm_dp_read_ahb(const struct msm_dp_ctrl_private *ctrl, u32 offset) in msm_dp_read_ahb() argument 147 return readl_relaxed(ctrl->ahb_base + offset); in msm_dp_read_ahb() 150 static inline void msm_dp_write_ahb(struct msm_dp_ctrl_private *ctrl, in msm_dp_write_ahb() argument 157 writel(data, ctrl->ahb_base + offset); in msm_dp_write_ahb() 160 static inline u32 msm_dp_read_link(struct msm_dp_ctrl_private *ctrl, u32 offset) in msm_dp_read_link() argument 162 return readl_relaxed(ctrl->link_base + offset); in msm_dp_read_link() 165 static inline void msm_dp_write_link(struct msm_dp_ctrl_private *ctrl, in msm_dp_write_link() argument 172 writel(data, ctrl->link_base + offset); in msm_dp_write_link() 207 struct msm_dp_ctrl_private *ctrl = in msm_dp_ctrl_reset() local 211 sw_reset = msm_dp_read_ahb(ctrl, REG_DP_SW_RESET); in msm_dp_ctrl_reset() [all …]
|
| /linux/drivers/pci/hotplug/ |
| H A D | pciehp_hpc.c | 49 static inline struct pci_dev *ctrl_dev(struct controller *ctrl) in ctrl_dev() argument 51 return ctrl->pcie->port; in ctrl_dev() 58 static inline int pciehp_request_irq(struct controller *ctrl) in pciehp_request_irq() argument 60 int retval, irq = ctrl->pcie->irq; in pciehp_request_irq() 63 ctrl->poll_thread = kthread_run(&pciehp_poll, ctrl, in pciehp_request_irq() 65 slot_name(ctrl)); in pciehp_request_irq() 66 return PTR_ERR_OR_ZERO(ctrl->poll_thread); in pciehp_request_irq() 71 IRQF_SHARED, "pciehp", ctrl); in pciehp_request_irq() 73 ctrl_err(ctrl, "Cannot get irq %d for the hotplug controller\n", in pciehp_request_irq() 78 static inline void pciehp_free_irq(struct controller *ctrl) in pciehp_free_irq() argument [all …]
|
| H A D | shpchp_hpc.c | 169 static void start_int_poll_timer(struct controller *ctrl, int sec); 171 static inline u8 shpc_readb(struct controller *ctrl, int reg) in shpc_readb() argument 173 return readb(ctrl->creg + reg); in shpc_readb() 176 static inline u16 shpc_readw(struct controller *ctrl, int reg) in shpc_readw() argument 178 return readw(ctrl->creg + reg); in shpc_readw() 181 static inline void shpc_writew(struct controller *ctrl, int reg, u16 val) in shpc_writew() argument 183 writew(val, ctrl->creg + reg); in shpc_writew() 186 static inline u32 shpc_readl(struct controller *ctrl, int reg) in shpc_readl() argument 188 return readl(ctrl->creg + reg); in shpc_readl() 191 static inline void shpc_writel(struct controller *ctrl, int reg, u32 val) in shpc_writel() argument [all …]
|
| H A D | cpqphp_core.c | 116 * @ctrl: controller to use 120 static int init_SERR(struct controller *ctrl) in init_SERR() argument 125 if (!ctrl) in init_SERR() 128 tempdword = ctrl->first_slot; in init_SERR() 130 number_of_slots = readb(ctrl->hpc_reg + SLOT_MASK) & 0x0F; in init_SERR() 133 writeb(0, ctrl->hpc_reg + SLOT_SERR); in init_SERR() 266 static int ctrl_slot_cleanup(struct controller *ctrl) in ctrl_slot_cleanup() argument 270 old_slot = ctrl->slot; in ctrl_slot_cleanup() 271 ctrl->slot = NULL; in ctrl_slot_cleanup() 280 cpqhp_remove_debugfs_files(ctrl); in ctrl_slot_cleanup() [all …]
|
| H A D | pciehp_core.c | 51 static int init_slot(struct controller *ctrl) in init_slot() argument 67 if (MRL_SENS(ctrl)) in init_slot() 69 if (ATTN_LED(ctrl)) { in init_slot() 72 } else if (ctrl->pcie->port->hotplug_user_indicators) { in init_slot() 78 ctrl->hotplug_slot.ops = ops; in init_slot() 79 snprintf(name, SLOT_NAME_SIZE, "%u", PSN(ctrl)); in init_slot() 81 retval = pci_hp_initialize(&ctrl->hotplug_slot, in init_slot() 82 ctrl->pcie->port->subordinate, 0, name); in init_slot() 84 ctrl_err(ctrl, "pci_hp_initialize failed: error %d\n", retval); in init_slot() 90 static void cleanup_slot(struct controller *ctrl) in cleanup_slot() argument [all …]
|
| H A D | shpchp_ctrl.c | 45 u8 shpchp_handle_attention_button(u8 hp_slot, struct controller *ctrl) in shpchp_handle_attention_button() argument 51 ctrl_dbg(ctrl, "Attention button interrupt received\n"); in shpchp_handle_attention_button() 53 p_slot = shpchp_find_slot(ctrl, hp_slot + ctrl->slot_device_offset); in shpchp_handle_attention_button() 59 ctrl_info(ctrl, "Button pressed on Slot(%s)\n", slot_name(p_slot)); in shpchp_handle_attention_button() 68 u8 shpchp_handle_switch_change(u8 hp_slot, struct controller *ctrl) in shpchp_handle_switch_change() argument 75 ctrl_dbg(ctrl, "Switch interrupt received\n"); in shpchp_handle_switch_change() 77 p_slot = shpchp_find_slot(ctrl, hp_slot + ctrl->slot_device_offset); in shpchp_handle_switch_change() 80 ctrl_dbg(ctrl, "Card present %x Power status %x\n", in shpchp_handle_switch_change() 87 ctrl_info(ctrl, "Latch open on Slot(%s)\n", slot_name(p_slot)); in shpchp_handle_switch_change() 91 ctrl_err(ctrl, "Surprise Removal of card\n"); in shpchp_handle_switch_change() [all …]
|
| /linux/drivers/media/v4l2-core/ |
| H A D | v4l2-ctrls-api.c | 41 struct v4l2_ctrl *ctrl, in ptr_to_user() argument 46 if (ctrl->is_ptr && !ctrl->is_string) in ptr_to_user() 50 switch (ctrl->type) { in ptr_to_user() 54 c->size = ctrl->elem_size; in ptr_to_user() 70 static int cur_to_user(struct v4l2_ext_control *c, struct v4l2_ctrl *ctrl) in cur_to_user() argument 72 return ptr_to_user(c, ctrl, ctrl->p_cur); in cur_to_user() 77 struct v4l2_ctrl *ctrl) in new_to_user() argument 79 return ptr_to_user(c, ctrl, ctrl->p_new); in new_to_user() 86 return ptr_to_user(c, ref->ctrl, ref->p_req); in req_to_user() 90 static int def_to_user(struct v4l2_ext_control *c, struct v4l2_ctrl *ctrl) in def_to_user() argument [all …]
|
| H A D | v4l2-ctrls-core.c | 19 static void fill_event(struct v4l2_event *ev, struct v4l2_ctrl *ctrl, in fill_event() argument 24 ev->id = ctrl->id; in fill_event() 25 ev->u.ctrl.changes = changes; in fill_event() 26 ev->u.ctrl.type = ctrl->type; in fill_event() 27 ev->u.ctrl.flags = user_flags(ctrl); in fill_event() 28 if (ctrl->is_ptr) in fill_event() 29 ev->u.ctrl.value64 = 0; in fill_event() 31 ev->u.ctrl.value64 = *ctrl->p_cur.p_s64; in fill_event() 32 ev->u.ctrl.minimum = ctrl->minimum; in fill_event() 33 ev->u.ctrl.maximum = ctrl->maximum; in fill_event() [all …]
|
| /linux/drivers/mtd/nand/raw/brcmnand/ |
| H A D | brcmnand.c | 346 struct brcmnand_controller *ctrl; member 666 static inline bool brcmnand_non_mmio_ops(struct brcmnand_controller *ctrl) in brcmnand_non_mmio_ops() argument 675 static inline u32 nand_readreg(struct brcmnand_controller *ctrl, u32 offs) in nand_readreg() argument 677 if (brcmnand_non_mmio_ops(ctrl)) in nand_readreg() 678 return brcmnand_soc_read(ctrl->soc, offs); in nand_readreg() 679 return brcmnand_readl(ctrl->nand_base + offs); in nand_readreg() 682 static inline void nand_writereg(struct brcmnand_controller *ctrl, u32 offs, in nand_writereg() argument 685 if (brcmnand_non_mmio_ops(ctrl)) in nand_writereg() 686 brcmnand_soc_write(ctrl->soc, val, offs); in nand_writereg() 688 brcmnand_writel(val, ctrl->nand_base + offs); in nand_writereg() [all …]
|
| /linux/drivers/tty/serdev/ |
| H A D | core.c | 79 struct serdev_controller *ctrl = to_serdev_controller(dev); in serdev_ctrl_release() local 80 ida_free(&ctrl_ida, ctrl->nr); in serdev_ctrl_release() 81 kfree(ctrl); in serdev_ctrl_release() 106 struct serdev_controller *ctrl = serdev->ctrl; in serdev_device_add() local 113 if (ctrl->serdev) { in serdev_device_add() 117 ctrl->serdev = serdev; in serdev_device_add() 130 ctrl->serdev = NULL; in serdev_device_add() 141 struct serdev_controller *ctrl = serdev->ctrl; in serdev_device_remove() local 144 ctrl->serdev = NULL; in serdev_device_remove() 150 struct serdev_controller *ctrl = serdev->ctrl; in serdev_device_open() local [all …]
|
| /linux/drivers/net/mdio/ |
| H A D | mdio-bitbang.c | 44 static void mdiobb_send_bit(struct mdiobb_ctrl *ctrl, int val) in mdiobb_send_bit() argument 46 const struct mdiobb_ops *ops = ctrl->ops; in mdiobb_send_bit() 48 ops->set_mdio_data(ctrl, val); in mdiobb_send_bit() 50 ops->set_mdc(ctrl, 1); in mdiobb_send_bit() 52 ops->set_mdc(ctrl, 0); in mdiobb_send_bit() 56 static int mdiobb_get_bit(struct mdiobb_ctrl *ctrl) in mdiobb_get_bit() argument 58 const struct mdiobb_ops *ops = ctrl->ops; in mdiobb_get_bit() 61 ops->set_mdc(ctrl, 1); in mdiobb_get_bit() 63 ops->set_mdc(ctrl, 0); in mdiobb_get_bit() 65 return ops->get_mdio_data(ctrl); in mdiobb_get_bit() [all …]
|
| /linux/drivers/clk/bcm/ |
| H A D | clk-iproc-pll.c | 68 const struct iproc_pll_ctrl *ctrl; member 76 const struct iproc_clk_ctrl *ctrl; member 150 const struct iproc_pll_ctrl *ctrl = pll->ctrl; in pll_wait_for_lock() local 153 u32 val = readl(pll->status_base + ctrl->status.offset); in pll_wait_for_lock() 155 if (val & (1 << ctrl->status.shift)) in pll_wait_for_lock() 166 const struct iproc_pll_ctrl *ctrl = pll->ctrl; in iproc_pll_write() local 170 if (unlikely(ctrl->flags & IPROC_CLK_NEEDS_READ_BACK && in iproc_pll_write() 177 const struct iproc_pll_ctrl *ctrl = pll->ctrl; in __pll_disable() local 180 if (ctrl->flags & IPROC_CLK_PLL_ASIU) { in __pll_disable() 181 val = readl(pll->asiu_base + ctrl->asiu.offset); in __pll_disable() [all …]
|
| /linux/drivers/watchdog/ |
| H A D | realtek_otto_wdt.c | 76 struct otto_wdt_ctrl *ctrl = watchdog_get_drvdata(wdev); in otto_wdt_start() local 79 v = ioread32(ctrl->base + OTTO_WDT_REG_CTRL); in otto_wdt_start() 81 iowrite32(v, ctrl->base + OTTO_WDT_REG_CTRL); in otto_wdt_start() 88 struct otto_wdt_ctrl *ctrl = watchdog_get_drvdata(wdev); in otto_wdt_stop() local 91 v = ioread32(ctrl->base + OTTO_WDT_REG_CTRL); in otto_wdt_stop() 93 iowrite32(v, ctrl->base + OTTO_WDT_REG_CTRL); in otto_wdt_stop() 100 struct otto_wdt_ctrl *ctrl = watchdog_get_drvdata(wdev); in otto_wdt_ping() local 102 iowrite32(OTTO_WDT_CNTR_PING, ctrl->base + OTTO_WDT_REG_CNTR); in otto_wdt_ping() 107 static int otto_wdt_tick_ms(struct otto_wdt_ctrl *ctrl, int prescale) in otto_wdt_tick_ms() argument 109 return DIV_ROUND_CLOSEST(1 << (25 + prescale), ctrl->clk_rate_khz); in otto_wdt_tick_ms() [all …]
|