Home
last modified time | relevance | path

Searched full:ctrl (Results 1 – 25 of 1980) sorted by relevance

12345678910>>...80

/linux/drivers/soundwire/
H A Dqcom.c220 int (*reg_read)(struct qcom_swrm_ctrl *ctrl, int reg, u32 *val);
221 int (*reg_write)(struct qcom_swrm_ctrl *ctrl, int reg, int val);
331 static int qcom_swrm_ahb_reg_read(struct qcom_swrm_ctrl *ctrl, int reg, in qcom_swrm_ahb_reg_read() argument
334 struct regmap *wcd_regmap = ctrl->regmap; in qcom_swrm_ahb_reg_read()
351 static int qcom_swrm_ahb_reg_write(struct qcom_swrm_ctrl *ctrl, in qcom_swrm_ahb_reg_write() argument
354 struct regmap *wcd_regmap = ctrl->regmap; in qcom_swrm_ahb_reg_write()
371 static int qcom_swrm_cpu_reg_read(struct qcom_swrm_ctrl *ctrl, int reg, in qcom_swrm_cpu_reg_read() argument
374 *val = readl(ctrl->mmio + reg); in qcom_swrm_cpu_reg_read()
378 static int qcom_swrm_cpu_reg_write(struct qcom_swrm_ctrl *ctrl, int reg, in qcom_swrm_cpu_reg_write() argument
381 writel(val, ctrl->mmio + reg); in qcom_swrm_cpu_reg_write()
[all …]
/linux/drivers/slimbus/
H A Dqcom-ngd-ctrl.c135 struct qcom_slim_ngd_ctrl *ctrl; member
150 struct slim_controller ctrl; member
347 static int qcom_slim_qmi_send_select_inst_req(struct qcom_slim_ngd_ctrl *ctrl, in qcom_slim_qmi_send_select_inst_req() argument
354 rc = qmi_txn_init(ctrl->qmi.handle, &txn, in qcom_slim_qmi_send_select_inst_req()
357 dev_err(ctrl->dev, "QMI TXN init fail: %d\n", rc); in qcom_slim_qmi_send_select_inst_req()
361 rc = qmi_send_request(ctrl->qmi.handle, NULL, &txn, in qcom_slim_qmi_send_select_inst_req()
366 dev_err(ctrl->dev, "QMI send req fail %d\n", rc); in qcom_slim_qmi_send_select_inst_req()
373 dev_err(ctrl->dev, "QMI TXN wait fail: %d\n", rc); in qcom_slim_qmi_send_select_inst_req()
378 dev_err(ctrl->dev, "QMI request failed 0x%x\n", in qcom_slim_qmi_send_select_inst_req()
400 static int qcom_slim_qmi_send_power_request(struct qcom_slim_ngd_ctrl *ctrl, in qcom_slim_qmi_send_power_request() argument
[all …]
H A Dcore.c153 static int slim_add_device(struct slim_controller *ctrl, in slim_add_device() argument
158 sbdev->dev.parent = ctrl->dev; in slim_add_device()
161 sbdev->ctrl = ctrl; in slim_add_device()
176 static struct slim_device *slim_alloc_device(struct slim_controller *ctrl, in slim_alloc_device() argument
188 ret = slim_add_device(ctrl, sbdev, node); in slim_alloc_device()
197 static void of_register_slim_devices(struct slim_controller *ctrl) in of_register_slim_devices() argument
199 struct device *dev = ctrl->dev; in of_register_slim_devices()
202 if (!ctrl->dev->of_node) in of_register_slim_devices()
205 for_each_child_of_node(ctrl->dev->of_node, node) { in of_register_slim_devices()
235 sbdev = slim_alloc_device(ctrl, &e_addr, node); in of_register_slim_devices()
[all …]
/linux/drivers/nvme/host/
H A Drdma.c89 struct nvme_rdma_ctrl *ctrl; member
125 struct nvme_ctrl ctrl; member
130 static inline struct nvme_rdma_ctrl *to_rdma_ctrl(struct nvme_ctrl *ctrl) in to_rdma_ctrl() argument
132 return container_of(ctrl, struct nvme_rdma_ctrl, ctrl); in to_rdma_ctrl()
161 return queue - queue->ctrl->queues; in nvme_rdma_queue_idx()
167 queue->ctrl->io_queues[HCTX_TYPE_DEFAULT] + in nvme_rdma_poll_queue()
168 queue->ctrl->io_queues[HCTX_TYPE_READ]; in nvme_rdma_poll_queue()
297 struct nvme_rdma_ctrl *ctrl = to_rdma_ctrl(set->driver_data); in nvme_rdma_init_request() local
299 int queue_idx = (set == &ctrl in nvme_rdma_init_request()
322 struct nvme_rdma_ctrl *ctrl = to_rdma_ctrl(data); nvme_rdma_init_hctx() local
334 struct nvme_rdma_ctrl *ctrl = to_rdma_ctrl(data); nvme_rdma_init_admin_hctx() local
569 nvme_rdma_alloc_queue(struct nvme_rdma_ctrl * ctrl,int idx,size_t queue_size) nvme_rdma_alloc_queue() argument
660 nvme_rdma_free_io_queues(struct nvme_rdma_ctrl * ctrl) nvme_rdma_free_io_queues() argument
668 nvme_rdma_stop_io_queues(struct nvme_rdma_ctrl * ctrl) nvme_rdma_stop_io_queues() argument
676 nvme_rdma_start_queue(struct nvme_rdma_ctrl * ctrl,int idx) nvme_rdma_start_queue() argument
697 nvme_rdma_start_io_queues(struct nvme_rdma_ctrl * ctrl,int first,int last) nvme_rdma_start_io_queues() argument
716 nvme_rdma_alloc_io_queues(struct nvme_rdma_ctrl * ctrl) nvme_rdma_alloc_io_queues() argument
754 nvme_rdma_alloc_tag_set(struct nvme_ctrl * ctrl) nvme_rdma_alloc_tag_set() argument
769 nvme_rdma_destroy_admin_queue(struct nvme_rdma_ctrl * ctrl) nvme_rdma_destroy_admin_queue() argument
780 nvme_rdma_configure_admin_queue(struct nvme_rdma_ctrl * ctrl,bool new) nvme_rdma_configure_admin_queue() argument
864 nvme_rdma_configure_io_queues(struct nvme_rdma_ctrl * ctrl,bool new) nvme_rdma_configure_io_queues() argument
930 nvme_rdma_teardown_admin_queue(struct nvme_rdma_ctrl * ctrl,bool remove) nvme_rdma_teardown_admin_queue() argument
944 nvme_rdma_teardown_io_queues(struct nvme_rdma_ctrl * ctrl,bool remove) nvme_rdma_teardown_io_queues() argument
962 struct nvme_rdma_ctrl *ctrl = to_rdma_ctrl(nctrl); nvme_rdma_stop_ctrl() local
970 struct nvme_rdma_ctrl *ctrl = to_rdma_ctrl(nctrl); nvme_rdma_free_ctrl() local
985 nvme_rdma_reconnect_or_remove(struct nvme_rdma_ctrl * ctrl,int status) nvme_rdma_reconnect_or_remove() argument
1006 nvme_rdma_setup_ctrl(struct nvme_rdma_ctrl * ctrl,bool new) nvme_rdma_setup_ctrl() argument
1100 struct nvme_rdma_ctrl *ctrl = container_of(to_delayed_work(work), nvme_rdma_reconnect_ctrl_work() local
1125 struct nvme_rdma_ctrl *ctrl = container_of(work, nvme_rdma_error_recovery_work() local
1148 nvme_rdma_error_recovery(struct nvme_rdma_ctrl * ctrl) nvme_rdma_error_recovery() argument
1171 struct nvme_rdma_ctrl *ctrl = queue->ctrl; nvme_rdma_wr_error() local
1665 struct nvme_rdma_ctrl *ctrl = to_rdma_ctrl(arg); nvme_rdma_submit_async_event() local
1815 struct nvme_ctrl *ctrl = &queue->ctrl->ctrl; nvme_rdma_addr_resolved() local
1840 struct nvme_rdma_ctrl *ctrl = queue->ctrl; nvme_rdma_route_resolved() local
1957 struct nvme_rdma_ctrl *ctrl = queue->ctrl; nvme_rdma_timeout() local
2128 struct nvme_rdma_ctrl *ctrl = to_rdma_ctrl(set->driver_data); nvme_rdma_map_queues() local
2153 nvme_rdma_shutdown_ctrl(struct nvme_rdma_ctrl * ctrl,bool shutdown) nvme_rdma_shutdown_ctrl() argument
2161 nvme_rdma_delete_ctrl(struct nvme_ctrl * ctrl) nvme_rdma_delete_ctrl() argument
2168 struct nvme_rdma_ctrl *ctrl = nvme_rdma_reset_ctrl_work() local
2223 struct nvme_rdma_ctrl *ctrl; nvme_rdma_existing_controller() local
2240 struct nvme_rdma_ctrl *ctrl; nvme_rdma_alloc_ctrl() local
2314 struct nvme_rdma_ctrl *ctrl; nvme_rdma_create_ctrl() local
2364 struct nvme_rdma_ctrl *ctrl; nvme_rdma_remove_one() local
2418 struct nvme_rdma_ctrl *ctrl; nvme_rdma_cleanup_module() local
[all...]
H A Dfc.c35 struct nvme_fc_ctrl *ctrl; member
99 struct nvme_fc_ctrl *ctrl; member
179 struct nvme_ctrl ctrl; member
183 to_fc_ctrl(struct nvme_ctrl *ctrl) in to_fc_ctrl() argument
185 return container_of(ctrl, struct nvme_fc_ctrl, ctrl); in to_fc_ctrl()
552 nvme_fc_resume_controller(struct nvme_fc_ctrl *ctrl) in nvme_fc_resume_controller() argument
554 switch (nvme_ctrl_state(&ctrl->ctrl)) { in nvme_fc_resume_controller()
561 dev_info(ctrl in nvme_fc_resume_controller()
587 struct nvme_fc_ctrl *ctrl; nvme_fc_attach_to_suspended_rport() local
784 nvme_fc_ctrl_connectivity_loss(struct nvme_fc_ctrl * ctrl) nvme_fc_ctrl_connectivity_loss() argument
809 struct nvme_fc_ctrl *ctrl; nvme_fc_unregister_remoteport() local
1130 nvme_fc_connect_admin_queue(struct nvme_fc_ctrl * ctrl,struct nvme_fc_queue * queue,u16 qsize,u16 ersp_ratio) nvme_fc_connect_admin_queue() argument
1247 nvme_fc_connect_queue(struct nvme_fc_ctrl * ctrl,struct nvme_fc_queue * queue,u16 qsize,u16 ersp_ratio) nvme_fc_connect_queue() argument
1378 nvme_fc_xmt_disconnect_assoc(struct nvme_fc_ctrl * ctrl) nvme_fc_xmt_disconnect_assoc() argument
1473 struct nvme_fc_ctrl *ctrl, *tmp, *ret = NULL; nvme_fc_match_disconn_ls() local
1530 struct nvme_fc_ctrl *ctrl = NULL; nvme_fc_ls_disconnect_assoc() local
1790 __nvme_fc_exit_request(struct nvme_fc_ctrl * ctrl,struct nvme_fc_fcp_op * op) __nvme_fc_exit_request() argument
1811 __nvme_fc_abort_op(struct nvme_fc_ctrl * ctrl,struct nvme_fc_fcp_op * op) __nvme_fc_abort_op() argument
1838 nvme_fc_abort_aen_ops(struct nvme_fc_ctrl * ctrl) nvme_fc_abort_aen_ops() argument
1852 __nvme_fc_fcpop_chk_teardowns(struct nvme_fc_ctrl * ctrl,struct nvme_fc_fcp_op * op,int opstate) __nvme_fc_fcpop_chk_teardowns() argument
1871 struct nvme_fc_ctrl *ctrl = nvme_fc_ctrl_ioerr_work() local
1901 struct nvme_fc_ctrl *ctrl = op->ctrl; nvme_fc_fcpio_done() local
2058 __nvme_fc_init_request(struct nvme_fc_ctrl * ctrl,struct nvme_fc_queue * queue,struct nvme_fc_fcp_op * op,struct request * rq,u32 rqno) __nvme_fc_init_request() argument
2114 struct nvme_fc_ctrl *ctrl = to_fc_ctrl(set->driver_data); nvme_fc_init_request() local
2131 nvme_fc_init_aen_ops(struct nvme_fc_ctrl * ctrl) nvme_fc_init_aen_ops() argument
2170 nvme_fc_term_aen_ops(struct nvme_fc_ctrl * ctrl) nvme_fc_term_aen_ops() argument
2188 struct nvme_fc_ctrl *ctrl = to_fc_ctrl(data); __nvme_fc_init_hctx() local
2210 nvme_fc_init_queue(struct nvme_fc_ctrl * ctrl,int idx) nvme_fc_init_queue() argument
2264 __nvme_fc_delete_hw_queue(struct nvme_fc_ctrl * ctrl,struct nvme_fc_queue * queue,unsigned int qidx) __nvme_fc_delete_hw_queue() argument
2274 nvme_fc_free_io_queues(struct nvme_fc_ctrl * ctrl) nvme_fc_free_io_queues() argument
2283 __nvme_fc_create_hw_queue(struct nvme_fc_ctrl * ctrl,struct nvme_fc_queue * queue,unsigned int qidx,u16 qsize) __nvme_fc_create_hw_queue() argument
2297 nvme_fc_delete_hw_io_queues(struct nvme_fc_ctrl * ctrl) nvme_fc_delete_hw_io_queues() argument
2307 nvme_fc_create_hw_io_queues(struct nvme_fc_ctrl * ctrl,u16 qsize) nvme_fc_create_hw_io_queues() argument
2327 nvme_fc_connect_io_queues(struct nvme_fc_ctrl * ctrl,u16 qsize) nvme_fc_connect_io_queues() argument
2347 nvme_fc_init_io_queues(struct nvme_fc_ctrl * ctrl) nvme_fc_init_io_queues() argument
2358 struct nvme_fc_ctrl *ctrl = nvme_fc_ctrl_free() local
2379 nvme_fc_ctrl_put(struct nvme_fc_ctrl * ctrl) nvme_fc_ctrl_put() argument
2385 nvme_fc_ctrl_get(struct nvme_fc_ctrl * ctrl) nvme_fc_ctrl_get() argument
2397 struct nvme_fc_ctrl *ctrl = to_fc_ctrl(nctrl); nvme_fc_free_ctrl() local
2420 struct nvme_fc_ctrl *ctrl = to_fc_ctrl(nctrl); nvme_fc_terminate_exchange() local
2438 __nvme_fc_abort_outstanding_ios(struct nvme_fc_ctrl * ctrl,bool start_queues) __nvme_fc_abort_outstanding_ios() argument
2499 nvme_fc_error_recovery(struct nvme_fc_ctrl * ctrl,char * errmsg) nvme_fc_error_recovery() argument
2534 struct nvme_fc_ctrl *ctrl = op->ctrl; nvme_fc_timeout() local
2561 nvme_fc_map_data(struct nvme_fc_ctrl * ctrl,struct request * rq,struct nvme_fc_fcp_op * op) nvme_fc_map_data() argument
2596 nvme_fc_unmap_data(struct nvme_fc_ctrl * ctrl,struct request * rq,struct nvme_fc_fcp_op * op) nvme_fc_unmap_data() argument
2636 nvme_fc_start_fcp_op(struct nvme_fc_ctrl * ctrl,struct nvme_fc_queue * queue,struct nvme_fc_fcp_op * op,u32 data_len,enum nvmefc_fcp_datadir io_dir) nvme_fc_start_fcp_op() argument
2757 struct nvme_fc_ctrl *ctrl = queue->ctrl; nvme_fc_queue_rq() local
2797 struct nvme_fc_ctrl *ctrl = to_fc_ctrl(arg); nvme_fc_submit_async_event() local
2817 struct nvme_fc_ctrl *ctrl = op->ctrl; nvme_fc_complete_rq() local
2829 struct nvme_fc_ctrl *ctrl = to_fc_ctrl(set->driver_data); nvme_fc_map_queues() local
2860 nvme_fc_create_io_queues(struct nvme_fc_ctrl * ctrl) nvme_fc_create_io_queues() argument
2913 nvme_fc_recreate_io_queues(struct nvme_fc_ctrl * ctrl) nvme_fc_recreate_io_queues() argument
2985 nvme_fc_ctlr_active_on_rport(struct nvme_fc_ctrl * ctrl) nvme_fc_ctlr_active_on_rport() argument
3001 nvme_fc_ctlr_inactive_on_rport(struct nvme_fc_ctrl * ctrl) nvme_fc_ctlr_inactive_on_rport() argument
3024 nvme_fc_create_association(struct nvme_fc_ctrl * ctrl) nvme_fc_create_association() argument
3191 nvme_fc_delete_association(struct nvme_fc_ctrl * ctrl) nvme_fc_delete_association() argument
3258 struct nvme_fc_ctrl *ctrl = to_fc_ctrl(nctrl); nvme_fc_delete_ctrl() local
3277 nvme_fc_reconnect_or_delete(struct nvme_fc_ctrl * ctrl,int status) nvme_fc_reconnect_or_delete() argument
3329 struct nvme_fc_ctrl *ctrl = nvme_fc_reset_ctrl_work() local
3376 struct nvme_fc_ctrl *ctrl = nvme_fc_connect_ctrl_work() local
3412 struct nvme_fc_ctrl *ctrl; nvme_fc_existing_controller() local
3431 struct nvme_fc_ctrl *ctrl; nvme_fc_alloc_ctrl() local
3538 struct nvme_fc_ctrl *ctrl; nvme_fc_init_ctrl() local
3681 struct nvme_ctrl *ctrl; nvme_fc_create_ctrl() local
3935 struct nvme_fc_ctrl *ctrl; nvme_fc_delete_controllers() local
[all...]
H A Dcore.c151 static void nvme_remove_invalid_namespaces(struct nvme_ctrl *ctrl,
153 static void nvme_update_keep_alive(struct nvme_ctrl *ctrl,
155 static int nvme_get_log_lsi(struct nvme_ctrl *ctrl, u32 nsid, u8 log_page,
158 void nvme_queue_scan(struct nvme_ctrl *ctrl) in nvme_queue_scan() argument
163 if (nvme_ctrl_state(ctrl) == NVME_CTRL_LIVE && ctrl->tagset) in nvme_queue_scan()
164 queue_work(nvme_wq, &ctrl->scan_work); in nvme_queue_scan()
173 int nvme_try_sched_reset(struct nvme_ctrl *ctrl) in nvme_try_sched_reset() argument
175 if (nvme_ctrl_state(ctrl) != NVME_CTRL_RESETTING) in nvme_try_sched_reset()
177 if (!queue_work(nvme_reset_wq, &ctrl in nvme_try_sched_reset()
185 struct nvme_ctrl *ctrl = container_of(to_delayed_work(work), nvme_failfast_work() local
196 nvme_start_failfast_work(struct nvme_ctrl * ctrl) nvme_start_failfast_work() argument
205 nvme_stop_failfast_work(struct nvme_ctrl * ctrl) nvme_stop_failfast_work() argument
215 nvme_reset_ctrl(struct nvme_ctrl * ctrl) nvme_reset_ctrl() argument
225 nvme_reset_ctrl_sync(struct nvme_ctrl * ctrl) nvme_reset_ctrl_sync() argument
239 nvme_do_delete_ctrl(struct nvme_ctrl * ctrl) nvme_do_delete_ctrl() argument
253 struct nvme_ctrl *ctrl = nvme_delete_ctrl_work() local
259 nvme_delete_ctrl(struct nvme_ctrl * ctrl) nvme_delete_ctrl() argument
269 nvme_delete_ctrl_sync(struct nvme_ctrl * ctrl) nvme_delete_ctrl_sync() argument
459 struct nvme_ctrl *ctrl = nvme_req(req)->ctrl; nvme_complete_rq() local
537 nvme_cancel_tagset(struct nvme_ctrl * ctrl) nvme_cancel_tagset() argument
547 nvme_cancel_admin_tagset(struct nvme_ctrl * ctrl) nvme_cancel_admin_tagset() argument
557 nvme_change_ctrl_state(struct nvme_ctrl * ctrl,enum nvme_ctrl_state new_state) nvme_change_ctrl_state() argument
656 nvme_wait_reset(struct nvme_ctrl * ctrl) nvme_wait_reset() argument
756 nvme_fail_nonready_command(struct nvme_ctrl * ctrl,struct request * rq) nvme_fail_nonready_command() argument
775 __nvme_check_ready(struct nvme_ctrl * ctrl,struct request * rq,bool queue_live,enum nvme_ctrl_state state) __nvme_check_ready() argument
1070 struct nvme_ctrl *ctrl = nvme_req(req)->ctrl; nvme_cleanup_cmd() local
1206 nvme_command_effects(struct nvme_ctrl * ctrl,struct nvme_ns * ns,u8 opcode) nvme_command_effects() argument
1235 nvme_passthru_start(struct nvme_ctrl * ctrl,struct nvme_ns * ns,u8 opcode) nvme_passthru_start() argument
1255 nvme_passthru_end(struct nvme_ctrl * ctrl,struct nvme_ns * ns,u32 effects,struct nvme_command * cmd,int status) nvme_passthru_end() argument
1306 nvme_keep_alive_work_period(struct nvme_ctrl * ctrl) nvme_keep_alive_work_period() argument
1321 nvme_queue_keep_alive_work(struct nvme_ctrl * ctrl) nvme_queue_keep_alive_work() argument
1339 struct nvme_ctrl *ctrl = rq->end_io_data; nvme_keep_alive_end_io() local
1374 struct nvme_ctrl *ctrl = container_of(to_delayed_work(work), nvme_keep_alive_work() local
1405 nvme_start_keep_alive(struct nvme_ctrl * ctrl) nvme_start_keep_alive() argument
1413 nvme_stop_keep_alive(struct nvme_ctrl * ctrl) nvme_stop_keep_alive() argument
1422 nvme_update_keep_alive(struct nvme_ctrl * ctrl,struct nvme_command * cmd) nvme_update_keep_alive() argument
1437 nvme_id_cns_ok(struct nvme_ctrl * ctrl,u8 cns) nvme_id_cns_ok() argument
1485 nvme_process_ns_desc(struct nvme_ctrl * ctrl,struct nvme_ns_ids * ids,struct nvme_ns_id_desc * cur,bool * csi_seen) nvme_process_ns_desc() argument
1537 nvme_identify_ns_descs(struct nvme_ctrl * ctrl,struct nvme_ns_info * info) nvme_identify_ns_descs() argument
1591 nvme_identify_ns(struct nvme_ctrl * ctrl,unsigned nsid,struct nvme_id_ns ** id) nvme_identify_ns() argument
1615 nvme_ns_info_from_identify(struct nvme_ctrl * ctrl,struct nvme_ns_info * info) nvme_ns_info_from_identify() argument
1655 nvme_ns_info_from_id_cs_indep(struct nvme_ctrl * ctrl,struct nvme_ns_info * info) nvme_ns_info_from_id_cs_indep() argument
1720 nvme_set_queue_count(struct nvme_ctrl * ctrl,int * count) nvme_set_queue_count() argument
1758 nvme_enable_aen(struct nvme_ctrl * ctrl) nvme_enable_aen() argument
1895 nvme_identify_ns_nvm(struct nvme_ctrl * ctrl,unsigned int nsid,struct nvme_id_ns_nvm ** nvmp) nvme_identify_ns_nvm() argument
1947 nvme_configure_metadata(struct nvme_ctrl * ctrl,struct nvme_ns_head * head,struct nvme_id_ns * id,struct nvme_id_ns_nvm * nvm,struct nvme_ns_info * info) nvme_configure_metadata() argument
2044 nvme_max_drv_segments(struct nvme_ctrl * ctrl) nvme_max_drv_segments() argument
2049 nvme_set_ctrl_limits(struct nvme_ctrl * ctrl,struct queue_limits * lim,bool is_admin) nvme_set_ctrl_limits() argument
2065 struct nvme_ctrl *ctrl = ns->ctrl; nvme_update_disk_info() local
2172 struct nvme_ctrl *ctrl = ns->ctrl; nvme_set_chunk_sectors() local
2222 nvme_query_fdp_granularity(struct nvme_ctrl * ctrl,struct nvme_ns_info * info,u8 fdp_idx) nvme_query_fdp_granularity() argument
2298 struct nvme_ctrl *ctrl = ns->ctrl; nvme_query_fdp_info() local
2595 struct nvme_ctrl *ctrl = data; nvme_sec_submit() local
2610 nvme_configure_opal(struct nvme_ctrl * ctrl,bool was_suspended) nvme_configure_opal() argument
2623 nvme_configure_opal(struct nvme_ctrl * ctrl,bool was_suspended) nvme_configure_opal() argument
2650 nvme_wait_ready(struct nvme_ctrl * ctrl,u32 mask,u32 val,u32 timeout,const char * op) nvme_wait_ready() argument
2677 nvme_disable_ctrl(struct nvme_ctrl * ctrl,bool shutdown) nvme_disable_ctrl() argument
2703 nvme_enable_ctrl(struct nvme_ctrl * ctrl) nvme_enable_ctrl() argument
2782 nvme_configure_timestamp(struct nvme_ctrl * ctrl) nvme_configure_timestamp() argument
2799 nvme_configure_host_options(struct nvme_ctrl * ctrl) nvme_configure_host_options() argument
2880 nvme_configure_apst(struct nvme_ctrl * ctrl) nvme_configure_apst() argument
2987 struct nvme_ctrl *ctrl = dev_get_drvdata(dev); nvme_set_latency_tolerance() local
3098 nvme_init_subnqn(struct nvme_subsystem * subsys,struct nvme_ctrl * ctrl,struct nvme_id_ctrl * id) nvme_init_subnqn() argument
3187 nvme_discovery_ctrl(struct nvme_ctrl * ctrl) nvme_discovery_ctrl() argument
3192 nvme_admin_ctrl(struct nvme_ctrl * ctrl) nvme_admin_ctrl() argument
3197 nvme_is_io_ctrl(struct nvme_ctrl * ctrl) nvme_is_io_ctrl() argument
3203 nvme_validate_cntlid(struct nvme_subsystem * subsys,struct nvme_ctrl * ctrl,struct nvme_id_ctrl * id) nvme_validate_cntlid() argument
3233 nvme_init_subsystem(struct nvme_ctrl * ctrl,struct nvme_id_ctrl * id) nvme_init_subsystem() argument
3319 nvme_get_log_lsi(struct nvme_ctrl * ctrl,u32 nsid,u8 log_page,u8 lsp,u8 csi,void * log,size_t size,u64 offset,u16 lsi) nvme_get_log_lsi() argument
3339 nvme_get_log(struct nvme_ctrl * ctrl,u32 nsid,u8 log_page,u8 lsp,u8 csi,void * log,size_t size,u64 offset) nvme_get_log() argument
3346 nvme_get_effects_log(struct nvme_ctrl * ctrl,u8 csi,struct nvme_effects_log ** log) nvme_get_effects_log() argument
3376 nvme_mps_to_sectors(struct nvme_ctrl * ctrl,u32 units) nvme_mps_to_sectors() argument
3385 nvme_init_non_mdts_limits(struct nvme_ctrl * ctrl) nvme_init_non_mdts_limits() argument
3432 nvme_init_effects_log(struct nvme_ctrl * ctrl,u8 csi,struct nvme_effects_log ** log) nvme_init_effects_log() argument
3451 nvme_init_known_nvm_effects(struct nvme_ctrl * ctrl) nvme_init_known_nvm_effects() argument
3483 nvme_init_effects(struct nvme_ctrl * ctrl,struct nvme_id_ctrl * id) nvme_init_effects() argument
3506 nvme_check_ctrl_fabric_info(struct nvme_ctrl * ctrl,struct nvme_id_ctrl * id) nvme_check_ctrl_fabric_info() argument
3548 nvme_init_identify(struct nvme_ctrl * ctrl) nvme_init_identify() argument
3695 nvme_init_ctrl_finish(struct nvme_ctrl * ctrl,bool was_suspended) nvme_init_ctrl_finish() argument
3760 struct nvme_ctrl *ctrl = nvme_dev_open() local
3782 struct nvme_ctrl *ctrl = nvme_dev_release() local
3799 nvme_find_ns_head(struct nvme_ctrl * ctrl,unsigned nsid) nvme_find_ns_head() argument
3912 nvme_alloc_ns_head(struct nvme_ctrl * ctrl,struct nvme_ns_info * info) nvme_alloc_ns_head() argument
3999 struct nvme_ctrl *ctrl = ns->ctrl; nvme_init_ns_head() local
4097 nvme_find_get_ns(struct nvme_ctrl * ctrl,unsigned nsid) nvme_find_get_ns() argument
4135 nvme_alloc_ns(struct nvme_ctrl * ctrl,struct nvme_ns_info * info) nvme_alloc_ns() argument
4300 nvme_ns_remove_by_nsid(struct nvme_ctrl * ctrl,u32 nsid) nvme_ns_remove_by_nsid() argument
4332 nvme_scan_ns(struct nvme_ctrl * ctrl,unsigned nsid) nvme_scan_ns() argument
4390 struct nvme_ctrl *ctrl; global() member
4407 nvme_remove_invalid_namespaces(struct nvme_ctrl * ctrl,unsigned nsid) nvme_remove_invalid_namespaces() argument
4427 nvme_scan_ns_list(struct nvme_ctrl * ctrl) nvme_scan_ns_list() argument
4478 nvme_scan_ns_sequential(struct nvme_ctrl * ctrl) nvme_scan_ns_sequential() argument
4494 nvme_clear_changed_ns_log(struct nvme_ctrl * ctrl) nvme_clear_changed_ns_log() argument
4521 struct nvme_ctrl *ctrl = nvme_scan_work() local
4578 nvme_remove_namespaces(struct nvme_ctrl * ctrl) nvme_remove_namespaces() argument
4623 const struct nvme_ctrl *ctrl = nvme_class_uevent() local
4653 nvme_change_uevent(struct nvme_ctrl * ctrl,char * envdata) nvme_change_uevent() argument
4660 nvme_aen_uevent(struct nvme_ctrl * ctrl) nvme_aen_uevent() argument
4678 struct nvme_ctrl *ctrl = nvme_async_event_work() local
4692 nvme_ctrl_pp_status(struct nvme_ctrl * ctrl) nvme_ctrl_pp_status() argument
4706 nvme_get_fw_slot_info(struct nvme_ctrl * ctrl) nvme_get_fw_slot_info() argument
4738 struct nvme_ctrl *ctrl = container_of(work, nvme_fw_act_work() local
4781 nvme_handle_aen_notice(struct nvme_ctrl * ctrl,u32 result) nvme_handle_aen_notice() argument
4818 nvme_handle_aer_persistent_error(struct nvme_ctrl * ctrl) nvme_handle_aer_persistent_error() argument
4825 nvme_complete_async_event(struct nvme_ctrl * ctrl,__le16 status,volatile union nvme_result * res) nvme_complete_async_event() argument
4865 nvme_alloc_admin_tag_set(struct nvme_ctrl * ctrl,struct blk_mq_tag_set * set,const struct blk_mq_ops * ops,unsigned int cmd_size) nvme_alloc_admin_tag_set() argument
4922 nvme_remove_admin_tag_set(struct nvme_ctrl * ctrl) nvme_remove_admin_tag_set() argument
4938 nvme_alloc_io_tag_set(struct nvme_ctrl * ctrl,struct blk_mq_tag_set * set,const struct blk_mq_ops * ops,unsigned int nr_maps,unsigned int cmd_size) nvme_alloc_io_tag_set() argument
4990 nvme_remove_io_tag_set(struct nvme_ctrl * ctrl) nvme_remove_io_tag_set() argument
5000 nvme_stop_ctrl(struct nvme_ctrl * ctrl) nvme_stop_ctrl() argument
5012 nvme_start_ctrl(struct nvme_ctrl * ctrl) nvme_start_ctrl() argument
5043 nvme_uninit_ctrl(struct nvme_ctrl * ctrl) nvme_uninit_ctrl() argument
5054 nvme_free_cels(struct nvme_ctrl * ctrl) nvme_free_cels() argument
5069 struct nvme_ctrl *ctrl = nvme_free_ctrl() local
5106 nvme_init_ctrl(struct nvme_ctrl * ctrl,struct device * dev,const struct nvme_ctrl_ops * ops,unsigned long quirks) nvme_init_ctrl() argument
5188 nvme_add_ctrl(struct nvme_ctrl * ctrl) nvme_add_ctrl() argument
5218 nvme_mark_namespaces_dead(struct nvme_ctrl * ctrl) nvme_mark_namespaces_dead() argument
5231 nvme_unfreeze(struct nvme_ctrl * ctrl) nvme_unfreeze() argument
5245 nvme_wait_freeze_timeout(struct nvme_ctrl * ctrl,long timeout) nvme_wait_freeze_timeout() argument
5262 nvme_wait_freeze(struct nvme_ctrl * ctrl) nvme_wait_freeze() argument
5275 nvme_start_freeze(struct nvme_ctrl * ctrl) nvme_start_freeze() argument
5294 nvme_quiesce_io_queues(struct nvme_ctrl * ctrl) nvme_quiesce_io_queues() argument
5305 nvme_unquiesce_io_queues(struct nvme_ctrl * ctrl) nvme_unquiesce_io_queues() argument
5314 nvme_quiesce_admin_queue(struct nvme_ctrl * ctrl) nvme_quiesce_admin_queue() argument
5323 nvme_unquiesce_admin_queue(struct nvme_ctrl * ctrl) nvme_unquiesce_admin_queue() argument
5330 nvme_sync_io_queues(struct nvme_ctrl * ctrl) nvme_sync_io_queues() argument
5343 nvme_sync_queues(struct nvme_ctrl * ctrl) nvme_sync_queues() argument
[all...]
H A Dtcp.c165 struct nvme_tcp_ctrl *ctrl; member
195 struct nvme_ctrl ctrl; member
210 static inline struct nvme_tcp_ctrl *to_tcp_ctrl(struct nvme_ctrl *ctrl) in to_tcp_ctrl() argument
212 return container_of(ctrl, struct nvme_tcp_ctrl, ctrl); in to_tcp_ctrl()
217 return queue - queue->ctrl->queues; in nvme_tcp_queue_id()
247 static inline bool nvme_tcp_tls_configured(struct nvme_ctrl *ctrl) in nvme_tcp_tls_configured() argument
252 return ctrl->opts->tls || ctrl->opts->concat; in nvme_tcp_tls_configured()
260 return queue->ctrl in nvme_tcp_tagset()
553 struct nvme_tcp_ctrl *ctrl = to_tcp_ctrl(set->driver_data); nvme_tcp_init_request() local
579 struct nvme_tcp_ctrl *ctrl = to_tcp_ctrl(data); nvme_tcp_init_hctx() local
589 struct nvme_tcp_ctrl *ctrl = to_tcp_ctrl(data); nvme_tcp_init_admin_hctx() local
613 nvme_tcp_error_recovery(struct nvme_ctrl * ctrl) nvme_tcp_error_recovery() argument
1414 nvme_tcp_free_async_req(struct nvme_tcp_ctrl * ctrl) nvme_tcp_free_async_req() argument
1421 nvme_tcp_alloc_async_req(struct nvme_tcp_ctrl * ctrl) nvme_tcp_alloc_async_req() argument
1439 struct nvme_tcp_ctrl *ctrl = to_tcp_ctrl(nctrl); nvme_tcp_free_queue() local
1597 struct nvme_tcp_ctrl *ctrl = queue->ctrl; nvme_tcp_default_queue() local
1606 struct nvme_tcp_ctrl *ctrl = queue->ctrl; nvme_tcp_read_queue() local
1617 struct nvme_tcp_ctrl *ctrl = queue->ctrl; nvme_tcp_poll_queue() local
1639 struct nvme_tcp_ctrl *ctrl = queue->ctrl; nvme_tcp_set_queue_io_cpu() local
1683 struct nvme_tcp_ctrl *ctrl = queue->ctrl; nvme_tcp_tls_done() local
1768 struct nvme_tcp_ctrl *ctrl = to_tcp_ctrl(nctrl); nvme_tcp_alloc_queue() local
1934 struct nvme_tcp_ctrl *ctrl = to_tcp_ctrl(nctrl); nvme_tcp_stop_queue_nowait() local
1953 struct nvme_tcp_ctrl *ctrl = to_tcp_ctrl(nctrl); nvme_tcp_wait_queue() local
1994 struct nvme_tcp_ctrl *ctrl = to_tcp_ctrl(nctrl); nvme_tcp_start_queue() local
2019 nvme_tcp_free_admin_queue(struct nvme_ctrl * ctrl) nvme_tcp_free_admin_queue() argument
2030 nvme_tcp_free_io_queues(struct nvme_ctrl * ctrl) nvme_tcp_free_io_queues() argument
2038 nvme_tcp_stop_io_queues(struct nvme_ctrl * ctrl) nvme_tcp_stop_io_queues() argument
2048 nvme_tcp_start_io_queues(struct nvme_ctrl * ctrl,int first,int last) nvme_tcp_start_io_queues() argument
2067 nvme_tcp_alloc_admin_queue(struct nvme_ctrl * ctrl) nvme_tcp_alloc_admin_queue() argument
2101 __nvme_tcp_alloc_io_queues(struct nvme_ctrl * ctrl) __nvme_tcp_alloc_io_queues() argument
2142 nvme_tcp_alloc_io_queues(struct nvme_ctrl * ctrl) nvme_tcp_alloc_io_queues() argument
2167 nvme_tcp_configure_io_queues(struct nvme_ctrl * ctrl,bool new) nvme_tcp_configure_io_queues() argument
2236 nvme_tcp_configure_admin_queue(struct nvme_ctrl * ctrl,bool new) nvme_tcp_configure_admin_queue() argument
2286 nvme_tcp_teardown_admin_queue(struct nvme_ctrl * ctrl,bool remove) nvme_tcp_teardown_admin_queue() argument
2305 nvme_tcp_teardown_io_queues(struct nvme_ctrl * ctrl,bool remove) nvme_tcp_teardown_io_queues() argument
2321 nvme_tcp_reconnect_or_remove(struct nvme_ctrl * ctrl,int status) nvme_tcp_reconnect_or_remove() argument
2360 nvme_tcp_key_revoke_needed(struct nvme_ctrl * ctrl) nvme_tcp_key_revoke_needed() argument
2365 nvme_tcp_setup_ctrl(struct nvme_ctrl * ctrl,bool new) nvme_tcp_setup_ctrl() argument
2452 struct nvme_ctrl *ctrl = &tcp_ctrl->ctrl; nvme_tcp_reconnect_ctrl_work() local
2478 struct nvme_ctrl *ctrl = &tcp_ctrl->ctrl; nvme_tcp_error_recovery_work() local
2503 nvme_tcp_teardown_ctrl(struct nvme_ctrl * ctrl,bool shutdown) nvme_tcp_teardown_ctrl() argument
2511 nvme_tcp_delete_ctrl(struct nvme_ctrl * ctrl) nvme_tcp_delete_ctrl() argument
2518 struct nvme_ctrl *ctrl = nvme_reset_ctrl_work() local
2547 nvme_tcp_stop_ctrl(struct nvme_ctrl * ctrl) nvme_tcp_stop_ctrl() argument
2555 struct nvme_tcp_ctrl *ctrl = to_tcp_ctrl(nctrl); nvme_tcp_free_ctrl() local
2603 struct nvme_tcp_ctrl *ctrl = to_tcp_ctrl(arg); nvme_tcp_submit_async_event() local
2634 struct nvme_ctrl *ctrl = &req->queue->ctrl->ctrl; nvme_tcp_complete_timed_out() local
2643 struct nvme_ctrl *ctrl = &req->queue->ctrl->ctrl; nvme_tcp_timeout() local
2787 struct nvme_tcp_ctrl *ctrl = to_tcp_ctrl(set->driver_data); nvme_tcp_map_queues() local
2809 nvme_tcp_get_address(struct nvme_ctrl * ctrl,char * buf,int size) nvme_tcp_get_address() argument
2875 struct nvme_tcp_ctrl *ctrl; nvme_tcp_existing_controller() local
2892 struct nvme_tcp_ctrl *ctrl; nvme_tcp_alloc_ctrl() local
2974 struct nvme_tcp_ctrl *ctrl; nvme_tcp_create_ctrl() local
3056 struct nvme_tcp_ctrl *ctrl; nvme_tcp_cleanup_module() local
[all...]
H A Dnvme.h251 struct nvme_ctrl *ctrl; member
471 static inline enum nvme_ctrl_state nvme_ctrl_state(struct nvme_ctrl *ctrl) in nvme_ctrl_state() argument
473 return READ_ONCE(ctrl->state); in nvme_ctrl_state()
586 struct nvme_ctrl *ctrl; member
616 static inline unsigned long nvme_get_virt_boundary(struct nvme_ctrl *ctrl, in nvme_get_virt_boundary() argument
631 int (*reg_read32)(struct nvme_ctrl *ctrl, u32 off, u32 *val);
632 int (*reg_write32)(struct nvme_ctrl *ctrl, u32 off, u32 val);
633 int (*reg_read64)(struct nvme_ctrl *ctrl, u32 off, u64 *val);
634 void (*free_ctrl)(struct nvme_ctrl *ctrl);
635 void (*submit_async_event)(struct nvme_ctrl *ctrl);
[all …]
/linux/drivers/pci/hotplug/
H A Dpciehp_hpc.c49 static inline struct pci_dev *ctrl_dev(struct controller *ctrl) in ctrl_dev() argument
51 return ctrl->pcie->port; in ctrl_dev()
58 static inline int pciehp_request_irq(struct controller *ctrl) in pciehp_request_irq() argument
60 int retval, irq = ctrl->pcie->irq; in pciehp_request_irq()
63 ctrl->poll_thread = kthread_run(&pciehp_poll, ctrl, in pciehp_request_irq()
65 slot_name(ctrl)); in pciehp_request_irq()
66 return PTR_ERR_OR_ZERO(ctrl->poll_thread); in pciehp_request_irq()
71 IRQF_SHARED, "pciehp", ctrl); in pciehp_request_irq()
73 ctrl_err(ctrl, "Cannot get irq %d for the hotplug controller\n", in pciehp_request_irq()
78 static inline void pciehp_free_irq(struct controller *ctrl) in pciehp_free_irq() argument
[all …]
H A Dshpchp_hpc.c169 static void start_int_poll_timer(struct controller *ctrl, int sec);
171 static inline u8 shpc_readb(struct controller *ctrl, int reg) in shpc_readb() argument
173 return readb(ctrl->creg + reg); in shpc_readb()
176 static inline u16 shpc_readw(struct controller *ctrl, int reg) in shpc_readw() argument
178 return readw(ctrl->creg + reg); in shpc_readw()
181 static inline void shpc_writew(struct controller *ctrl, int reg, u16 val) in shpc_writew() argument
183 writew(val, ctrl->creg + reg); in shpc_writew()
186 static inline u32 shpc_readl(struct controller *ctrl, int reg) in shpc_readl() argument
188 return readl(ctrl->creg + reg); in shpc_readl()
191 static inline void shpc_writel(struct controller *ctrl, int reg, u32 val) in shpc_writel() argument
[all …]
H A Dcpqphp_core.c116 * @ctrl: controller to use
120 static int init_SERR(struct controller *ctrl) in init_SERR() argument
125 if (!ctrl) in init_SERR()
128 tempdword = ctrl->first_slot; in init_SERR()
130 number_of_slots = readb(ctrl->hpc_reg + SLOT_MASK) & 0x0F; in init_SERR()
133 writeb(0, ctrl->hpc_reg + SLOT_SERR); in init_SERR()
266 static int ctrl_slot_cleanup(struct controller *ctrl) in ctrl_slot_cleanup() argument
270 old_slot = ctrl->slot; in ctrl_slot_cleanup()
271 ctrl->slot = NULL; in ctrl_slot_cleanup()
280 cpqhp_remove_debugfs_files(ctrl); in ctrl_slot_cleanup()
[all …]
H A Dshpchp_ctrl.c45 u8 shpchp_handle_attention_button(u8 hp_slot, struct controller *ctrl) in shpchp_handle_attention_button() argument
51 ctrl_dbg(ctrl, "Attention button interrupt received\n"); in shpchp_handle_attention_button()
53 p_slot = shpchp_find_slot(ctrl, hp_slot + ctrl->slot_device_offset); in shpchp_handle_attention_button()
59 ctrl_info(ctrl, "Button pressed on Slot(%s)\n", slot_name(p_slot)); in shpchp_handle_attention_button()
68 u8 shpchp_handle_switch_change(u8 hp_slot, struct controller *ctrl) in shpchp_handle_switch_change() argument
75 ctrl_dbg(ctrl, "Switch interrupt received\n"); in shpchp_handle_switch_change()
77 p_slot = shpchp_find_slot(ctrl, hp_slot + ctrl->slot_device_offset); in shpchp_handle_switch_change()
80 ctrl_dbg(ctrl, "Card present %x Power status %x\n", in shpchp_handle_switch_change()
87 ctrl_info(ctrl, "Latch open on Slot(%s)\n", slot_name(p_slot)); in shpchp_handle_switch_change()
91 ctrl_err(ctrl, "Surprise Removal of card\n"); in shpchp_handle_switch_change()
[all …]
H A Dcpqphp_ctrl.c28 static u32 configure_new_device(struct controller *ctrl, struct pci_func *func,
30 static int configure_new_function(struct controller *ctrl, struct pci_func *func,
32 static void interrupt_event_handler(struct controller *ctrl);
53 static u8 handle_switch_change(u8 change, struct controller *ctrl) in handle_switch_change() argument
72 func = cpqhp_slot_find(ctrl->bus, in handle_switch_change()
73 (hp_slot + ctrl->slot_device_offset), 0); in handle_switch_change()
78 taskInfo = &(ctrl->event_queue[ctrl->next_event]); in handle_switch_change()
79 ctrl->next_event = (ctrl->next_event + 1) % 10; in handle_switch_change()
84 temp_word = ctrl->ctrl_int_comp >> 16; in handle_switch_change()
88 if (ctrl->ctrl_int_comp & (0x1L << hp_slot)) { in handle_switch_change()
[all …]
H A Dpciehp.h35 #define ctrl_dbg(ctrl, format, arg...) \ argument
36 pci_dbg(ctrl->pcie->port, format, ## arg)
37 #define ctrl_err(ctrl, format, arg...) \ argument
38 pci_err(ctrl->pcie->port, format, ## arg)
39 #define ctrl_info(ctrl, format, arg...) \ argument
40 pci_info(ctrl->pcie->port, format, ## arg)
41 #define ctrl_warn(ctrl, format, arg...) \ argument
42 pci_warn(ctrl->pcie->port, format, ## arg)
156 #define ATTN_BUTTN(ctrl) ((ctrl)->slot_cap & PCI_EXP_SLTCAP_ABP) argument
157 #define POWER_CTRL(ctrl) ((ctrl)->slot_cap & PCI_EXP_SLTCAP_PCP) argument
[all …]
/linux/drivers/media/v4l2-core/
H A Dv4l2-ctrls-api.c41 struct v4l2_ctrl *ctrl, in ptr_to_user() argument
46 if (ctrl->is_ptr && !ctrl->is_string) in ptr_to_user()
50 switch (ctrl->type) { in ptr_to_user()
54 c->size = ctrl->elem_size; in ptr_to_user()
70 static int cur_to_user(struct v4l2_ext_control *c, struct v4l2_ctrl *ctrl) in cur_to_user() argument
72 return ptr_to_user(c, ctrl, ctrl->p_cur); in cur_to_user()
77 struct v4l2_ctrl *ctrl) in new_to_user() argument
79 return ptr_to_user(c, ctrl, ctrl->p_new); in new_to_user()
86 return ptr_to_user(c, ref->ctrl, ref->p_req); in req_to_user()
90 static int def_to_user(struct v4l2_ext_control *c, struct v4l2_ctrl *ctrl) in def_to_user() argument
[all …]
H A Dv4l2-ctrls-core.c19 static void fill_event(struct v4l2_event *ev, struct v4l2_ctrl *ctrl, in fill_event() argument
24 ev->id = ctrl->id; in fill_event()
25 ev->u.ctrl.changes = changes; in fill_event()
26 ev->u.ctrl.type = ctrl->type; in fill_event()
27 ev->u.ctrl.flags = user_flags(ctrl); in fill_event()
28 if (ctrl->is_ptr) in fill_event()
29 ev->u.ctrl.value64 = 0; in fill_event()
31 ev->u.ctrl.value64 = *ctrl->p_cur.p_s64; in fill_event()
32 ev->u.ctrl.minimum = ctrl->minimum; in fill_event()
33 ev->u.ctrl.maximum = ctrl->maximum; in fill_event()
[all …]
/linux/drivers/mtd/nand/raw/brcmnand/
H A Dbrcmnand.c346 struct brcmnand_controller *ctrl; member
666 static inline bool brcmnand_non_mmio_ops(struct brcmnand_controller *ctrl) in brcmnand_non_mmio_ops() argument
675 static inline u32 nand_readreg(struct brcmnand_controller *ctrl, u32 offs) in nand_readreg() argument
677 if (brcmnand_non_mmio_ops(ctrl)) in nand_readreg()
678 return brcmnand_soc_read(ctrl->soc, offs); in nand_readreg()
679 return brcmnand_readl(ctrl->nand_base + offs); in nand_readreg()
682 static inline void nand_writereg(struct brcmnand_controller *ctrl, u32 offs, in nand_writereg() argument
685 if (brcmnand_non_mmio_ops(ctrl)) in nand_writereg()
686 brcmnand_soc_write(ctrl->soc, val, offs); in nand_writereg()
688 brcmnand_writel(val, ctrl->nand_base + offs); in nand_writereg()
[all …]
/linux/drivers/net/mdio/
H A Dmdio-bitbang.c44 static void mdiobb_send_bit(struct mdiobb_ctrl *ctrl, int val) in mdiobb_send_bit() argument
46 const struct mdiobb_ops *ops = ctrl->ops; in mdiobb_send_bit()
48 ops->set_mdio_data(ctrl, val); in mdiobb_send_bit()
50 ops->set_mdc(ctrl, 1); in mdiobb_send_bit()
52 ops->set_mdc(ctrl, 0); in mdiobb_send_bit()
56 static int mdiobb_get_bit(struct mdiobb_ctrl *ctrl) in mdiobb_get_bit() argument
58 const struct mdiobb_ops *ops = ctrl->ops; in mdiobb_get_bit()
61 ops->set_mdc(ctrl, 1); in mdiobb_get_bit()
63 ops->set_mdc(ctrl, 0); in mdiobb_get_bit()
65 return ops->get_mdio_data(ctrl); in mdiobb_get_bit()
[all …]
/linux/drivers/nvme/target/
H A Dcore.c137 static void nvmet_async_events_failall(struct nvmet_ctrl *ctrl) in nvmet_async_events_failall() argument
141 mutex_lock(&ctrl->lock); in nvmet_async_events_failall()
142 while (ctrl->nr_async_event_cmds) { in nvmet_async_events_failall()
143 req = ctrl->async_event_cmds[--ctrl->nr_async_event_cmds]; in nvmet_async_events_failall()
144 mutex_unlock(&ctrl->lock); in nvmet_async_events_failall()
146 mutex_lock(&ctrl->lock); in nvmet_async_events_failall()
148 mutex_unlock(&ctrl->lock); in nvmet_async_events_failall()
151 static void nvmet_async_events_process(struct nvmet_ctrl *ctrl) in nvmet_async_events_process() argument
156 mutex_lock(&ctrl in nvmet_async_events_process()
174 nvmet_async_events_free(struct nvmet_ctrl * ctrl) nvmet_async_events_free() argument
188 struct nvmet_ctrl *ctrl = nvmet_async_event_work() local
194 nvmet_add_async_event(struct nvmet_ctrl * ctrl,u8 event_type,u8 event_info,u8 log_page) nvmet_add_async_event() argument
214 nvmet_add_to_changed_ns_log(struct nvmet_ctrl * ctrl,__le32 nsid) nvmet_add_to_changed_ns_log() argument
240 struct nvmet_ctrl *ctrl; nvmet_ns_changed() local
257 struct nvmet_ctrl *ctrl; nvmet_send_ana_event() local
306 struct nvmet_ctrl *ctrl; nvmet_port_del_ctrls() local
398 struct nvmet_ctrl *ctrl = container_of(to_delayed_work(work), nvmet_keep_alive_timer() local
416 nvmet_start_keep_alive_timer(struct nvmet_ctrl * ctrl) nvmet_start_keep_alive_timer() argument
427 nvmet_stop_keep_alive_timer(struct nvmet_ctrl * ctrl) nvmet_stop_keep_alive_timer() argument
519 nvmet_p2pmem_ns_add_p2p(struct nvmet_ctrl * ctrl,struct nvmet_ns * ns) nvmet_p2pmem_ns_add_p2p() argument
572 struct nvmet_ctrl *ctrl; nvmet_ns_enable() local
629 struct nvmet_ctrl *ctrl; nvmet_ns_disable() local
750 struct nvmet_ctrl *ctrl = req->sq->ctrl; nvmet_set_error() local
828 nvmet_cq_setup(struct nvmet_ctrl * ctrl,struct nvmet_cq * cq,u16 qid,u16 size) nvmet_cq_setup() argument
839 struct nvmet_ctrl *ctrl = cq->ctrl; nvmet_cq_destroy() local
848 nvmet_sq_setup(struct nvmet_ctrl * ctrl,struct nvmet_sq * sq,u16 qid,u16 size) nvmet_sq_setup() argument
865 nvmet_check_cqid(struct nvmet_ctrl * ctrl,u16 cqid,bool create) nvmet_check_cqid() argument
879 nvmet_check_io_cqid(struct nvmet_ctrl * ctrl,u16 cqid,bool create) nvmet_check_io_cqid() argument
892 nvmet_cq_create(struct nvmet_ctrl * ctrl,struct nvmet_cq * cq,u16 qid,u16 size) nvmet_cq_create() argument
912 nvmet_check_sqid(struct nvmet_ctrl * ctrl,u16 sqid,bool create) nvmet_check_sqid() argument
928 nvmet_sq_create(struct nvmet_ctrl * ctrl,struct nvmet_sq * sq,struct nvmet_cq * cq,u16 sqid,u16 size) nvmet_sq_create() argument
960 struct nvmet_ctrl *ctrl = sq->ctrl; nvmet_sq_destroy() local
1375 nvmet_start_ctrl(struct nvmet_ctrl * ctrl) nvmet_start_ctrl() argument
1411 nvmet_clear_ctrl(struct nvmet_ctrl * ctrl) nvmet_clear_ctrl() argument
1420 nvmet_update_cc(struct nvmet_ctrl * ctrl,u32 new) nvmet_update_cc() argument
1442 nvmet_init_cap(struct nvmet_ctrl * ctrl) nvmet_init_cap() argument
1465 struct nvmet_ctrl *ctrl = NULL; nvmet_ctrl_find_get() local
1544 nvmet_setup_p2p_ns_map(struct nvmet_ctrl * ctrl,struct device * p2p_client) nvmet_setup_p2p_ns_map() argument
1561 nvmet_release_p2p_ns_map(struct nvmet_ctrl * ctrl) nvmet_release_p2p_ns_map() argument
1576 struct nvmet_ctrl *ctrl = nvmet_fatal_error_handler() local
1586 struct nvmet_ctrl *ctrl; nvmet_alloc_ctrl() local
1735 struct nvmet_ctrl *ctrl = container_of(ref, struct nvmet_ctrl, ref); nvmet_ctrl_free() local
1764 nvmet_ctrl_put(struct nvmet_ctrl * ctrl) nvmet_ctrl_put() argument
1770 nvmet_ctrl_fatal_error(struct nvmet_ctrl * ctrl) nvmet_ctrl_fatal_error() argument
1781 nvmet_ctrl_host_traddr(struct nvmet_ctrl * ctrl,char * traddr,size_t traddr_len) nvmet_ctrl_host_traddr() argument
1922 struct nvmet_ctrl *ctrl; nvmet_subsys_del_ctrls() local
[all...]
H A Dpci-epf.c84 struct nvmet_pci_epf_ctrl *ctrl; member
129 struct nvmet_pci_epf_ctrl *ctrl; member
209 struct nvmet_pci_epf_ctrl ctrl; member
226 static inline u32 nvmet_pci_epf_bar_read32(struct nvmet_pci_epf_ctrl *ctrl, in nvmet_pci_epf_bar_read32() argument
229 __le32 *bar_reg = ctrl->bar + off; in nvmet_pci_epf_bar_read32()
234 static inline void nvmet_pci_epf_bar_write32(struct nvmet_pci_epf_ctrl *ctrl, in nvmet_pci_epf_bar_write32() argument
237 __le32 *bar_reg = ctrl->bar + off; in nvmet_pci_epf_bar_write32()
242 static inline u64 nvmet_pci_epf_bar_read64(struct nvmet_pci_epf_ctrl *ctrl, in nvmet_pci_epf_bar_read64() argument
245 return (u64)nvmet_pci_epf_bar_read32(ctrl, off) | in nvmet_pci_epf_bar_read64()
246 ((u64)nvmet_pci_epf_bar_read32(ctrl, off + 4) << 32); in nvmet_pci_epf_bar_read64()
[all …]
/linux/drivers/clk/bcm/
H A Dclk-iproc-pll.c68 const struct iproc_pll_ctrl *ctrl; member
76 const struct iproc_clk_ctrl *ctrl; member
150 const struct iproc_pll_ctrl *ctrl = pll->ctrl; in pll_wait_for_lock() local
153 u32 val = readl(pll->status_base + ctrl->status.offset); in pll_wait_for_lock()
155 if (val & (1 << ctrl->status.shift)) in pll_wait_for_lock()
166 const struct iproc_pll_ctrl *ctrl = pll->ctrl; in iproc_pll_write() local
170 if (unlikely(ctrl->flags & IPROC_CLK_NEEDS_READ_BACK && in iproc_pll_write()
177 const struct iproc_pll_ctrl *ctrl = pll->ctrl; in __pll_disable() local
180 if (ctrl->flags & IPROC_CLK_PLL_ASIU) { in __pll_disable()
181 val = readl(pll->asiu_base + ctrl->asiu.offset); in __pll_disable()
[all …]
/linux/drivers/watchdog/
H A Drealtek_otto_wdt.c76 struct otto_wdt_ctrl *ctrl = watchdog_get_drvdata(wdev); in otto_wdt_start() local
79 v = ioread32(ctrl->base + OTTO_WDT_REG_CTRL); in otto_wdt_start()
81 iowrite32(v, ctrl->base + OTTO_WDT_REG_CTRL); in otto_wdt_start()
88 struct otto_wdt_ctrl *ctrl = watchdog_get_drvdata(wdev); in otto_wdt_stop() local
91 v = ioread32(ctrl->base + OTTO_WDT_REG_CTRL); in otto_wdt_stop()
93 iowrite32(v, ctrl->base + OTTO_WDT_REG_CTRL); in otto_wdt_stop()
100 struct otto_wdt_ctrl *ctrl = watchdog_get_drvdata(wdev); in otto_wdt_ping() local
102 iowrite32(OTTO_WDT_CNTR_PING, ctrl->base + OTTO_WDT_REG_CNTR); in otto_wdt_ping()
107 static int otto_wdt_tick_ms(struct otto_wdt_ctrl *ctrl, int prescale) in otto_wdt_tick_ms() argument
109 return DIV_ROUND_CLOSEST(1 << (25 + prescale), ctrl->clk_rate_khz); in otto_wdt_tick_ms()
[all …]
/linux/sound/soc/codecs/
H A Dwcd-clsh-v2.c119 static inline void wcd_enable_clsh_block(struct wcd_clsh_ctrl *ctrl, in wcd_enable_clsh_block() argument
122 struct snd_soc_component *comp = ctrl->comp; in wcd_enable_clsh_block()
124 if ((enable && ++ctrl->clsh_users == 1) || in wcd_enable_clsh_block()
125 (!enable && --ctrl->clsh_users == 0)) in wcd_enable_clsh_block()
129 if (ctrl->clsh_users < 0) in wcd_enable_clsh_block()
130 ctrl->clsh_users = 0; in wcd_enable_clsh_block()
175 static void wcd_clsh_buck_ctrl(struct wcd_clsh_ctrl *ctrl, in wcd_clsh_buck_ctrl() argument
179 struct snd_soc_component *comp = ctrl->comp; in wcd_clsh_buck_ctrl()
182 if ((enable && (++ctrl->buck_users == 1)) || in wcd_clsh_buck_ctrl()
183 (!enable && (--ctrl in wcd_clsh_buck_ctrl()
195 wcd_clsh_v3_buck_ctrl(struct snd_soc_component * component,struct wcd_clsh_ctrl * ctrl,int mode,bool enable) wcd_clsh_v3_buck_ctrl() argument
224 wcd_clsh_flyback_ctrl(struct wcd_clsh_ctrl * ctrl,int mode,bool enable) wcd_clsh_flyback_ctrl() argument
246 wcd_clsh_set_gain_path(struct wcd_clsh_ctrl * ctrl,int mode) wcd_clsh_set_gain_path() argument
348 wcd_clsh_set_hph_mode(struct wcd_clsh_ctrl * ctrl,int mode) wcd_clsh_set_hph_mode() argument
453 wcd_clsh_v3_flyback_ctrl(struct snd_soc_component * component,struct wcd_clsh_ctrl * ctrl,int mode,bool enable) wcd_clsh_v3_flyback_ctrl() argument
490 wcd_clsh_v3_state_aux(struct wcd_clsh_ctrl * ctrl,int req_state,bool is_enable,int mode) wcd_clsh_v3_state_aux() argument
509 wcd_clsh_state_lo(struct wcd_clsh_ctrl * ctrl,int req_state,bool is_enable,int mode) wcd_clsh_state_lo() argument
536 wcd_clsh_v3_state_hph_r(struct wcd_clsh_ctrl * ctrl,int req_state,bool is_enable,int mode) wcd_clsh_v3_state_hph_r() argument
568 wcd_clsh_state_hph_r(struct wcd_clsh_ctrl * ctrl,int req_state,bool is_enable,int mode) wcd_clsh_state_hph_r() argument
626 wcd_clsh_v3_state_hph_l(struct wcd_clsh_ctrl * ctrl,int req_state,bool is_enable,int mode) wcd_clsh_v3_state_hph_l() argument
658 wcd_clsh_state_hph_l(struct wcd_clsh_ctrl * ctrl,int req_state,bool is_enable,int mode) wcd_clsh_state_hph_l() argument
716 wcd_clsh_v3_state_ear(struct wcd_clsh_ctrl * ctrl,int req_state,bool is_enable,int mode) wcd_clsh_v3_state_ear() argument
742 wcd_clsh_state_ear(struct wcd_clsh_ctrl * ctrl,int req_state,bool is_enable,int mode) wcd_clsh_state_ear() argument
777 _wcd_clsh_ctrl_set_state(struct wcd_clsh_ctrl * ctrl,int req_state,bool is_enable,int mode) _wcd_clsh_ctrl_set_state() argument
844 wcd_clsh_ctrl_set_state(struct wcd_clsh_ctrl * ctrl,enum wcd_clsh_event clsh_event,int nstate,enum wcd_clsh_mode mode) wcd_clsh_ctrl_set_state() argument
875 wcd_clsh_ctrl_get_state(struct wcd_clsh_ctrl * ctrl) wcd_clsh_ctrl_get_state() argument
884 struct wcd_clsh_ctrl *ctrl; wcd_clsh_ctrl_alloc() local
898 wcd_clsh_ctrl_free(struct wcd_clsh_ctrl * ctrl) wcd_clsh_ctrl_free() argument
[all...]
/linux/drivers/gpio/
H A Dgpio-en7523.c33 struct airoha_gpio_ctrl *ctrl = gpiochip_get_data(gc); in airoha_dir_set() local
34 u32 dir = ioread32(ctrl->dir[gpio / 16]); in airoha_dir_set()
35 u32 output = ioread32(ctrl->output); in airoha_dir_set()
46 iowrite32(dir, ctrl->dir[gpio / 16]); in airoha_dir_set()
49 gpio_generic_chip_set(&ctrl->gen_gc, gpio, val); in airoha_dir_set()
51 iowrite32(output, ctrl->output); in airoha_dir_set()
69 struct airoha_gpio_ctrl *ctrl = gpiochip_get_data(gc); in airoha_get_dir() local
70 u32 dir = ioread32(ctrl->dir[gpio / 16]); in airoha_get_dir()
80 struct airoha_gpio_ctrl *ctrl; in airoha_gpio_probe() local
83 ctrl = devm_kzalloc(dev, sizeof(*ctrl), GFP_KERNEL); in airoha_gpio_probe()
[all …]
/linux/arch/arm/kernel/
H A Dunwind.c211 static unsigned long unwind_get_byte(struct unwind_ctrl_block *ctrl) in unwind_get_byte() argument
215 if (ctrl->entries <= 0) { in unwind_get_byte()
220 ret = (*ctrl->insn >> (ctrl->byte * 8)) & 0xff; in unwind_get_byte()
222 if (ctrl->byte == 0) { in unwind_get_byte()
223 ctrl->insn++; in unwind_get_byte()
224 ctrl->entries--; in unwind_get_byte()
225 ctrl->byte = 3; in unwind_get_byte()
227 ctrl->byte--; in unwind_get_byte()
233 static int unwind_pop_register(struct unwind_ctrl_block *ctrl, in unwind_pop_register() argument
236 if (unlikely(ctrl->check_each_pop)) in unwind_pop_register()
[all …]

12345678910>>...80