/linux/drivers/s390/char/ |
H A D | tape_std.c | 39 struct tape_request * request = timer_container_of(request, t, in tape_std_assign_timeout() local 41 struct tape_device * device = request->device; in tape_std_assign_timeout() 48 rc = tape_cancel_io(device, request); in tape_std_assign_timeout() 58 struct tape_request *request; in tape_std_assign() local 60 request = tape_alloc_request(2, 11); in tape_std_assign() 61 if (IS_ERR(request)) in tape_std_assign() 62 return PTR_ERR(request); in tape_std_assign() 64 request->op = TO_ASSIGN; in tape_std_assign() 65 tape_ccw_cc(request->cpaddr, ASSIGN, 11, request->cpdata); in tape_std_assign() 66 tape_ccw_end(request->cpaddr + 1, NOP, 0, NULL); in tape_std_assign() [all …]
|
H A D | tape_core.c | 293 __tape_cancel_io(struct tape_device *device, struct tape_request *request) in __tape_cancel_io() argument 299 if (request->callback == NULL) in __tape_cancel_io() 304 rc = ccw_device_clear(device->cdev, (long) request); in __tape_cancel_io() 308 request->status = TAPE_REQUEST_DONE; in __tape_cancel_io() 311 request->status = TAPE_REQUEST_CANCEL; in __tape_cancel_io() 435 * request. We may prevent this by returning an error. 593 struct tape_request * request; in __tape_discard_requests() local 597 request = list_entry(l, struct tape_request, list); in __tape_discard_requests() 598 if (request->status == TAPE_REQUEST_IN_IO) in __tape_discard_requests() 599 request->status = TAPE_REQUEST_DONE; in __tape_discard_requests() [all …]
|
H A D | tape_34xx.c | 57 static void __tape_34xx_medium_sense(struct tape_request *request) in __tape_34xx_medium_sense() argument 59 struct tape_device *device = request->device; in __tape_34xx_medium_sense() 62 if (request->rc == 0) { in __tape_34xx_medium_sense() 63 sense = request->cpdata; in __tape_34xx_medium_sense() 82 request->rc); in __tape_34xx_medium_sense() 83 tape_free_request(request); in __tape_34xx_medium_sense() 88 struct tape_request *request; in tape_34xx_medium_sense() local 91 request = tape_alloc_request(1, 32); in tape_34xx_medium_sense() 92 if (IS_ERR(request)) { in tape_34xx_medium_sense() 94 return PTR_ERR(request); in tape_34xx_medium_sense() [all …]
|
H A D | tape_3590.c | 205 struct tape_request *request; in tape_3592_kekl_query() local 214 request = tape_alloc_request(2, sizeof(*order)); in tape_3592_kekl_query() 215 if (IS_ERR(request)) { in tape_3592_kekl_query() 216 rc = PTR_ERR(request); in tape_3592_kekl_query() 219 order = request->cpdata; in tape_3592_kekl_query() 223 request->op = TO_KEKL_QUERY; in tape_3592_kekl_query() 224 tape_ccw_cc(request->cpaddr, PERF_SUBSYS_FUNC, sizeof(*order), order); in tape_3592_kekl_query() 225 tape_ccw_end(request->cpaddr + 1, READ_SS_DATA, sizeof(*int_kekls), in tape_3592_kekl_query() 227 rc = tape_do_io(device, request); in tape_3592_kekl_query() 234 tape_free_request(request); in tape_3592_kekl_query() [all …]
|
/linux/drivers/greybus/ |
H A D | svc.c | 176 struct gb_svc_pwrmon_sample_get_request request; in gb_svc_pwrmon_sample_get() local 180 request.rail_id = rail_id; in gb_svc_pwrmon_sample_get() 181 request.measurement_type = measurement_type; in gb_svc_pwrmon_sample_get() 184 &request, sizeof(request), in gb_svc_pwrmon_sample_get() 213 struct gb_svc_pwrmon_intf_sample_get_request request; in gb_svc_pwrmon_intf_sample_get() local 217 request.intf_id = intf_id; in gb_svc_pwrmon_intf_sample_get() 218 request.measurement_type = measurement_type; in gb_svc_pwrmon_intf_sample_get() 222 &request, sizeof(request), in gb_svc_pwrmon_intf_sample_get() 260 struct gb_svc_intf_device_id_request request; in gb_svc_intf_device_id() local 262 request.intf_id = intf_id; in gb_svc_intf_device_id() [all …]
|
/linux/include/linux/surface_aggregator/ |
H A D | controller.h | 54 * Specifies that the request expects a response. If not set, the request 56 * transmitted. If set, the request transport system waits for a response 57 * of the request. 60 * Specifies that the request should be transmitted via an unsequenced 61 * packet. If set, the request must not have a response, meaning that this 70 * struct ssam_request - SAM request description. 71 * @target_category: Category of the request's target. See &enum ssam_ssh_tc. 72 * @target_id: ID of the request's target. 73 * @command_id: Command ID of the request. 74 * @instance_id: Instance ID of the request's target. [all …]
|
/linux/Documentation/userspace-api/media/mediactl/ |
H A D | request-api.rst | 4 .. _media-request-api: 6 Request API 9 The Request API has been designed to allow V4L2 to deal with requirements of 19 Supporting these features without the Request API is not always possible and if 26 The Request API allows a specific configuration of the pipeline (media 31 of request completion are also available for reading. 36 The Request API extends the Media Controller API and cooperates with 37 subsystem-specific APIs to support request usage. At the Media Controller 39 node. Their life cycle is then managed through the request file descriptors in 42 request support, such as V4L2 APIs that take an explicit ``request_fd`` [all …]
|
H A D | media-request-ioc-queue.rst | 13 MEDIA_REQUEST_IOC_QUEUE - Queue a request 31 If the media device supports :ref:`requests <media-request-api>`, then 32 this request ioctl can be used to queue a previously allocated request. 34 If the request was successfully queued, then the file descriptor can be 35 :ref:`polled <request-func-poll>` to wait for the request to complete. 37 If the request was already queued before, then ``EBUSY`` is returned. 38 Other errors can be returned if the contents of the request contained 40 common error codes. On error both the request and driver state are unchanged. 42 Once a request is queued, then the driver is required to gracefully handle 43 errors that occur when the request is applied to the hardware. The [all …]
|
/linux/sound/soc/intel/catpt/ |
H A D | messages.c | 17 struct catpt_ipc_msg request = {{0}}, reply; in catpt_ipc_get_fw_version() local 20 request.header = msg.val; in catpt_ipc_get_fw_version() 24 ret = catpt_dsp_send_msg(cdev, request, &reply); in catpt_ipc_get_fw_version() 58 struct catpt_ipc_msg request, reply; in catpt_ipc_alloc_stream() local 91 request.header = msg.val; in catpt_ipc_alloc_stream() 92 request.size = size; in catpt_ipc_alloc_stream() 93 request.data = payload; in catpt_ipc_alloc_stream() 97 ret = catpt_dsp_send_msg(cdev, request, &reply); in catpt_ipc_alloc_stream() 109 struct catpt_ipc_msg request; in catpt_ipc_free_stream() local 112 request.header = msg.val; in catpt_ipc_free_stream() [all …]
|
/linux/include/linux/ |
H A D | blk-mq.h | 25 typedef enum rq_end_io_ret (rq_end_io_fn)(struct request *, blk_status_t); 28 * request flags */ 35 /* request for flush sequence */ 43 /* use an I/O scheduler for this request */ 51 /* runtime pm request */ 60 /* request completion needs to be signaled to zone write plugging. */ 103 struct request { struct 125 struct request *rq_next; argument 130 /* Time that the first bio started allocating this request. */ argument 133 /* Time that this request was allocated for this IO. */ argument [all …]
|
/linux/fs/nfsd/ |
H A D | xdr4.h | 212 u32 ac_req_access; /* request */ 218 u32 cl_seqid; /* request */ 219 stateid_t cl_stateid; /* request+response */ 223 u64 co_offset; /* request */ 224 u32 co_count; /* request */ 229 u32 cr_namelen; /* request */ 230 char * cr_name; /* request */ 231 u32 cr_type; /* request */ 232 union { /* request */ 243 u32 cr_bmval[3]; /* request */ [all …]
|
/linux/block/ |
H A D | elevator.h | 42 bool (*allow_merge)(struct request_queue *, struct request *, struct bio *); 44 int (*request_merge)(struct request_queue *q, struct request **, struct bio *); 45 void (*request_merged)(struct request_queue *, struct request *, enum elv_merge); 46 void (*requests_merged)(struct request_queue *, struct request *, struct request *); 48 void (*prepare_request)(struct request *); 49 void (*finish_request)(struct request *); 52 struct request *(*dispatch_request)(struct blk_mq_hw_ctx *); 54 void (*completed_request)(struct request *, u64); 55 void (*requeue_request)(struct request *); 56 struct request *(*former_request)(struct request_queue *, struct request *); [all …]
|
H A D | blk-crypto-internal.h | 32 bool bio_crypt_rq_ctx_compatible(struct request *rq, struct bio *bio); 37 static inline bool bio_crypt_ctx_back_mergeable(struct request *req, in bio_crypt_ctx_back_mergeable() 44 static inline bool bio_crypt_ctx_front_mergeable(struct request *req, in bio_crypt_ctx_front_mergeable() 51 static inline bool bio_crypt_ctx_merge_rq(struct request *req, in bio_crypt_ctx_merge_rq() 52 struct request *next) in bio_crypt_ctx_merge_rq() 58 static inline void blk_crypto_rq_set_defaults(struct request *rq) in blk_crypto_rq_set_defaults() 64 static inline bool blk_crypto_rq_is_encrypted(struct request *rq) in blk_crypto_rq_is_encrypted() 69 static inline bool blk_crypto_rq_has_keyslot(struct request *rq) in blk_crypto_rq_has_keyslot() 100 static inline bool bio_crypt_rq_ctx_compatible(struct request *rq, in bio_crypt_rq_ctx_compatible() 106 static inline bool bio_crypt_ctx_front_mergeable(struct request *req, in bio_crypt_ctx_front_mergeable() [all …]
|
/linux/tools/perf/pmu-events/arch/riscv/ |
H A D | riscv-sbi-firmware.json | 51 "PublicDescription": "Sent FENCE.I request to other HART event", 54 "BriefDescription": "Sent FENCE.I request to other HART event" 57 "PublicDescription": "Received FENCE.I request from other HART event", 60 "BriefDescription": "Received FENCE.I request from other HART event" 63 "PublicDescription": "Sent SFENCE.VMA request to other HART event", 66 "BriefDescription": "Sent SFENCE.VMA request to other HART event" 69 "PublicDescription": "Received SFENCE.VMA request from other HART event", 72 "BriefDescription": "Received SFENCE.VMA request from other HART event" 75 "PublicDescription": "Sent SFENCE.VMA with ASID request to other HART event", 78 "BriefDescription": "Sent SFENCE.VMA with ASID request to other HART event" [all …]
|
/linux/rust/kernel/block/mq/ |
H A D | request.rs | 3 //! This module provides a wrapper for the C `struct request` type. 19 /// A wrapper around a blk-mq [`struct request`]. This represents an IO request. 23 /// There are four states for a request that the Rust bindings care about: 25 /// 1. Request is owned by block layer (refcount 0). 26 /// 2. Request is owned by driver but with zero [`ARef`]s in existence 28 /// 3. Request is owned by driver with exactly one [`ARef`] in existence 30 /// 4. Request is owned by driver with more than one [`ARef`] in existence 34 /// We need to track 1 and 2 to ensure we fail tag to request conversions for 37 /// We need to track 3 and 4 to ensure that it is safe to end the request and hand 42 /// [`struct request`]. [all …]
|
/linux/drivers/usb/musb/ |
H A D | musb_gadget.c | 33 static inline void map_dma_buffer(struct musb_request *request, in map_dma_buffer() argument 39 request->map_state = UN_MAPPED; in map_dma_buffer() 44 /* Check if DMA engine can handle this request. in map_dma_buffer() 45 * DMA code must reject the USB request explicitly. in map_dma_buffer() 46 * Default behaviour is to map the request. in map_dma_buffer() 50 musb_ep->packet_sz, request->request.buf, in map_dma_buffer() 51 request->request.length); in map_dma_buffer() 55 if (request->request.dma == DMA_ADDR_INVALID) { in map_dma_buffer() 61 request->request.buf, in map_dma_buffer() 62 request->request.length, in map_dma_buffer() [all …]
|
/linux/drivers/staging/greybus/ |
H A D | gpio.c | 64 struct gb_gpio_activate_request request; in gb_gpio_activate_operation() local 72 request.which = which; in gb_gpio_activate_operation() 74 &request, sizeof(request), NULL, 0); in gb_gpio_activate_operation() 90 struct gb_gpio_deactivate_request request; in gb_gpio_deactivate_operation() local 93 request.which = which; in gb_gpio_deactivate_operation() 95 &request, sizeof(request), NULL, 0); in gb_gpio_deactivate_operation() 111 struct gb_gpio_get_direction_request request; in gb_gpio_get_direction_operation() local 116 request.which = which; in gb_gpio_get_direction_operation() 118 &request, sizeof(request), in gb_gpio_get_direction_operation() 135 struct gb_gpio_direction_in_request request; in gb_gpio_direction_in_operation() local [all …]
|
/linux/net/ethtool/ |
H A D | eeprom.c | 29 static int fallback_set_params(struct eeprom_req_info *request, in fallback_set_params() argument 33 u32 offset = request->offset; in fallback_set_params() 34 u32 length = request->length; in fallback_set_params() 36 if (request->page) in fallback_set_params() 37 offset = request->page * ETH_MODULE_EEPROM_PAGE_LEN + offset; in fallback_set_params() 40 request->i2c_address == 0x51) in fallback_set_params() 53 static int eeprom_fallback(struct eeprom_req_info *request, in eeprom_fallback() argument 67 err = fallback_set_params(request, &modinfo, &eeprom); in eeprom_fallback() 114 struct eeprom_req_info *request = MODULE_EEPROM_REQINFO(req_base); in eeprom_prepare_data() local 119 page_data.offset = request->offset; in eeprom_prepare_data() [all …]
|
/linux/tools/perf/pmu-events/arch/x86/westmereep-dp/ |
H A D | memory.json | 11 "BriefDescription": "REQUEST = ANY_DATA read and RESPONSE = ANY_DRAM AND REMOTE_FWD", 21 "BriefDescription": "REQUEST = ANY_DATA read and RESPONSE = ANY_LLC_MISS", 31 "BriefDescription": "REQUEST = ANY_DATA read and RESPONSE = OTHER_LOCAL_DRAM", 41 "BriefDescription": "REQUEST = ANY_DATA read and RESPONSE = REMOTE_DRAM", 51 "BriefDescription": "REQUEST = ANY IFETCH and RESPONSE = ANY_DRAM AND REMOTE_FWD", 61 "BriefDescription": "REQUEST = ANY IFETCH and RESPONSE = ANY_LLC_MISS", 71 "BriefDescription": "REQUEST = ANY IFETCH and RESPONSE = OTHER_LOCAL_DRAM", 81 "BriefDescription": "REQUEST = ANY IFETCH and RESPONSE = REMOTE_DRAM", 91 "BriefDescription": "REQUEST = ANY_REQUEST and RESPONSE = ANY_DRAM AND REMOTE_FWD", 101 "BriefDescription": "REQUEST = ANY_REQUEST and RESPONSE = ANY_LLC_MISS", [all …]
|
/linux/drivers/media/usb/as102/ |
H A D | as10x_cmd.h | 26 /* context request types */ 83 /* request */ 85 /* request identifier */ 98 /* request */ 100 /* request identifier */ 113 /* request */ 115 /* request identifier */ 130 /* request */ 132 /* request identifier */ 147 /* request */ [all …]
|
/linux/drivers/md/dm-vdo/indexer/ |
H A D | index.c | 43 * request wants to add a chapter to the sparse cache, it sends a barrier message to each zone 89 struct uds_request *request; in launch_zone_message() local 91 result = vdo_allocate(1, struct uds_request, __func__, &request); in launch_zone_message() 95 request->index = index; in launch_zone_message() 96 request->unbatched = true; in launch_zone_message() 97 request->zone_number = zone; in launch_zone_message() 98 request->zone_message = message; in launch_zone_message() 100 uds_enqueue_request(request, STAGE_MESSAGE); in launch_zone_message() 120 * Determine whether this request should trigger a sparse cache barrier message to change the 124 static u64 triage_index_request(struct uds_index *index, struct uds_request *request) in triage_index_request() argument [all …]
|
/linux/drivers/scsi/fnic/ |
H A D | fcpio.h | 29 * Initiator request types 38 * Target request types 52 * Misc request types 72 FCPIO_SUCCESS = 0, /* request was successful */ 75 * If a request to the firmware is rejected, the original request 79 FCPIO_OUT_OF_RESOURCE, /* out of resources to complete request */ 80 FCPIO_INVALID_PARAM, /* some parameter in request is invalid */ 81 FCPIO_REQ_NOT_SUPPORTED, /* request type is not supported */ 85 * Once a request is processed, the firmware will usually return 89 FCPIO_ABORTED = 0x41, /* request was aborted */ [all …]
|
/linux/Documentation/driver-api/surface_aggregator/ |
H A D | internal.rst | 70 Above this sits the *request transport layer (RTL)*. This layer is centered 73 It, specifically, distinguishes events from request responses, matches 74 responses to their corresponding requests, and implements request timeouts. 77 how request responses and, especially, events are dealt with. It provides an 79 workqueue for event and asynchronous request completion, and also manages 217 the request transport layer. 267 somewhat special. It is either set when the upper layer request is submitted 288 Request Transport Layer 291 The request transport layer is represented via |ssh_rtl| and builds on top 297 the corresponding callback. The request transport layer is structured around [all …]
|
/linux/tools/perf/pmu-events/arch/x86/broadwell/ |
H A D | uncore-cache.json | 3 … "BriefDescription": "L3 Lookup any request that access cache and found line in E or S-state", 8 … "PublicDescription": "L3 Lookup any request that access cache and found line in E or S-state.", 13 "BriefDescription": "L3 Lookup any request that access cache and found line in I-state", 18 "PublicDescription": "L3 Lookup any request that access cache and found line in I-state.", 23 "BriefDescription": "L3 Lookup any request that access cache and found line in M-state", 28 "PublicDescription": "L3 Lookup any request that access cache and found line in M-state.", 33 "BriefDescription": "L3 Lookup any request that access cache and found line in MESI-state", 38 … "PublicDescription": "L3 Lookup any request that access cache and found line in MESI-state.", 43 … "BriefDescription": "L3 Lookup read request that access cache and found line in E or S-state", 48 … "PublicDescription": "L3 Lookup read request that access cache and found line in E or S-state.", [all …]
|
/linux/drivers/firewire/ |
H A D | core-transaction.c | 97 * If the request packet has already been sent, we need to see in fw_cancel_transaction() 279 * __fw_send_request() - submit a request packet for transmission to generate callback for response 281 * @card: interface to send the request at 282 * @t: transaction instance to which the request belongs 285 * @generation: bus generation in which request and response are valid 288 * @payload: data payload for the request subaction 295 * Submit a request packet into the asynchronous request transmission queue. 303 * @generation. Otherwise the request is in danger to be sent to a wrong node. 314 * In case of request types without payload, @data is NULL and @length is 0. 321 * %RCODE_NO_ACK) denote transaction timeout, busy responder, stale request [all …]
|