Lines Matching full:agent

278 	struct data_vio *agent = as_data_vio(completion);  in finish_compressed_write()  local
281 assert_data_vio_in_allocated_zone(agent); in finish_compressed_write()
284 * Process all the non-agent waiters first to ensure that the pbn lock can not be released in finish_compressed_write()
287 for (client = agent->compression.next_in_batch; client != NULL; client = next) { in finish_compressed_write()
289 release_compressed_write_waiter(client, &agent->allocation); in finish_compressed_write()
293 release_compressed_write_waiter(agent, &agent->allocation); in finish_compressed_write()
298 struct data_vio *agent = as_data_vio(completion); in handle_compressed_write_error() local
299 struct allocation *allocation = &agent->allocation; in handle_compressed_write_error()
309 for (client = agent->compression.next_in_batch; client != NULL; client = next) { in handle_compressed_write_error()
317 write_data_vio(agent); in handle_compressed_write_error()
361 * @size: The size of the agent's fragment.
363 * This method initializes the compressed block in the compressed write agent. Because the
364 * compressor already put the agent's compressed fragment at the start of the compressed block's
366 * the agent's fragment.
383 * @compression: The agent's compression_state to pack in to.
431 struct data_vio *agent = remove_from_bin(packer, bin); in write_bin() local
435 if (agent == NULL) in write_bin()
438 compression = &agent->compression; in write_bin()
452 abort_packing(agent); in write_bin()
462 agent->vio.completion.error_handler = handle_compressed_write_error; in write_bin()
463 if (vdo_is_read_only(vdo_from_data_vio(agent))) { in write_bin()
464 continue_data_vio_with_error(agent, VDO_READ_ONLY); in write_bin()
468 result = vio_reset_bio(&agent->vio, (char *) block, compressed_write_end_io, in write_bin()
469 REQ_OP_WRITE, agent->allocation.pbn); in write_bin()
471 continue_data_vio_with_error(agent, result); in write_bin()
487 vdo_submit_data_vio(agent); in write_bin()