Lines Matching full:agent

277 	struct data_vio *agent = as_data_vio(completion);  in finish_compressed_write()  local
280 assert_data_vio_in_allocated_zone(agent); in finish_compressed_write()
283 * Process all the non-agent waiters first to ensure that the pbn lock can not be released in finish_compressed_write()
286 for (client = agent->compression.next_in_batch; client != NULL; client = next) { in finish_compressed_write()
288 release_compressed_write_waiter(client, &agent->allocation); in finish_compressed_write()
292 release_compressed_write_waiter(agent, &agent->allocation); in finish_compressed_write()
297 struct data_vio *agent = as_data_vio(completion); in handle_compressed_write_error() local
298 struct allocation *allocation = &agent->allocation; in handle_compressed_write_error()
308 for (client = agent->compression.next_in_batch; client != NULL; client = next) { in handle_compressed_write_error()
316 write_data_vio(agent); in handle_compressed_write_error()
360 * @size: The size of the agent's fragment.
362 * This method initializes the compressed block in the compressed write agent. Because the
363 * compressor already put the agent's compressed fragment at the start of the compressed block's
365 * the agent's fragment.
382 * @compression: The agent's compression_state to pack in to.
430 struct data_vio *agent = remove_from_bin(packer, bin); in write_bin() local
434 if (agent == NULL) in write_bin()
437 compression = &agent->compression; in write_bin()
451 abort_packing(agent); in write_bin()
461 agent->vio.completion.error_handler = handle_compressed_write_error; in write_bin()
462 if (vdo_is_read_only(vdo_from_data_vio(agent))) { in write_bin()
463 continue_data_vio_with_error(agent, VDO_READ_ONLY); in write_bin()
467 result = vio_reset_bio(&agent->vio, (char *) block, compressed_write_end_io, in write_bin()
468 REQ_OP_WRITE, agent->allocation.pbn); in write_bin()
470 continue_data_vio_with_error(agent, result); in write_bin()
486 vdo_submit_data_vio(agent); in write_bin()