1 #ifndef TARGET_CORE_BASE_H 2 #define TARGET_CORE_BASE_H 3 4 #include <linux/in.h> 5 #include <linux/configfs.h> 6 #include <linux/dma-mapping.h> 7 #include <linux/blkdev.h> 8 #include <scsi/scsi_cmnd.h> 9 #include <net/sock.h> 10 #include <net/tcp.h> 11 12 #define TARGET_CORE_MOD_VERSION "v4.0.0-rc7-ml" 13 #define SHUTDOWN_SIGS (sigmask(SIGKILL)|sigmask(SIGINT)|sigmask(SIGABRT)) 14 15 /* Used by transport_generic_allocate_iovecs() */ 16 #define TRANSPORT_IOV_DATA_BUFFER 5 17 /* Maximum Number of LUNs per Target Portal Group */ 18 #define TRANSPORT_MAX_LUNS_PER_TPG 256 19 /* 20 * By default we use 32-byte CDBs in TCM Core and subsystem plugin code. 21 * 22 * Note that both include/scsi/scsi_cmnd.h:MAX_COMMAND_SIZE and 23 * include/linux/blkdev.h:BLOCK_MAX_CDB as of v2.6.36-rc4 still use 24 * 16-byte CDBs by default and require an extra allocation for 25 * 32-byte CDBs to because of legacy issues. 26 * 27 * Within TCM Core there are no such legacy limitiations, so we go ahead 28 * use 32-byte CDBs by default and use include/scsi/scsi.h:scsi_command_size() 29 * within all TCM Core and subsystem plugin code. 30 */ 31 #define TCM_MAX_COMMAND_SIZE 32 32 /* 33 * From include/scsi/scsi_cmnd.h:SCSI_SENSE_BUFFERSIZE, currently 34 * defined 96, but the real limit is 252 (or 260 including the header) 35 */ 36 #define TRANSPORT_SENSE_BUFFER SCSI_SENSE_BUFFERSIZE 37 /* Used by transport_send_check_condition_and_sense() */ 38 #define SPC_SENSE_KEY_OFFSET 2 39 #define SPC_ASC_KEY_OFFSET 12 40 #define SPC_ASCQ_KEY_OFFSET 13 41 #define TRANSPORT_IQN_LEN 224 42 /* Used by target_core_store_alua_lu_gp() and target_core_alua_lu_gp_show_attr_members() */ 43 #define LU_GROUP_NAME_BUF 256 44 /* Used by core_alua_store_tg_pt_gp_info() and target_core_alua_tg_pt_gp_show_attr_members() */ 45 #define TG_PT_GROUP_NAME_BUF 256 46 /* Used to parse VPD into struct t10_vpd */ 47 #define VPD_TMP_BUF_SIZE 128 48 /* Used by transport_generic_cmd_sequencer() */ 49 #define READ_BLOCK_LEN 6 50 #define READ_CAP_LEN 8 51 #define READ_POSITION_LEN 20 52 #define INQUIRY_LEN 36 53 /* Used by transport_get_inquiry_vpd_serial() */ 54 #define INQUIRY_VPD_SERIAL_LEN 254 55 /* Used by transport_get_inquiry_vpd_device_ident() */ 56 #define INQUIRY_VPD_DEVICE_IDENTIFIER_LEN 254 57 58 /* struct se_hba->hba_flags */ 59 enum hba_flags_table { 60 HBA_FLAGS_INTERNAL_USE = 0x01, 61 HBA_FLAGS_PSCSI_MODE = 0x02, 62 }; 63 64 /* struct se_lun->lun_status */ 65 enum transport_lun_status_table { 66 TRANSPORT_LUN_STATUS_FREE = 0, 67 TRANSPORT_LUN_STATUS_ACTIVE = 1, 68 }; 69 70 /* struct se_portal_group->se_tpg_type */ 71 enum transport_tpg_type_table { 72 TRANSPORT_TPG_TYPE_NORMAL = 0, 73 TRANSPORT_TPG_TYPE_DISCOVERY = 1, 74 }; 75 76 /* Used for generate timer flags */ 77 enum timer_flags_table { 78 TF_RUNNING = 0x01, 79 TF_STOP = 0x02, 80 }; 81 82 /* Special transport agnostic struct se_cmd->t_states */ 83 enum transport_state_table { 84 TRANSPORT_NO_STATE = 0, 85 TRANSPORT_NEW_CMD = 1, 86 TRANSPORT_DEFERRED_CMD = 2, 87 TRANSPORT_WRITE_PENDING = 3, 88 TRANSPORT_PROCESS_WRITE = 4, 89 TRANSPORT_PROCESSING = 5, 90 TRANSPORT_COMPLETE_OK = 6, 91 TRANSPORT_COMPLETE_FAILURE = 7, 92 TRANSPORT_COMPLETE_TIMEOUT = 8, 93 TRANSPORT_PROCESS_TMR = 9, 94 TRANSPORT_TMR_COMPLETE = 10, 95 TRANSPORT_ISTATE_PROCESSING = 11, 96 TRANSPORT_ISTATE_PROCESSED = 12, 97 TRANSPORT_KILL = 13, 98 TRANSPORT_REMOVE = 14, 99 TRANSPORT_FREE = 15, 100 TRANSPORT_NEW_CMD_MAP = 16, 101 TRANSPORT_FREE_CMD_INTR = 17, 102 }; 103 104 /* Used for struct se_cmd->se_cmd_flags */ 105 enum se_cmd_flags_table { 106 SCF_SUPPORTED_SAM_OPCODE = 0x00000001, 107 SCF_TRANSPORT_TASK_SENSE = 0x00000002, 108 SCF_EMULATED_TASK_SENSE = 0x00000004, 109 SCF_SCSI_DATA_SG_IO_CDB = 0x00000008, 110 SCF_SCSI_CONTROL_SG_IO_CDB = 0x00000010, 111 SCF_SCSI_CONTROL_NONSG_IO_CDB = 0x00000020, 112 SCF_SCSI_NON_DATA_CDB = 0x00000040, 113 SCF_SCSI_CDB_EXCEPTION = 0x00000080, 114 SCF_SCSI_RESERVATION_CONFLICT = 0x00000100, 115 SCF_CMD_PASSTHROUGH_NOALLOC = 0x00000200, 116 SCF_SE_CMD_FAILED = 0x00000400, 117 SCF_SE_LUN_CMD = 0x00000800, 118 SCF_SE_ALLOW_EOO = 0x00001000, 119 SCF_SE_DISABLE_ONLINE_CHECK = 0x00002000, 120 SCF_SENT_CHECK_CONDITION = 0x00004000, 121 SCF_OVERFLOW_BIT = 0x00008000, 122 SCF_UNDERFLOW_BIT = 0x00010000, 123 SCF_SENT_DELAYED_TAS = 0x00020000, 124 SCF_ALUA_NON_OPTIMIZED = 0x00040000, 125 SCF_DELAYED_CMD_FROM_SAM_ATTR = 0x00080000, 126 SCF_PASSTHROUGH_SG_TO_MEM = 0x00100000, 127 SCF_PASSTHROUGH_CONTIG_TO_SG = 0x00200000, 128 SCF_PASSTHROUGH_SG_TO_MEM_NOALLOC = 0x00400000, 129 SCF_EMULATE_SYNC_CACHE = 0x00800000, 130 SCF_EMULATE_CDB_ASYNC = 0x01000000, 131 SCF_EMULATE_SYNC_UNMAP = 0x02000000 132 }; 133 134 /* struct se_dev_entry->lun_flags and struct se_lun->lun_access */ 135 enum transport_lunflags_table { 136 TRANSPORT_LUNFLAGS_NO_ACCESS = 0x00, 137 TRANSPORT_LUNFLAGS_INITIATOR_ACCESS = 0x01, 138 TRANSPORT_LUNFLAGS_READ_ONLY = 0x02, 139 TRANSPORT_LUNFLAGS_READ_WRITE = 0x04, 140 }; 141 142 /* struct se_device->dev_status */ 143 enum transport_device_status_table { 144 TRANSPORT_DEVICE_ACTIVATED = 0x01, 145 TRANSPORT_DEVICE_DEACTIVATED = 0x02, 146 TRANSPORT_DEVICE_QUEUE_FULL = 0x04, 147 TRANSPORT_DEVICE_SHUTDOWN = 0x08, 148 TRANSPORT_DEVICE_OFFLINE_ACTIVATED = 0x10, 149 TRANSPORT_DEVICE_OFFLINE_DEACTIVATED = 0x20, 150 }; 151 152 /* 153 * Used by transport_send_check_condition_and_sense() and se_cmd->scsi_sense_reason 154 * to signal which ASC/ASCQ sense payload should be built. 155 */ 156 enum tcm_sense_reason_table { 157 TCM_NON_EXISTENT_LUN = 0x01, 158 TCM_UNSUPPORTED_SCSI_OPCODE = 0x02, 159 TCM_INCORRECT_AMOUNT_OF_DATA = 0x03, 160 TCM_UNEXPECTED_UNSOLICITED_DATA = 0x04, 161 TCM_SERVICE_CRC_ERROR = 0x05, 162 TCM_SNACK_REJECTED = 0x06, 163 TCM_SECTOR_COUNT_TOO_MANY = 0x07, 164 TCM_INVALID_CDB_FIELD = 0x08, 165 TCM_INVALID_PARAMETER_LIST = 0x09, 166 TCM_LOGICAL_UNIT_COMMUNICATION_FAILURE = 0x0a, 167 TCM_UNKNOWN_MODE_PAGE = 0x0b, 168 TCM_WRITE_PROTECTED = 0x0c, 169 TCM_CHECK_CONDITION_ABORT_CMD = 0x0d, 170 TCM_CHECK_CONDITION_UNIT_ATTENTION = 0x0e, 171 TCM_CHECK_CONDITION_NOT_READY = 0x0f, 172 }; 173 174 struct se_obj { 175 atomic_t obj_access_count; 176 } ____cacheline_aligned; 177 178 /* 179 * Used by TCM Core internally to signal if ALUA emulation is enabled or 180 * disabled, or running in with TCM/pSCSI passthrough mode 181 */ 182 typedef enum { 183 SPC_ALUA_PASSTHROUGH, 184 SPC2_ALUA_DISABLED, 185 SPC3_ALUA_EMULATED 186 } t10_alua_index_t; 187 188 /* 189 * Used by TCM Core internally to signal if SAM Task Attribute emulation 190 * is enabled or disabled, or running in with TCM/pSCSI passthrough mode 191 */ 192 typedef enum { 193 SAM_TASK_ATTR_PASSTHROUGH, 194 SAM_TASK_ATTR_UNTAGGED, 195 SAM_TASK_ATTR_EMULATED 196 } t10_task_attr_index_t; 197 198 /* 199 * Used for target SCSI statistics 200 */ 201 typedef enum { 202 SCSI_INST_INDEX, 203 SCSI_DEVICE_INDEX, 204 SCSI_AUTH_INTR_INDEX, 205 SCSI_INDEX_TYPE_MAX 206 } scsi_index_t; 207 208 struct scsi_index_table { 209 spinlock_t lock; 210 u32 scsi_mib_index[SCSI_INDEX_TYPE_MAX]; 211 } ____cacheline_aligned; 212 213 struct se_cmd; 214 215 struct t10_alua { 216 t10_alua_index_t alua_type; 217 /* ALUA Target Port Group ID */ 218 u16 alua_tg_pt_gps_counter; 219 u32 alua_tg_pt_gps_count; 220 spinlock_t tg_pt_gps_lock; 221 struct se_subsystem_dev *t10_sub_dev; 222 /* Used for default ALUA Target Port Group */ 223 struct t10_alua_tg_pt_gp *default_tg_pt_gp; 224 /* Used for default ALUA Target Port Group ConfigFS group */ 225 struct config_group alua_tg_pt_gps_group; 226 int (*alua_state_check)(struct se_cmd *, unsigned char *, u8 *); 227 struct list_head tg_pt_gps_list; 228 } ____cacheline_aligned; 229 230 struct t10_alua_lu_gp { 231 u16 lu_gp_id; 232 int lu_gp_valid_id; 233 u32 lu_gp_members; 234 atomic_t lu_gp_shutdown; 235 atomic_t lu_gp_ref_cnt; 236 spinlock_t lu_gp_lock; 237 struct config_group lu_gp_group; 238 struct list_head lu_gp_list; 239 struct list_head lu_gp_mem_list; 240 } ____cacheline_aligned; 241 242 struct t10_alua_lu_gp_member { 243 bool lu_gp_assoc; 244 atomic_t lu_gp_mem_ref_cnt; 245 spinlock_t lu_gp_mem_lock; 246 struct t10_alua_lu_gp *lu_gp; 247 struct se_device *lu_gp_mem_dev; 248 struct list_head lu_gp_mem_list; 249 } ____cacheline_aligned; 250 251 struct t10_alua_tg_pt_gp { 252 u16 tg_pt_gp_id; 253 int tg_pt_gp_valid_id; 254 int tg_pt_gp_alua_access_status; 255 int tg_pt_gp_alua_access_type; 256 int tg_pt_gp_nonop_delay_msecs; 257 int tg_pt_gp_trans_delay_msecs; 258 int tg_pt_gp_pref; 259 int tg_pt_gp_write_metadata; 260 /* Used by struct t10_alua_tg_pt_gp->tg_pt_gp_md_buf_len */ 261 #define ALUA_MD_BUF_LEN 1024 262 u32 tg_pt_gp_md_buf_len; 263 u32 tg_pt_gp_members; 264 atomic_t tg_pt_gp_alua_access_state; 265 atomic_t tg_pt_gp_ref_cnt; 266 spinlock_t tg_pt_gp_lock; 267 struct mutex tg_pt_gp_md_mutex; 268 struct se_subsystem_dev *tg_pt_gp_su_dev; 269 struct config_group tg_pt_gp_group; 270 struct list_head tg_pt_gp_list; 271 struct list_head tg_pt_gp_mem_list; 272 } ____cacheline_aligned; 273 274 struct t10_alua_tg_pt_gp_member { 275 bool tg_pt_gp_assoc; 276 atomic_t tg_pt_gp_mem_ref_cnt; 277 spinlock_t tg_pt_gp_mem_lock; 278 struct t10_alua_tg_pt_gp *tg_pt_gp; 279 struct se_port *tg_pt; 280 struct list_head tg_pt_gp_mem_list; 281 } ____cacheline_aligned; 282 283 struct t10_vpd { 284 unsigned char device_identifier[INQUIRY_VPD_DEVICE_IDENTIFIER_LEN]; 285 int protocol_identifier_set; 286 u32 protocol_identifier; 287 u32 device_identifier_code_set; 288 u32 association; 289 u32 device_identifier_type; 290 struct list_head vpd_list; 291 } ____cacheline_aligned; 292 293 struct t10_wwn { 294 unsigned char vendor[8]; 295 unsigned char model[16]; 296 unsigned char revision[4]; 297 unsigned char unit_serial[INQUIRY_VPD_SERIAL_LEN]; 298 spinlock_t t10_vpd_lock; 299 struct se_subsystem_dev *t10_sub_dev; 300 struct config_group t10_wwn_group; 301 struct list_head t10_vpd_list; 302 } ____cacheline_aligned; 303 304 305 /* 306 * Used by TCM Core internally to signal if >= SPC-3 persistent reservations 307 * emulation is enabled or disabled, or running in with TCM/pSCSI passthrough 308 * mode 309 */ 310 typedef enum { 311 SPC_PASSTHROUGH, 312 SPC2_RESERVATIONS, 313 SPC3_PERSISTENT_RESERVATIONS 314 } t10_reservations_index_t; 315 316 struct t10_pr_registration { 317 /* Used for fabrics that contain WWN+ISID */ 318 #define PR_REG_ISID_LEN 16 319 /* PR_REG_ISID_LEN + ',i,0x' */ 320 #define PR_REG_ISID_ID_LEN (PR_REG_ISID_LEN + 5) 321 char pr_reg_isid[PR_REG_ISID_LEN]; 322 /* Used during APTPL metadata reading */ 323 #define PR_APTPL_MAX_IPORT_LEN 256 324 unsigned char pr_iport[PR_APTPL_MAX_IPORT_LEN]; 325 /* Used during APTPL metadata reading */ 326 #define PR_APTPL_MAX_TPORT_LEN 256 327 unsigned char pr_tport[PR_APTPL_MAX_TPORT_LEN]; 328 /* For writing out live meta data */ 329 unsigned char *pr_aptpl_buf; 330 u16 pr_aptpl_rpti; 331 u16 pr_reg_tpgt; 332 /* Reservation effects all target ports */ 333 int pr_reg_all_tg_pt; 334 /* Activate Persistence across Target Power Loss */ 335 int pr_reg_aptpl; 336 int pr_res_holder; 337 int pr_res_type; 338 int pr_res_scope; 339 /* Used for fabric initiator WWPNs using a ISID */ 340 bool isid_present_at_reg; 341 u32 pr_res_mapped_lun; 342 u32 pr_aptpl_target_lun; 343 u32 pr_res_generation; 344 u64 pr_reg_bin_isid; 345 u64 pr_res_key; 346 atomic_t pr_res_holders; 347 struct se_node_acl *pr_reg_nacl; 348 struct se_dev_entry *pr_reg_deve; 349 struct se_lun *pr_reg_tg_pt_lun; 350 struct list_head pr_reg_list; 351 struct list_head pr_reg_abort_list; 352 struct list_head pr_reg_aptpl_list; 353 struct list_head pr_reg_atp_list; 354 struct list_head pr_reg_atp_mem_list; 355 } ____cacheline_aligned; 356 357 /* 358 * This set of function pointer ops is set based upon SPC3_PERSISTENT_RESERVATIONS, 359 * SPC2_RESERVATIONS or SPC_PASSTHROUGH in drivers/target/target_core_pr.c: 360 * core_setup_reservations() 361 */ 362 struct t10_reservation_ops { 363 int (*t10_reservation_check)(struct se_cmd *, u32 *); 364 int (*t10_seq_non_holder)(struct se_cmd *, unsigned char *, u32); 365 int (*t10_pr_register)(struct se_cmd *); 366 int (*t10_pr_clear)(struct se_cmd *); 367 }; 368 369 struct t10_reservation_template { 370 /* Reservation effects all target ports */ 371 int pr_all_tg_pt; 372 /* Activate Persistence across Target Power Loss enabled 373 * for SCSI device */ 374 int pr_aptpl_active; 375 /* Used by struct t10_reservation_template->pr_aptpl_buf_len */ 376 #define PR_APTPL_BUF_LEN 8192 377 u32 pr_aptpl_buf_len; 378 u32 pr_generation; 379 t10_reservations_index_t res_type; 380 spinlock_t registration_lock; 381 spinlock_t aptpl_reg_lock; 382 /* 383 * This will always be set by one individual I_T Nexus. 384 * However with all_tg_pt=1, other I_T Nexus from the 385 * same initiator can access PR reg/res info on a different 386 * target port. 387 * 388 * There is also the 'All Registrants' case, where there is 389 * a single *pr_res_holder of the reservation, but all 390 * registrations are considered reservation holders. 391 */ 392 struct se_node_acl *pr_res_holder; 393 struct list_head registration_list; 394 struct list_head aptpl_reg_list; 395 struct t10_reservation_ops pr_ops; 396 } ____cacheline_aligned; 397 398 struct se_queue_req { 399 int state; 400 void *cmd; 401 struct list_head qr_list; 402 } ____cacheline_aligned; 403 404 struct se_queue_obj { 405 atomic_t queue_cnt; 406 spinlock_t cmd_queue_lock; 407 struct list_head qobj_list; 408 wait_queue_head_t thread_wq; 409 } ____cacheline_aligned; 410 411 /* 412 * Used one per struct se_cmd to hold all extra struct se_task 413 * metadata. This structure is setup and allocated in 414 * drivers/target/target_core_transport.c:__transport_alloc_se_cmd() 415 */ 416 struct se_transport_task { 417 unsigned char *t_task_cdb; 418 unsigned char __t_task_cdb[TCM_MAX_COMMAND_SIZE]; 419 unsigned long long t_task_lba; 420 int t_tasks_failed; 421 int t_tasks_fua; 422 bool t_tasks_bidi; 423 u32 t_task_cdbs; 424 u32 t_tasks_check; 425 u32 t_tasks_no; 426 u32 t_tasks_sectors; 427 u32 t_tasks_se_num; 428 u32 t_tasks_se_bidi_num; 429 u32 t_tasks_sg_chained_no; 430 atomic_t t_fe_count; 431 atomic_t t_se_count; 432 atomic_t t_task_cdbs_left; 433 atomic_t t_task_cdbs_ex_left; 434 atomic_t t_task_cdbs_timeout_left; 435 atomic_t t_task_cdbs_sent; 436 atomic_t t_transport_aborted; 437 atomic_t t_transport_active; 438 atomic_t t_transport_complete; 439 atomic_t t_transport_queue_active; 440 atomic_t t_transport_sent; 441 atomic_t t_transport_stop; 442 atomic_t t_transport_timeout; 443 atomic_t transport_dev_active; 444 atomic_t transport_lun_active; 445 atomic_t transport_lun_fe_stop; 446 atomic_t transport_lun_stop; 447 spinlock_t t_state_lock; 448 struct completion t_transport_stop_comp; 449 struct completion transport_lun_fe_stop_comp; 450 struct completion transport_lun_stop_comp; 451 struct scatterlist *t_tasks_sg_chained; 452 struct scatterlist t_tasks_sg_bounce; 453 void *t_task_buf; 454 /* 455 * Used for pre-registered fabric SGL passthrough WRITE and READ 456 * with the special SCF_PASSTHROUGH_CONTIG_TO_SG case for TCM_Loop 457 * and other HW target mode fabric modules. 458 */ 459 struct scatterlist *t_task_pt_sgl; 460 struct list_head *t_mem_list; 461 /* Used for BIDI READ */ 462 struct list_head *t_mem_bidi_list; 463 struct list_head t_task_list; 464 } ____cacheline_aligned; 465 466 struct se_task { 467 unsigned char task_sense; 468 struct scatterlist *task_sg; 469 struct scatterlist *task_sg_bidi; 470 u8 task_scsi_status; 471 u8 task_flags; 472 int task_error_status; 473 int task_state_flags; 474 bool task_padded_sg; 475 unsigned long long task_lba; 476 u32 task_no; 477 u32 task_sectors; 478 u32 task_size; 479 u32 task_sg_num; 480 u32 task_sg_offset; 481 enum dma_data_direction task_data_direction; 482 struct se_cmd *task_se_cmd; 483 struct se_device *se_dev; 484 struct completion task_stop_comp; 485 atomic_t task_active; 486 atomic_t task_execute_queue; 487 atomic_t task_timeout; 488 atomic_t task_sent; 489 atomic_t task_stop; 490 atomic_t task_state_active; 491 struct timer_list task_timer; 492 struct se_device *se_obj_ptr; 493 struct list_head t_list; 494 struct list_head t_execute_list; 495 struct list_head t_state_list; 496 } ____cacheline_aligned; 497 498 #define TASK_CMD(task) ((task)->task_se_cmd) 499 #define TASK_DEV(task) ((task)->se_dev) 500 501 struct se_cmd { 502 /* SAM response code being sent to initiator */ 503 u8 scsi_status; 504 u8 scsi_asc; 505 u8 scsi_ascq; 506 u8 scsi_sense_reason; 507 u16 scsi_sense_length; 508 /* Delay for ALUA Active/NonOptimized state access in milliseconds */ 509 int alua_nonop_delay; 510 /* See include/linux/dma-mapping.h */ 511 enum dma_data_direction data_direction; 512 /* For SAM Task Attribute */ 513 int sam_task_attr; 514 /* Transport protocol dependent state, see transport_state_table */ 515 enum transport_state_table t_state; 516 /* Transport protocol dependent state for out of order CmdSNs */ 517 int deferred_t_state; 518 /* Transport specific error status */ 519 int transport_error_status; 520 /* See se_cmd_flags_table */ 521 u32 se_cmd_flags; 522 u32 se_ordered_id; 523 /* Total size in bytes associated with command */ 524 u32 data_length; 525 /* SCSI Presented Data Transfer Length */ 526 u32 cmd_spdtl; 527 u32 residual_count; 528 u32 orig_fe_lun; 529 /* Persistent Reservation key */ 530 u64 pr_res_key; 531 atomic_t transport_sent; 532 /* Used for sense data */ 533 void *sense_buffer; 534 struct list_head se_delayed_list; 535 struct list_head se_ordered_list; 536 struct list_head se_lun_list; 537 struct se_device *se_dev; 538 struct se_dev_entry *se_deve; 539 struct se_device *se_obj_ptr; 540 struct se_device *se_orig_obj_ptr; 541 struct se_lun *se_lun; 542 /* Only used for internal passthrough and legacy TCM fabric modules */ 543 struct se_session *se_sess; 544 struct se_tmr_req *se_tmr_req; 545 /* t_task is setup to t_task_backstore in transport_init_se_cmd() */ 546 struct se_transport_task *t_task; 547 struct se_transport_task t_task_backstore; 548 struct target_core_fabric_ops *se_tfo; 549 int (*transport_emulate_cdb)(struct se_cmd *); 550 void (*transport_split_cdb)(unsigned long long, u32 *, unsigned char *); 551 void (*transport_wait_for_tasks)(struct se_cmd *, int, int); 552 void (*transport_complete_callback)(struct se_cmd *); 553 } ____cacheline_aligned; 554 555 #define T_TASK(cmd) ((cmd)->t_task) 556 #define CMD_TFO(cmd) ((cmd)->se_tfo) 557 558 struct se_tmr_req { 559 /* Task Management function to be preformed */ 560 u8 function; 561 /* Task Management response to send */ 562 u8 response; 563 int call_transport; 564 /* Reference to ITT that Task Mgmt should be preformed */ 565 u32 ref_task_tag; 566 /* 64-bit encoded SAM LUN from $FABRIC_MOD TMR header */ 567 u64 ref_task_lun; 568 void *fabric_tmr_ptr; 569 struct se_cmd *task_cmd; 570 struct se_cmd *ref_cmd; 571 struct se_device *tmr_dev; 572 struct se_lun *tmr_lun; 573 struct list_head tmr_list; 574 } ____cacheline_aligned; 575 576 struct se_ua { 577 u8 ua_asc; 578 u8 ua_ascq; 579 struct se_node_acl *ua_nacl; 580 struct list_head ua_dev_list; 581 struct list_head ua_nacl_list; 582 } ____cacheline_aligned; 583 584 struct se_node_acl { 585 char initiatorname[TRANSPORT_IQN_LEN]; 586 /* Used to signal demo mode created ACL, disabled by default */ 587 bool dynamic_node_acl; 588 u32 queue_depth; 589 u32 acl_index; 590 u64 num_cmds; 591 u64 read_bytes; 592 u64 write_bytes; 593 spinlock_t stats_lock; 594 /* Used for PR SPEC_I_PT=1 and REGISTER_AND_MOVE */ 595 atomic_t acl_pr_ref_count; 596 struct se_dev_entry *device_list; 597 struct se_session *nacl_sess; 598 struct se_portal_group *se_tpg; 599 spinlock_t device_list_lock; 600 spinlock_t nacl_sess_lock; 601 struct config_group acl_group; 602 struct config_group acl_attrib_group; 603 struct config_group acl_auth_group; 604 struct config_group acl_param_group; 605 struct config_group acl_fabric_stat_group; 606 struct config_group *acl_default_groups[5]; 607 struct list_head acl_list; 608 struct list_head acl_sess_list; 609 } ____cacheline_aligned; 610 611 struct se_session { 612 u64 sess_bin_isid; 613 struct se_node_acl *se_node_acl; 614 struct se_portal_group *se_tpg; 615 void *fabric_sess_ptr; 616 struct list_head sess_list; 617 struct list_head sess_acl_list; 618 } ____cacheline_aligned; 619 620 #define SE_SESS(cmd) ((cmd)->se_sess) 621 #define SE_NODE_ACL(sess) ((sess)->se_node_acl) 622 623 struct se_device; 624 struct se_transform_info; 625 struct scatterlist; 626 627 struct se_ml_stat_grps { 628 struct config_group stat_group; 629 struct config_group scsi_auth_intr_group; 630 struct config_group scsi_att_intr_port_group; 631 }; 632 633 struct se_lun_acl { 634 char initiatorname[TRANSPORT_IQN_LEN]; 635 u32 mapped_lun; 636 struct se_node_acl *se_lun_nacl; 637 struct se_lun *se_lun; 638 struct list_head lacl_list; 639 struct config_group se_lun_group; 640 struct se_ml_stat_grps ml_stat_grps; 641 } ____cacheline_aligned; 642 643 #define ML_STAT_GRPS(lacl) (&(lacl)->ml_stat_grps) 644 645 struct se_dev_entry { 646 bool def_pr_registered; 647 /* See transport_lunflags_table */ 648 u32 lun_flags; 649 u32 deve_cmds; 650 u32 mapped_lun; 651 u32 average_bytes; 652 u32 last_byte_count; 653 u32 total_cmds; 654 u32 total_bytes; 655 u64 pr_res_key; 656 u64 creation_time; 657 u32 attach_count; 658 u64 read_bytes; 659 u64 write_bytes; 660 atomic_t ua_count; 661 /* Used for PR SPEC_I_PT=1 and REGISTER_AND_MOVE */ 662 atomic_t pr_ref_count; 663 struct se_lun_acl *se_lun_acl; 664 spinlock_t ua_lock; 665 struct se_lun *se_lun; 666 struct list_head alua_port_list; 667 struct list_head ua_list; 668 } ____cacheline_aligned; 669 670 struct se_dev_limits { 671 /* Max supported HW queue depth */ 672 u32 hw_queue_depth; 673 /* Max supported virtual queue depth */ 674 u32 queue_depth; 675 /* From include/linux/blkdev.h for the other HW/SW limits. */ 676 struct queue_limits limits; 677 } ____cacheline_aligned; 678 679 struct se_dev_attrib { 680 int emulate_dpo; 681 int emulate_fua_write; 682 int emulate_fua_read; 683 int emulate_write_cache; 684 int emulate_ua_intlck_ctrl; 685 int emulate_tas; 686 int emulate_tpu; 687 int emulate_tpws; 688 int emulate_reservations; 689 int emulate_alua; 690 int enforce_pr_isids; 691 u32 hw_block_size; 692 u32 block_size; 693 u32 hw_max_sectors; 694 u32 max_sectors; 695 u32 optimal_sectors; 696 u32 hw_queue_depth; 697 u32 queue_depth; 698 u32 task_timeout; 699 u32 max_unmap_lba_count; 700 u32 max_unmap_block_desc_count; 701 u32 unmap_granularity; 702 u32 unmap_granularity_alignment; 703 struct se_subsystem_dev *da_sub_dev; 704 struct config_group da_group; 705 } ____cacheline_aligned; 706 707 struct se_dev_stat_grps { 708 struct config_group stat_group; 709 struct config_group scsi_dev_group; 710 struct config_group scsi_tgt_dev_group; 711 struct config_group scsi_lu_group; 712 }; 713 714 struct se_subsystem_dev { 715 /* Used for struct se_subsystem_dev-->se_dev_alias, must be less than PAGE_SIZE */ 716 #define SE_DEV_ALIAS_LEN 512 717 unsigned char se_dev_alias[SE_DEV_ALIAS_LEN]; 718 /* Used for struct se_subsystem_dev->se_dev_udev_path[], must be less than PAGE_SIZE */ 719 #define SE_UDEV_PATH_LEN 512 720 unsigned char se_dev_udev_path[SE_UDEV_PATH_LEN]; 721 u32 su_dev_flags; 722 struct se_hba *se_dev_hba; 723 struct se_device *se_dev_ptr; 724 struct se_dev_attrib se_dev_attrib; 725 /* T10 Asymmetric Logical Unit Assignment for Target Ports */ 726 struct t10_alua t10_alua; 727 /* T10 Inquiry and VPD WWN Information */ 728 struct t10_wwn t10_wwn; 729 /* T10 SPC-2 + SPC-3 Reservations */ 730 struct t10_reservation_template t10_reservation; 731 spinlock_t se_dev_lock; 732 void *se_dev_su_ptr; 733 struct list_head g_se_dev_list; 734 struct config_group se_dev_group; 735 /* For T10 Reservations */ 736 struct config_group se_dev_pr_group; 737 /* For target_core_stat.c groups */ 738 struct se_dev_stat_grps dev_stat_grps; 739 } ____cacheline_aligned; 740 741 #define T10_ALUA(su_dev) (&(su_dev)->t10_alua) 742 #define T10_RES(su_dev) (&(su_dev)->t10_reservation) 743 #define T10_PR_OPS(su_dev) (&(su_dev)->t10_reservation.pr_ops) 744 #define DEV_STAT_GRP(dev) (&(dev)->dev_stat_grps) 745 746 struct se_device { 747 /* Set to 1 if thread is NOT sleeping on thread_sem */ 748 u8 thread_active; 749 u8 dev_status_timer_flags; 750 /* RELATIVE TARGET PORT IDENTIFER Counter */ 751 u16 dev_rpti_counter; 752 /* Used for SAM Task Attribute ordering */ 753 u32 dev_cur_ordered_id; 754 u32 dev_flags; 755 u32 dev_port_count; 756 /* See transport_device_status_table */ 757 u32 dev_status; 758 u32 dev_tcq_window_closed; 759 /* Physical device queue depth */ 760 u32 queue_depth; 761 /* Used for SPC-2 reservations enforce of ISIDs */ 762 u64 dev_res_bin_isid; 763 t10_task_attr_index_t dev_task_attr_type; 764 /* Pointer to transport specific device structure */ 765 void *dev_ptr; 766 u32 dev_index; 767 u64 creation_time; 768 u32 num_resets; 769 u64 num_cmds; 770 u64 read_bytes; 771 u64 write_bytes; 772 spinlock_t stats_lock; 773 /* Active commands on this virtual SE device */ 774 atomic_t active_cmds; 775 atomic_t simple_cmds; 776 atomic_t depth_left; 777 atomic_t dev_ordered_id; 778 atomic_t dev_tur_active; 779 atomic_t execute_tasks; 780 atomic_t dev_status_thr_count; 781 atomic_t dev_hoq_count; 782 atomic_t dev_ordered_sync; 783 struct se_obj dev_obj; 784 struct se_obj dev_access_obj; 785 struct se_obj dev_export_obj; 786 struct se_queue_obj *dev_queue_obj; 787 struct se_queue_obj *dev_status_queue_obj; 788 spinlock_t delayed_cmd_lock; 789 spinlock_t ordered_cmd_lock; 790 spinlock_t execute_task_lock; 791 spinlock_t state_task_lock; 792 spinlock_t dev_alua_lock; 793 spinlock_t dev_reservation_lock; 794 spinlock_t dev_state_lock; 795 spinlock_t dev_status_lock; 796 spinlock_t dev_status_thr_lock; 797 spinlock_t se_port_lock; 798 spinlock_t se_tmr_lock; 799 /* Used for legacy SPC-2 reservationsa */ 800 struct se_node_acl *dev_reserved_node_acl; 801 /* Used for ALUA Logical Unit Group membership */ 802 struct t10_alua_lu_gp_member *dev_alua_lu_gp_mem; 803 /* Used for SPC-3 Persistent Reservations */ 804 struct t10_pr_registration *dev_pr_res_holder; 805 struct list_head dev_sep_list; 806 struct list_head dev_tmr_list; 807 struct timer_list dev_status_timer; 808 /* Pointer to descriptor for processing thread */ 809 struct task_struct *process_thread; 810 pid_t process_thread_pid; 811 struct task_struct *dev_mgmt_thread; 812 struct list_head delayed_cmd_list; 813 struct list_head ordered_cmd_list; 814 struct list_head execute_task_list; 815 struct list_head state_task_list; 816 /* Pointer to associated SE HBA */ 817 struct se_hba *se_hba; 818 struct se_subsystem_dev *se_sub_dev; 819 /* Pointer to template of function pointers for transport */ 820 struct se_subsystem_api *transport; 821 /* Linked list for struct se_hba struct se_device list */ 822 struct list_head dev_list; 823 /* Linked list for struct se_global->g_se_dev_list */ 824 struct list_head g_se_dev_list; 825 } ____cacheline_aligned; 826 827 #define SE_DEV(cmd) ((cmd)->se_lun->lun_se_dev) 828 #define SU_DEV(dev) ((dev)->se_sub_dev) 829 #define DEV_ATTRIB(dev) (&(dev)->se_sub_dev->se_dev_attrib) 830 #define DEV_T10_WWN(dev) (&(dev)->se_sub_dev->t10_wwn) 831 832 struct se_hba { 833 u16 hba_tpgt; 834 u32 hba_id; 835 /* See hba_flags_table */ 836 u32 hba_flags; 837 /* Virtual iSCSI devices attached. */ 838 u32 dev_count; 839 u32 hba_index; 840 atomic_t load_balance_queue; 841 atomic_t left_queue_depth; 842 /* Maximum queue depth the HBA can handle. */ 843 atomic_t max_queue_depth; 844 /* Pointer to transport specific host structure. */ 845 void *hba_ptr; 846 /* Linked list for struct se_device */ 847 struct list_head hba_dev_list; 848 struct list_head hba_list; 849 spinlock_t device_lock; 850 spinlock_t hba_queue_lock; 851 struct config_group hba_group; 852 struct mutex hba_access_mutex; 853 struct se_subsystem_api *transport; 854 } ____cacheline_aligned; 855 856 #define SE_HBA(dev) ((dev)->se_hba) 857 858 struct se_port_stat_grps { 859 struct config_group stat_group; 860 struct config_group scsi_port_group; 861 struct config_group scsi_tgt_port_group; 862 struct config_group scsi_transport_group; 863 }; 864 865 struct se_lun { 866 /* See transport_lun_status_table */ 867 enum transport_lun_status_table lun_status; 868 u32 lun_access; 869 u32 lun_flags; 870 u32 unpacked_lun; 871 atomic_t lun_acl_count; 872 spinlock_t lun_acl_lock; 873 spinlock_t lun_cmd_lock; 874 spinlock_t lun_sep_lock; 875 struct completion lun_shutdown_comp; 876 struct list_head lun_cmd_list; 877 struct list_head lun_acl_list; 878 struct se_device *lun_se_dev; 879 struct se_port *lun_sep; 880 struct config_group lun_group; 881 struct se_port_stat_grps port_stat_grps; 882 } ____cacheline_aligned; 883 884 #define SE_LUN(cmd) ((cmd)->se_lun) 885 #define PORT_STAT_GRP(lun) (&(lun)->port_stat_grps) 886 887 struct scsi_port_stats { 888 u64 cmd_pdus; 889 u64 tx_data_octets; 890 u64 rx_data_octets; 891 } ____cacheline_aligned; 892 893 struct se_port { 894 /* RELATIVE TARGET PORT IDENTIFER */ 895 u16 sep_rtpi; 896 int sep_tg_pt_secondary_stat; 897 int sep_tg_pt_secondary_write_md; 898 u32 sep_index; 899 struct scsi_port_stats sep_stats; 900 /* Used for ALUA Target Port Groups membership */ 901 atomic_t sep_tg_pt_gp_active; 902 atomic_t sep_tg_pt_secondary_offline; 903 /* Used for PR ALL_TG_PT=1 */ 904 atomic_t sep_tg_pt_ref_cnt; 905 spinlock_t sep_alua_lock; 906 struct mutex sep_tg_pt_md_mutex; 907 struct t10_alua_tg_pt_gp_member *sep_alua_tg_pt_gp_mem; 908 struct se_lun *sep_lun; 909 struct se_portal_group *sep_tpg; 910 struct list_head sep_alua_list; 911 struct list_head sep_list; 912 } ____cacheline_aligned; 913 914 struct se_tpg_np { 915 struct se_portal_group *tpg_np_parent; 916 struct config_group tpg_np_group; 917 } ____cacheline_aligned; 918 919 struct se_portal_group { 920 /* Type of target portal group, see transport_tpg_type_table */ 921 enum transport_tpg_type_table se_tpg_type; 922 /* Number of ACLed Initiator Nodes for this TPG */ 923 u32 num_node_acls; 924 /* Used for PR SPEC_I_PT=1 and REGISTER_AND_MOVE */ 925 atomic_t tpg_pr_ref_count; 926 /* Spinlock for adding/removing ACLed Nodes */ 927 spinlock_t acl_node_lock; 928 /* Spinlock for adding/removing sessions */ 929 spinlock_t session_lock; 930 spinlock_t tpg_lun_lock; 931 /* Pointer to $FABRIC_MOD portal group */ 932 void *se_tpg_fabric_ptr; 933 struct list_head se_tpg_list; 934 /* linked list for initiator ACL list */ 935 struct list_head acl_node_list; 936 struct se_lun *tpg_lun_list; 937 struct se_lun tpg_virt_lun0; 938 /* List of TCM sessions associated wth this TPG */ 939 struct list_head tpg_sess_list; 940 /* Pointer to $FABRIC_MOD dependent code */ 941 struct target_core_fabric_ops *se_tpg_tfo; 942 struct se_wwn *se_tpg_wwn; 943 struct config_group tpg_group; 944 struct config_group *tpg_default_groups[6]; 945 struct config_group tpg_lun_group; 946 struct config_group tpg_np_group; 947 struct config_group tpg_acl_group; 948 struct config_group tpg_attrib_group; 949 struct config_group tpg_param_group; 950 } ____cacheline_aligned; 951 952 #define TPG_TFO(se_tpg) ((se_tpg)->se_tpg_tfo) 953 954 struct se_wwn { 955 struct target_fabric_configfs *wwn_tf; 956 struct config_group wwn_group; 957 struct config_group *wwn_default_groups[2]; 958 struct config_group fabric_stat_group; 959 } ____cacheline_aligned; 960 961 struct se_global { 962 u16 alua_lu_gps_counter; 963 int g_sub_api_initialized; 964 u32 in_shutdown; 965 u32 alua_lu_gps_count; 966 u32 g_hba_id_counter; 967 struct config_group target_core_hbagroup; 968 struct config_group alua_group; 969 struct config_group alua_lu_gps_group; 970 struct list_head g_lu_gps_list; 971 struct list_head g_se_tpg_list; 972 struct list_head g_hba_list; 973 struct list_head g_se_dev_list; 974 struct se_hba *g_lun0_hba; 975 struct se_subsystem_dev *g_lun0_su_dev; 976 struct se_device *g_lun0_dev; 977 struct t10_alua_lu_gp *default_lu_gp; 978 spinlock_t g_device_lock; 979 spinlock_t hba_lock; 980 spinlock_t se_tpg_lock; 981 spinlock_t lu_gps_lock; 982 spinlock_t plugin_class_lock; 983 } ____cacheline_aligned; 984 985 #endif /* TARGET_CORE_BASE_H */ 986