mirror of
https://github.com/FEX-Emu/linux.git
synced 2024-12-27 03:47:43 +00:00
e70beee783
In order to support local WRITE_INSERT + READ_STRIP operations for non PI enabled fabrics, the fabric driver needs to be able signal what protection offload operations are supported. This is done at session initialization time so the modes can be signaled by individual se_wwn + se_portal_group endpoints, as well as optionally across different transports on the same endpoint. For iser-target, set TARGET_PROT_ALL if the underlying ib_device has already signaled PI offload support, and allow this to be exposed via a new iscsit_transport->iscsit_get_sup_prot_ops() callback. For loopback, set TARGET_PROT_ALL to signal SCSI initiator mode operation. For all other drivers, set TARGET_PROT_NORMAL to disable fabric level PI. Cc: Martin K. Petersen <martin.petersen@oracle.com> Cc: Sagi Grimberg <sagig@mellanox.com> Cc: Or Gerlitz <ogerlitz@mellanox.com> Cc: Quinn Tran <quinn.tran@qlogic.com> Cc: Giridhar Malavali <giridhar.malavali@qlogic.com> Signed-off-by: Nicholas Bellinger <nab@linux-iscsi.org>
214 lines
8.9 KiB
C
214 lines
8.9 KiB
C
#ifndef TARGET_CORE_FABRIC_H
|
|
#define TARGET_CORE_FABRIC_H
|
|
|
|
struct target_core_fabric_ops {
|
|
struct configfs_subsystem *tf_subsys;
|
|
char *(*get_fabric_name)(void);
|
|
u8 (*get_fabric_proto_ident)(struct se_portal_group *);
|
|
char *(*tpg_get_wwn)(struct se_portal_group *);
|
|
u16 (*tpg_get_tag)(struct se_portal_group *);
|
|
u32 (*tpg_get_default_depth)(struct se_portal_group *);
|
|
u32 (*tpg_get_pr_transport_id)(struct se_portal_group *,
|
|
struct se_node_acl *,
|
|
struct t10_pr_registration *, int *,
|
|
unsigned char *);
|
|
u32 (*tpg_get_pr_transport_id_len)(struct se_portal_group *,
|
|
struct se_node_acl *,
|
|
struct t10_pr_registration *, int *);
|
|
char *(*tpg_parse_pr_out_transport_id)(struct se_portal_group *,
|
|
const char *, u32 *, char **);
|
|
int (*tpg_check_demo_mode)(struct se_portal_group *);
|
|
int (*tpg_check_demo_mode_cache)(struct se_portal_group *);
|
|
int (*tpg_check_demo_mode_write_protect)(struct se_portal_group *);
|
|
int (*tpg_check_prod_mode_write_protect)(struct se_portal_group *);
|
|
/*
|
|
* Optionally used by fabrics to allow demo-mode login, but not
|
|
* expose any TPG LUNs, and return 'not connected' in standard
|
|
* inquiry response
|
|
*/
|
|
int (*tpg_check_demo_mode_login_only)(struct se_portal_group *);
|
|
struct se_node_acl *(*tpg_alloc_fabric_acl)(
|
|
struct se_portal_group *);
|
|
void (*tpg_release_fabric_acl)(struct se_portal_group *,
|
|
struct se_node_acl *);
|
|
u32 (*tpg_get_inst_index)(struct se_portal_group *);
|
|
/*
|
|
* Optional to release struct se_cmd and fabric dependent allocated
|
|
* I/O descriptor in transport_cmd_check_stop().
|
|
*
|
|
* Returning 1 will signal a descriptor has been released.
|
|
* Returning 0 will signal a descriptor has not been released.
|
|
*/
|
|
int (*check_stop_free)(struct se_cmd *);
|
|
void (*release_cmd)(struct se_cmd *);
|
|
void (*put_session)(struct se_session *);
|
|
/*
|
|
* Called with spin_lock_bh(struct se_portal_group->session_lock held.
|
|
*/
|
|
int (*shutdown_session)(struct se_session *);
|
|
void (*close_session)(struct se_session *);
|
|
u32 (*sess_get_index)(struct se_session *);
|
|
/*
|
|
* Used only for SCSI fabrics that contain multi-value TransportIDs
|
|
* (like iSCSI). All other SCSI fabrics should set this to NULL.
|
|
*/
|
|
u32 (*sess_get_initiator_sid)(struct se_session *,
|
|
unsigned char *, u32);
|
|
int (*write_pending)(struct se_cmd *);
|
|
int (*write_pending_status)(struct se_cmd *);
|
|
void (*set_default_node_attributes)(struct se_node_acl *);
|
|
u32 (*get_task_tag)(struct se_cmd *);
|
|
int (*get_cmd_state)(struct se_cmd *);
|
|
int (*queue_data_in)(struct se_cmd *);
|
|
int (*queue_status)(struct se_cmd *);
|
|
void (*queue_tm_rsp)(struct se_cmd *);
|
|
void (*aborted_task)(struct se_cmd *);
|
|
/*
|
|
* fabric module calls for target_core_fabric_configfs.c
|
|
*/
|
|
struct se_wwn *(*fabric_make_wwn)(struct target_fabric_configfs *,
|
|
struct config_group *, const char *);
|
|
void (*fabric_drop_wwn)(struct se_wwn *);
|
|
struct se_portal_group *(*fabric_make_tpg)(struct se_wwn *,
|
|
struct config_group *, const char *);
|
|
void (*fabric_drop_tpg)(struct se_portal_group *);
|
|
int (*fabric_post_link)(struct se_portal_group *,
|
|
struct se_lun *);
|
|
void (*fabric_pre_unlink)(struct se_portal_group *,
|
|
struct se_lun *);
|
|
struct se_tpg_np *(*fabric_make_np)(struct se_portal_group *,
|
|
struct config_group *, const char *);
|
|
void (*fabric_drop_np)(struct se_tpg_np *);
|
|
struct se_node_acl *(*fabric_make_nodeacl)(struct se_portal_group *,
|
|
struct config_group *, const char *);
|
|
void (*fabric_drop_nodeacl)(struct se_node_acl *);
|
|
};
|
|
|
|
struct se_session *transport_init_session(enum target_prot_op);
|
|
int transport_alloc_session_tags(struct se_session *, unsigned int,
|
|
unsigned int);
|
|
struct se_session *transport_init_session_tags(unsigned int, unsigned int,
|
|
enum target_prot_op);
|
|
void __transport_register_session(struct se_portal_group *,
|
|
struct se_node_acl *, struct se_session *, void *);
|
|
void transport_register_session(struct se_portal_group *,
|
|
struct se_node_acl *, struct se_session *, void *);
|
|
void target_get_session(struct se_session *);
|
|
void target_put_session(struct se_session *);
|
|
void transport_free_session(struct se_session *);
|
|
void target_put_nacl(struct se_node_acl *);
|
|
void transport_deregister_session_configfs(struct se_session *);
|
|
void transport_deregister_session(struct se_session *);
|
|
|
|
|
|
void transport_init_se_cmd(struct se_cmd *, struct target_core_fabric_ops *,
|
|
struct se_session *, u32, int, int, unsigned char *);
|
|
sense_reason_t transport_lookup_cmd_lun(struct se_cmd *, u32);
|
|
sense_reason_t target_setup_cmd_from_cdb(struct se_cmd *, unsigned char *);
|
|
int target_submit_cmd_map_sgls(struct se_cmd *, struct se_session *,
|
|
unsigned char *, unsigned char *, u32, u32, int, int, int,
|
|
struct scatterlist *, u32, struct scatterlist *, u32,
|
|
struct scatterlist *, u32);
|
|
int target_submit_cmd(struct se_cmd *, struct se_session *, unsigned char *,
|
|
unsigned char *, u32, u32, int, int, int);
|
|
int target_submit_tmr(struct se_cmd *se_cmd, struct se_session *se_sess,
|
|
unsigned char *sense, u32 unpacked_lun,
|
|
void *fabric_tmr_ptr, unsigned char tm_type,
|
|
gfp_t, unsigned int, int);
|
|
int transport_handle_cdb_direct(struct se_cmd *);
|
|
sense_reason_t transport_generic_new_cmd(struct se_cmd *);
|
|
|
|
void target_execute_cmd(struct se_cmd *cmd);
|
|
|
|
int transport_generic_free_cmd(struct se_cmd *, int);
|
|
|
|
bool transport_wait_for_tasks(struct se_cmd *);
|
|
int transport_check_aborted_status(struct se_cmd *, int);
|
|
int transport_send_check_condition_and_sense(struct se_cmd *,
|
|
sense_reason_t, int);
|
|
int target_get_sess_cmd(struct se_session *, struct se_cmd *, bool);
|
|
int target_put_sess_cmd(struct se_session *, struct se_cmd *);
|
|
void target_sess_cmd_list_set_waiting(struct se_session *);
|
|
void target_wait_for_sess_cmds(struct se_session *);
|
|
|
|
int core_alua_check_nonop_delay(struct se_cmd *);
|
|
|
|
int core_tmr_alloc_req(struct se_cmd *, void *, u8, gfp_t);
|
|
void core_tmr_release_req(struct se_tmr_req *);
|
|
int transport_generic_handle_tmr(struct se_cmd *);
|
|
void transport_generic_request_failure(struct se_cmd *, sense_reason_t);
|
|
void __target_execute_cmd(struct se_cmd *);
|
|
int transport_lookup_tmr_lun(struct se_cmd *, u32);
|
|
|
|
struct se_node_acl *core_tpg_get_initiator_node_acl(struct se_portal_group *tpg,
|
|
unsigned char *);
|
|
struct se_node_acl *core_tpg_check_initiator_node_acl(struct se_portal_group *,
|
|
unsigned char *);
|
|
void core_tpg_clear_object_luns(struct se_portal_group *);
|
|
struct se_node_acl *core_tpg_add_initiator_node_acl(struct se_portal_group *,
|
|
struct se_node_acl *, const char *, u32);
|
|
int core_tpg_del_initiator_node_acl(struct se_portal_group *,
|
|
struct se_node_acl *, int);
|
|
int core_tpg_set_initiator_node_queue_depth(struct se_portal_group *,
|
|
unsigned char *, u32, int);
|
|
int core_tpg_set_initiator_node_tag(struct se_portal_group *,
|
|
struct se_node_acl *, const char *);
|
|
int core_tpg_register(struct target_core_fabric_ops *, struct se_wwn *,
|
|
struct se_portal_group *, void *, int);
|
|
int core_tpg_deregister(struct se_portal_group *);
|
|
|
|
/* SAS helpers */
|
|
u8 sas_get_fabric_proto_ident(struct se_portal_group *);
|
|
u32 sas_get_pr_transport_id(struct se_portal_group *, struct se_node_acl *,
|
|
struct t10_pr_registration *, int *, unsigned char *);
|
|
u32 sas_get_pr_transport_id_len(struct se_portal_group *, struct se_node_acl *,
|
|
struct t10_pr_registration *, int *);
|
|
char *sas_parse_pr_out_transport_id(struct se_portal_group *, const char *,
|
|
u32 *, char **);
|
|
|
|
/* FC helpers */
|
|
u8 fc_get_fabric_proto_ident(struct se_portal_group *);
|
|
u32 fc_get_pr_transport_id(struct se_portal_group *, struct se_node_acl *,
|
|
struct t10_pr_registration *, int *, unsigned char *);
|
|
u32 fc_get_pr_transport_id_len(struct se_portal_group *, struct se_node_acl *,
|
|
struct t10_pr_registration *, int *);
|
|
char *fc_parse_pr_out_transport_id(struct se_portal_group *, const char *,
|
|
u32 *, char **);
|
|
|
|
/* iSCSI helpers */
|
|
u8 iscsi_get_fabric_proto_ident(struct se_portal_group *);
|
|
u32 iscsi_get_pr_transport_id(struct se_portal_group *, struct se_node_acl *,
|
|
struct t10_pr_registration *, int *, unsigned char *);
|
|
u32 iscsi_get_pr_transport_id_len(struct se_portal_group *, struct se_node_acl *,
|
|
struct t10_pr_registration *, int *);
|
|
char *iscsi_parse_pr_out_transport_id(struct se_portal_group *, const char *,
|
|
u32 *, char **);
|
|
|
|
/*
|
|
* The LIO target core uses DMA_TO_DEVICE to mean that data is going
|
|
* to the target (eg handling a WRITE) and DMA_FROM_DEVICE to mean
|
|
* that data is coming from the target (eg handling a READ). However,
|
|
* this is just the opposite of what we have to tell the DMA mapping
|
|
* layer -- eg when handling a READ, the HBA will have to DMA the data
|
|
* out of memory so it can send it to the initiator, which means we
|
|
* need to use DMA_TO_DEVICE when we map the data.
|
|
*/
|
|
static inline enum dma_data_direction
|
|
target_reverse_dma_direction(struct se_cmd *se_cmd)
|
|
{
|
|
if (se_cmd->se_cmd_flags & SCF_BIDI)
|
|
return DMA_BIDIRECTIONAL;
|
|
|
|
switch (se_cmd->data_direction) {
|
|
case DMA_TO_DEVICE:
|
|
return DMA_FROM_DEVICE;
|
|
case DMA_FROM_DEVICE:
|
|
return DMA_TO_DEVICE;
|
|
case DMA_NONE:
|
|
default:
|
|
return DMA_NONE;
|
|
}
|
|
}
|
|
|
|
#endif /* TARGET_CORE_FABRICH */
|