target_core_fabric.h 8.7 KB
Newer Older
1
/* SPDX-License-Identifier: GPL-2.0 */
2 3
#ifndef TARGET_CORE_FABRIC_H
#define TARGET_CORE_FABRIC_H
4

5 6 7 8
#include <linux/configfs.h>
#include <linux/types.h>
#include <target/target_core_base.h>

9
struct target_core_fabric_ops {
10 11
	struct module *module;
	const char *name;
12
	size_t node_acl_size;
13 14 15 16 17 18 19 20 21 22 23 24 25
	/*
	 * Limits number of scatterlist entries per SCF_SCSI_DATA_CDB payload.
	 * Setting this value tells target-core to enforce this limit, and
	 * report as INQUIRY EVPD=b0 MAXIMUM TRANSFER LENGTH.
	 *
	 * target-core will currently reset se_cmd->data_length to this
	 * maximum size, and set UNDERFLOW residual count if length exceeds
	 * this limit.
	 *
	 * XXX: Not all initiator hosts honor this block-limit EVPD
	 * XXX: Currently assumes single PAGE_SIZE per scatterlist entry
	 */
	u32 max_data_sg_nents;
26 27 28 29 30 31 32 33
	char *(*get_fabric_name)(void);
	char *(*tpg_get_wwn)(struct se_portal_group *);
	u16 (*tpg_get_tag)(struct se_portal_group *);
	u32 (*tpg_get_default_depth)(struct se_portal_group *);
	int (*tpg_check_demo_mode)(struct se_portal_group *);
	int (*tpg_check_demo_mode_cache)(struct se_portal_group *);
	int (*tpg_check_demo_mode_write_protect)(struct se_portal_group *);
	int (*tpg_check_prod_mode_write_protect)(struct se_portal_group *);
34 35 36 37 38 39
	/*
	 * Optionally used by fabrics to allow demo-mode login, but not
	 * expose any TPG LUNs, and return 'not connected' in standard
	 * inquiry response
	 */
	int (*tpg_check_demo_mode_login_only)(struct se_portal_group *);
40 41 42 43 44 45 46 47
	/*
	 * Optionally used as a configfs tunable to determine when
	 * target-core should signal the PROTECT=1 feature bit for
	 * backends that don't support T10-PI, so that either fabric
	 * HW offload or target-core emulation performs the associated
	 * WRITE_STRIP and READ_INSERT operations.
	 */
	int (*tpg_check_prot_fabric_only)(struct se_portal_group *);
48 49 50
	u32 (*tpg_get_inst_index)(struct se_portal_group *);
	/*
	 * Optional to release struct se_cmd and fabric dependent allocated
51
	 * I/O descriptor after command execution has finished.
52 53 54
	 *
	 * Returning 1 will signal a descriptor has been released.
	 * Returning 0 will signal a descriptor has not been released.
55
	 */
56
	int (*check_stop_free)(struct se_cmd *);
57
	void (*release_cmd)(struct se_cmd *);
58 59 60 61 62 63 64 65 66 67 68 69 70 71
	void (*close_session)(struct se_session *);
	u32 (*sess_get_index)(struct se_session *);
	/*
	 * Used only for SCSI fabrics that contain multi-value TransportIDs
	 * (like iSCSI).  All other SCSI fabrics should set this to NULL.
	 */
	u32 (*sess_get_initiator_sid)(struct se_session *,
				      unsigned char *, u32);
	int (*write_pending)(struct se_cmd *);
	int (*write_pending_status)(struct se_cmd *);
	void (*set_default_node_attributes)(struct se_node_acl *);
	int (*get_cmd_state)(struct se_cmd *);
	int (*queue_data_in)(struct se_cmd *);
	int (*queue_status)(struct se_cmd *);
72
	void (*queue_tm_rsp)(struct se_cmd *);
73
	void (*aborted_task)(struct se_cmd *);
74 75 76 77 78 79
	/*
	 * fabric module calls for target_core_fabric_configfs.c
	 */
	struct se_wwn *(*fabric_make_wwn)(struct target_fabric_configfs *,
				struct config_group *, const char *);
	void (*fabric_drop_wwn)(struct se_wwn *);
80
	void (*add_wwn_groups)(struct se_wwn *);
81
	struct se_portal_group *(*fabric_make_tpg)(struct se_wwn *,
82
						   const char *);
83 84 85 86 87 88 89 90
	void (*fabric_drop_tpg)(struct se_portal_group *);
	int (*fabric_post_link)(struct se_portal_group *,
				struct se_lun *);
	void (*fabric_pre_unlink)(struct se_portal_group *,
				struct se_lun *);
	struct se_tpg_np *(*fabric_make_np)(struct se_portal_group *,
				struct config_group *, const char *);
	void (*fabric_drop_np)(struct se_tpg_np *);
91
	int (*fabric_init_nodeacl)(struct se_node_acl *, const char *);
92 93 94 95 96 97 98 99 100 101 102 103

	struct configfs_attribute **tfc_discovery_attrs;
	struct configfs_attribute **tfc_wwn_attrs;
	struct configfs_attribute **tfc_tpg_base_attrs;
	struct configfs_attribute **tfc_tpg_np_base_attrs;
	struct configfs_attribute **tfc_tpg_attrib_attrs;
	struct configfs_attribute **tfc_tpg_auth_attrs;
	struct configfs_attribute **tfc_tpg_param_attrs;
	struct configfs_attribute **tfc_tpg_nacl_base_attrs;
	struct configfs_attribute **tfc_tpg_nacl_attrib_attrs;
	struct configfs_attribute **tfc_tpg_nacl_auth_attrs;
	struct configfs_attribute **tfc_tpg_nacl_param_attrs;
104
};
105

106 107 108
int target_register_template(const struct target_core_fabric_ops *fo);
void target_unregister_template(const struct target_core_fabric_ops *fo);

109 110 111
int target_depend_item(struct config_item *item);
void target_undepend_item(struct config_item *item);

112 113 114 115 116 117
struct se_session *target_alloc_session(struct se_portal_group *,
		unsigned int, unsigned int, enum target_prot_op prot_op,
		const char *, void *,
		int (*callback)(struct se_portal_group *,
				struct se_session *, void *));

118
void transport_init_session(struct se_session *);
119
struct se_session *transport_alloc_session(enum target_prot_op);
120 121
int transport_alloc_session_tags(struct se_session *, unsigned int,
		unsigned int);
122 123 124 125
void	__transport_register_session(struct se_portal_group *,
		struct se_node_acl *, struct se_session *, void *);
void	transport_register_session(struct se_portal_group *,
		struct se_node_acl *, struct se_session *, void *);
126
ssize_t	target_show_dynamic_sessions(struct se_portal_group *, char *);
127
void	transport_free_session(struct se_session *);
128
void	target_put_nacl(struct se_node_acl *);
129 130 131 132
void	transport_deregister_session_configfs(struct se_session *);
void	transport_deregister_session(struct se_session *);


133 134
void	transport_init_se_cmd(struct se_cmd *,
		const struct target_core_fabric_ops *,
135
		struct se_session *, u32, int, int, unsigned char *);
H
Hannes Reinecke 已提交
136
sense_reason_t transport_lookup_cmd_lun(struct se_cmd *, u64);
137
sense_reason_t target_setup_cmd_from_cdb(struct se_cmd *, unsigned char *);
138
int	target_submit_cmd_map_sgls(struct se_cmd *, struct se_session *,
H
Hannes Reinecke 已提交
139
		unsigned char *, unsigned char *, u64, u32, int, int, int,
140 141
		struct scatterlist *, u32, struct scatterlist *, u32,
		struct scatterlist *, u32);
142
int	target_submit_cmd(struct se_cmd *, struct se_session *, unsigned char *,
H
Hannes Reinecke 已提交
143
		unsigned char *, u64, u32, int, int, int);
144
int	target_submit_tmr(struct se_cmd *se_cmd, struct se_session *se_sess,
H
Hannes Reinecke 已提交
145
		unsigned char *sense, u64 unpacked_lun,
146
		void *fabric_tmr_ptr, unsigned char tm_type,
147
		gfp_t, u64, int);
148
int	transport_handle_cdb_direct(struct se_cmd *);
149
sense_reason_t	transport_generic_new_cmd(struct se_cmd *);
150

151
void	target_execute_cmd(struct se_cmd *cmd);
152

153
int	transport_generic_free_cmd(struct se_cmd *, int);
154 155 156

bool	transport_wait_for_tasks(struct se_cmd *);
int	transport_check_aborted_status(struct se_cmd *, int);
157 158
int	transport_send_check_condition_and_sense(struct se_cmd *,
		sense_reason_t, int);
159 160
int	target_get_sess_cmd(struct se_cmd *, bool);
int	target_put_sess_cmd(struct se_cmd *);
161
void	target_sess_cmd_list_set_waiting(struct se_session *);
162
void	target_wait_for_sess_cmds(struct se_session *);
163
void	target_show_cmd(const char *pfx, struct se_cmd *cmd);
164 165 166

int	core_alua_check_nonop_delay(struct se_cmd *);

167
int	core_tmr_alloc_req(struct se_cmd *, void *, u8, gfp_t);
168 169
void	core_tmr_release_req(struct se_tmr_req *);
int	transport_generic_handle_tmr(struct se_cmd *);
170
void	transport_generic_request_failure(struct se_cmd *, sense_reason_t);
H
Hannes Reinecke 已提交
171
int	transport_lookup_tmr_lun(struct se_cmd *, u64);
172
void	core_allocate_nexus_loss_ua(struct se_node_acl *acl);
173

174 175
struct se_node_acl *core_tpg_get_initiator_node_acl(struct se_portal_group *tpg,
		unsigned char *);
176 177
bool	target_tpg_has_node_acl(struct se_portal_group *tpg,
		const char *);
178 179
struct se_node_acl *core_tpg_check_initiator_node_acl(struct se_portal_group *,
		unsigned char *);
180
int	core_tpg_set_initiator_node_queue_depth(struct se_node_acl *, u32);
181 182
int	core_tpg_set_initiator_node_tag(struct se_portal_group *,
		struct se_node_acl *, const char *);
183
int	core_tpg_register(struct se_wwn *, struct se_portal_group *, int);
184 185
int	core_tpg_deregister(struct se_portal_group *);

186 187 188 189
int	target_alloc_sgl(struct scatterlist **sgl, unsigned int *nents,
		u32 length, bool zero_page, bool chainable);
void	target_free_sgl(struct scatterlist *sgl, int nents);

190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215
/*
 * The LIO target core uses DMA_TO_DEVICE to mean that data is going
 * to the target (eg handling a WRITE) and DMA_FROM_DEVICE to mean
 * that data is coming from the target (eg handling a READ).  However,
 * this is just the opposite of what we have to tell the DMA mapping
 * layer -- eg when handling a READ, the HBA will have to DMA the data
 * out of memory so it can send it to the initiator, which means we
 * need to use DMA_TO_DEVICE when we map the data.
 */
static inline enum dma_data_direction
target_reverse_dma_direction(struct se_cmd *se_cmd)
{
	if (se_cmd->se_cmd_flags & SCF_BIDI)
		return DMA_BIDIRECTIONAL;

	switch (se_cmd->data_direction) {
	case DMA_TO_DEVICE:
		return DMA_FROM_DEVICE;
	case DMA_FROM_DEVICE:
		return DMA_TO_DEVICE;
	case DMA_NONE:
	default:
		return DMA_NONE;
	}
}

216
#endif /* TARGET_CORE_FABRICH */