提交 03f6fb93 编写于 作者: B Bart Van Assche 提交者: Doug Ledford

IB/srp: Create an insecure all physical rkey only if needed

The SRP initiator only needs this if the insecure register_always=N
performance optimization is enabled, or if FRWR/FMR is not supported
in the driver.

Do not create an all physical MR unless it is needed to support
either of those modes. Default register_always to true so the out of
the box configuration does not create an insecure all physical MR.
Signed-off-by: NJason Gunthorpe <jgunthorpe@obsidianresearch.com>
[bvanassche: reworked and rebased this patch]
Signed-off-by: NBart Van Assche <bart.vanassche@sandisk.com>
Signed-off-by: NDoug Ledford <dledford@redhat.com>
上级 330179f2
...@@ -68,8 +68,8 @@ static unsigned int srp_sg_tablesize; ...@@ -68,8 +68,8 @@ static unsigned int srp_sg_tablesize;
static unsigned int cmd_sg_entries; static unsigned int cmd_sg_entries;
static unsigned int indirect_sg_entries; static unsigned int indirect_sg_entries;
static bool allow_ext_sg; static bool allow_ext_sg;
static bool prefer_fr; static bool prefer_fr = true;
static bool register_always; static bool register_always = true;
static int topspin_workarounds = 1; static int topspin_workarounds = 1;
module_param(srp_sg_tablesize, uint, 0444); module_param(srp_sg_tablesize, uint, 0444);
...@@ -1353,9 +1353,9 @@ static int srp_finish_mapping(struct srp_map_state *state, ...@@ -1353,9 +1353,9 @@ static int srp_finish_mapping(struct srp_map_state *state,
if (state->npages == 0) if (state->npages == 0)
return 0; return 0;
if (state->npages == 1 && !register_always) if (state->npages == 1 && target->global_mr)
srp_map_desc(state, state->base_dma_addr, state->dma_len, srp_map_desc(state, state->base_dma_addr, state->dma_len,
target->rkey); target->global_mr->rkey);
else else
ret = dev->use_fast_reg ? srp_map_finish_fr(state, ch) : ret = dev->use_fast_reg ? srp_map_finish_fr(state, ch) :
srp_map_finish_fmr(state, ch); srp_map_finish_fmr(state, ch);
...@@ -1442,7 +1442,8 @@ static int srp_map_sg(struct srp_map_state *state, struct srp_rdma_ch *ch, ...@@ -1442,7 +1442,8 @@ static int srp_map_sg(struct srp_map_state *state, struct srp_rdma_ch *ch,
} else { } else {
for_each_sg(scat, sg, count, i) { for_each_sg(scat, sg, count, i) {
srp_map_desc(state, ib_sg_dma_address(dev->dev, sg), srp_map_desc(state, ib_sg_dma_address(dev->dev, sg),
ib_sg_dma_len(dev->dev, sg), target->rkey); ib_sg_dma_len(dev->dev, sg),
target->global_mr->rkey);
} }
} }
...@@ -1531,7 +1532,7 @@ static int srp_map_data(struct scsi_cmnd *scmnd, struct srp_rdma_ch *ch, ...@@ -1531,7 +1532,7 @@ static int srp_map_data(struct scsi_cmnd *scmnd, struct srp_rdma_ch *ch,
fmt = SRP_DATA_DESC_DIRECT; fmt = SRP_DATA_DESC_DIRECT;
len = sizeof (struct srp_cmd) + sizeof (struct srp_direct_buf); len = sizeof (struct srp_cmd) + sizeof (struct srp_direct_buf);
if (count == 1 && !register_always) { if (count == 1 && target->global_mr) {
/* /*
* The midlayer only generated a single gather/scatter * The midlayer only generated a single gather/scatter
* entry, or DMA mapping coalesced everything to a * entry, or DMA mapping coalesced everything to a
...@@ -1541,7 +1542,7 @@ static int srp_map_data(struct scsi_cmnd *scmnd, struct srp_rdma_ch *ch, ...@@ -1541,7 +1542,7 @@ static int srp_map_data(struct scsi_cmnd *scmnd, struct srp_rdma_ch *ch,
struct srp_direct_buf *buf = (void *) cmd->add_data; struct srp_direct_buf *buf = (void *) cmd->add_data;
buf->va = cpu_to_be64(ib_sg_dma_address(ibdev, scat)); buf->va = cpu_to_be64(ib_sg_dma_address(ibdev, scat));
buf->key = cpu_to_be32(target->rkey); buf->key = cpu_to_be32(target->global_mr->rkey);
buf->len = cpu_to_be32(ib_sg_dma_len(ibdev, scat)); buf->len = cpu_to_be32(ib_sg_dma_len(ibdev, scat));
req->nmdesc = 0; req->nmdesc = 0;
...@@ -1595,14 +1596,14 @@ static int srp_map_data(struct scsi_cmnd *scmnd, struct srp_rdma_ch *ch, ...@@ -1595,14 +1596,14 @@ static int srp_map_data(struct scsi_cmnd *scmnd, struct srp_rdma_ch *ch,
memcpy(indirect_hdr->desc_list, req->indirect_desc, memcpy(indirect_hdr->desc_list, req->indirect_desc,
count * sizeof (struct srp_direct_buf)); count * sizeof (struct srp_direct_buf));
if (register_always && (dev->use_fast_reg || dev->use_fmr)) { if (!target->global_mr) {
ret = srp_map_idb(ch, req, state.gen.next, state.gen.end, ret = srp_map_idb(ch, req, state.gen.next, state.gen.end,
idb_len, &idb_rkey); idb_len, &idb_rkey);
if (ret < 0) if (ret < 0)
return ret; return ret;
req->nmdesc++; req->nmdesc++;
} else { } else {
idb_rkey = target->rkey; idb_rkey = target->global_mr->rkey;
} }
indirect_hdr->table_desc.va = cpu_to_be64(req->indirect_dma_addr); indirect_hdr->table_desc.va = cpu_to_be64(req->indirect_dma_addr);
...@@ -3157,7 +3158,7 @@ static ssize_t srp_create_target(struct device *dev, ...@@ -3157,7 +3158,7 @@ static ssize_t srp_create_target(struct device *dev,
target->scsi_host = target_host; target->scsi_host = target_host;
target->srp_host = host; target->srp_host = host;
target->lkey = host->srp_dev->pd->local_dma_lkey; target->lkey = host->srp_dev->pd->local_dma_lkey;
target->rkey = host->srp_dev->mr->rkey; target->global_mr = host->srp_dev->global_mr;
target->cmd_sg_cnt = cmd_sg_entries; target->cmd_sg_cnt = cmd_sg_entries;
target->sg_tablesize = indirect_sg_entries ? : cmd_sg_entries; target->sg_tablesize = indirect_sg_entries ? : cmd_sg_entries;
target->allow_ext_sg = allow_ext_sg; target->allow_ext_sg = allow_ext_sg;
...@@ -3447,12 +3448,16 @@ static void srp_add_one(struct ib_device *device) ...@@ -3447,12 +3448,16 @@ static void srp_add_one(struct ib_device *device)
if (IS_ERR(srp_dev->pd)) if (IS_ERR(srp_dev->pd))
goto free_dev; goto free_dev;
srp_dev->mr = ib_get_dma_mr(srp_dev->pd, if (!register_always || (!srp_dev->has_fmr && !srp_dev->has_fr)) {
IB_ACCESS_LOCAL_WRITE | srp_dev->global_mr = ib_get_dma_mr(srp_dev->pd,
IB_ACCESS_REMOTE_READ | IB_ACCESS_LOCAL_WRITE |
IB_ACCESS_REMOTE_WRITE); IB_ACCESS_REMOTE_READ |
if (IS_ERR(srp_dev->mr)) IB_ACCESS_REMOTE_WRITE);
goto err_pd; if (IS_ERR(srp_dev->global_mr))
goto err_pd;
} else {
srp_dev->global_mr = NULL;
}
for (p = rdma_start_port(device); p <= rdma_end_port(device); ++p) { for (p = rdma_start_port(device); p <= rdma_end_port(device); ++p) {
host = srp_add_port(srp_dev, p); host = srp_add_port(srp_dev, p);
...@@ -3509,7 +3514,8 @@ static void srp_remove_one(struct ib_device *device, void *client_data) ...@@ -3509,7 +3514,8 @@ static void srp_remove_one(struct ib_device *device, void *client_data)
kfree(host); kfree(host);
} }
ib_dereg_mr(srp_dev->mr); if (srp_dev->global_mr)
ib_dereg_mr(srp_dev->global_mr);
ib_dealloc_pd(srp_dev->pd); ib_dealloc_pd(srp_dev->pd);
kfree(srp_dev); kfree(srp_dev);
......
...@@ -95,7 +95,7 @@ struct srp_device { ...@@ -95,7 +95,7 @@ struct srp_device {
struct list_head dev_list; struct list_head dev_list;
struct ib_device *dev; struct ib_device *dev;
struct ib_pd *pd; struct ib_pd *pd;
struct ib_mr *mr; struct ib_mr *global_mr;
u64 mr_page_mask; u64 mr_page_mask;
int mr_page_size; int mr_page_size;
int mr_max_size; int mr_max_size;
...@@ -183,10 +183,10 @@ struct srp_target_port { ...@@ -183,10 +183,10 @@ struct srp_target_port {
spinlock_t lock; spinlock_t lock;
/* read only in the hot path */ /* read only in the hot path */
struct ib_mr *global_mr;
struct srp_rdma_ch *ch; struct srp_rdma_ch *ch;
u32 ch_count; u32 ch_count;
u32 lkey; u32 lkey;
u32 rkey;
enum srp_target_state state; enum srp_target_state state;
unsigned int max_iu_len; unsigned int max_iu_len;
unsigned int cmd_sg_cnt; unsigned int cmd_sg_cnt;
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册