提交 f6a9d47a 编写于 作者: J Jason Gunthorpe

RDMA/cma: Execute rdma_cm destruction from a handler properly

When a rdma_cm_id needs to be destroyed after a handler callback fails,
part of the destruction pattern is open coded into each call site.

Unfortunately the blind assignment to state discards important information
needed to do cma_cancel_operation(). This results in active operations
being left running after rdma_destroy_id() completes, and the
use-after-free bugs from KASAN.

Consolidate this entire pattern into destroy_id_handler_unlock() and
manage the locking correctly. The state should be set to
RDMA_CM_DESTROYING under the handler_lock to atomically ensure no futher
handlers are called.

Link: https://lore.kernel.org/r/20200723070707.1771101-5-leon@kernel.org
Reported-by: syzbot+08092148130652a6faae@syzkaller.appspotmail.com
Reported-by: syzbot+a929647172775e335941@syzkaller.appspotmail.com
Signed-off-by: NLeon Romanovsky <leonro@mellanox.com>
Signed-off-by: NJason Gunthorpe <jgg@nvidia.com>
上级 cc9c0373
...@@ -428,19 +428,6 @@ static int cma_comp_exch(struct rdma_id_private *id_priv, ...@@ -428,19 +428,6 @@ static int cma_comp_exch(struct rdma_id_private *id_priv,
return ret; return ret;
} }
static enum rdma_cm_state cma_exch(struct rdma_id_private *id_priv,
enum rdma_cm_state exch)
{
unsigned long flags;
enum rdma_cm_state old;
spin_lock_irqsave(&id_priv->lock, flags);
old = id_priv->state;
id_priv->state = exch;
spin_unlock_irqrestore(&id_priv->lock, flags);
return old;
}
static inline u8 cma_get_ip_ver(const struct cma_hdr *hdr) static inline u8 cma_get_ip_ver(const struct cma_hdr *hdr)
{ {
return hdr->ip_version >> 4; return hdr->ip_version >> 4;
...@@ -1825,21 +1812,9 @@ static void cma_leave_mc_groups(struct rdma_id_private *id_priv) ...@@ -1825,21 +1812,9 @@ static void cma_leave_mc_groups(struct rdma_id_private *id_priv)
} }
} }
void rdma_destroy_id(struct rdma_cm_id *id) static void _destroy_id(struct rdma_id_private *id_priv,
enum rdma_cm_state state)
{ {
struct rdma_id_private *id_priv =
container_of(id, struct rdma_id_private, id);
enum rdma_cm_state state;
/*
* Wait for any active callback to finish. New callbacks will find
* the id_priv state set to destroying and abort.
*/
mutex_lock(&id_priv->handler_mutex);
trace_cm_id_destroy(id_priv);
state = cma_exch(id_priv, RDMA_CM_DESTROYING);
mutex_unlock(&id_priv->handler_mutex);
cma_cancel_operation(id_priv, state); cma_cancel_operation(id_priv, state);
rdma_restrack_del(&id_priv->res); rdma_restrack_del(&id_priv->res);
...@@ -1870,6 +1845,42 @@ void rdma_destroy_id(struct rdma_cm_id *id) ...@@ -1870,6 +1845,42 @@ void rdma_destroy_id(struct rdma_cm_id *id)
put_net(id_priv->id.route.addr.dev_addr.net); put_net(id_priv->id.route.addr.dev_addr.net);
kfree(id_priv); kfree(id_priv);
} }
/*
* destroy an ID from within the handler_mutex. This ensures that no other
* handlers can start running concurrently.
*/
static void destroy_id_handler_unlock(struct rdma_id_private *id_priv)
__releases(&idprv->handler_mutex)
{
enum rdma_cm_state state;
unsigned long flags;
trace_cm_id_destroy(id_priv);
/*
* Setting the state to destroyed under the handler mutex provides a
* fence against calling handler callbacks. If this is invoked due to
* the failure of a handler callback then it guarentees that no future
* handlers will be called.
*/
lockdep_assert_held(&id_priv->handler_mutex);
spin_lock_irqsave(&id_priv->lock, flags);
state = id_priv->state;
id_priv->state = RDMA_CM_DESTROYING;
spin_unlock_irqrestore(&id_priv->lock, flags);
mutex_unlock(&id_priv->handler_mutex);
_destroy_id(id_priv, state);
}
void rdma_destroy_id(struct rdma_cm_id *id)
{
struct rdma_id_private *id_priv =
container_of(id, struct rdma_id_private, id);
mutex_lock(&id_priv->handler_mutex);
destroy_id_handler_unlock(id_priv);
}
EXPORT_SYMBOL(rdma_destroy_id); EXPORT_SYMBOL(rdma_destroy_id);
static int cma_rep_recv(struct rdma_id_private *id_priv) static int cma_rep_recv(struct rdma_id_private *id_priv)
...@@ -1934,7 +1945,7 @@ static int cma_ib_handler(struct ib_cm_id *cm_id, ...@@ -1934,7 +1945,7 @@ static int cma_ib_handler(struct ib_cm_id *cm_id,
{ {
struct rdma_id_private *id_priv = cm_id->context; struct rdma_id_private *id_priv = cm_id->context;
struct rdma_cm_event event = {}; struct rdma_cm_event event = {};
int ret = 0; int ret;
mutex_lock(&id_priv->handler_mutex); mutex_lock(&id_priv->handler_mutex);
if ((ib_event->event != IB_CM_TIMEWAIT_EXIT && if ((ib_event->event != IB_CM_TIMEWAIT_EXIT &&
...@@ -2003,14 +2014,12 @@ static int cma_ib_handler(struct ib_cm_id *cm_id, ...@@ -2003,14 +2014,12 @@ static int cma_ib_handler(struct ib_cm_id *cm_id,
if (ret) { if (ret) {
/* Destroy the CM ID by returning a non-zero value. */ /* Destroy the CM ID by returning a non-zero value. */
id_priv->cm_id.ib = NULL; id_priv->cm_id.ib = NULL;
cma_exch(id_priv, RDMA_CM_DESTROYING); destroy_id_handler_unlock(id_priv);
mutex_unlock(&id_priv->handler_mutex);
rdma_destroy_id(&id_priv->id);
return ret; return ret;
} }
out: out:
mutex_unlock(&id_priv->handler_mutex); mutex_unlock(&id_priv->handler_mutex);
return ret; return 0;
} }
static struct rdma_id_private * static struct rdma_id_private *
...@@ -2172,7 +2181,7 @@ static int cma_ib_req_handler(struct ib_cm_id *cm_id, ...@@ -2172,7 +2181,7 @@ static int cma_ib_req_handler(struct ib_cm_id *cm_id,
mutex_lock(&listen_id->handler_mutex); mutex_lock(&listen_id->handler_mutex);
if (listen_id->state != RDMA_CM_LISTEN) { if (listen_id->state != RDMA_CM_LISTEN) {
ret = -ECONNABORTED; ret = -ECONNABORTED;
goto err1; goto err_unlock;
} }
offset = cma_user_data_offset(listen_id); offset = cma_user_data_offset(listen_id);
...@@ -2189,43 +2198,38 @@ static int cma_ib_req_handler(struct ib_cm_id *cm_id, ...@@ -2189,43 +2198,38 @@ static int cma_ib_req_handler(struct ib_cm_id *cm_id,
} }
if (!conn_id) { if (!conn_id) {
ret = -ENOMEM; ret = -ENOMEM;
goto err1; goto err_unlock;
} }
mutex_lock_nested(&conn_id->handler_mutex, SINGLE_DEPTH_NESTING); mutex_lock_nested(&conn_id->handler_mutex, SINGLE_DEPTH_NESTING);
ret = cma_ib_acquire_dev(conn_id, listen_id, &req); ret = cma_ib_acquire_dev(conn_id, listen_id, &req);
if (ret) if (ret) {
goto err2; destroy_id_handler_unlock(conn_id);
goto err_unlock;
}
conn_id->cm_id.ib = cm_id; conn_id->cm_id.ib = cm_id;
cm_id->context = conn_id; cm_id->context = conn_id;
cm_id->cm_handler = cma_ib_handler; cm_id->cm_handler = cma_ib_handler;
ret = cma_cm_event_handler(conn_id, &event); ret = cma_cm_event_handler(conn_id, &event);
if (ret) if (ret) {
goto err3; /* Destroy the CM ID by returning a non-zero value. */
conn_id->cm_id.ib = NULL;
mutex_unlock(&listen_id->handler_mutex);
destroy_id_handler_unlock(conn_id);
goto net_dev_put;
}
if (cma_comp(conn_id, RDMA_CM_CONNECT) && if (cma_comp(conn_id, RDMA_CM_CONNECT) &&
(conn_id->id.qp_type != IB_QPT_UD)) { (conn_id->id.qp_type != IB_QPT_UD)) {
trace_cm_send_mra(cm_id->context); trace_cm_send_mra(cm_id->context);
ib_send_cm_mra(cm_id, CMA_CM_MRA_SETTING, NULL, 0); ib_send_cm_mra(cm_id, CMA_CM_MRA_SETTING, NULL, 0);
} }
mutex_unlock(&lock);
mutex_unlock(&conn_id->handler_mutex); mutex_unlock(&conn_id->handler_mutex);
mutex_unlock(&listen_id->handler_mutex);
if (net_dev)
dev_put(net_dev);
return 0;
err3: err_unlock:
/* Destroy the CM ID by returning a non-zero value. */
conn_id->cm_id.ib = NULL;
err2:
cma_exch(conn_id, RDMA_CM_DESTROYING);
mutex_unlock(&conn_id->handler_mutex);
err1:
mutex_unlock(&listen_id->handler_mutex); mutex_unlock(&listen_id->handler_mutex);
if (conn_id)
rdma_destroy_id(&conn_id->id);
net_dev_put: net_dev_put:
if (net_dev) if (net_dev)
...@@ -2325,9 +2329,7 @@ static int cma_iw_handler(struct iw_cm_id *iw_id, struct iw_cm_event *iw_event) ...@@ -2325,9 +2329,7 @@ static int cma_iw_handler(struct iw_cm_id *iw_id, struct iw_cm_event *iw_event)
if (ret) { if (ret) {
/* Destroy the CM ID by returning a non-zero value. */ /* Destroy the CM ID by returning a non-zero value. */
id_priv->cm_id.iw = NULL; id_priv->cm_id.iw = NULL;
cma_exch(id_priv, RDMA_CM_DESTROYING); destroy_id_handler_unlock(id_priv);
mutex_unlock(&id_priv->handler_mutex);
rdma_destroy_id(&id_priv->id);
return ret; return ret;
} }
...@@ -2374,16 +2376,16 @@ static int iw_conn_req_handler(struct iw_cm_id *cm_id, ...@@ -2374,16 +2376,16 @@ static int iw_conn_req_handler(struct iw_cm_id *cm_id,
ret = rdma_translate_ip(laddr, &conn_id->id.route.addr.dev_addr); ret = rdma_translate_ip(laddr, &conn_id->id.route.addr.dev_addr);
if (ret) { if (ret) {
mutex_unlock(&conn_id->handler_mutex); mutex_unlock(&listen_id->handler_mutex);
rdma_destroy_id(new_cm_id); destroy_id_handler_unlock(conn_id);
goto out; return ret;
} }
ret = cma_iw_acquire_dev(conn_id, listen_id); ret = cma_iw_acquire_dev(conn_id, listen_id);
if (ret) { if (ret) {
mutex_unlock(&conn_id->handler_mutex); mutex_unlock(&listen_id->handler_mutex);
rdma_destroy_id(new_cm_id); destroy_id_handler_unlock(conn_id);
goto out; return ret;
} }
conn_id->cm_id.iw = cm_id; conn_id->cm_id.iw = cm_id;
...@@ -2397,10 +2399,8 @@ static int iw_conn_req_handler(struct iw_cm_id *cm_id, ...@@ -2397,10 +2399,8 @@ static int iw_conn_req_handler(struct iw_cm_id *cm_id,
if (ret) { if (ret) {
/* User wants to destroy the CM ID */ /* User wants to destroy the CM ID */
conn_id->cm_id.iw = NULL; conn_id->cm_id.iw = NULL;
cma_exch(conn_id, RDMA_CM_DESTROYING);
mutex_unlock(&conn_id->handler_mutex);
mutex_unlock(&listen_id->handler_mutex); mutex_unlock(&listen_id->handler_mutex);
rdma_destroy_id(&conn_id->id); destroy_id_handler_unlock(conn_id);
return ret; return ret;
} }
...@@ -2638,21 +2638,21 @@ static void cma_work_handler(struct work_struct *_work) ...@@ -2638,21 +2638,21 @@ static void cma_work_handler(struct work_struct *_work)
{ {
struct cma_work *work = container_of(_work, struct cma_work, work); struct cma_work *work = container_of(_work, struct cma_work, work);
struct rdma_id_private *id_priv = work->id; struct rdma_id_private *id_priv = work->id;
int destroy = 0;
mutex_lock(&id_priv->handler_mutex); mutex_lock(&id_priv->handler_mutex);
if (!cma_comp_exch(id_priv, work->old_state, work->new_state)) if (!cma_comp_exch(id_priv, work->old_state, work->new_state))
goto out; goto out_unlock;
if (cma_cm_event_handler(id_priv, &work->event)) { if (cma_cm_event_handler(id_priv, &work->event)) {
cma_exch(id_priv, RDMA_CM_DESTROYING); cma_id_put(id_priv);
destroy = 1; destroy_id_handler_unlock(id_priv);
goto out_free;
} }
out:
out_unlock:
mutex_unlock(&id_priv->handler_mutex); mutex_unlock(&id_priv->handler_mutex);
cma_id_put(id_priv); cma_id_put(id_priv);
if (destroy) out_free:
rdma_destroy_id(&id_priv->id);
kfree(work); kfree(work);
} }
...@@ -2660,23 +2660,22 @@ static void cma_ndev_work_handler(struct work_struct *_work) ...@@ -2660,23 +2660,22 @@ static void cma_ndev_work_handler(struct work_struct *_work)
{ {
struct cma_ndev_work *work = container_of(_work, struct cma_ndev_work, work); struct cma_ndev_work *work = container_of(_work, struct cma_ndev_work, work);
struct rdma_id_private *id_priv = work->id; struct rdma_id_private *id_priv = work->id;
int destroy = 0;
mutex_lock(&id_priv->handler_mutex); mutex_lock(&id_priv->handler_mutex);
if (id_priv->state == RDMA_CM_DESTROYING || if (id_priv->state == RDMA_CM_DESTROYING ||
id_priv->state == RDMA_CM_DEVICE_REMOVAL) id_priv->state == RDMA_CM_DEVICE_REMOVAL)
goto out; goto out_unlock;
if (cma_cm_event_handler(id_priv, &work->event)) { if (cma_cm_event_handler(id_priv, &work->event)) {
cma_exch(id_priv, RDMA_CM_DESTROYING); cma_id_put(id_priv);
destroy = 1; destroy_id_handler_unlock(id_priv);
goto out_free;
} }
out: out_unlock:
mutex_unlock(&id_priv->handler_mutex); mutex_unlock(&id_priv->handler_mutex);
cma_id_put(id_priv); cma_id_put(id_priv);
if (destroy) out_free:
rdma_destroy_id(&id_priv->id);
kfree(work); kfree(work);
} }
...@@ -3152,9 +3151,7 @@ static void addr_handler(int status, struct sockaddr *src_addr, ...@@ -3152,9 +3151,7 @@ static void addr_handler(int status, struct sockaddr *src_addr,
event.event = RDMA_CM_EVENT_ADDR_RESOLVED; event.event = RDMA_CM_EVENT_ADDR_RESOLVED;
if (cma_cm_event_handler(id_priv, &event)) { if (cma_cm_event_handler(id_priv, &event)) {
cma_exch(id_priv, RDMA_CM_DESTROYING); destroy_id_handler_unlock(id_priv);
mutex_unlock(&id_priv->handler_mutex);
rdma_destroy_id(&id_priv->id);
return; return;
} }
out: out:
...@@ -3759,7 +3756,7 @@ static int cma_sidr_rep_handler(struct ib_cm_id *cm_id, ...@@ -3759,7 +3756,7 @@ static int cma_sidr_rep_handler(struct ib_cm_id *cm_id,
struct rdma_cm_event event = {}; struct rdma_cm_event event = {};
const struct ib_cm_sidr_rep_event_param *rep = const struct ib_cm_sidr_rep_event_param *rep =
&ib_event->param.sidr_rep_rcvd; &ib_event->param.sidr_rep_rcvd;
int ret = 0; int ret;
mutex_lock(&id_priv->handler_mutex); mutex_lock(&id_priv->handler_mutex);
if (id_priv->state != RDMA_CM_CONNECT) if (id_priv->state != RDMA_CM_CONNECT)
...@@ -3809,14 +3806,12 @@ static int cma_sidr_rep_handler(struct ib_cm_id *cm_id, ...@@ -3809,14 +3806,12 @@ static int cma_sidr_rep_handler(struct ib_cm_id *cm_id,
if (ret) { if (ret) {
/* Destroy the CM ID by returning a non-zero value. */ /* Destroy the CM ID by returning a non-zero value. */
id_priv->cm_id.ib = NULL; id_priv->cm_id.ib = NULL;
cma_exch(id_priv, RDMA_CM_DESTROYING); destroy_id_handler_unlock(id_priv);
mutex_unlock(&id_priv->handler_mutex);
rdma_destroy_id(&id_priv->id);
return ret; return ret;
} }
out: out:
mutex_unlock(&id_priv->handler_mutex); mutex_unlock(&id_priv->handler_mutex);
return ret; return 0;
} }
static int cma_resolve_ib_udp(struct rdma_id_private *id_priv, static int cma_resolve_ib_udp(struct rdma_id_private *id_priv,
...@@ -4341,9 +4336,7 @@ static int cma_ib_mc_handler(int status, struct ib_sa_multicast *multicast) ...@@ -4341,9 +4336,7 @@ static int cma_ib_mc_handler(int status, struct ib_sa_multicast *multicast)
rdma_destroy_ah_attr(&event.param.ud.ah_attr); rdma_destroy_ah_attr(&event.param.ud.ah_attr);
if (ret) { if (ret) {
cma_exch(id_priv, RDMA_CM_DESTROYING); destroy_id_handler_unlock(id_priv);
mutex_unlock(&id_priv->handler_mutex);
rdma_destroy_id(&id_priv->id);
return 0; return 0;
} }
...@@ -4784,7 +4777,8 @@ static void cma_send_device_removal_put(struct rdma_id_private *id_priv) ...@@ -4784,7 +4777,8 @@ static void cma_send_device_removal_put(struct rdma_id_private *id_priv)
*/ */
cma_id_put(id_priv); cma_id_put(id_priv);
mutex_unlock(&id_priv->handler_mutex); mutex_unlock(&id_priv->handler_mutex);
rdma_destroy_id(&id_priv->id); trace_cm_id_destroy(id_priv);
_destroy_id(id_priv, state);
return; return;
} }
mutex_unlock(&id_priv->handler_mutex); mutex_unlock(&id_priv->handler_mutex);
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册