Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
OpenHarmony
kernel_linux
提交
2b76c057
K
kernel_linux
项目概览
OpenHarmony
/
kernel_linux
上一次同步 4 年多
通知
15
Star
8
Fork
2
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
DevOps
流水线
流水线任务
计划
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
K
kernel_linux
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
DevOps
DevOps
流水线
流水线任务
计划
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
流水线任务
提交
Issue看板
提交
2b76c057
编写于
1月 10, 2011
作者:
R
Roland Dreier
浏览文件
操作
浏览文件
下载
差异文件
Merge branches 'cxgb4', 'ipath', 'ipoib', 'mlx4', 'mthca', 'nes', 'qib' and 'srp' into for-next
上级
db8b1016
1eba27e8
8ae31e5b
13974909
d0444f15
601d87b0
4db62d47
9af76271
变更
32
展开全部
隐藏空白更改
内联
并排
Showing
32 changed file
with
887 addition
and
531 deletion
+887
-531
drivers/infiniband/hw/ipath/ipath_driver.c
drivers/infiniband/hw/ipath/ipath_driver.c
+2
-3
drivers/infiniband/hw/mlx4/cq.c
drivers/infiniband/hw/mlx4/cq.c
+8
-1
drivers/infiniband/hw/mlx4/mad.c
drivers/infiniband/hw/mlx4/mad.c
+2
-0
drivers/infiniband/hw/mthca/mthca_mad.c
drivers/infiniband/hw/mthca/mthca_mad.c
+2
-0
drivers/infiniband/hw/nes/nes_nic.c
drivers/infiniband/hw/nes/nes_nic.c
+2
-2
drivers/infiniband/hw/qib/qib.h
drivers/infiniband/hw/qib/qib.h
+1
-1
drivers/infiniband/hw/qib/qib_cq.c
drivers/infiniband/hw/qib/qib_cq.c
+2
-1
drivers/infiniband/hw/qib/qib_driver.c
drivers/infiniband/hw/qib/qib_driver.c
+148
-7
drivers/infiniband/hw/qib/qib_file_ops.c
drivers/infiniband/hw/qib/qib_file_ops.c
+5
-5
drivers/infiniband/hw/qib/qib_iba6120.c
drivers/infiniband/hw/qib/qib_iba6120.c
+1
-1
drivers/infiniband/hw/qib/qib_iba7220.c
drivers/infiniband/hw/qib/qib_iba7220.c
+2
-2
drivers/infiniband/hw/qib/qib_iba7322.c
drivers/infiniband/hw/qib/qib_iba7322.c
+339
-34
drivers/infiniband/hw/qib/qib_init.c
drivers/infiniband/hw/qib/qib_init.c
+4
-2
drivers/infiniband/hw/qib/qib_intr.c
drivers/infiniband/hw/qib/qib_intr.c
+2
-1
drivers/infiniband/hw/qib/qib_keys.c
drivers/infiniband/hw/qib/qib_keys.c
+56
-24
drivers/infiniband/hw/qib/qib_mad.c
drivers/infiniband/hw/qib/qib_mad.c
+25
-20
drivers/infiniband/hw/qib/qib_mr.c
drivers/infiniband/hw/qib/qib_mr.c
+5
-3
drivers/infiniband/hw/qib/qib_qp.c
drivers/infiniband/hw/qib/qib_qp.c
+15
-17
drivers/infiniband/hw/qib/qib_rc.c
drivers/infiniband/hw/qib/qib_rc.c
+24
-0
drivers/infiniband/hw/qib/qib_ud.c
drivers/infiniband/hw/qib/qib_ud.c
+25
-32
drivers/infiniband/hw/qib/qib_user_sdma.c
drivers/infiniband/hw/qib/qib_user_sdma.c
+1
-0
drivers/infiniband/hw/qib/qib_verbs.h
drivers/infiniband/hw/qib/qib_verbs.h
+3
-8
drivers/infiniband/ulp/ipoib/Kconfig
drivers/infiniband/ulp/ipoib/Kconfig
+0
-1
drivers/infiniband/ulp/ipoib/ipoib.h
drivers/infiniband/ulp/ipoib/ipoib.h
+1
-11
drivers/infiniband/ulp/ipoib/ipoib_cm.c
drivers/infiniband/ulp/ipoib/ipoib_cm.c
+1
-0
drivers/infiniband/ulp/ipoib/ipoib_ethtool.c
drivers/infiniband/ulp/ipoib/ipoib_ethtool.c
+0
-51
drivers/infiniband/ulp/ipoib/ipoib_ib.c
drivers/infiniband/ulp/ipoib/ipoib_ib.c
+1
-7
drivers/infiniband/ulp/ipoib/ipoib_main.c
drivers/infiniband/ulp/ipoib/ipoib_main.c
+1
-61
drivers/infiniband/ulp/srp/ib_srp.c
drivers/infiniband/ulp/srp/ib_srp.c
+181
-211
drivers/infiniband/ulp/srp/ib_srp.h
drivers/infiniband/ulp/srp/ib_srp.h
+25
-21
drivers/net/mlx4/alloc.c
drivers/net/mlx4/alloc.c
+2
-1
drivers/net/mlx4/fw.c
drivers/net/mlx4/fw.c
+1
-3
未找到文件。
drivers/infiniband/hw/ipath/ipath_driver.c
浏览文件 @
2b76c057
...
@@ -530,9 +530,8 @@ static int __devinit ipath_init_one(struct pci_dev *pdev,
...
@@ -530,9 +530,8 @@ static int __devinit ipath_init_one(struct pci_dev *pdev,
for
(
j
=
0
;
j
<
6
;
j
++
)
{
for
(
j
=
0
;
j
<
6
;
j
++
)
{
if
(
!
pdev
->
resource
[
j
].
start
)
if
(
!
pdev
->
resource
[
j
].
start
)
continue
;
continue
;
ipath_cdbg
(
VERBOSE
,
"BAR %d start %llx, end %llx, len %llx
\n
"
,
ipath_cdbg
(
VERBOSE
,
"BAR %d %pR, len %llx
\n
"
,
j
,
(
unsigned
long
long
)
pdev
->
resource
[
j
].
start
,
j
,
&
pdev
->
resource
[
j
],
(
unsigned
long
long
)
pdev
->
resource
[
j
].
end
,
(
unsigned
long
long
)
pci_resource_len
(
pdev
,
j
));
(
unsigned
long
long
)
pci_resource_len
(
pdev
,
j
));
}
}
...
...
drivers/infiniband/hw/mlx4/cq.c
浏览文件 @
2b76c057
...
@@ -397,10 +397,14 @@ int mlx4_ib_resize_cq(struct ib_cq *ibcq, int entries, struct ib_udata *udata)
...
@@ -397,10 +397,14 @@ int mlx4_ib_resize_cq(struct ib_cq *ibcq, int entries, struct ib_udata *udata)
cq
->
resize_buf
=
NULL
;
cq
->
resize_buf
=
NULL
;
cq
->
resize_umem
=
NULL
;
cq
->
resize_umem
=
NULL
;
}
else
{
}
else
{
struct
mlx4_ib_cq_buf
tmp_buf
;
int
tmp_cqe
=
0
;
spin_lock_irq
(
&
cq
->
lock
);
spin_lock_irq
(
&
cq
->
lock
);
if
(
cq
->
resize_buf
)
{
if
(
cq
->
resize_buf
)
{
mlx4_ib_cq_resize_copy_cqes
(
cq
);
mlx4_ib_cq_resize_copy_cqes
(
cq
);
mlx4_ib_free_cq_buf
(
dev
,
&
cq
->
buf
,
cq
->
ibcq
.
cqe
);
tmp_buf
=
cq
->
buf
;
tmp_cqe
=
cq
->
ibcq
.
cqe
;
cq
->
buf
=
cq
->
resize_buf
->
buf
;
cq
->
buf
=
cq
->
resize_buf
->
buf
;
cq
->
ibcq
.
cqe
=
cq
->
resize_buf
->
cqe
;
cq
->
ibcq
.
cqe
=
cq
->
resize_buf
->
cqe
;
...
@@ -408,6 +412,9 @@ int mlx4_ib_resize_cq(struct ib_cq *ibcq, int entries, struct ib_udata *udata)
...
@@ -408,6 +412,9 @@ int mlx4_ib_resize_cq(struct ib_cq *ibcq, int entries, struct ib_udata *udata)
cq
->
resize_buf
=
NULL
;
cq
->
resize_buf
=
NULL
;
}
}
spin_unlock_irq
(
&
cq
->
lock
);
spin_unlock_irq
(
&
cq
->
lock
);
if
(
tmp_cqe
)
mlx4_ib_free_cq_buf
(
dev
,
&
tmp_buf
,
tmp_cqe
);
}
}
goto
out
;
goto
out
;
...
...
drivers/infiniband/hw/mlx4/mad.c
浏览文件 @
2b76c057
...
@@ -211,6 +211,8 @@ static void forward_trap(struct mlx4_ib_dev *dev, u8 port_num, struct ib_mad *ma
...
@@ -211,6 +211,8 @@ static void forward_trap(struct mlx4_ib_dev *dev, u8 port_num, struct ib_mad *ma
if
(
agent
)
{
if
(
agent
)
{
send_buf
=
ib_create_send_mad
(
agent
,
qpn
,
0
,
0
,
IB_MGMT_MAD_HDR
,
send_buf
=
ib_create_send_mad
(
agent
,
qpn
,
0
,
0
,
IB_MGMT_MAD_HDR
,
IB_MGMT_MAD_DATA
,
GFP_ATOMIC
);
IB_MGMT_MAD_DATA
,
GFP_ATOMIC
);
if
(
IS_ERR
(
send_buf
))
return
;
/*
/*
* We rely here on the fact that MLX QPs don't use the
* We rely here on the fact that MLX QPs don't use the
* address handle after the send is posted (this is
* address handle after the send is posted (this is
...
...
drivers/infiniband/hw/mthca/mthca_mad.c
浏览文件 @
2b76c057
...
@@ -171,6 +171,8 @@ static void forward_trap(struct mthca_dev *dev,
...
@@ -171,6 +171,8 @@ static void forward_trap(struct mthca_dev *dev,
if
(
agent
)
{
if
(
agent
)
{
send_buf
=
ib_create_send_mad
(
agent
,
qpn
,
0
,
0
,
IB_MGMT_MAD_HDR
,
send_buf
=
ib_create_send_mad
(
agent
,
qpn
,
0
,
0
,
IB_MGMT_MAD_HDR
,
IB_MGMT_MAD_DATA
,
GFP_ATOMIC
);
IB_MGMT_MAD_DATA
,
GFP_ATOMIC
);
if
(
IS_ERR
(
send_buf
))
return
;
/*
/*
* We rely here on the fact that MLX QPs don't use the
* We rely here on the fact that MLX QPs don't use the
* address handle after the send is posted (this is
* address handle after the send is posted (this is
...
...
drivers/infiniband/hw/nes/nes_nic.c
浏览文件 @
2b76c057
...
@@ -908,8 +908,8 @@ static void nes_netdev_set_multicast_list(struct net_device *netdev)
...
@@ -908,8 +908,8 @@ static void nes_netdev_set_multicast_list(struct net_device *netdev)
nesvnic
->
nic_index
&&
nesvnic
->
nic_index
&&
mc_index
<
max_pft_entries_avaiable
)
{
mc_index
<
max_pft_entries_avaiable
)
{
nes_debug
(
NES_DBG_NIC_RX
,
nes_debug
(
NES_DBG_NIC_RX
,
"mc_index=%d skipping nic_index=%d,
\
"mc_index=%d skipping nic_index=%d,
"
used for=%d
\n
"
,
mc_index
,
"
used for=%d
\n
"
,
mc_index
,
nesvnic
->
nic_index
,
nesvnic
->
nic_index
,
nesadapter
->
pft_mcast_map
[
mc_index
]);
nesadapter
->
pft_mcast_map
[
mc_index
]);
mc_index
++
;
mc_index
++
;
...
...
drivers/infiniband/hw/qib/qib.h
浏览文件 @
2b76c057
...
@@ -766,7 +766,7 @@ struct qib_devdata {
...
@@ -766,7 +766,7 @@ struct qib_devdata {
void
(
*
f_sdma_hw_start_up
)(
struct
qib_pportdata
*
);
void
(
*
f_sdma_hw_start_up
)(
struct
qib_pportdata
*
);
void
(
*
f_sdma_init_early
)(
struct
qib_pportdata
*
);
void
(
*
f_sdma_init_early
)(
struct
qib_pportdata
*
);
void
(
*
f_set_cntr_sample
)(
struct
qib_pportdata
*
,
u32
,
u32
);
void
(
*
f_set_cntr_sample
)(
struct
qib_pportdata
*
,
u32
,
u32
);
void
(
*
f_update_usrhead
)(
struct
qib_ctxtdata
*
,
u64
,
u32
,
u32
);
void
(
*
f_update_usrhead
)(
struct
qib_ctxtdata
*
,
u64
,
u32
,
u32
,
u32
);
u32
(
*
f_hdrqempty
)(
struct
qib_ctxtdata
*
);
u32
(
*
f_hdrqempty
)(
struct
qib_ctxtdata
*
);
u64
(
*
f_portcntr
)(
struct
qib_pportdata
*
,
u32
);
u64
(
*
f_portcntr
)(
struct
qib_pportdata
*
,
u32
);
u32
(
*
f_read_cntrs
)(
struct
qib_devdata
*
,
loff_t
,
char
**
,
u32
(
*
f_read_cntrs
)(
struct
qib_devdata
*
,
loff_t
,
char
**
,
...
...
drivers/infiniband/hw/qib/qib_cq.c
浏览文件 @
2b76c057
...
@@ -100,7 +100,8 @@ void qib_cq_enter(struct qib_cq *cq, struct ib_wc *entry, int solicited)
...
@@ -100,7 +100,8 @@ void qib_cq_enter(struct qib_cq *cq, struct ib_wc *entry, int solicited)
wc
->
head
=
next
;
wc
->
head
=
next
;
if
(
cq
->
notify
==
IB_CQ_NEXT_COMP
||
if
(
cq
->
notify
==
IB_CQ_NEXT_COMP
||
(
cq
->
notify
==
IB_CQ_SOLICITED
&&
solicited
))
{
(
cq
->
notify
==
IB_CQ_SOLICITED
&&
(
solicited
||
entry
->
status
!=
IB_WC_SUCCESS
)))
{
cq
->
notify
=
IB_CQ_NONE
;
cq
->
notify
=
IB_CQ_NONE
;
cq
->
triggered
++
;
cq
->
triggered
++
;
/*
/*
...
...
drivers/infiniband/hw/qib/qib_driver.c
浏览文件 @
2b76c057
...
@@ -71,6 +71,11 @@ MODULE_DESCRIPTION("QLogic IB driver");
...
@@ -71,6 +71,11 @@ MODULE_DESCRIPTION("QLogic IB driver");
*/
*/
#define QIB_PIO_MAXIBHDR 128
#define QIB_PIO_MAXIBHDR 128
/*
* QIB_MAX_PKT_RCV is the max # if packets processed per receive interrupt.
*/
#define QIB_MAX_PKT_RECV 64
struct
qlogic_ib_stats
qib_stats
;
struct
qlogic_ib_stats
qib_stats
;
const
char
*
qib_get_unit_name
(
int
unit
)
const
char
*
qib_get_unit_name
(
int
unit
)
...
@@ -284,14 +289,147 @@ static inline void *qib_get_egrbuf(const struct qib_ctxtdata *rcd, u32 etail)
...
@@ -284,14 +289,147 @@ static inline void *qib_get_egrbuf(const struct qib_ctxtdata *rcd, u32 etail)
* Returns 1 if error was a CRC, else 0.
* Returns 1 if error was a CRC, else 0.
* Needed for some chip's synthesized error counters.
* Needed for some chip's synthesized error counters.
*/
*/
static
u32
qib_rcv_hdrerr
(
struct
qib_
pportdata
*
ppd
,
u32
ctxt
,
static
u32
qib_rcv_hdrerr
(
struct
qib_
ctxtdata
*
rcd
,
struct
qib_pportdata
*
ppd
,
u32
eflags
,
u32
l
,
u32
etail
,
__le32
*
rhf_addr
,
u32
ctxt
,
u32
eflags
,
u32
l
,
u32
etail
,
struct
qib_message_header
*
hdr
)
__le32
*
rhf_addr
,
struct
qib_message_header
*
r
hdr
)
{
{
u32
ret
=
0
;
u32
ret
=
0
;
if
(
eflags
&
(
QLOGIC_IB_RHF_H_ICRCERR
|
QLOGIC_IB_RHF_H_VCRCERR
))
if
(
eflags
&
(
QLOGIC_IB_RHF_H_ICRCERR
|
QLOGIC_IB_RHF_H_VCRCERR
))
ret
=
1
;
ret
=
1
;
else
if
(
eflags
==
QLOGIC_IB_RHF_H_TIDERR
)
{
/* For TIDERR and RC QPs premptively schedule a NAK */
struct
qib_ib_header
*
hdr
=
(
struct
qib_ib_header
*
)
rhdr
;
struct
qib_other_headers
*
ohdr
=
NULL
;
struct
qib_ibport
*
ibp
=
&
ppd
->
ibport_data
;
struct
qib_qp
*
qp
=
NULL
;
u32
tlen
=
qib_hdrget_length_in_bytes
(
rhf_addr
);
u16
lid
=
be16_to_cpu
(
hdr
->
lrh
[
1
]);
int
lnh
=
be16_to_cpu
(
hdr
->
lrh
[
0
])
&
3
;
u32
qp_num
;
u32
opcode
;
u32
psn
;
int
diff
;
unsigned
long
flags
;
/* Sanity check packet */
if
(
tlen
<
24
)
goto
drop
;
if
(
lid
<
QIB_MULTICAST_LID_BASE
)
{
lid
&=
~
((
1
<<
ppd
->
lmc
)
-
1
);
if
(
unlikely
(
lid
!=
ppd
->
lid
))
goto
drop
;
}
/* Check for GRH */
if
(
lnh
==
QIB_LRH_BTH
)
ohdr
=
&
hdr
->
u
.
oth
;
else
if
(
lnh
==
QIB_LRH_GRH
)
{
u32
vtf
;
ohdr
=
&
hdr
->
u
.
l
.
oth
;
if
(
hdr
->
u
.
l
.
grh
.
next_hdr
!=
IB_GRH_NEXT_HDR
)
goto
drop
;
vtf
=
be32_to_cpu
(
hdr
->
u
.
l
.
grh
.
version_tclass_flow
);
if
((
vtf
>>
IB_GRH_VERSION_SHIFT
)
!=
IB_GRH_VERSION
)
goto
drop
;
}
else
goto
drop
;
/* Get opcode and PSN from packet */
opcode
=
be32_to_cpu
(
ohdr
->
bth
[
0
]);
opcode
>>=
24
;
psn
=
be32_to_cpu
(
ohdr
->
bth
[
2
]);
/* Get the destination QP number. */
qp_num
=
be32_to_cpu
(
ohdr
->
bth
[
1
])
&
QIB_QPN_MASK
;
if
(
qp_num
!=
QIB_MULTICAST_QPN
)
{
int
ruc_res
;
qp
=
qib_lookup_qpn
(
ibp
,
qp_num
);
if
(
!
qp
)
goto
drop
;
/*
* Handle only RC QPs - for other QP types drop error
* packet.
*/
spin_lock
(
&
qp
->
r_lock
);
/* Check for valid receive state. */
if
(
!
(
ib_qib_state_ops
[
qp
->
state
]
&
QIB_PROCESS_RECV_OK
))
{
ibp
->
n_pkt_drops
++
;
goto
unlock
;
}
switch
(
qp
->
ibqp
.
qp_type
)
{
case
IB_QPT_RC
:
spin_lock_irqsave
(
&
qp
->
s_lock
,
flags
);
ruc_res
=
qib_ruc_check_hdr
(
ibp
,
hdr
,
lnh
==
QIB_LRH_GRH
,
qp
,
be32_to_cpu
(
ohdr
->
bth
[
0
]));
if
(
ruc_res
)
{
spin_unlock_irqrestore
(
&
qp
->
s_lock
,
flags
);
goto
unlock
;
}
spin_unlock_irqrestore
(
&
qp
->
s_lock
,
flags
);
/* Only deal with RDMA Writes for now */
if
(
opcode
<
IB_OPCODE_RC_RDMA_READ_RESPONSE_FIRST
)
{
diff
=
qib_cmp24
(
psn
,
qp
->
r_psn
);
if
(
!
qp
->
r_nak_state
&&
diff
>=
0
)
{
ibp
->
n_rc_seqnak
++
;
qp
->
r_nak_state
=
IB_NAK_PSN_ERROR
;
/* Use the expected PSN. */
qp
->
r_ack_psn
=
qp
->
r_psn
;
/*
* Wait to send the sequence
* NAK until all packets
* in the receive queue have
* been processed.
* Otherwise, we end up
* propagating congestion.
*/
if
(
list_empty
(
&
qp
->
rspwait
))
{
qp
->
r_flags
|=
QIB_R_RSP_NAK
;
atomic_inc
(
&
qp
->
refcount
);
list_add_tail
(
&
qp
->
rspwait
,
&
rcd
->
qp_wait_list
);
}
}
/* Out of sequence NAK */
}
/* QP Request NAKs */
break
;
case
IB_QPT_SMI
:
case
IB_QPT_GSI
:
case
IB_QPT_UD
:
case
IB_QPT_UC
:
default:
/* For now don't handle any other QP types */
break
;
}
unlock:
spin_unlock
(
&
qp
->
r_lock
);
/*
* Notify qib_destroy_qp() if it is waiting
* for us to finish.
*/
if
(
atomic_dec_and_test
(
&
qp
->
refcount
))
wake_up
(
&
qp
->
wait
);
}
/* Unicast QP */
}
/* Valid packet with TIDErr */
drop:
return
ret
;
return
ret
;
}
}
...
@@ -335,7 +473,7 @@ u32 qib_kreceive(struct qib_ctxtdata *rcd, u32 *llic, u32 *npkts)
...
@@ -335,7 +473,7 @@ u32 qib_kreceive(struct qib_ctxtdata *rcd, u32 *llic, u32 *npkts)
smp_rmb
();
/* prevent speculative reads of dma'ed hdrq */
smp_rmb
();
/* prevent speculative reads of dma'ed hdrq */
}
}
for
(
last
=
0
,
i
=
1
;
!
last
&&
i
<=
64
;
i
+=
!
last
)
{
for
(
last
=
0
,
i
=
1
;
!
last
;
i
+=
!
last
)
{
hdr
=
dd
->
f_get_msgheader
(
dd
,
rhf_addr
);
hdr
=
dd
->
f_get_msgheader
(
dd
,
rhf_addr
);
eflags
=
qib_hdrget_err_flags
(
rhf_addr
);
eflags
=
qib_hdrget_err_flags
(
rhf_addr
);
etype
=
qib_hdrget_rcv_type
(
rhf_addr
);
etype
=
qib_hdrget_rcv_type
(
rhf_addr
);
...
@@ -371,7 +509,7 @@ u32 qib_kreceive(struct qib_ctxtdata *rcd, u32 *llic, u32 *npkts)
...
@@ -371,7 +509,7 @@ u32 qib_kreceive(struct qib_ctxtdata *rcd, u32 *llic, u32 *npkts)
* packets; only qibhdrerr should be set.
* packets; only qibhdrerr should be set.
*/
*/
if
(
unlikely
(
eflags
))
if
(
unlikely
(
eflags
))
crcs
+=
qib_rcv_hdrerr
(
ppd
,
rcd
->
ctxt
,
eflags
,
l
,
crcs
+=
qib_rcv_hdrerr
(
rcd
,
ppd
,
rcd
->
ctxt
,
eflags
,
l
,
etail
,
rhf_addr
,
hdr
);
etail
,
rhf_addr
,
hdr
);
else
if
(
etype
==
RCVHQ_RCV_TYPE_NON_KD
)
{
else
if
(
etype
==
RCVHQ_RCV_TYPE_NON_KD
)
{
qib_ib_rcv
(
rcd
,
hdr
,
ebuf
,
tlen
);
qib_ib_rcv
(
rcd
,
hdr
,
ebuf
,
tlen
);
...
@@ -384,6 +522,9 @@ u32 qib_kreceive(struct qib_ctxtdata *rcd, u32 *llic, u32 *npkts)
...
@@ -384,6 +522,9 @@ u32 qib_kreceive(struct qib_ctxtdata *rcd, u32 *llic, u32 *npkts)
l
+=
rsize
;
l
+=
rsize
;
if
(
l
>=
maxcnt
)
if
(
l
>=
maxcnt
)
l
=
0
;
l
=
0
;
if
(
i
==
QIB_MAX_PKT_RECV
)
last
=
1
;
rhf_addr
=
(
__le32
*
)
rcd
->
rcvhdrq
+
l
+
dd
->
rhf_offset
;
rhf_addr
=
(
__le32
*
)
rcd
->
rcvhdrq
+
l
+
dd
->
rhf_offset
;
if
(
dd
->
flags
&
QIB_NODMA_RTAIL
)
{
if
(
dd
->
flags
&
QIB_NODMA_RTAIL
)
{
u32
seq
=
qib_hdrget_seq
(
rhf_addr
);
u32
seq
=
qib_hdrget_seq
(
rhf_addr
);
...
@@ -402,7 +543,7 @@ u32 qib_kreceive(struct qib_ctxtdata *rcd, u32 *llic, u32 *npkts)
...
@@ -402,7 +543,7 @@ u32 qib_kreceive(struct qib_ctxtdata *rcd, u32 *llic, u32 *npkts)
*/
*/
lval
=
l
;
lval
=
l
;
if
(
!
last
&&
!
(
i
&
0xf
))
{
if
(
!
last
&&
!
(
i
&
0xf
))
{
dd
->
f_update_usrhead
(
rcd
,
lval
,
updegr
,
etail
);
dd
->
f_update_usrhead
(
rcd
,
lval
,
updegr
,
etail
,
i
);
updegr
=
0
;
updegr
=
0
;
}
}
}
}
...
@@ -444,7 +585,7 @@ u32 qib_kreceive(struct qib_ctxtdata *rcd, u32 *llic, u32 *npkts)
...
@@ -444,7 +585,7 @@ u32 qib_kreceive(struct qib_ctxtdata *rcd, u32 *llic, u32 *npkts)
* if no packets were processed.
* if no packets were processed.
*/
*/
lval
=
(
u64
)
rcd
->
head
|
dd
->
rhdrhead_intr_off
;
lval
=
(
u64
)
rcd
->
head
|
dd
->
rhdrhead_intr_off
;
dd
->
f_update_usrhead
(
rcd
,
lval
,
updegr
,
etail
);
dd
->
f_update_usrhead
(
rcd
,
lval
,
updegr
,
etail
,
i
);
return
crcs
;
return
crcs
;
}
}
...
...
drivers/infiniband/hw/qib/qib_file_ops.c
浏览文件 @
2b76c057
...
@@ -1379,17 +1379,17 @@ static int get_a_ctxt(struct file *fp, const struct qib_user_info *uinfo,
...
@@ -1379,17 +1379,17 @@ static int get_a_ctxt(struct file *fp, const struct qib_user_info *uinfo,
/* find device (with ACTIVE ports) with fewest ctxts in use */
/* find device (with ACTIVE ports) with fewest ctxts in use */
for
(
ndev
=
0
;
ndev
<
devmax
;
ndev
++
)
{
for
(
ndev
=
0
;
ndev
<
devmax
;
ndev
++
)
{
struct
qib_devdata
*
dd
=
qib_lookup
(
ndev
);
struct
qib_devdata
*
dd
=
qib_lookup
(
ndev
);
unsigned
cused
=
0
,
cfree
=
0
;
unsigned
cused
=
0
,
cfree
=
0
,
pusable
=
0
;
if
(
!
dd
)
if
(
!
dd
)
continue
;
continue
;
if
(
port
&&
port
<=
dd
->
num_pports
&&
if
(
port
&&
port
<=
dd
->
num_pports
&&
usable
(
dd
->
pport
+
port
-
1
))
usable
(
dd
->
pport
+
port
-
1
))
d
usable
=
1
;
p
usable
=
1
;
else
else
for
(
i
=
0
;
i
<
dd
->
num_pports
;
i
++
)
for
(
i
=
0
;
i
<
dd
->
num_pports
;
i
++
)
if
(
usable
(
dd
->
pport
+
i
))
if
(
usable
(
dd
->
pport
+
i
))
d
usable
++
;
p
usable
++
;
if
(
!
d
usable
)
if
(
!
p
usable
)
continue
;
continue
;
for
(
ctxt
=
dd
->
first_user_ctxt
;
ctxt
<
dd
->
cfgctxts
;
for
(
ctxt
=
dd
->
first_user_ctxt
;
ctxt
<
dd
->
cfgctxts
;
ctxt
++
)
ctxt
++
)
...
@@ -1397,7 +1397,7 @@ static int get_a_ctxt(struct file *fp, const struct qib_user_info *uinfo,
...
@@ -1397,7 +1397,7 @@ static int get_a_ctxt(struct file *fp, const struct qib_user_info *uinfo,
cused
++
;
cused
++
;
else
else
cfree
++
;
cfree
++
;
if
(
cfree
&&
cused
<
inuse
)
{
if
(
pusable
&&
cfree
&&
cused
<
inuse
)
{
udd
=
dd
;
udd
=
dd
;
inuse
=
cused
;
inuse
=
cused
;
}
}
...
...
drivers/infiniband/hw/qib/qib_iba6120.c
浏览文件 @
2b76c057
...
@@ -2074,7 +2074,7 @@ static void qib_6120_config_ctxts(struct qib_devdata *dd)
...
@@ -2074,7 +2074,7 @@ static void qib_6120_config_ctxts(struct qib_devdata *dd)
}
}
static
void
qib_update_6120_usrhead
(
struct
qib_ctxtdata
*
rcd
,
u64
hd
,
static
void
qib_update_6120_usrhead
(
struct
qib_ctxtdata
*
rcd
,
u64
hd
,
u32
updegr
,
u32
egrhd
)
u32
updegr
,
u32
egrhd
,
u32
npkts
)
{
{
qib_write_ureg
(
rcd
->
dd
,
ur_rcvhdrhead
,
hd
,
rcd
->
ctxt
);
qib_write_ureg
(
rcd
->
dd
,
ur_rcvhdrhead
,
hd
,
rcd
->
ctxt
);
if
(
updegr
)
if
(
updegr
)
...
...
drivers/infiniband/hw/qib/qib_iba7220.c
浏览文件 @
2b76c057
...
@@ -2297,7 +2297,7 @@ static void qib_7220_config_ctxts(struct qib_devdata *dd)
...
@@ -2297,7 +2297,7 @@ static void qib_7220_config_ctxts(struct qib_devdata *dd)
nchipctxts
=
qib_read_kreg32
(
dd
,
kr_portcnt
);
nchipctxts
=
qib_read_kreg32
(
dd
,
kr_portcnt
);
dd
->
cspec
->
numctxts
=
nchipctxts
;
dd
->
cspec
->
numctxts
=
nchipctxts
;
if
(
qib_n_krcv_queues
>
1
)
{
if
(
qib_n_krcv_queues
>
1
)
{
dd
->
qpn_mask
=
0x3
f
;
dd
->
qpn_mask
=
0x3
e
;
dd
->
first_user_ctxt
=
qib_n_krcv_queues
*
dd
->
num_pports
;
dd
->
first_user_ctxt
=
qib_n_krcv_queues
*
dd
->
num_pports
;
if
(
dd
->
first_user_ctxt
>
nchipctxts
)
if
(
dd
->
first_user_ctxt
>
nchipctxts
)
dd
->
first_user_ctxt
=
nchipctxts
;
dd
->
first_user_ctxt
=
nchipctxts
;
...
@@ -2703,7 +2703,7 @@ static int qib_7220_set_loopback(struct qib_pportdata *ppd, const char *what)
...
@@ -2703,7 +2703,7 @@ static int qib_7220_set_loopback(struct qib_pportdata *ppd, const char *what)
}
}
static
void
qib_update_7220_usrhead
(
struct
qib_ctxtdata
*
rcd
,
u64
hd
,
static
void
qib_update_7220_usrhead
(
struct
qib_ctxtdata
*
rcd
,
u64
hd
,
u32
updegr
,
u32
egrhd
)
u32
updegr
,
u32
egrhd
,
u32
npkts
)
{
{
qib_write_ureg
(
rcd
->
dd
,
ur_rcvhdrhead
,
hd
,
rcd
->
ctxt
);
qib_write_ureg
(
rcd
->
dd
,
ur_rcvhdrhead
,
hd
,
rcd
->
ctxt
);
if
(
updegr
)
if
(
updegr
)
...
...
drivers/infiniband/hw/qib/qib_iba7322.c
浏览文件 @
2b76c057
此差异已折叠。
点击以展开。
drivers/infiniband/hw/qib/qib_init.c
浏览文件 @
2b76c057
...
@@ -92,9 +92,11 @@ unsigned long *qib_cpulist;
...
@@ -92,9 +92,11 @@ unsigned long *qib_cpulist;
/* set number of contexts we'll actually use */
/* set number of contexts we'll actually use */
void
qib_set_ctxtcnt
(
struct
qib_devdata
*
dd
)
void
qib_set_ctxtcnt
(
struct
qib_devdata
*
dd
)
{
{
if
(
!
qib_cfgctxts
)
if
(
!
qib_cfgctxts
)
{
dd
->
cfgctxts
=
dd
->
first_user_ctxt
+
num_online_cpus
();
dd
->
cfgctxts
=
dd
->
first_user_ctxt
+
num_online_cpus
();
else
if
(
qib_cfgctxts
<
dd
->
num_pports
)
if
(
dd
->
cfgctxts
>
dd
->
ctxtcnt
)
dd
->
cfgctxts
=
dd
->
ctxtcnt
;
}
else
if
(
qib_cfgctxts
<
dd
->
num_pports
)
dd
->
cfgctxts
=
dd
->
ctxtcnt
;
dd
->
cfgctxts
=
dd
->
ctxtcnt
;
else
if
(
qib_cfgctxts
<=
dd
->
ctxtcnt
)
else
if
(
qib_cfgctxts
<=
dd
->
ctxtcnt
)
dd
->
cfgctxts
=
qib_cfgctxts
;
dd
->
cfgctxts
=
qib_cfgctxts
;
...
...
drivers/infiniband/hw/qib/qib_intr.c
浏览文件 @
2b76c057
...
@@ -131,7 +131,8 @@ void qib_handle_e_ibstatuschanged(struct qib_pportdata *ppd, u64 ibcs)
...
@@ -131,7 +131,8 @@ void qib_handle_e_ibstatuschanged(struct qib_pportdata *ppd, u64 ibcs)
/* start a 75msec timer to clear symbol errors */
/* start a 75msec timer to clear symbol errors */
mod_timer
(
&
ppd
->
symerr_clear_timer
,
mod_timer
(
&
ppd
->
symerr_clear_timer
,
msecs_to_jiffies
(
75
));
msecs_to_jiffies
(
75
));
}
else
if
(
ltstate
==
IB_PHYSPORTSTATE_LINKUP
)
{
}
else
if
(
ltstate
==
IB_PHYSPORTSTATE_LINKUP
&&
!
(
ppd
->
lflags
&
QIBL_LINKACTIVE
))
{
/* active, but not active defered */
/* active, but not active defered */
qib_hol_up
(
ppd
);
/* useful only for 6120 now */
qib_hol_up
(
ppd
);
/* useful only for 6120 now */
*
ppd
->
statusp
|=
*
ppd
->
statusp
|=
...
...
drivers/infiniband/hw/qib/qib_keys.c
浏览文件 @
2b76c057
...
@@ -136,7 +136,6 @@ int qib_lkey_ok(struct qib_lkey_table *rkt, struct qib_pd *pd,
...
@@ -136,7 +136,6 @@ int qib_lkey_ok(struct qib_lkey_table *rkt, struct qib_pd *pd,
struct
qib_mregion
*
mr
;
struct
qib_mregion
*
mr
;
unsigned
n
,
m
;
unsigned
n
,
m
;
size_t
off
;
size_t
off
;
int
ret
=
0
;
unsigned
long
flags
;
unsigned
long
flags
;
/*
/*
...
@@ -152,6 +151,8 @@ int qib_lkey_ok(struct qib_lkey_table *rkt, struct qib_pd *pd,
...
@@ -152,6 +151,8 @@ int qib_lkey_ok(struct qib_lkey_table *rkt, struct qib_pd *pd,
if
(
!
dev
->
dma_mr
)
if
(
!
dev
->
dma_mr
)
goto
bail
;
goto
bail
;
atomic_inc
(
&
dev
->
dma_mr
->
refcount
);
atomic_inc
(
&
dev
->
dma_mr
->
refcount
);
spin_unlock_irqrestore
(
&
rkt
->
lock
,
flags
);
isge
->
mr
=
dev
->
dma_mr
;
isge
->
mr
=
dev
->
dma_mr
;
isge
->
vaddr
=
(
void
*
)
sge
->
addr
;
isge
->
vaddr
=
(
void
*
)
sge
->
addr
;
isge
->
length
=
sge
->
length
;
isge
->
length
=
sge
->
length
;
...
@@ -170,19 +171,34 @@ int qib_lkey_ok(struct qib_lkey_table *rkt, struct qib_pd *pd,
...
@@ -170,19 +171,34 @@ int qib_lkey_ok(struct qib_lkey_table *rkt, struct qib_pd *pd,
off
+
sge
->
length
>
mr
->
length
||
off
+
sge
->
length
>
mr
->
length
||
(
mr
->
access_flags
&
acc
)
!=
acc
))
(
mr
->
access_flags
&
acc
)
!=
acc
))
goto
bail
;
goto
bail
;
atomic_inc
(
&
mr
->
refcount
);
spin_unlock_irqrestore
(
&
rkt
->
lock
,
flags
);
off
+=
mr
->
offset
;
off
+=
mr
->
offset
;
m
=
0
;
if
(
mr
->
page_shift
)
{
n
=
0
;
/*
while
(
off
>=
mr
->
map
[
m
]
->
segs
[
n
].
length
)
{
page sizes are uniform power of 2 so no loop is necessary
off
-=
mr
->
map
[
m
]
->
segs
[
n
].
length
;
entries_spanned_by_off is the number of times the loop below
n
++
;
would have executed.
if
(
n
>=
QIB_SEGSZ
)
{
*/
m
++
;
size_t
entries_spanned_by_off
;
n
=
0
;
entries_spanned_by_off
=
off
>>
mr
->
page_shift
;
off
-=
(
entries_spanned_by_off
<<
mr
->
page_shift
);
m
=
entries_spanned_by_off
/
QIB_SEGSZ
;
n
=
entries_spanned_by_off
%
QIB_SEGSZ
;
}
else
{
m
=
0
;
n
=
0
;
while
(
off
>=
mr
->
map
[
m
]
->
segs
[
n
].
length
)
{
off
-=
mr
->
map
[
m
]
->
segs
[
n
].
length
;
n
++
;
if
(
n
>=
QIB_SEGSZ
)
{
m
++
;
n
=
0
;
}
}
}
}
}
atomic_inc
(
&
mr
->
refcount
);
isge
->
mr
=
mr
;
isge
->
mr
=
mr
;
isge
->
vaddr
=
mr
->
map
[
m
]
->
segs
[
n
].
vaddr
+
off
;
isge
->
vaddr
=
mr
->
map
[
m
]
->
segs
[
n
].
vaddr
+
off
;
isge
->
length
=
mr
->
map
[
m
]
->
segs
[
n
].
length
-
off
;
isge
->
length
=
mr
->
map
[
m
]
->
segs
[
n
].
length
-
off
;
...
@@ -190,10 +206,10 @@ int qib_lkey_ok(struct qib_lkey_table *rkt, struct qib_pd *pd,
...
@@ -190,10 +206,10 @@ int qib_lkey_ok(struct qib_lkey_table *rkt, struct qib_pd *pd,
isge
->
m
=
m
;
isge
->
m
=
m
;
isge
->
n
=
n
;
isge
->
n
=
n
;
ok:
ok:
ret
=
1
;
ret
urn
1
;
bail:
bail:
spin_unlock_irqrestore
(
&
rkt
->
lock
,
flags
);
spin_unlock_irqrestore
(
&
rkt
->
lock
,
flags
);
return
ret
;
return
0
;
}
}
/**
/**
...
@@ -214,7 +230,6 @@ int qib_rkey_ok(struct qib_qp *qp, struct qib_sge *sge,
...
@@ -214,7 +230,6 @@ int qib_rkey_ok(struct qib_qp *qp, struct qib_sge *sge,
struct
qib_mregion
*
mr
;
struct
qib_mregion
*
mr
;
unsigned
n
,
m
;
unsigned
n
,
m
;
size_t
off
;
size_t
off
;
int
ret
=
0
;
unsigned
long
flags
;
unsigned
long
flags
;
/*
/*
...
@@ -231,6 +246,8 @@ int qib_rkey_ok(struct qib_qp *qp, struct qib_sge *sge,
...
@@ -231,6 +246,8 @@ int qib_rkey_ok(struct qib_qp *qp, struct qib_sge *sge,
if
(
!
dev
->
dma_mr
)
if
(
!
dev
->
dma_mr
)
goto
bail
;
goto
bail
;
atomic_inc
(
&
dev
->
dma_mr
->
refcount
);
atomic_inc
(
&
dev
->
dma_mr
->
refcount
);
spin_unlock_irqrestore
(
&
rkt
->
lock
,
flags
);
sge
->
mr
=
dev
->
dma_mr
;
sge
->
mr
=
dev
->
dma_mr
;
sge
->
vaddr
=
(
void
*
)
vaddr
;
sge
->
vaddr
=
(
void
*
)
vaddr
;
sge
->
length
=
len
;
sge
->
length
=
len
;
...
@@ -248,19 +265,34 @@ int qib_rkey_ok(struct qib_qp *qp, struct qib_sge *sge,
...
@@ -248,19 +265,34 @@ int qib_rkey_ok(struct qib_qp *qp, struct qib_sge *sge,
if
(
unlikely
(
vaddr
<
mr
->
iova
||
off
+
len
>
mr
->
length
||
if
(
unlikely
(
vaddr
<
mr
->
iova
||
off
+
len
>
mr
->
length
||
(
mr
->
access_flags
&
acc
)
==
0
))
(
mr
->
access_flags
&
acc
)
==
0
))
goto
bail
;
goto
bail
;
atomic_inc
(
&
mr
->
refcount
);
spin_unlock_irqrestore
(
&
rkt
->
lock
,
flags
);
off
+=
mr
->
offset
;
off
+=
mr
->
offset
;
m
=
0
;
if
(
mr
->
page_shift
)
{
n
=
0
;
/*
while
(
off
>=
mr
->
map
[
m
]
->
segs
[
n
].
length
)
{
page sizes are uniform power of 2 so no loop is necessary
off
-=
mr
->
map
[
m
]
->
segs
[
n
].
length
;
entries_spanned_by_off is the number of times the loop below
n
++
;
would have executed.
if
(
n
>=
QIB_SEGSZ
)
{
*/
m
++
;
size_t
entries_spanned_by_off
;
n
=
0
;
entries_spanned_by_off
=
off
>>
mr
->
page_shift
;
off
-=
(
entries_spanned_by_off
<<
mr
->
page_shift
);
m
=
entries_spanned_by_off
/
QIB_SEGSZ
;
n
=
entries_spanned_by_off
%
QIB_SEGSZ
;
}
else
{
m
=
0
;
n
=
0
;
while
(
off
>=
mr
->
map
[
m
]
->
segs
[
n
].
length
)
{
off
-=
mr
->
map
[
m
]
->
segs
[
n
].
length
;
n
++
;
if
(
n
>=
QIB_SEGSZ
)
{
m
++
;
n
=
0
;
}
}
}
}
}
atomic_inc
(
&
mr
->
refcount
);
sge
->
mr
=
mr
;
sge
->
mr
=
mr
;
sge
->
vaddr
=
mr
->
map
[
m
]
->
segs
[
n
].
vaddr
+
off
;
sge
->
vaddr
=
mr
->
map
[
m
]
->
segs
[
n
].
vaddr
+
off
;
sge
->
length
=
mr
->
map
[
m
]
->
segs
[
n
].
length
-
off
;
sge
->
length
=
mr
->
map
[
m
]
->
segs
[
n
].
length
-
off
;
...
@@ -268,10 +300,10 @@ int qib_rkey_ok(struct qib_qp *qp, struct qib_sge *sge,
...
@@ -268,10 +300,10 @@ int qib_rkey_ok(struct qib_qp *qp, struct qib_sge *sge,
sge
->
m
=
m
;
sge
->
m
=
m
;
sge
->
n
=
n
;
sge
->
n
=
n
;
ok:
ok:
ret
=
1
;
ret
urn
1
;
bail:
bail:
spin_unlock_irqrestore
(
&
rkt
->
lock
,
flags
);
spin_unlock_irqrestore
(
&
rkt
->
lock
,
flags
);
return
ret
;
return
0
;
}
}
/*
/*
...
...
drivers/infiniband/hw/qib/qib_mad.c
浏览文件 @
2b76c057
...
@@ -668,8 +668,8 @@ static int subn_set_portinfo(struct ib_smp *smp, struct ib_device *ibdev,
...
@@ -668,8 +668,8 @@ static int subn_set_portinfo(struct ib_smp *smp, struct ib_device *ibdev,
lid
=
be16_to_cpu
(
pip
->
lid
);
lid
=
be16_to_cpu
(
pip
->
lid
);
/* Must be a valid unicast LID address. */
/* Must be a valid unicast LID address. */
if
(
lid
==
0
||
lid
>=
QIB_MULTICAST_LID_BASE
)
if
(
lid
==
0
||
lid
>=
QIB_MULTICAST_LID_BASE
)
goto
err
;
smp
->
status
|=
IB_SMP_INVALID_FIELD
;
if
(
ppd
->
lid
!=
lid
||
ppd
->
lmc
!=
(
pip
->
mkeyprot_resv_lmc
&
7
))
{
else
if
(
ppd
->
lid
!=
lid
||
ppd
->
lmc
!=
(
pip
->
mkeyprot_resv_lmc
&
7
))
{
if
(
ppd
->
lid
!=
lid
)
if
(
ppd
->
lid
!=
lid
)
qib_set_uevent_bits
(
ppd
,
_QIB_EVENT_LID_CHANGE_BIT
);
qib_set_uevent_bits
(
ppd
,
_QIB_EVENT_LID_CHANGE_BIT
);
if
(
ppd
->
lmc
!=
(
pip
->
mkeyprot_resv_lmc
&
7
))
if
(
ppd
->
lmc
!=
(
pip
->
mkeyprot_resv_lmc
&
7
))
...
@@ -683,8 +683,8 @@ static int subn_set_portinfo(struct ib_smp *smp, struct ib_device *ibdev,
...
@@ -683,8 +683,8 @@ static int subn_set_portinfo(struct ib_smp *smp, struct ib_device *ibdev,
msl
=
pip
->
neighbormtu_mastersmsl
&
0xF
;
msl
=
pip
->
neighbormtu_mastersmsl
&
0xF
;
/* Must be a valid unicast LID address. */
/* Must be a valid unicast LID address. */
if
(
smlid
==
0
||
smlid
>=
QIB_MULTICAST_LID_BASE
)
if
(
smlid
==
0
||
smlid
>=
QIB_MULTICAST_LID_BASE
)
goto
err
;
smp
->
status
|=
IB_SMP_INVALID_FIELD
;
if
(
smlid
!=
ibp
->
sm_lid
||
msl
!=
ibp
->
sm_sl
)
{
else
if
(
smlid
!=
ibp
->
sm_lid
||
msl
!=
ibp
->
sm_sl
)
{
spin_lock_irqsave
(
&
ibp
->
lock
,
flags
);
spin_lock_irqsave
(
&
ibp
->
lock
,
flags
);
if
(
ibp
->
sm_ah
)
{
if
(
ibp
->
sm_ah
)
{
if
(
smlid
!=
ibp
->
sm_lid
)
if
(
smlid
!=
ibp
->
sm_lid
)
...
@@ -707,8 +707,9 @@ static int subn_set_portinfo(struct ib_smp *smp, struct ib_device *ibdev,
...
@@ -707,8 +707,9 @@ static int subn_set_portinfo(struct ib_smp *smp, struct ib_device *ibdev,
if
(
lwe
==
0xFF
)
if
(
lwe
==
0xFF
)
lwe
=
ppd
->
link_width_supported
;
lwe
=
ppd
->
link_width_supported
;
else
if
(
lwe
>=
16
||
(
lwe
&
~
ppd
->
link_width_supported
))
else
if
(
lwe
>=
16
||
(
lwe
&
~
ppd
->
link_width_supported
))
goto
err
;
smp
->
status
|=
IB_SMP_INVALID_FIELD
;
set_link_width_enabled
(
ppd
,
lwe
);
else
if
(
lwe
!=
ppd
->
link_width_enabled
)
set_link_width_enabled
(
ppd
,
lwe
);
}
}
lse
=
pip
->
linkspeedactive_enabled
&
0xF
;
lse
=
pip
->
linkspeedactive_enabled
&
0xF
;
...
@@ -721,8 +722,9 @@ static int subn_set_portinfo(struct ib_smp *smp, struct ib_device *ibdev,
...
@@ -721,8 +722,9 @@ static int subn_set_portinfo(struct ib_smp *smp, struct ib_device *ibdev,
if
(
lse
==
15
)
if
(
lse
==
15
)
lse
=
ppd
->
link_speed_supported
;
lse
=
ppd
->
link_speed_supported
;
else
if
(
lse
>=
8
||
(
lse
&
~
ppd
->
link_speed_supported
))
else
if
(
lse
>=
8
||
(
lse
&
~
ppd
->
link_speed_supported
))
goto
err
;
smp
->
status
|=
IB_SMP_INVALID_FIELD
;
set_link_speed_enabled
(
ppd
,
lse
);
else
if
(
lse
!=
ppd
->
link_speed_enabled
)
set_link_speed_enabled
(
ppd
,
lse
);
}
}
/* Set link down default state. */
/* Set link down default state. */
...
@@ -738,7 +740,7 @@ static int subn_set_portinfo(struct ib_smp *smp, struct ib_device *ibdev,
...
@@ -738,7 +740,7 @@ static int subn_set_portinfo(struct ib_smp *smp, struct ib_device *ibdev,
IB_LINKINITCMD_POLL
);
IB_LINKINITCMD_POLL
);
break
;
break
;
default:
default:
goto
err
;
smp
->
status
|=
IB_SMP_INVALID_FIELD
;
}
}
ibp
->
mkeyprot
=
pip
->
mkeyprot_resv_lmc
>>
6
;
ibp
->
mkeyprot
=
pip
->
mkeyprot_resv_lmc
>>
6
;
...
@@ -748,15 +750,17 @@ static int subn_set_portinfo(struct ib_smp *smp, struct ib_device *ibdev,
...
@@ -748,15 +750,17 @@ static int subn_set_portinfo(struct ib_smp *smp, struct ib_device *ibdev,
mtu
=
ib_mtu_enum_to_int
((
pip
->
neighbormtu_mastersmsl
>>
4
)
&
0xF
);
mtu
=
ib_mtu_enum_to_int
((
pip
->
neighbormtu_mastersmsl
>>
4
)
&
0xF
);
if
(
mtu
==
-
1
)
if
(
mtu
==
-
1
)
goto
err
;
smp
->
status
|=
IB_SMP_INVALID_FIELD
;
qib_set_mtu
(
ppd
,
mtu
);
else
qib_set_mtu
(
ppd
,
mtu
);
/* Set operational VLs */
/* Set operational VLs */
vls
=
(
pip
->
operationalvl_pei_peo_fpi_fpo
>>
4
)
&
0xF
;
vls
=
(
pip
->
operationalvl_pei_peo_fpi_fpo
>>
4
)
&
0xF
;
if
(
vls
)
{
if
(
vls
)
{
if
(
vls
>
ppd
->
vls_supported
)
if
(
vls
>
ppd
->
vls_supported
)
goto
err
;
smp
->
status
|=
IB_SMP_INVALID_FIELD
;
(
void
)
dd
->
f_set_ib_cfg
(
ppd
,
QIB_IB_CFG_OP_VLS
,
vls
);
else
(
void
)
dd
->
f_set_ib_cfg
(
ppd
,
QIB_IB_CFG_OP_VLS
,
vls
);
}
}
if
(
pip
->
mkey_violations
==
0
)
if
(
pip
->
mkey_violations
==
0
)
...
@@ -770,10 +774,10 @@ static int subn_set_portinfo(struct ib_smp *smp, struct ib_device *ibdev,
...
@@ -770,10 +774,10 @@ static int subn_set_portinfo(struct ib_smp *smp, struct ib_device *ibdev,
ore
=
pip
->
localphyerrors_overrunerrors
;
ore
=
pip
->
localphyerrors_overrunerrors
;
if
(
set_phyerrthreshold
(
ppd
,
(
ore
>>
4
)
&
0xF
))
if
(
set_phyerrthreshold
(
ppd
,
(
ore
>>
4
)
&
0xF
))
goto
err
;
smp
->
status
|=
IB_SMP_INVALID_FIELD
;
if
(
set_overrunthreshold
(
ppd
,
(
ore
&
0xF
)))
if
(
set_overrunthreshold
(
ppd
,
(
ore
&
0xF
)))
goto
err
;
smp
->
status
|=
IB_SMP_INVALID_FIELD
;
ibp
->
subnet_timeout
=
pip
->
clientrereg_resv_subnetto
&
0x1F
;
ibp
->
subnet_timeout
=
pip
->
clientrereg_resv_subnetto
&
0x1F
;
...
@@ -792,7 +796,7 @@ static int subn_set_portinfo(struct ib_smp *smp, struct ib_device *ibdev,
...
@@ -792,7 +796,7 @@ static int subn_set_portinfo(struct ib_smp *smp, struct ib_device *ibdev,
state
=
pip
->
linkspeed_portstate
&
0xF
;
state
=
pip
->
linkspeed_portstate
&
0xF
;
lstate
=
(
pip
->
portphysstate_linkdown
>>
4
)
&
0xF
;
lstate
=
(
pip
->
portphysstate_linkdown
>>
4
)
&
0xF
;
if
(
lstate
&&
!
(
state
==
IB_PORT_DOWN
||
state
==
IB_PORT_NOP
))
if
(
lstate
&&
!
(
state
==
IB_PORT_DOWN
||
state
==
IB_PORT_NOP
))
goto
err
;
smp
->
status
|=
IB_SMP_INVALID_FIELD
;
/*
/*
* Only state changes of DOWN, ARM, and ACTIVE are valid
* Only state changes of DOWN, ARM, and ACTIVE are valid
...
@@ -812,8 +816,10 @@ static int subn_set_portinfo(struct ib_smp *smp, struct ib_device *ibdev,
...
@@ -812,8 +816,10 @@ static int subn_set_portinfo(struct ib_smp *smp, struct ib_device *ibdev,
lstate
=
QIB_IB_LINKDOWN
;
lstate
=
QIB_IB_LINKDOWN
;
else
if
(
lstate
==
3
)
else
if
(
lstate
==
3
)
lstate
=
QIB_IB_LINKDOWN_DISABLE
;
lstate
=
QIB_IB_LINKDOWN_DISABLE
;
else
else
{
goto
err
;
smp
->
status
|=
IB_SMP_INVALID_FIELD
;
break
;
}
spin_lock_irqsave
(
&
ppd
->
lflags_lock
,
flags
);
spin_lock_irqsave
(
&
ppd
->
lflags_lock
,
flags
);
ppd
->
lflags
&=
~
QIBL_LINKV
;
ppd
->
lflags
&=
~
QIBL_LINKV
;
spin_unlock_irqrestore
(
&
ppd
->
lflags_lock
,
flags
);
spin_unlock_irqrestore
(
&
ppd
->
lflags_lock
,
flags
);
...
@@ -835,8 +841,7 @@ static int subn_set_portinfo(struct ib_smp *smp, struct ib_device *ibdev,
...
@@ -835,8 +841,7 @@ static int subn_set_portinfo(struct ib_smp *smp, struct ib_device *ibdev,
qib_set_linkstate
(
ppd
,
QIB_IB_LINKACTIVE
);
qib_set_linkstate
(
ppd
,
QIB_IB_LINKACTIVE
);
break
;
break
;
default:
default:
/* XXX We have already partially updated our state! */
smp
->
status
|=
IB_SMP_INVALID_FIELD
;
goto
err
;
}
}
ret
=
subn_get_portinfo
(
smp
,
ibdev
,
port
);
ret
=
subn_get_portinfo
(
smp
,
ibdev
,
port
);
...
...
drivers/infiniband/hw/qib/qib_mr.c
浏览文件 @
2b76c057
...
@@ -39,7 +39,6 @@
...
@@ -39,7 +39,6 @@
/* Fast memory region */
/* Fast memory region */
struct
qib_fmr
{
struct
qib_fmr
{
struct
ib_fmr
ibfmr
;
struct
ib_fmr
ibfmr
;
u8
page_shift
;
struct
qib_mregion
mr
;
/* must be last */
struct
qib_mregion
mr
;
/* must be last */
};
};
...
@@ -107,6 +106,7 @@ static struct qib_mr *alloc_mr(int count, struct qib_lkey_table *lk_table)
...
@@ -107,6 +106,7 @@ static struct qib_mr *alloc_mr(int count, struct qib_lkey_table *lk_table)
goto
bail
;
goto
bail
;
}
}
mr
->
mr
.
mapsz
=
m
;
mr
->
mr
.
mapsz
=
m
;
mr
->
mr
.
page_shift
=
0
;
mr
->
mr
.
max_segs
=
count
;
mr
->
mr
.
max_segs
=
count
;
/*
/*
...
@@ -231,6 +231,8 @@ struct ib_mr *qib_reg_user_mr(struct ib_pd *pd, u64 start, u64 length,
...
@@ -231,6 +231,8 @@ struct ib_mr *qib_reg_user_mr(struct ib_pd *pd, u64 start, u64 length,
mr
->
mr
.
access_flags
=
mr_access_flags
;
mr
->
mr
.
access_flags
=
mr_access_flags
;
mr
->
umem
=
umem
;
mr
->
umem
=
umem
;
if
(
is_power_of_2
(
umem
->
page_size
))
mr
->
mr
.
page_shift
=
ilog2
(
umem
->
page_size
);
m
=
0
;
m
=
0
;
n
=
0
;
n
=
0
;
list_for_each_entry
(
chunk
,
&
umem
->
chunk_list
,
list
)
{
list_for_each_entry
(
chunk
,
&
umem
->
chunk_list
,
list
)
{
...
@@ -390,7 +392,7 @@ struct ib_fmr *qib_alloc_fmr(struct ib_pd *pd, int mr_access_flags,
...
@@ -390,7 +392,7 @@ struct ib_fmr *qib_alloc_fmr(struct ib_pd *pd, int mr_access_flags,
fmr
->
mr
.
offset
=
0
;
fmr
->
mr
.
offset
=
0
;
fmr
->
mr
.
access_flags
=
mr_access_flags
;
fmr
->
mr
.
access_flags
=
mr_access_flags
;
fmr
->
mr
.
max_segs
=
fmr_attr
->
max_pages
;
fmr
->
mr
.
max_segs
=
fmr_attr
->
max_pages
;
fmr
->
page_shift
=
fmr_attr
->
page_shift
;
fmr
->
mr
.
page_shift
=
fmr_attr
->
page_shift
;
atomic_set
(
&
fmr
->
mr
.
refcount
,
0
);
atomic_set
(
&
fmr
->
mr
.
refcount
,
0
);
ret
=
&
fmr
->
ibfmr
;
ret
=
&
fmr
->
ibfmr
;
...
@@ -437,7 +439,7 @@ int qib_map_phys_fmr(struct ib_fmr *ibfmr, u64 *page_list,
...
@@ -437,7 +439,7 @@ int qib_map_phys_fmr(struct ib_fmr *ibfmr, u64 *page_list,
spin_lock_irqsave
(
&
rkt
->
lock
,
flags
);
spin_lock_irqsave
(
&
rkt
->
lock
,
flags
);
fmr
->
mr
.
user_base
=
iova
;
fmr
->
mr
.
user_base
=
iova
;
fmr
->
mr
.
iova
=
iova
;
fmr
->
mr
.
iova
=
iova
;
ps
=
1
<<
fmr
->
page_shift
;
ps
=
1
<<
fmr
->
mr
.
page_shift
;
fmr
->
mr
.
length
=
list_len
*
ps
;
fmr
->
mr
.
length
=
list_len
*
ps
;
m
=
0
;
m
=
0
;
n
=
0
;
n
=
0
;
...
...
drivers/infiniband/hw/qib/qib_qp.c
浏览文件 @
2b76c057
...
@@ -48,13 +48,12 @@ static inline unsigned mk_qpn(struct qib_qpn_table *qpt,
...
@@ -48,13 +48,12 @@ static inline unsigned mk_qpn(struct qib_qpn_table *qpt,
static
inline
unsigned
find_next_offset
(
struct
qib_qpn_table
*
qpt
,
static
inline
unsigned
find_next_offset
(
struct
qib_qpn_table
*
qpt
,
struct
qpn_map
*
map
,
unsigned
off
,
struct
qpn_map
*
map
,
unsigned
off
,
unsigned
r
)
unsigned
n
)
{
{
if
(
qpt
->
mask
)
{
if
(
qpt
->
mask
)
{
off
++
;
off
++
;
if
((
off
&
qpt
->
mask
)
>>
1
!=
r
)
if
(((
off
&
qpt
->
mask
)
>>
1
)
>=
n
)
off
=
((
off
&
qpt
->
mask
)
?
off
=
(
off
|
qpt
->
mask
)
+
2
;
(
off
|
qpt
->
mask
)
+
1
:
off
)
|
(
r
<<
1
);
}
else
}
else
off
=
find_next_zero_bit
(
map
->
page
,
BITS_PER_PAGE
,
off
);
off
=
find_next_zero_bit
(
map
->
page
,
BITS_PER_PAGE
,
off
);
return
off
;
return
off
;
...
@@ -123,7 +122,6 @@ static int alloc_qpn(struct qib_devdata *dd, struct qib_qpn_table *qpt,
...
@@ -123,7 +122,6 @@ static int alloc_qpn(struct qib_devdata *dd, struct qib_qpn_table *qpt,
u32
i
,
offset
,
max_scan
,
qpn
;
u32
i
,
offset
,
max_scan
,
qpn
;
struct
qpn_map
*
map
;
struct
qpn_map
*
map
;
u32
ret
;
u32
ret
;
int
r
;
if
(
type
==
IB_QPT_SMI
||
type
==
IB_QPT_GSI
)
{
if
(
type
==
IB_QPT_SMI
||
type
==
IB_QPT_GSI
)
{
unsigned
n
;
unsigned
n
;
...
@@ -139,15 +137,11 @@ static int alloc_qpn(struct qib_devdata *dd, struct qib_qpn_table *qpt,
...
@@ -139,15 +137,11 @@ static int alloc_qpn(struct qib_devdata *dd, struct qib_qpn_table *qpt,
goto
bail
;
goto
bail
;
}
}
r
=
smp_processor_id
();
qpn
=
qpt
->
last
+
2
;
if
(
r
>=
dd
->
n_krcv_queues
)
r
%=
dd
->
n_krcv_queues
;
qpn
=
qpt
->
last
+
1
;
if
(
qpn
>=
QPN_MAX
)
if
(
qpn
>=
QPN_MAX
)
qpn
=
2
;
qpn
=
2
;
if
(
qpt
->
mask
&&
((
qpn
&
qpt
->
mask
)
>>
1
)
!=
r
)
if
(
qpt
->
mask
&&
((
qpn
&
qpt
->
mask
)
>>
1
)
>=
dd
->
n_krcv_queues
)
qpn
=
((
qpn
&
qpt
->
mask
)
?
(
qpn
|
qpt
->
mask
)
+
1
:
qpn
)
|
qpn
=
(
qpn
|
qpt
->
mask
)
+
2
;
(
r
<<
1
);
offset
=
qpn
&
BITS_PER_PAGE_MASK
;
offset
=
qpn
&
BITS_PER_PAGE_MASK
;
map
=
&
qpt
->
map
[
qpn
/
BITS_PER_PAGE
];
map
=
&
qpt
->
map
[
qpn
/
BITS_PER_PAGE
];
max_scan
=
qpt
->
nmaps
-
!
offset
;
max_scan
=
qpt
->
nmaps
-
!
offset
;
...
@@ -163,7 +157,8 @@ static int alloc_qpn(struct qib_devdata *dd, struct qib_qpn_table *qpt,
...
@@ -163,7 +157,8 @@ static int alloc_qpn(struct qib_devdata *dd, struct qib_qpn_table *qpt,
ret
=
qpn
;
ret
=
qpn
;
goto
bail
;
goto
bail
;
}
}
offset
=
find_next_offset
(
qpt
,
map
,
offset
,
r
);
offset
=
find_next_offset
(
qpt
,
map
,
offset
,
dd
->
n_krcv_queues
);
qpn
=
mk_qpn
(
qpt
,
map
,
offset
);
qpn
=
mk_qpn
(
qpt
,
map
,
offset
);
/*
/*
* This test differs from alloc_pidmap().
* This test differs from alloc_pidmap().
...
@@ -183,13 +178,13 @@ static int alloc_qpn(struct qib_devdata *dd, struct qib_qpn_table *qpt,
...
@@ -183,13 +178,13 @@ static int alloc_qpn(struct qib_devdata *dd, struct qib_qpn_table *qpt,
if
(
qpt
->
nmaps
==
QPNMAP_ENTRIES
)
if
(
qpt
->
nmaps
==
QPNMAP_ENTRIES
)
break
;
break
;
map
=
&
qpt
->
map
[
qpt
->
nmaps
++
];
map
=
&
qpt
->
map
[
qpt
->
nmaps
++
];
offset
=
qpt
->
mask
?
(
r
<<
1
)
:
0
;
offset
=
0
;
}
else
if
(
map
<
&
qpt
->
map
[
qpt
->
nmaps
])
{
}
else
if
(
map
<
&
qpt
->
map
[
qpt
->
nmaps
])
{
++
map
;
++
map
;
offset
=
qpt
->
mask
?
(
r
<<
1
)
:
0
;
offset
=
0
;
}
else
{
}
else
{
map
=
&
qpt
->
map
[
0
];
map
=
&
qpt
->
map
[
0
];
offset
=
qpt
->
mask
?
(
r
<<
1
)
:
2
;
offset
=
2
;
}
}
qpn
=
mk_qpn
(
qpt
,
map
,
offset
);
qpn
=
mk_qpn
(
qpt
,
map
,
offset
);
}
}
...
@@ -468,6 +463,10 @@ int qib_error_qp(struct qib_qp *qp, enum ib_wc_status err)
...
@@ -468,6 +463,10 @@ int qib_error_qp(struct qib_qp *qp, enum ib_wc_status err)
qp
->
s_flags
&=
~
(
QIB_S_TIMER
|
QIB_S_WAIT_RNR
);
qp
->
s_flags
&=
~
(
QIB_S_TIMER
|
QIB_S_WAIT_RNR
);
del_timer
(
&
qp
->
s_timer
);
del_timer
(
&
qp
->
s_timer
);
}
}
if
(
qp
->
s_flags
&
QIB_S_ANY_WAIT_SEND
)
qp
->
s_flags
&=
~
QIB_S_ANY_WAIT_SEND
;
spin_lock
(
&
dev
->
pending_lock
);
spin_lock
(
&
dev
->
pending_lock
);
if
(
!
list_empty
(
&
qp
->
iowait
)
&&
!
(
qp
->
s_flags
&
QIB_S_BUSY
))
{
if
(
!
list_empty
(
&
qp
->
iowait
)
&&
!
(
qp
->
s_flags
&
QIB_S_BUSY
))
{
qp
->
s_flags
&=
~
QIB_S_ANY_WAIT_IO
;
qp
->
s_flags
&=
~
QIB_S_ANY_WAIT_IO
;
...
@@ -1061,7 +1060,6 @@ struct ib_qp *qib_create_qp(struct ib_pd *ibpd,
...
@@ -1061,7 +1060,6 @@ struct ib_qp *qib_create_qp(struct ib_pd *ibpd,
}
}
qp
->
ibqp
.
qp_num
=
err
;
qp
->
ibqp
.
qp_num
=
err
;
qp
->
port_num
=
init_attr
->
port_num
;
qp
->
port_num
=
init_attr
->
port_num
;
qp
->
processor_id
=
smp_processor_id
();
qib_reset_qp
(
qp
,
init_attr
->
qp_type
);
qib_reset_qp
(
qp
,
init_attr
->
qp_type
);
break
;
break
;
...
...
drivers/infiniband/hw/qib/qib_rc.c
浏览文件 @
2b76c057
...
@@ -1407,6 +1407,7 @@ static void qib_rc_rcv_resp(struct qib_ibport *ibp,
...
@@ -1407,6 +1407,7 @@ static void qib_rc_rcv_resp(struct qib_ibport *ibp,
struct
qib_ctxtdata
*
rcd
)
struct
qib_ctxtdata
*
rcd
)
{
{
struct
qib_swqe
*
wqe
;
struct
qib_swqe
*
wqe
;
struct
qib_pportdata
*
ppd
=
ppd_from_ibp
(
ibp
);
enum
ib_wc_status
status
;
enum
ib_wc_status
status
;
unsigned
long
flags
;
unsigned
long
flags
;
int
diff
;
int
diff
;
...
@@ -1414,6 +1415,29 @@ static void qib_rc_rcv_resp(struct qib_ibport *ibp,
...
@@ -1414,6 +1415,29 @@ static void qib_rc_rcv_resp(struct qib_ibport *ibp,
u32
aeth
;
u32
aeth
;
u64
val
;
u64
val
;
if
(
opcode
!=
OP
(
RDMA_READ_RESPONSE_MIDDLE
))
{
/*
* If ACK'd PSN on SDMA busy list try to make progress to
* reclaim SDMA credits.
*/
if
((
qib_cmp24
(
psn
,
qp
->
s_sending_psn
)
>=
0
)
&&
(
qib_cmp24
(
qp
->
s_sending_psn
,
qp
->
s_sending_hpsn
)
<=
0
))
{
/*
* If send tasklet not running attempt to progress
* SDMA queue.
*/
if
(
!
(
qp
->
s_flags
&
QIB_S_BUSY
))
{
/* Acquire SDMA Lock */
spin_lock_irqsave
(
&
ppd
->
sdma_lock
,
flags
);
/* Invoke sdma make progress */
qib_sdma_make_progress
(
ppd
);
/* Release SDMA Lock */
spin_unlock_irqrestore
(
&
ppd
->
sdma_lock
,
flags
);
}
}
}
spin_lock_irqsave
(
&
qp
->
s_lock
,
flags
);
spin_lock_irqsave
(
&
qp
->
s_lock
,
flags
);
/* Ignore invalid responses. */
/* Ignore invalid responses. */
...
...
drivers/infiniband/hw/qib/qib_ud.c
浏览文件 @
2b76c057
...
@@ -445,13 +445,14 @@ void qib_ud_rcv(struct qib_ibport *ibp, struct qib_ib_header *hdr,
...
@@ -445,13 +445,14 @@ void qib_ud_rcv(struct qib_ibport *ibp, struct qib_ib_header *hdr,
qkey
=
be32_to_cpu
(
ohdr
->
u
.
ud
.
deth
[
0
]);
qkey
=
be32_to_cpu
(
ohdr
->
u
.
ud
.
deth
[
0
]);
src_qp
=
be32_to_cpu
(
ohdr
->
u
.
ud
.
deth
[
1
])
&
QIB_QPN_MASK
;
src_qp
=
be32_to_cpu
(
ohdr
->
u
.
ud
.
deth
[
1
])
&
QIB_QPN_MASK
;
/* Get the number of bytes the message was padded by. */
/*
* Get the number of bytes the message was padded by
* and drop incomplete packets.
*/
pad
=
(
be32_to_cpu
(
ohdr
->
bth
[
0
])
>>
20
)
&
3
;
pad
=
(
be32_to_cpu
(
ohdr
->
bth
[
0
])
>>
20
)
&
3
;
if
(
unlikely
(
tlen
<
(
hdrsize
+
pad
+
4
)))
{
if
(
unlikely
(
tlen
<
(
hdrsize
+
pad
+
4
)))
/* Drop incomplete packets. */
goto
drop
;
ibp
->
n_pkt_drops
++
;
goto
bail
;
}
tlen
-=
hdrsize
+
pad
+
4
;
tlen
-=
hdrsize
+
pad
+
4
;
/*
/*
...
@@ -460,10 +461,8 @@ void qib_ud_rcv(struct qib_ibport *ibp, struct qib_ib_header *hdr,
...
@@ -460,10 +461,8 @@ void qib_ud_rcv(struct qib_ibport *ibp, struct qib_ib_header *hdr,
*/
*/
if
(
qp
->
ibqp
.
qp_num
)
{
if
(
qp
->
ibqp
.
qp_num
)
{
if
(
unlikely
(
hdr
->
lrh
[
1
]
==
IB_LID_PERMISSIVE
||
if
(
unlikely
(
hdr
->
lrh
[
1
]
==
IB_LID_PERMISSIVE
||
hdr
->
lrh
[
3
]
==
IB_LID_PERMISSIVE
))
{
hdr
->
lrh
[
3
]
==
IB_LID_PERMISSIVE
))
ibp
->
n_pkt_drops
++
;
goto
drop
;
goto
bail
;
}
if
(
qp
->
ibqp
.
qp_num
>
1
)
{
if
(
qp
->
ibqp
.
qp_num
>
1
)
{
u16
pkey1
,
pkey2
;
u16
pkey1
,
pkey2
;
...
@@ -476,7 +475,7 @@ void qib_ud_rcv(struct qib_ibport *ibp, struct qib_ib_header *hdr,
...
@@ -476,7 +475,7 @@ void qib_ud_rcv(struct qib_ibport *ibp, struct qib_ib_header *hdr,
0xF
,
0xF
,
src_qp
,
qp
->
ibqp
.
qp_num
,
src_qp
,
qp
->
ibqp
.
qp_num
,
hdr
->
lrh
[
3
],
hdr
->
lrh
[
1
]);
hdr
->
lrh
[
3
],
hdr
->
lrh
[
1
]);
goto
bail
;
return
;
}
}
}
}
if
(
unlikely
(
qkey
!=
qp
->
qkey
))
{
if
(
unlikely
(
qkey
!=
qp
->
qkey
))
{
...
@@ -484,30 +483,24 @@ void qib_ud_rcv(struct qib_ibport *ibp, struct qib_ib_header *hdr,
...
@@ -484,30 +483,24 @@ void qib_ud_rcv(struct qib_ibport *ibp, struct qib_ib_header *hdr,
(
be16_to_cpu
(
hdr
->
lrh
[
0
])
>>
4
)
&
0xF
,
(
be16_to_cpu
(
hdr
->
lrh
[
0
])
>>
4
)
&
0xF
,
src_qp
,
qp
->
ibqp
.
qp_num
,
src_qp
,
qp
->
ibqp
.
qp_num
,
hdr
->
lrh
[
3
],
hdr
->
lrh
[
1
]);
hdr
->
lrh
[
3
],
hdr
->
lrh
[
1
]);
goto
bail
;
return
;
}
}
/* Drop invalid MAD packets (see 13.5.3.1). */
/* Drop invalid MAD packets (see 13.5.3.1). */
if
(
unlikely
(
qp
->
ibqp
.
qp_num
==
1
&&
if
(
unlikely
(
qp
->
ibqp
.
qp_num
==
1
&&
(
tlen
!=
256
||
(
tlen
!=
256
||
(
be16_to_cpu
(
hdr
->
lrh
[
0
])
>>
12
)
==
15
)))
{
(
be16_to_cpu
(
hdr
->
lrh
[
0
])
>>
12
)
==
15
)))
ibp
->
n_pkt_drops
++
;
goto
drop
;
goto
bail
;
}
}
else
{
}
else
{
struct
ib_smp
*
smp
;
struct
ib_smp
*
smp
;
/* Drop invalid MAD packets (see 13.5.3.1). */
/* Drop invalid MAD packets (see 13.5.3.1). */
if
(
tlen
!=
256
||
(
be16_to_cpu
(
hdr
->
lrh
[
0
])
>>
12
)
!=
15
)
{
if
(
tlen
!=
256
||
(
be16_to_cpu
(
hdr
->
lrh
[
0
])
>>
12
)
!=
15
)
ibp
->
n_pkt_drops
++
;
goto
drop
;
goto
bail
;
}
smp
=
(
struct
ib_smp
*
)
data
;
smp
=
(
struct
ib_smp
*
)
data
;
if
((
hdr
->
lrh
[
1
]
==
IB_LID_PERMISSIVE
||
if
((
hdr
->
lrh
[
1
]
==
IB_LID_PERMISSIVE
||
hdr
->
lrh
[
3
]
==
IB_LID_PERMISSIVE
)
&&
hdr
->
lrh
[
3
]
==
IB_LID_PERMISSIVE
)
&&
smp
->
mgmt_class
!=
IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE
)
{
smp
->
mgmt_class
!=
IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE
)
ibp
->
n_pkt_drops
++
;
goto
drop
;
goto
bail
;
}
}
}
/*
/*
...
@@ -519,14 +512,12 @@ void qib_ud_rcv(struct qib_ibport *ibp, struct qib_ib_header *hdr,
...
@@ -519,14 +512,12 @@ void qib_ud_rcv(struct qib_ibport *ibp, struct qib_ib_header *hdr,
opcode
==
IB_OPCODE_UD_SEND_ONLY_WITH_IMMEDIATE
)
{
opcode
==
IB_OPCODE_UD_SEND_ONLY_WITH_IMMEDIATE
)
{
wc
.
ex
.
imm_data
=
ohdr
->
u
.
ud
.
imm_data
;
wc
.
ex
.
imm_data
=
ohdr
->
u
.
ud
.
imm_data
;
wc
.
wc_flags
=
IB_WC_WITH_IMM
;
wc
.
wc_flags
=
IB_WC_WITH_IMM
;
hdrsize
+
=
sizeof
(
u32
);
tlen
-
=
sizeof
(
u32
);
}
else
if
(
opcode
==
IB_OPCODE_UD_SEND_ONLY
)
{
}
else
if
(
opcode
==
IB_OPCODE_UD_SEND_ONLY
)
{
wc
.
ex
.
imm_data
=
0
;
wc
.
ex
.
imm_data
=
0
;
wc
.
wc_flags
=
0
;
wc
.
wc_flags
=
0
;
}
else
{
}
else
ibp
->
n_pkt_drops
++
;
goto
drop
;
goto
bail
;
}
/*
/*
* A GRH is expected to preceed the data even if not
* A GRH is expected to preceed the data even if not
...
@@ -556,8 +547,7 @@ void qib_ud_rcv(struct qib_ibport *ibp, struct qib_ib_header *hdr,
...
@@ -556,8 +547,7 @@ void qib_ud_rcv(struct qib_ibport *ibp, struct qib_ib_header *hdr,
/* Silently drop packets which are too big. */
/* Silently drop packets which are too big. */
if
(
unlikely
(
wc
.
byte_len
>
qp
->
r_len
))
{
if
(
unlikely
(
wc
.
byte_len
>
qp
->
r_len
))
{
qp
->
r_flags
|=
QIB_R_REUSE_SGE
;
qp
->
r_flags
|=
QIB_R_REUSE_SGE
;
ibp
->
n_pkt_drops
++
;
goto
drop
;
return
;
}
}
if
(
has_grh
)
{
if
(
has_grh
)
{
qib_copy_sge
(
&
qp
->
r_sge
,
&
hdr
->
u
.
l
.
grh
,
qib_copy_sge
(
&
qp
->
r_sge
,
&
hdr
->
u
.
l
.
grh
,
...
@@ -594,5 +584,8 @@ void qib_ud_rcv(struct qib_ibport *ibp, struct qib_ib_header *hdr,
...
@@ -594,5 +584,8 @@ void qib_ud_rcv(struct qib_ibport *ibp, struct qib_ib_header *hdr,
qib_cq_enter
(
to_icq
(
qp
->
ibqp
.
recv_cq
),
&
wc
,
qib_cq_enter
(
to_icq
(
qp
->
ibqp
.
recv_cq
),
&
wc
,
(
ohdr
->
bth
[
0
]
&
(
ohdr
->
bth
[
0
]
&
cpu_to_be32
(
IB_BTH_SOLICITED
))
!=
0
);
cpu_to_be32
(
IB_BTH_SOLICITED
))
!=
0
);
bail:
;
return
;
drop:
ibp
->
n_pkt_drops
++
;
}
}
drivers/infiniband/hw/qib/qib_user_sdma.c
浏览文件 @
2b76c057
...
@@ -382,6 +382,7 @@ static void qib_user_sdma_free_pkt_list(struct device *dev,
...
@@ -382,6 +382,7 @@ static void qib_user_sdma_free_pkt_list(struct device *dev,
kmem_cache_free
(
pq
->
pkt_slab
,
pkt
);
kmem_cache_free
(
pq
->
pkt_slab
,
pkt
);
}
}
INIT_LIST_HEAD
(
list
);
}
}
/*
/*
...
...
drivers/infiniband/hw/qib/qib_verbs.h
浏览文件 @
2b76c057
...
@@ -301,6 +301,7 @@ struct qib_mregion {
...
@@ -301,6 +301,7 @@ struct qib_mregion {
int
access_flags
;
int
access_flags
;
u32
max_segs
;
/* number of qib_segs in all the arrays */
u32
max_segs
;
/* number of qib_segs in all the arrays */
u32
mapsz
;
/* size of the map array */
u32
mapsz
;
/* size of the map array */
u8
page_shift
;
/* 0 - non unform/non powerof2 sizes */
atomic_t
refcount
;
atomic_t
refcount
;
struct
qib_segarray
*
map
[
0
];
/* the segments */
struct
qib_segarray
*
map
[
0
];
/* the segments */
};
};
...
@@ -435,7 +436,6 @@ struct qib_qp {
...
@@ -435,7 +436,6 @@ struct qib_qp {
spinlock_t
r_lock
;
/* used for APM */
spinlock_t
r_lock
;
/* used for APM */
spinlock_t
s_lock
;
spinlock_t
s_lock
;
atomic_t
s_dma_busy
;
atomic_t
s_dma_busy
;
unsigned
processor_id
;
/* Processor ID QP is bound to */
u32
s_flags
;
u32
s_flags
;
u32
s_cur_size
;
/* size of send packet in bytes */
u32
s_cur_size
;
/* size of send packet in bytes */
u32
s_len
;
/* total length of s_sge */
u32
s_len
;
/* total length of s_sge */
...
@@ -813,13 +813,8 @@ extern struct workqueue_struct *qib_cq_wq;
...
@@ -813,13 +813,8 @@ extern struct workqueue_struct *qib_cq_wq;
*/
*/
static
inline
void
qib_schedule_send
(
struct
qib_qp
*
qp
)
static
inline
void
qib_schedule_send
(
struct
qib_qp
*
qp
)
{
{
if
(
qib_send_ok
(
qp
))
{
if
(
qib_send_ok
(
qp
))
if
(
qp
->
processor_id
==
smp_processor_id
())
queue_work
(
qib_wq
,
&
qp
->
s_work
);
queue_work
(
qib_wq
,
&
qp
->
s_work
);
else
queue_work_on
(
qp
->
processor_id
,
qib_wq
,
&
qp
->
s_work
);
}
}
}
static
inline
int
qib_pkey_ok
(
u16
pkey1
,
u16
pkey2
)
static
inline
int
qib_pkey_ok
(
u16
pkey1
,
u16
pkey2
)
...
...
drivers/infiniband/ulp/ipoib/Kconfig
浏览文件 @
2b76c057
config INFINIBAND_IPOIB
config INFINIBAND_IPOIB
tristate "IP-over-InfiniBand"
tristate "IP-over-InfiniBand"
depends on NETDEVICES && INET && (IPV6 || IPV6=n)
depends on NETDEVICES && INET && (IPV6 || IPV6=n)
select INET_LRO
---help---
---help---
Support for the IP-over-InfiniBand protocol (IPoIB). This
Support for the IP-over-InfiniBand protocol (IPoIB). This
transports IP packets over InfiniBand so you can use your IB
transports IP packets over InfiniBand so you can use your IB
...
...
drivers/infiniband/ulp/ipoib/ipoib.h
浏览文件 @
2b76c057
...
@@ -50,7 +50,7 @@
...
@@ -50,7 +50,7 @@
#include <rdma/ib_verbs.h>
#include <rdma/ib_verbs.h>
#include <rdma/ib_pack.h>
#include <rdma/ib_pack.h>
#include <rdma/ib_sa.h>
#include <rdma/ib_sa.h>
#include <linux/
inet_lro
.h>
#include <linux/
sched
.h>
/* constants */
/* constants */
...
@@ -100,9 +100,6 @@ enum {
...
@@ -100,9 +100,6 @@ enum {
IPOIB_MCAST_FLAG_BUSY
=
2
,
/* joining or already joined */
IPOIB_MCAST_FLAG_BUSY
=
2
,
/* joining or already joined */
IPOIB_MCAST_FLAG_ATTACHED
=
3
,
IPOIB_MCAST_FLAG_ATTACHED
=
3
,
IPOIB_MAX_LRO_DESCRIPTORS
=
8
,
IPOIB_LRO_MAX_AGGR
=
64
,
MAX_SEND_CQE
=
16
,
MAX_SEND_CQE
=
16
,
IPOIB_CM_COPYBREAK
=
256
,
IPOIB_CM_COPYBREAK
=
256
,
};
};
...
@@ -262,11 +259,6 @@ struct ipoib_ethtool_st {
...
@@ -262,11 +259,6 @@ struct ipoib_ethtool_st {
u16
max_coalesced_frames
;
u16
max_coalesced_frames
;
};
};
struct
ipoib_lro
{
struct
net_lro_mgr
lro_mgr
;
struct
net_lro_desc
lro_desc
[
IPOIB_MAX_LRO_DESCRIPTORS
];
};
/*
/*
* Device private locking: network stack tx_lock protects members used
* Device private locking: network stack tx_lock protects members used
* in TX fast path, lock protects everything else. lock nests inside
* in TX fast path, lock protects everything else. lock nests inside
...
@@ -352,8 +344,6 @@ struct ipoib_dev_priv {
...
@@ -352,8 +344,6 @@ struct ipoib_dev_priv {
int
hca_caps
;
int
hca_caps
;
struct
ipoib_ethtool_st
ethtool
;
struct
ipoib_ethtool_st
ethtool
;
struct
timer_list
poll_timer
;
struct
timer_list
poll_timer
;
struct
ipoib_lro
lro
;
};
};
struct
ipoib_ah
{
struct
ipoib_ah
{
...
...
drivers/infiniband/ulp/ipoib/ipoib_cm.c
浏览文件 @
2b76c057
...
@@ -1480,6 +1480,7 @@ static ssize_t set_mode(struct device *d, struct device_attribute *attr,
...
@@ -1480,6 +1480,7 @@ static ssize_t set_mode(struct device *d, struct device_attribute *attr,
if
(
test_bit
(
IPOIB_FLAG_CSUM
,
&
priv
->
flags
))
{
if
(
test_bit
(
IPOIB_FLAG_CSUM
,
&
priv
->
flags
))
{
dev
->
features
|=
NETIF_F_IP_CSUM
|
NETIF_F_SG
;
dev
->
features
|=
NETIF_F_IP_CSUM
|
NETIF_F_SG
;
priv
->
dev
->
features
|=
NETIF_F_GRO
;
if
(
priv
->
hca_caps
&
IB_DEVICE_UD_TSO
)
if
(
priv
->
hca_caps
&
IB_DEVICE_UD_TSO
)
dev
->
features
|=
NETIF_F_TSO
;
dev
->
features
|=
NETIF_F_TSO
;
}
}
...
...
drivers/infiniband/ulp/ipoib/ipoib_ethtool.c
浏览文件 @
2b76c057
...
@@ -106,63 +106,12 @@ static int ipoib_set_coalesce(struct net_device *dev,
...
@@ -106,63 +106,12 @@ static int ipoib_set_coalesce(struct net_device *dev,
return
0
;
return
0
;
}
}
static
const
char
ipoib_stats_keys
[][
ETH_GSTRING_LEN
]
=
{
"LRO aggregated"
,
"LRO flushed"
,
"LRO avg aggr"
,
"LRO no desc"
};
static
void
ipoib_get_strings
(
struct
net_device
*
netdev
,
u32
stringset
,
u8
*
data
)
{
switch
(
stringset
)
{
case
ETH_SS_STATS
:
memcpy
(
data
,
*
ipoib_stats_keys
,
sizeof
(
ipoib_stats_keys
));
break
;
}
}
static
int
ipoib_get_sset_count
(
struct
net_device
*
dev
,
int
sset
)
{
switch
(
sset
)
{
case
ETH_SS_STATS
:
return
ARRAY_SIZE
(
ipoib_stats_keys
);
default:
return
-
EOPNOTSUPP
;
}
}
static
void
ipoib_get_ethtool_stats
(
struct
net_device
*
dev
,
struct
ethtool_stats
*
stats
,
uint64_t
*
data
)
{
struct
ipoib_dev_priv
*
priv
=
netdev_priv
(
dev
);
int
index
=
0
;
/* Get LRO statistics */
data
[
index
++
]
=
priv
->
lro
.
lro_mgr
.
stats
.
aggregated
;
data
[
index
++
]
=
priv
->
lro
.
lro_mgr
.
stats
.
flushed
;
if
(
priv
->
lro
.
lro_mgr
.
stats
.
flushed
)
data
[
index
++
]
=
priv
->
lro
.
lro_mgr
.
stats
.
aggregated
/
priv
->
lro
.
lro_mgr
.
stats
.
flushed
;
else
data
[
index
++
]
=
0
;
data
[
index
++
]
=
priv
->
lro
.
lro_mgr
.
stats
.
no_desc
;
}
static
int
ipoib_set_flags
(
struct
net_device
*
dev
,
u32
flags
)
{
return
ethtool_op_set_flags
(
dev
,
flags
,
ETH_FLAG_LRO
);
}
static
const
struct
ethtool_ops
ipoib_ethtool_ops
=
{
static
const
struct
ethtool_ops
ipoib_ethtool_ops
=
{
.
get_drvinfo
=
ipoib_get_drvinfo
,
.
get_drvinfo
=
ipoib_get_drvinfo
,
.
get_rx_csum
=
ipoib_get_rx_csum
,
.
get_rx_csum
=
ipoib_get_rx_csum
,
.
set_tso
=
ipoib_set_tso
,
.
set_tso
=
ipoib_set_tso
,
.
get_coalesce
=
ipoib_get_coalesce
,
.
get_coalesce
=
ipoib_get_coalesce
,
.
set_coalesce
=
ipoib_set_coalesce
,
.
set_coalesce
=
ipoib_set_coalesce
,
.
get_flags
=
ethtool_op_get_flags
,
.
set_flags
=
ipoib_set_flags
,
.
get_strings
=
ipoib_get_strings
,
.
get_sset_count
=
ipoib_get_sset_count
,
.
get_ethtool_stats
=
ipoib_get_ethtool_stats
,
};
};
void
ipoib_set_ethtool_ops
(
struct
net_device
*
dev
)
void
ipoib_set_ethtool_ops
(
struct
net_device
*
dev
)
...
...
drivers/infiniband/ulp/ipoib/ipoib_ib.c
浏览文件 @
2b76c057
...
@@ -295,10 +295,7 @@ static void ipoib_ib_handle_rx_wc(struct net_device *dev, struct ib_wc *wc)
...
@@ -295,10 +295,7 @@ static void ipoib_ib_handle_rx_wc(struct net_device *dev, struct ib_wc *wc)
if
(
test_bit
(
IPOIB_FLAG_CSUM
,
&
priv
->
flags
)
&&
likely
(
wc
->
csum_ok
))
if
(
test_bit
(
IPOIB_FLAG_CSUM
,
&
priv
->
flags
)
&&
likely
(
wc
->
csum_ok
))
skb
->
ip_summed
=
CHECKSUM_UNNECESSARY
;
skb
->
ip_summed
=
CHECKSUM_UNNECESSARY
;
if
(
dev
->
features
&
NETIF_F_LRO
)
napi_gro_receive
(
&
priv
->
napi
,
skb
);
lro_receive_skb
(
&
priv
->
lro
.
lro_mgr
,
skb
,
NULL
);
else
netif_receive_skb
(
skb
);
repost:
repost:
if
(
unlikely
(
ipoib_ib_post_receive
(
dev
,
wr_id
)))
if
(
unlikely
(
ipoib_ib_post_receive
(
dev
,
wr_id
)))
...
@@ -450,9 +447,6 @@ int ipoib_poll(struct napi_struct *napi, int budget)
...
@@ -450,9 +447,6 @@ int ipoib_poll(struct napi_struct *napi, int budget)
}
}
if
(
done
<
budget
)
{
if
(
done
<
budget
)
{
if
(
dev
->
features
&
NETIF_F_LRO
)
lro_flush_all
(
&
priv
->
lro
.
lro_mgr
);
napi_complete
(
napi
);
napi_complete
(
napi
);
if
(
unlikely
(
ib_req_notify_cq
(
priv
->
recv_cq
,
if
(
unlikely
(
ib_req_notify_cq
(
priv
->
recv_cq
,
IB_CQ_NEXT_COMP
|
IB_CQ_NEXT_COMP
|
...
...
drivers/infiniband/ulp/ipoib/ipoib_main.c
浏览文件 @
2b76c057
...
@@ -60,15 +60,6 @@ MODULE_PARM_DESC(send_queue_size, "Number of descriptors in send queue");
...
@@ -60,15 +60,6 @@ MODULE_PARM_DESC(send_queue_size, "Number of descriptors in send queue");
module_param_named
(
recv_queue_size
,
ipoib_recvq_size
,
int
,
0444
);
module_param_named
(
recv_queue_size
,
ipoib_recvq_size
,
int
,
0444
);
MODULE_PARM_DESC
(
recv_queue_size
,
"Number of descriptors in receive queue"
);
MODULE_PARM_DESC
(
recv_queue_size
,
"Number of descriptors in receive queue"
);
static
int
lro
;
module_param
(
lro
,
bool
,
0444
);
MODULE_PARM_DESC
(
lro
,
"Enable LRO (Large Receive Offload)"
);
static
int
lro_max_aggr
=
IPOIB_LRO_MAX_AGGR
;
module_param
(
lro_max_aggr
,
int
,
0644
);
MODULE_PARM_DESC
(
lro_max_aggr
,
"LRO: Max packets to be aggregated "
"(default = 64)"
);
#ifdef CONFIG_INFINIBAND_IPOIB_DEBUG
#ifdef CONFIG_INFINIBAND_IPOIB_DEBUG
int
ipoib_debug_level
;
int
ipoib_debug_level
;
...
@@ -976,54 +967,6 @@ static const struct header_ops ipoib_header_ops = {
...
@@ -976,54 +967,6 @@ static const struct header_ops ipoib_header_ops = {
.
create
=
ipoib_hard_header
,
.
create
=
ipoib_hard_header
,
};
};
static
int
get_skb_hdr
(
struct
sk_buff
*
skb
,
void
**
iphdr
,
void
**
tcph
,
u64
*
hdr_flags
,
void
*
priv
)
{
unsigned
int
ip_len
;
struct
iphdr
*
iph
;
if
(
unlikely
(
skb
->
protocol
!=
htons
(
ETH_P_IP
)))
return
-
1
;
/*
* In the future we may add an else clause that verifies the
* checksum and allows devices which do not calculate checksum
* to use LRO.
*/
if
(
unlikely
(
skb
->
ip_summed
!=
CHECKSUM_UNNECESSARY
))
return
-
1
;
/* Check for non-TCP packet */
skb_reset_network_header
(
skb
);
iph
=
ip_hdr
(
skb
);
if
(
iph
->
protocol
!=
IPPROTO_TCP
)
return
-
1
;
ip_len
=
ip_hdrlen
(
skb
);
skb_set_transport_header
(
skb
,
ip_len
);
*
tcph
=
tcp_hdr
(
skb
);
/* check if IP header and TCP header are complete */
if
(
ntohs
(
iph
->
tot_len
)
<
ip_len
+
tcp_hdrlen
(
skb
))
return
-
1
;
*
hdr_flags
=
LRO_IPV4
|
LRO_TCP
;
*
iphdr
=
iph
;
return
0
;
}
static
void
ipoib_lro_setup
(
struct
ipoib_dev_priv
*
priv
)
{
priv
->
lro
.
lro_mgr
.
max_aggr
=
lro_max_aggr
;
priv
->
lro
.
lro_mgr
.
max_desc
=
IPOIB_MAX_LRO_DESCRIPTORS
;
priv
->
lro
.
lro_mgr
.
lro_arr
=
priv
->
lro
.
lro_desc
;
priv
->
lro
.
lro_mgr
.
get_skb_header
=
get_skb_hdr
;
priv
->
lro
.
lro_mgr
.
features
=
LRO_F_NAPI
;
priv
->
lro
.
lro_mgr
.
dev
=
priv
->
dev
;
priv
->
lro
.
lro_mgr
.
ip_summed_aggr
=
CHECKSUM_UNNECESSARY
;
}
static
const
struct
net_device_ops
ipoib_netdev_ops
=
{
static
const
struct
net_device_ops
ipoib_netdev_ops
=
{
.
ndo_open
=
ipoib_open
,
.
ndo_open
=
ipoib_open
,
.
ndo_stop
=
ipoib_stop
,
.
ndo_stop
=
ipoib_stop
,
...
@@ -1067,8 +1010,6 @@ static void ipoib_setup(struct net_device *dev)
...
@@ -1067,8 +1010,6 @@ static void ipoib_setup(struct net_device *dev)
priv
->
dev
=
dev
;
priv
->
dev
=
dev
;
ipoib_lro_setup
(
priv
);
spin_lock_init
(
&
priv
->
lock
);
spin_lock_init
(
&
priv
->
lock
);
mutex_init
(
&
priv
->
vlan_mutex
);
mutex_init
(
&
priv
->
vlan_mutex
);
...
@@ -1218,8 +1159,7 @@ int ipoib_set_dev_features(struct ipoib_dev_priv *priv, struct ib_device *hca)
...
@@ -1218,8 +1159,7 @@ int ipoib_set_dev_features(struct ipoib_dev_priv *priv, struct ib_device *hca)
priv
->
dev
->
features
|=
NETIF_F_SG
|
NETIF_F_IP_CSUM
;
priv
->
dev
->
features
|=
NETIF_F_SG
|
NETIF_F_IP_CSUM
;
}
}
if
(
lro
)
priv
->
dev
->
features
|=
NETIF_F_GRO
;
priv
->
dev
->
features
|=
NETIF_F_LRO
;
if
(
priv
->
dev
->
features
&
NETIF_F_SG
&&
priv
->
hca_caps
&
IB_DEVICE_UD_TSO
)
if
(
priv
->
dev
->
features
&
NETIF_F_SG
&&
priv
->
hca_caps
&
IB_DEVICE_UD_TSO
)
priv
->
dev
->
features
|=
NETIF_F_TSO
;
priv
->
dev
->
features
|=
NETIF_F_TSO
;
...
...
drivers/infiniband/ulp/srp/ib_srp.c
浏览文件 @
2b76c057
此差异已折叠。
点击以展开。
drivers/infiniband/ulp/srp/ib_srp.h
浏览文件 @
2b76c057
...
@@ -59,16 +59,15 @@ enum {
...
@@ -59,16 +59,15 @@ enum {
SRP_RQ_SHIFT
=
6
,
SRP_RQ_SHIFT
=
6
,
SRP_RQ_SIZE
=
1
<<
SRP_RQ_SHIFT
,
SRP_RQ_SIZE
=
1
<<
SRP_RQ_SHIFT
,
SRP_RQ_MASK
=
SRP_RQ_SIZE
-
1
,
SRP_SQ_SIZE
=
SRP_RQ_SIZE
,
SRP_SQ_SIZE
=
SRP_RQ_SIZE
,
SRP_SQ_MASK
=
SRP_SQ_SIZE
-
1
,
SRP_RSP_SQ_SIZE
=
1
,
SRP_RSP_SQ_SIZE
=
1
,
SRP_REQ_SQ_SIZE
=
SRP_SQ_SIZE
-
SRP_RSP_SQ_SIZE
,
SRP_REQ_SQ_SIZE
=
SRP_SQ_SIZE
-
SRP_RSP_SQ_SIZE
,
SRP_TSK_MGMT_SQ_SIZE
=
1
,
SRP_TSK_MGMT_SQ_SIZE
=
1
,
SRP_CMD_SQ_SIZE
=
SRP_REQ_SQ_SIZE
-
SRP_TSK_MGMT_SQ_SIZE
,
SRP_CMD_SQ_SIZE
=
SRP_REQ_SQ_SIZE
-
SRP_TSK_MGMT_SQ_SIZE
,
SRP_TAG_TSK_MGMT
=
1
<<
(
SRP_RQ_SHIFT
+
1
),
SRP_TAG_NO_REQ
=
~
0U
,
SRP_TAG_TSK_MGMT
=
1U
<<
31
,
SRP_FMR_SIZE
=
256
,
SRP_FMR_SIZE
=
256
,
SRP_FMR_POOL_SIZE
=
1024
,
SRP_FMR_POOL_SIZE
=
1024
,
...
@@ -113,15 +112,29 @@ struct srp_request {
...
@@ -113,15 +112,29 @@ struct srp_request {
struct
list_head
list
;
struct
list_head
list
;
struct
scsi_cmnd
*
scmnd
;
struct
scsi_cmnd
*
scmnd
;
struct
srp_iu
*
cmd
;
struct
srp_iu
*
cmd
;
struct
srp_iu
*
tsk_mgmt
;
struct
ib_pool_fmr
*
fmr
;
struct
ib_pool_fmr
*
fmr
;
struct
completion
done
;
short
index
;
short
index
;
u8
cmd_done
;
u8
tsk_status
;
};
};
struct
srp_target_port
{
struct
srp_target_port
{
/* These are RW in the hot path, and commonly used together */
struct
list_head
free_tx
;
struct
list_head
free_reqs
;
spinlock_t
lock
;
s32
req_lim
;
/* These are read-only in the hot path */
struct
ib_cq
*
send_cq
____cacheline_aligned_in_smp
;
struct
ib_cq
*
recv_cq
;
struct
ib_qp
*
qp
;
u32
lkey
;
u32
rkey
;
enum
srp_target_state
state
;
/* Everything above this point is used in the hot path of
* command processing. Try to keep them packed into cachelines.
*/
__be64
id_ext
;
__be64
id_ext
;
__be64
ioc_guid
;
__be64
ioc_guid
;
__be64
service_id
;
__be64
service_id
;
...
@@ -138,24 +151,13 @@ struct srp_target_port {
...
@@ -138,24 +151,13 @@ struct srp_target_port {
int
path_query_id
;
int
path_query_id
;
struct
ib_cm_id
*
cm_id
;
struct
ib_cm_id
*
cm_id
;
struct
ib_cq
*
recv_cq
;
struct
ib_cq
*
send_cq
;
struct
ib_qp
*
qp
;
int
max_ti_iu_len
;
int
max_ti_iu_len
;
s32
req_lim
;
int
zero_req_lim
;
int
zero_req_lim
;
unsigned
rx_head
;
struct
srp_iu
*
rx_ring
[
SRP_RQ_SIZE
];
unsigned
tx_head
;
unsigned
tx_tail
;
struct
srp_iu
*
tx_ring
[
SRP_SQ_SIZE
];
struct
srp_iu
*
tx_ring
[
SRP_SQ_SIZE
];
struct
srp_iu
*
rx_ring
[
SRP_RQ_SIZE
];
struct
list_head
free_reqs
;
struct
list_head
req_queue
;
struct
srp_request
req_ring
[
SRP_CMD_SQ_SIZE
];
struct
srp_request
req_ring
[
SRP_CMD_SQ_SIZE
];
struct
work_struct
work
;
struct
work_struct
work
;
...
@@ -163,16 +165,18 @@ struct srp_target_port {
...
@@ -163,16 +165,18 @@ struct srp_target_port {
struct
list_head
list
;
struct
list_head
list
;
struct
completion
done
;
struct
completion
done
;
int
status
;
int
status
;
enum
srp_target_state
state
;
int
qp_in_error
;
int
qp_in_error
;
struct
completion
tsk_mgmt_done
;
u8
tsk_mgmt_status
;
};
};
struct
srp_iu
{
struct
srp_iu
{
struct
list_head
list
;
u64
dma
;
u64
dma
;
void
*
buf
;
void
*
buf
;
size_t
size
;
size_t
size
;
enum
dma_data_direction
direction
;
enum
dma_data_direction
direction
;
enum
srp_iu_type
type
;
};
};
#endif
/* IB_SRP_H */
#endif
/* IB_SRP_H */
drivers/net/mlx4/alloc.c
浏览文件 @
2b76c057
...
@@ -178,6 +178,7 @@ int mlx4_buf_alloc(struct mlx4_dev *dev, int size, int max_direct,
...
@@ -178,6 +178,7 @@ int mlx4_buf_alloc(struct mlx4_dev *dev, int size, int max_direct,
}
else
{
}
else
{
int
i
;
int
i
;
buf
->
direct
.
buf
=
NULL
;
buf
->
nbufs
=
(
size
+
PAGE_SIZE
-
1
)
/
PAGE_SIZE
;
buf
->
nbufs
=
(
size
+
PAGE_SIZE
-
1
)
/
PAGE_SIZE
;
buf
->
npages
=
buf
->
nbufs
;
buf
->
npages
=
buf
->
nbufs
;
buf
->
page_shift
=
PAGE_SHIFT
;
buf
->
page_shift
=
PAGE_SHIFT
;
...
@@ -229,7 +230,7 @@ void mlx4_buf_free(struct mlx4_dev *dev, int size, struct mlx4_buf *buf)
...
@@ -229,7 +230,7 @@ void mlx4_buf_free(struct mlx4_dev *dev, int size, struct mlx4_buf *buf)
dma_free_coherent
(
&
dev
->
pdev
->
dev
,
size
,
buf
->
direct
.
buf
,
dma_free_coherent
(
&
dev
->
pdev
->
dev
,
size
,
buf
->
direct
.
buf
,
buf
->
direct
.
map
);
buf
->
direct
.
map
);
else
{
else
{
if
(
BITS_PER_LONG
==
64
)
if
(
BITS_PER_LONG
==
64
&&
buf
->
direct
.
buf
)
vunmap
(
buf
->
direct
.
buf
);
vunmap
(
buf
->
direct
.
buf
);
for
(
i
=
0
;
i
<
buf
->
nbufs
;
++
i
)
for
(
i
=
0
;
i
<
buf
->
nbufs
;
++
i
)
...
...
drivers/net/mlx4/fw.c
浏览文件 @
2b76c057
...
@@ -289,10 +289,8 @@ int mlx4_QUERY_DEV_CAP(struct mlx4_dev *dev, struct mlx4_dev_cap *dev_cap)
...
@@ -289,10 +289,8 @@ int mlx4_QUERY_DEV_CAP(struct mlx4_dev *dev, struct mlx4_dev_cap *dev_cap)
MLX4_GET
(
field
,
outbox
,
QUERY_DEV_CAP_LOG_BF_REG_SZ_OFFSET
);
MLX4_GET
(
field
,
outbox
,
QUERY_DEV_CAP_LOG_BF_REG_SZ_OFFSET
);
dev_cap
->
bf_reg_size
=
1
<<
(
field
&
0x1f
);
dev_cap
->
bf_reg_size
=
1
<<
(
field
&
0x1f
);
MLX4_GET
(
field
,
outbox
,
QUERY_DEV_CAP_LOG_MAX_BF_REGS_PER_PAGE_OFFSET
);
MLX4_GET
(
field
,
outbox
,
QUERY_DEV_CAP_LOG_MAX_BF_REGS_PER_PAGE_OFFSET
);
if
((
1
<<
(
field
&
0x3f
))
>
(
PAGE_SIZE
/
dev_cap
->
bf_reg_size
))
{
if
((
1
<<
(
field
&
0x3f
))
>
(
PAGE_SIZE
/
dev_cap
->
bf_reg_size
))
mlx4_warn
(
dev
,
"firmware bug: log2 # of blue flame regs is invalid (%d), forcing 3
\n
"
,
field
&
0x1f
);
field
=
3
;
field
=
3
;
}
dev_cap
->
bf_regs_per_page
=
1
<<
(
field
&
0x3f
);
dev_cap
->
bf_regs_per_page
=
1
<<
(
field
&
0x3f
);
mlx4_dbg
(
dev
,
"BlueFlame available (reg size %d, regs/page %d)
\n
"
,
mlx4_dbg
(
dev
,
"BlueFlame available (reg size %d, regs/page %d)
\n
"
,
dev_cap
->
bf_reg_size
,
dev_cap
->
bf_regs_per_page
);
dev_cap
->
bf_reg_size
,
dev_cap
->
bf_regs_per_page
);
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录