Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
L
linux
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
Analytics
Analytics
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Commits
Issue Boards
Open sidebar
nexedi
linux
Commits
2a0d8366
Commit
2a0d8366
authored
Dec 24, 2008
by
Roland Dreier
Browse files
Options
Browse Files
Download
Plain Diff
Merge branches 'cma', 'ehca', 'ipath', 'iser', 'mlx4' and 'nes' into for-next
parents
1f5175ad
139cdab0
3d089098
bba7ebba
7798dbf4
e189062a
Changes
40
Expand all
Show whitespace changes
Inline
Side-by-side
Showing
40 changed files
with
754 additions
and
408 deletions
+754
-408
drivers/infiniband/hw/ehca/ehca_classes.h
drivers/infiniband/hw/ehca/ehca_classes.h
+7
-0
drivers/infiniband/hw/ehca/ehca_eq.c
drivers/infiniband/hw/ehca/ehca_eq.c
+1
-1
drivers/infiniband/hw/ehca/ehca_main.c
drivers/infiniband/hw/ehca/ehca_main.c
+10
-7
drivers/infiniband/hw/ehca/ehca_qp.c
drivers/infiniband/hw/ehca/ehca_qp.c
+6
-6
drivers/infiniband/hw/ehca/ehca_reqs.c
drivers/infiniband/hw/ehca/ehca_reqs.c
+6
-7
drivers/infiniband/hw/ipath/ipath_driver.c
drivers/infiniband/hw/ipath/ipath_driver.c
+30
-19
drivers/infiniband/hw/ipath/ipath_file_ops.c
drivers/infiniband/hw/ipath/ipath_file_ops.c
+17
-13
drivers/infiniband/hw/ipath/ipath_fs.c
drivers/infiniband/hw/ipath/ipath_fs.c
+1
-1
drivers/infiniband/hw/ipath/ipath_iba6120.c
drivers/infiniband/hw/ipath/ipath_iba6120.c
+61
-0
drivers/infiniband/hw/ipath/ipath_iba7220.c
drivers/infiniband/hw/ipath/ipath_iba7220.c
+77
-6
drivers/infiniband/hw/ipath/ipath_init_chip.c
drivers/infiniband/hw/ipath/ipath_init_chip.c
+1
-0
drivers/infiniband/hw/ipath/ipath_kernel.h
drivers/infiniband/hw/ipath/ipath_kernel.h
+15
-0
drivers/infiniband/hw/ipath/ipath_keys.c
drivers/infiniband/hw/ipath/ipath_keys.c
+2
-0
drivers/infiniband/hw/ipath/ipath_mad.c
drivers/infiniband/hw/ipath/ipath_mad.c
+2
-0
drivers/infiniband/hw/ipath/ipath_qp.c
drivers/infiniband/hw/ipath/ipath_qp.c
+24
-8
drivers/infiniband/hw/ipath/ipath_rc.c
drivers/infiniband/hw/ipath/ipath_rc.c
+2
-3
drivers/infiniband/hw/ipath/ipath_sdma.c
drivers/infiniband/hw/ipath/ipath_sdma.c
+16
-5
drivers/infiniband/hw/ipath/ipath_stats.c
drivers/infiniband/hw/ipath/ipath_stats.c
+8
-0
drivers/infiniband/hw/ipath/ipath_ud.c
drivers/infiniband/hw/ipath/ipath_ud.c
+1
-18
drivers/infiniband/hw/ipath/ipath_verbs.c
drivers/infiniband/hw/ipath/ipath_verbs.c
+2
-1
drivers/infiniband/hw/ipath/ipath_verbs.h
drivers/infiniband/hw/ipath/ipath_verbs.h
+1
-0
drivers/infiniband/hw/mlx4/cq.c
drivers/infiniband/hw/mlx4/cq.c
+7
-5
drivers/infiniband/hw/mlx4/main.c
drivers/infiniband/hw/mlx4/main.c
+1
-1
drivers/infiniband/hw/nes/nes.h
drivers/infiniband/hw/nes/nes.h
+11
-7
drivers/infiniband/hw/nes/nes_cm.c
drivers/infiniband/hw/nes/nes_cm.c
+145
-134
drivers/infiniband/hw/nes/nes_cm.h
drivers/infiniband/hw/nes/nes_cm.h
+10
-4
drivers/infiniband/hw/nes/nes_hw.c
drivers/infiniband/hw/nes/nes_hw.c
+24
-18
drivers/infiniband/hw/nes/nes_utils.c
drivers/infiniband/hw/nes/nes_utils.c
+6
-3
drivers/infiniband/hw/nes/nes_verbs.c
drivers/infiniband/hw/nes/nes_verbs.c
+10
-35
drivers/infiniband/ulp/iser/iscsi_iser.h
drivers/infiniband/ulp/iser/iscsi_iser.h
+3
-0
drivers/infiniband/ulp/iser/iser_initiator.c
drivers/infiniband/ulp/iser/iser_initiator.c
+91
-41
drivers/infiniband/ulp/iser/iser_verbs.c
drivers/infiniband/ulp/iser/iser_verbs.c
+1
-0
drivers/net/mlx4/cq.c
drivers/net/mlx4/cq.c
+8
-3
drivers/net/mlx4/en_cq.c
drivers/net/mlx4/en_cq.c
+6
-3
drivers/net/mlx4/en_main.c
drivers/net/mlx4/en_main.c
+2
-2
drivers/net/mlx4/eq.c
drivers/net/mlx4/eq.c
+88
-33
drivers/net/mlx4/main.c
drivers/net/mlx4/main.c
+39
-14
drivers/net/mlx4/mlx4.h
drivers/net/mlx4/mlx4.h
+6
-8
drivers/net/mlx4/profile.c
drivers/net/mlx4/profile.c
+3
-1
include/linux/mlx4/device.h
include/linux/mlx4/device.h
+3
-1
No files found.
drivers/infiniband/hw/ehca/ehca_classes.h
View file @
2a0d8366
...
...
@@ -175,6 +175,13 @@ struct ehca_queue_map {
unsigned
int
next_wqe_idx
;
/* Idx to first wqe to be flushed */
};
/* function to calculate the next index for the qmap */
static
inline
unsigned
int
next_index
(
unsigned
int
cur_index
,
unsigned
int
limit
)
{
unsigned
int
temp
=
cur_index
+
1
;
return
(
temp
==
limit
)
?
0
:
temp
;
}
struct
ehca_qp
{
union
{
struct
ib_qp
ib_qp
;
...
...
drivers/infiniband/hw/ehca/ehca_eq.c
View file @
2a0d8366
...
...
@@ -113,7 +113,7 @@ int ehca_create_eq(struct ehca_shca *shca,
if
(
h_ret
!=
H_SUCCESS
||
vpage
)
goto
create_eq_exit2
;
}
else
{
if
(
h_ret
!=
H_PAGE_REGISTERED
||
!
vpage
)
if
(
h_ret
!=
H_PAGE_REGISTERED
)
goto
create_eq_exit2
;
}
}
...
...
drivers/infiniband/hw/ehca/ehca_main.c
View file @
2a0d8366
...
...
@@ -717,6 +717,7 @@ static int __devinit ehca_probe(struct of_device *dev,
const
u64
*
handle
;
struct
ib_pd
*
ibpd
;
int
ret
,
i
,
eq_size
;
unsigned
long
flags
;
handle
=
of_get_property
(
dev
->
node
,
"ibm,hca-handle"
,
NULL
);
if
(
!
handle
)
{
...
...
@@ -830,9 +831,9 @@ static int __devinit ehca_probe(struct of_device *dev,
ehca_err
(
&
shca
->
ib_device
,
"Cannot create device attributes ret=%d"
,
ret
);
spin_lock
(
&
shca_list_lock
);
spin_lock
_irqsave
(
&
shca_list_lock
,
flags
);
list_add
(
&
shca
->
shca_list
,
&
shca_list
);
spin_unlock
(
&
shca_list_lock
);
spin_unlock
_irqrestore
(
&
shca_list_lock
,
flags
);
return
0
;
...
...
@@ -878,6 +879,7 @@ static int __devinit ehca_probe(struct of_device *dev,
static
int
__devexit
ehca_remove
(
struct
of_device
*
dev
)
{
struct
ehca_shca
*
shca
=
dev
->
dev
.
driver_data
;
unsigned
long
flags
;
int
ret
;
sysfs_remove_group
(
&
dev
->
dev
.
kobj
,
&
ehca_dev_attr_grp
);
...
...
@@ -915,9 +917,9 @@ static int __devexit ehca_remove(struct of_device *dev)
ib_dealloc_device
(
&
shca
->
ib_device
);
spin_lock
(
&
shca_list_lock
);
spin_lock
_irqsave
(
&
shca_list_lock
,
flags
);
list_del
(
&
shca
->
shca_list
);
spin_unlock
(
&
shca_list_lock
);
spin_unlock
_irqrestore
(
&
shca_list_lock
,
flags
);
return
ret
;
}
...
...
@@ -975,6 +977,7 @@ static int ehca_mem_notifier(struct notifier_block *nb,
unsigned
long
action
,
void
*
data
)
{
static
unsigned
long
ehca_dmem_warn_time
;
unsigned
long
flags
;
switch
(
action
)
{
case
MEM_CANCEL_OFFLINE
:
...
...
@@ -985,12 +988,12 @@ static int ehca_mem_notifier(struct notifier_block *nb,
case
MEM_GOING_ONLINE
:
case
MEM_GOING_OFFLINE
:
/* only ok if no hca is attached to the lpar */
spin_lock
(
&
shca_list_lock
);
spin_lock
_irqsave
(
&
shca_list_lock
,
flags
);
if
(
list_empty
(
&
shca_list
))
{
spin_unlock
(
&
shca_list_lock
);
spin_unlock
_irqrestore
(
&
shca_list_lock
,
flags
);
return
NOTIFY_OK
;
}
else
{
spin_unlock
(
&
shca_list_lock
);
spin_unlock
_irqrestore
(
&
shca_list_lock
,
flags
);
if
(
printk_timed_ratelimit
(
&
ehca_dmem_warn_time
,
30
*
1000
))
ehca_gen_err
(
"DMEM operations are not allowed"
...
...
drivers/infiniband/hw/ehca/ehca_qp.c
View file @
2a0d8366
...
...
@@ -1138,14 +1138,14 @@ static int calc_left_cqes(u64 wqe_p, struct ipz_queue *ipz_queue,
return
-
EFAULT
;
}
tail_idx
=
(
qmap
->
tail
+
1
)
%
qmap
->
entries
;
tail_idx
=
next_index
(
qmap
->
tail
,
qmap
->
entries
)
;
wqe_idx
=
q_ofs
/
ipz_queue
->
qe_size
;
/* check all processed wqes, whether a cqe is requested or not */
while
(
tail_idx
!=
wqe_idx
)
{
if
(
qmap
->
map
[
tail_idx
].
cqe_req
)
qmap
->
left_to_poll
++
;
tail_idx
=
(
tail_idx
+
1
)
%
qmap
->
entries
;
tail_idx
=
next_index
(
tail_idx
,
qmap
->
entries
)
;
}
/* save index in queue, where we have to start flushing */
qmap
->
next_wqe_idx
=
wqe_idx
;
...
...
@@ -1195,14 +1195,14 @@ static int check_for_left_cqes(struct ehca_qp *my_qp, struct ehca_shca *shca)
}
else
{
spin_lock_irqsave
(
&
my_qp
->
send_cq
->
spinlock
,
flags
);
my_qp
->
sq_map
.
left_to_poll
=
0
;
my_qp
->
sq_map
.
next_wqe_idx
=
(
my_qp
->
sq_map
.
tail
+
1
)
%
my_qp
->
sq_map
.
entries
;
my_qp
->
sq_map
.
next_wqe_idx
=
next_index
(
my_qp
->
sq_map
.
tail
,
my_qp
->
sq_map
.
entries
)
;
spin_unlock_irqrestore
(
&
my_qp
->
send_cq
->
spinlock
,
flags
);
spin_lock_irqsave
(
&
my_qp
->
recv_cq
->
spinlock
,
flags
);
my_qp
->
rq_map
.
left_to_poll
=
0
;
my_qp
->
rq_map
.
next_wqe_idx
=
(
my_qp
->
rq_map
.
tail
+
1
)
%
my_qp
->
rq_map
.
entries
;
my_qp
->
rq_map
.
next_wqe_idx
=
next_index
(
my_qp
->
rq_map
.
tail
,
my_qp
->
rq_map
.
entries
)
;
spin_unlock_irqrestore
(
&
my_qp
->
recv_cq
->
spinlock
,
flags
);
}
...
...
drivers/infiniband/hw/ehca/ehca_reqs.c
View file @
2a0d8366
...
...
@@ -726,13 +726,13 @@ static inline int ehca_poll_cq_one(struct ib_cq *cq, struct ib_wc *wc)
* set left_to_poll to 0 because in error state, we will not
* get any additional CQEs
*/
my_qp
->
sq_map
.
next_wqe_idx
=
(
my_qp
->
sq_map
.
tail
+
1
)
%
my_qp
->
sq_map
.
entries
;
my_qp
->
sq_map
.
next_wqe_idx
=
next_index
(
my_qp
->
sq_map
.
tail
,
my_qp
->
sq_map
.
entries
)
;
my_qp
->
sq_map
.
left_to_poll
=
0
;
ehca_add_to_err_list
(
my_qp
,
1
);
my_qp
->
rq_map
.
next_wqe_idx
=
(
my_qp
->
rq_map
.
tail
+
1
)
%
my_qp
->
rq_map
.
entries
;
my_qp
->
rq_map
.
next_wqe_idx
=
next_index
(
my_qp
->
rq_map
.
tail
,
my_qp
->
rq_map
.
entries
)
;
my_qp
->
rq_map
.
left_to_poll
=
0
;
if
(
HAS_RQ
(
my_qp
))
ehca_add_to_err_list
(
my_qp
,
0
);
...
...
@@ -860,9 +860,8 @@ static int generate_flush_cqes(struct ehca_qp *my_qp, struct ib_cq *cq,
/* mark as reported and advance next_wqe pointer */
qmap_entry
->
reported
=
1
;
qmap
->
next_wqe_idx
++
;
if
(
qmap
->
next_wqe_idx
==
qmap
->
entries
)
qmap
->
next_wqe_idx
=
0
;
qmap
->
next_wqe_idx
=
next_index
(
qmap
->
next_wqe_idx
,
qmap
->
entries
);
qmap_entry
=
&
qmap
->
map
[
qmap
->
next_wqe_idx
];
wc
++
;
nr
++
;
...
...
drivers/infiniband/hw/ipath/ipath_driver.c
View file @
2a0d8366
...
...
@@ -661,6 +661,8 @@ static int __devinit ipath_init_one(struct pci_dev *pdev,
static
void
__devexit
cleanup_device
(
struct
ipath_devdata
*
dd
)
{
int
port
;
struct
ipath_portdata
**
tmp
;
unsigned
long
flags
;
if
(
*
dd
->
ipath_statusp
&
IPATH_STATUS_CHIP_PRESENT
)
{
/* can't do anything more with chip; needs re-init */
...
...
@@ -742,20 +744,21 @@ static void __devexit cleanup_device(struct ipath_devdata *dd)
/*
* free any resources still in use (usually just kernel ports)
* at unload; we do for portcnt, not cfgports, because cfgports
* could have changed while we were loaded.
* at unload; we do for portcnt, because that's what we allocate.
* We acquire lock to be really paranoid that ipath_pd isn't being
* accessed from some interrupt-related code (that should not happen,
* but best to be sure).
*/
spin_lock_irqsave
(
&
dd
->
ipath_uctxt_lock
,
flags
);
tmp
=
dd
->
ipath_pd
;
dd
->
ipath_pd
=
NULL
;
spin_unlock_irqrestore
(
&
dd
->
ipath_uctxt_lock
,
flags
);
for
(
port
=
0
;
port
<
dd
->
ipath_portcnt
;
port
++
)
{
struct
ipath_portdata
*
pd
=
dd
->
ipath_pd
[
port
];
dd
->
ipath_pd
[
port
]
=
NULL
;
struct
ipath_portdata
*
pd
=
tmp
[
port
];
tmp
[
port
]
=
NULL
;
/* debugging paranoia */
ipath_free_pddata
(
dd
,
pd
);
}
kfree
(
dd
->
ipath_pd
);
/*
* debuggability, in case some cleanup path tries to use it
* after this
*/
dd
->
ipath_pd
=
NULL
;
kfree
(
tmp
);
}
static
void
__devexit
ipath_remove_one
(
struct
pci_dev
*
pdev
)
...
...
@@ -2586,6 +2589,7 @@ int ipath_reset_device(int unit)
{
int
ret
,
i
;
struct
ipath_devdata
*
dd
=
ipath_lookup
(
unit
);
unsigned
long
flags
;
if
(
!
dd
)
{
ret
=
-
ENODEV
;
...
...
@@ -2611,9 +2615,12 @@ int ipath_reset_device(int unit)
goto
bail
;
}
spin_lock_irqsave
(
&
dd
->
ipath_uctxt_lock
,
flags
);
if
(
dd
->
ipath_pd
)
for
(
i
=
1
;
i
<
dd
->
ipath_cfgports
;
i
++
)
{
if
(
dd
->
ipath_pd
[
i
]
&&
dd
->
ipath_pd
[
i
]
->
port_cnt
)
{
if
(
!
dd
->
ipath_pd
[
i
]
||
!
dd
->
ipath_pd
[
i
]
->
port_cnt
)
continue
;
spin_unlock_irqrestore
(
&
dd
->
ipath_uctxt_lock
,
flags
);
ipath_dbg
(
"unit %u port %d is in use "
"(PID %u cmd %s), can't reset
\n
"
,
unit
,
i
,
...
...
@@ -2622,7 +2629,7 @@ int ipath_reset_device(int unit)
ret
=
-
EBUSY
;
goto
bail
;
}
}
spin_unlock_irqrestore
(
&
dd
->
ipath_uctxt_lock
,
flags
);
if
(
dd
->
ipath_flags
&
IPATH_HAS_SEND_DMA
)
teardown_sdma
(
dd
);
...
...
@@ -2656,9 +2663,12 @@ static int ipath_signal_procs(struct ipath_devdata *dd, int sig)
{
int
i
,
sub
,
any
=
0
;
struct
pid
*
pid
;
unsigned
long
flags
;
if
(
!
dd
->
ipath_pd
)
return
0
;
spin_lock_irqsave
(
&
dd
->
ipath_uctxt_lock
,
flags
);
for
(
i
=
1
;
i
<
dd
->
ipath_cfgports
;
i
++
)
{
if
(
!
dd
->
ipath_pd
[
i
]
||
!
dd
->
ipath_pd
[
i
]
->
port_cnt
)
continue
;
...
...
@@ -2682,6 +2692,7 @@ static int ipath_signal_procs(struct ipath_devdata *dd, int sig)
any
++
;
}
}
spin_unlock_irqrestore
(
&
dd
->
ipath_uctxt_lock
,
flags
);
return
any
;
}
...
...
drivers/infiniband/hw/ipath/ipath_file_ops.c
View file @
2a0d8366
...
...
@@ -223,8 +223,13 @@ static int ipath_get_base_info(struct file *fp,
(
unsigned
long
long
)
kinfo
->
spi_subport_rcvhdr_base
);
}
kinfo
->
spi_pioindex
=
(
kinfo
->
spi_piobufbase
-
dd
->
ipath_piobufbase
)
/
dd
->
ipath_palign
;
/*
* All user buffers are 2KB buffers. If we ever support
* giving 4KB buffers to user processes, this will need some
* work.
*/
kinfo
->
spi_pioindex
=
(
kinfo
->
spi_piobufbase
-
(
dd
->
ipath_piobufbase
&
0xffffffff
))
/
dd
->
ipath_palign
;
kinfo
->
spi_pioalign
=
dd
->
ipath_palign
;
kinfo
->
spi_qpair
=
IPATH_KD_QP
;
...
...
@@ -2041,7 +2046,9 @@ static int ipath_close(struct inode *in, struct file *fp)
struct
ipath_filedata
*
fd
;
struct
ipath_portdata
*
pd
;
struct
ipath_devdata
*
dd
;
unsigned
long
flags
;
unsigned
port
;
struct
pid
*
pid
;
ipath_cdbg
(
VERBOSE
,
"close on dev %lx, private data %p
\n
"
,
(
long
)
in
->
i_rdev
,
fp
->
private_data
);
...
...
@@ -2074,14 +2081,13 @@ static int ipath_close(struct inode *in, struct file *fp)
mutex_unlock
(
&
ipath_mutex
);
goto
bail
;
}
/* early; no interrupt users after this */
spin_lock_irqsave
(
&
dd
->
ipath_uctxt_lock
,
flags
);
port
=
pd
->
port_port
;
if
(
pd
->
port_hdrqfull
)
{
ipath_cdbg
(
PROC
,
"%s[%u] had %u rcvhdrqfull errors "
"during run
\n
"
,
pd
->
port_comm
,
pid_nr
(
pd
->
port_pid
),
pd
->
port_hdrqfull
);
pd
->
port_hdrqfull
=
0
;
}
dd
->
ipath_pd
[
port
]
=
NULL
;
pid
=
pd
->
port_pid
;
pd
->
port_pid
=
NULL
;
spin_unlock_irqrestore
(
&
dd
->
ipath_uctxt_lock
,
flags
);
if
(
pd
->
port_rcvwait_to
||
pd
->
port_piowait_to
||
pd
->
port_rcvnowait
||
pd
->
port_pionowait
)
{
...
...
@@ -2138,13 +2144,11 @@ static int ipath_close(struct inode *in, struct file *fp)
unlock_expected_tids
(
pd
);
ipath_stats
.
sps_ports
--
;
ipath_cdbg
(
PROC
,
"%s[%u] closed port %u:%u
\n
"
,
pd
->
port_comm
,
pid_nr
(
p
d
->
port_p
id
),
pd
->
port_comm
,
pid_nr
(
pid
),
dd
->
ipath_unit
,
port
);
}
put_pid
(
pd
->
port_pid
);
pd
->
port_pid
=
NULL
;
dd
->
ipath_pd
[
pd
->
port_port
]
=
NULL
;
/* before releasing mutex */
put_pid
(
pid
);
mutex_unlock
(
&
ipath_mutex
);
ipath_free_pddata
(
dd
,
pd
);
/* after releasing the mutex */
...
...
drivers/infiniband/hw/ipath/ipath_fs.c
View file @
2a0d8366
...
...
@@ -86,7 +86,7 @@ static int create_file(const char *name, mode_t mode,
*
dentry
=
NULL
;
mutex_lock
(
&
parent
->
d_inode
->
i_mutex
);
*
dentry
=
lookup_one_len
(
name
,
parent
,
strlen
(
name
));
if
(
!
IS_ERR
(
dentry
))
if
(
!
IS_ERR
(
*
dentry
))
error
=
ipathfs_mknod
(
parent
->
d_inode
,
*
dentry
,
mode
,
fops
,
data
);
else
...
...
drivers/infiniband/hw/ipath/ipath_iba6120.c
View file @
2a0d8366
...
...
@@ -721,6 +721,12 @@ static int ipath_pe_bringup_serdes(struct ipath_devdata *dd)
INFINIPATH_HWE_SERDESPLLFAILED
);
}
dd
->
ibdeltainprog
=
1
;
dd
->
ibsymsnap
=
ipath_read_creg32
(
dd
,
dd
->
ipath_cregs
->
cr_ibsymbolerrcnt
);
dd
->
iblnkerrsnap
=
ipath_read_creg32
(
dd
,
dd
->
ipath_cregs
->
cr_iblinkerrrecovcnt
);
val
=
ipath_read_kreg64
(
dd
,
dd
->
ipath_kregs
->
kr_serdesconfig0
);
config1
=
ipath_read_kreg64
(
dd
,
dd
->
ipath_kregs
->
kr_serdesconfig1
);
...
...
@@ -810,6 +816,36 @@ static void ipath_pe_quiet_serdes(struct ipath_devdata *dd)
{
u64
val
=
ipath_read_kreg64
(
dd
,
dd
->
ipath_kregs
->
kr_serdesconfig0
);
if
(
dd
->
ibsymdelta
||
dd
->
iblnkerrdelta
||
dd
->
ibdeltainprog
)
{
u64
diagc
;
/* enable counter writes */
diagc
=
ipath_read_kreg64
(
dd
,
dd
->
ipath_kregs
->
kr_hwdiagctrl
);
ipath_write_kreg
(
dd
,
dd
->
ipath_kregs
->
kr_hwdiagctrl
,
diagc
|
INFINIPATH_DC_COUNTERWREN
);
if
(
dd
->
ibsymdelta
||
dd
->
ibdeltainprog
)
{
val
=
ipath_read_creg32
(
dd
,
dd
->
ipath_cregs
->
cr_ibsymbolerrcnt
);
if
(
dd
->
ibdeltainprog
)
val
-=
val
-
dd
->
ibsymsnap
;
val
-=
dd
->
ibsymdelta
;
ipath_write_creg
(
dd
,
dd
->
ipath_cregs
->
cr_ibsymbolerrcnt
,
val
);
}
if
(
dd
->
iblnkerrdelta
||
dd
->
ibdeltainprog
)
{
val
=
ipath_read_creg32
(
dd
,
dd
->
ipath_cregs
->
cr_iblinkerrrecovcnt
);
if
(
dd
->
ibdeltainprog
)
val
-=
val
-
dd
->
iblnkerrsnap
;
val
-=
dd
->
iblnkerrdelta
;
ipath_write_creg
(
dd
,
dd
->
ipath_cregs
->
cr_iblinkerrrecovcnt
,
val
);
}
/* and disable counter writes */
ipath_write_kreg
(
dd
,
dd
->
ipath_kregs
->
kr_hwdiagctrl
,
diagc
);
}
val
|=
INFINIPATH_SERDC0_TXIDLE
;
ipath_dbg
(
"Setting TxIdleEn on serdes (config0 = %llx)
\n
"
,
(
unsigned
long
long
)
val
);
...
...
@@ -1749,6 +1785,31 @@ static void ipath_pe_config_jint(struct ipath_devdata *dd, u16 a, u16 b)
static
int
ipath_pe_ib_updown
(
struct
ipath_devdata
*
dd
,
int
ibup
,
u64
ibcs
)
{
if
(
ibup
)
{
if
(
dd
->
ibdeltainprog
)
{
dd
->
ibdeltainprog
=
0
;
dd
->
ibsymdelta
+=
ipath_read_creg32
(
dd
,
dd
->
ipath_cregs
->
cr_ibsymbolerrcnt
)
-
dd
->
ibsymsnap
;
dd
->
iblnkerrdelta
+=
ipath_read_creg32
(
dd
,
dd
->
ipath_cregs
->
cr_iblinkerrrecovcnt
)
-
dd
->
iblnkerrsnap
;
}
}
else
{
dd
->
ipath_lli_counter
=
0
;
if
(
!
dd
->
ibdeltainprog
)
{
dd
->
ibdeltainprog
=
1
;
dd
->
ibsymsnap
=
ipath_read_creg32
(
dd
,
dd
->
ipath_cregs
->
cr_ibsymbolerrcnt
);
dd
->
iblnkerrsnap
=
ipath_read_creg32
(
dd
,
dd
->
ipath_cregs
->
cr_iblinkerrrecovcnt
);
}
}
ipath_setup_pe_setextled
(
dd
,
ipath_ib_linkstate
(
dd
,
ibcs
),
ipath_ib_linktrstate
(
dd
,
ibcs
));
return
0
;
...
...
drivers/infiniband/hw/ipath/ipath_iba7220.c
View file @
2a0d8366
...
...
@@ -951,6 +951,12 @@ static int ipath_7220_bringup_serdes(struct ipath_devdata *dd)
INFINIPATH_HWE_SERDESPLLFAILED
);
}
dd
->
ibdeltainprog
=
1
;
dd
->
ibsymsnap
=
ipath_read_creg32
(
dd
,
dd
->
ipath_cregs
->
cr_ibsymbolerrcnt
);
dd
->
iblnkerrsnap
=
ipath_read_creg32
(
dd
,
dd
->
ipath_cregs
->
cr_iblinkerrrecovcnt
);
if
(
!
dd
->
ipath_ibcddrctrl
)
{
/* not on re-init after reset */
dd
->
ipath_ibcddrctrl
=
...
...
@@ -1084,6 +1090,37 @@ static void ipath_7220_config_jint(struct ipath_devdata *dd,
static
void
ipath_7220_quiet_serdes
(
struct
ipath_devdata
*
dd
)
{
u64
val
;
if
(
dd
->
ibsymdelta
||
dd
->
iblnkerrdelta
||
dd
->
ibdeltainprog
)
{
u64
diagc
;
/* enable counter writes */
diagc
=
ipath_read_kreg64
(
dd
,
dd
->
ipath_kregs
->
kr_hwdiagctrl
);
ipath_write_kreg
(
dd
,
dd
->
ipath_kregs
->
kr_hwdiagctrl
,
diagc
|
INFINIPATH_DC_COUNTERWREN
);
if
(
dd
->
ibsymdelta
||
dd
->
ibdeltainprog
)
{
val
=
ipath_read_creg32
(
dd
,
dd
->
ipath_cregs
->
cr_ibsymbolerrcnt
);
if
(
dd
->
ibdeltainprog
)
val
-=
val
-
dd
->
ibsymsnap
;
val
-=
dd
->
ibsymdelta
;
ipath_write_creg
(
dd
,
dd
->
ipath_cregs
->
cr_ibsymbolerrcnt
,
val
);
}
if
(
dd
->
iblnkerrdelta
||
dd
->
ibdeltainprog
)
{
val
=
ipath_read_creg32
(
dd
,
dd
->
ipath_cregs
->
cr_iblinkerrrecovcnt
);
if
(
dd
->
ibdeltainprog
)
val
-=
val
-
dd
->
iblnkerrsnap
;
val
-=
dd
->
iblnkerrdelta
;
ipath_write_creg
(
dd
,
dd
->
ipath_cregs
->
cr_iblinkerrrecovcnt
,
val
);
}
/* and disable counter writes */
ipath_write_kreg
(
dd
,
dd
->
ipath_kregs
->
kr_hwdiagctrl
,
diagc
);
}
dd
->
ipath_flags
&=
~
IPATH_IB_AUTONEG_INPROG
;
wake_up
(
&
dd
->
ipath_autoneg_wait
);
cancel_delayed_work
(
&
dd
->
ipath_autoneg_work
);
...
...
@@ -2325,7 +2362,7 @@ static void try_auto_neg(struct ipath_devdata *dd)
static
int
ipath_7220_ib_updown
(
struct
ipath_devdata
*
dd
,
int
ibup
,
u64
ibcs
)
{
int
ret
=
0
;
int
ret
=
0
,
symadj
=
0
;
u32
ltstate
=
ipath_ib_linkstate
(
dd
,
ibcs
);
dd
->
ipath_link_width_active
=
...
...
@@ -2368,6 +2405,13 @@ static int ipath_7220_ib_updown(struct ipath_devdata *dd, int ibup, u64 ibcs)
ipath_dbg
(
"DDR negotiation try, %u/%u
\n
"
,
dd
->
ipath_autoneg_tries
,
IPATH_AUTONEG_TRIES
);
if
(
!
dd
->
ibdeltainprog
)
{
dd
->
ibdeltainprog
=
1
;
dd
->
ibsymsnap
=
ipath_read_creg32
(
dd
,
dd
->
ipath_cregs
->
cr_ibsymbolerrcnt
);
dd
->
iblnkerrsnap
=
ipath_read_creg32
(
dd
,
dd
->
ipath_cregs
->
cr_iblinkerrrecovcnt
);
}
try_auto_neg
(
dd
);
ret
=
1
;
/* no other IB status change processing */
}
else
if
((
dd
->
ipath_flags
&
IPATH_IB_AUTONEG_INPROG
)
...
...
@@ -2388,6 +2432,7 @@ static int ipath_7220_ib_updown(struct ipath_devdata *dd, int ibup, u64 ibcs)
set_speed_fast
(
dd
,
dd
->
ipath_link_speed_enabled
);
wake_up
(
&
dd
->
ipath_autoneg_wait
);
symadj
=
1
;
}
else
if
(
dd
->
ipath_flags
&
IPATH_IB_AUTONEG_FAILED
)
{
/*
* clear autoneg failure flag, and do setup
...
...
@@ -2403,22 +2448,28 @@ static int ipath_7220_ib_updown(struct ipath_devdata *dd, int ibup, u64 ibcs)
IBA7220_IBC_IBTA_1_2_MASK
;
ipath_write_kreg
(
dd
,
IPATH_KREG_OFFSET
(
IBNCModeCtrl
),
0
);
symadj
=
1
;
}
}
/*
* if we are in 1X
, and are in autoneg width, it
* could be due to an xgxs problem, so if we haven't
* if we are in 1X
on rev1 only, and are in autoneg width,
*
it
could be due to an xgxs problem, so if we haven't
* already tried, try twice to get to 4X; if we
* tried, and couldn't, report it, since it will
* probably not be what is desired.
*/
if
((
dd
->
ipath_link_width_enabled
&
(
IB_WIDTH_1X
|
if
(
dd
->
ipath_minrev
==
1
&&
(
dd
->
ipath_link_width_enabled
&
(
IB_WIDTH_1X
|
IB_WIDTH_4X
))
==
(
IB_WIDTH_1X
|
IB_WIDTH_4X
)
&&
dd
->
ipath_link_width_active
==
IB_WIDTH_1X
&&
dd
->
ipath_x1_fix_tries
<
3
)
{
if
(
++
dd
->
ipath_x1_fix_tries
==
3
)
if
(
++
dd
->
ipath_x1_fix_tries
==
3
)
{
dev_info
(
&
dd
->
pcidev
->
dev
,
"IB link is in 1X mode
\n
"
);
if
(
!
(
dd
->
ipath_flags
&
IPATH_IB_AUTONEG_INPROG
))
symadj
=
1
;
}
else
{
ipath_cdbg
(
VERBOSE
,
"IB 1X in "
"auto-width, try %u to be "
...
...
@@ -2429,7 +2480,8 @@ static int ipath_7220_ib_updown(struct ipath_devdata *dd, int ibup, u64 ibcs)
dd
->
ipath_f_xgxs_reset
(
dd
);
ret
=
1
;
/* skip other processing */
}
}
}
else
if
(
!
(
dd
->
ipath_flags
&
IPATH_IB_AUTONEG_INPROG
))
symadj
=
1
;
if
(
!
ret
)
{
dd
->
delay_mult
=
rate_to_delay
...
...
@@ -2440,6 +2492,25 @@ static int ipath_7220_ib_updown(struct ipath_devdata *dd, int ibup, u64 ibcs)
}
}
if
(
symadj
)
{
if
(
dd
->
ibdeltainprog
)
{
dd
->
ibdeltainprog
=
0
;
dd
->
ibsymdelta
+=
ipath_read_creg32
(
dd
,
dd
->
ipath_cregs
->
cr_ibsymbolerrcnt
)
-
dd
->
ibsymsnap
;
dd
->
iblnkerrdelta
+=
ipath_read_creg32
(
dd
,
dd
->
ipath_cregs
->
cr_iblinkerrrecovcnt
)
-
dd
->
iblnkerrsnap
;
}
}
else
if
(
!
ibup
&&
!
dd
->
ibdeltainprog
&&
!
(
dd
->
ipath_flags
&
IPATH_IB_AUTONEG_INPROG
))
{
dd
->
ibdeltainprog
=
1
;
dd
->
ibsymsnap
=
ipath_read_creg32
(
dd
,
dd
->
ipath_cregs
->
cr_ibsymbolerrcnt
);
dd
->
iblnkerrsnap
=
ipath_read_creg32
(
dd
,
dd
->
ipath_cregs
->
cr_iblinkerrrecovcnt
);
}
if
(
!
ret
)
ipath_setup_7220_setextled
(
dd
,
ipath_ib_linkstate
(
dd
,
ibcs
),
ltstate
);
...
...
drivers/infiniband/hw/ipath/ipath_init_chip.c
View file @
2a0d8366
...
...
@@ -229,6 +229,7 @@ static int init_chip_first(struct ipath_devdata *dd)
spin_lock_init
(
&
dd
->
ipath_kernel_tid_lock
);
spin_lock_init
(
&
dd
->
ipath_user_tid_lock
);
spin_lock_init
(
&
dd
->
ipath_sendctrl_lock
);
spin_lock_init
(
&
dd
->
ipath_uctxt_lock
);
spin_lock_init
(
&
dd
->
ipath_sdma_lock
);
spin_lock_init
(
&
dd
->
ipath_gpio_lock
);
spin_lock_init
(
&
dd
->
ipath_eep_st_lock
);
...
...
drivers/infiniband/hw/ipath/ipath_kernel.h
View file @
2a0d8366
...
...
@@ -355,6 +355,19 @@ struct ipath_devdata {
/* errors masked because they occur too fast */
ipath_err_t
ipath_maskederrs
;
u64
ipath_lastlinkrecov
;
/* link recoveries at last ACTIVE */
/* these 5 fields are used to establish deltas for IB Symbol
* errors and linkrecovery errors. They can be reported on
* some chips during link negotiation prior to INIT, and with
* DDR when faking DDR negotiations with non-IBTA switches.
* The chip counters are adjusted at driver unload if there is
* a non-zero delta.
*/
u64
ibdeltainprog
;
u64
ibsymdelta
;
u64
ibsymsnap
;
u64
iblnkerrdelta
;
u64
iblnkerrsnap
;
/* time in jiffies at which to re-enable maskederrs */
unsigned
long
ipath_unmasktime
;
/* count of egrfull errors, combined for all ports */
...
...
@@ -464,6 +477,8 @@ struct ipath_devdata {
spinlock_t
ipath_kernel_tid_lock
;
spinlock_t
ipath_user_tid_lock
;
spinlock_t
ipath_sendctrl_lock
;
/* around ipath_pd and (user ports) port_cnt use (intr vs free) */
spinlock_t
ipath_uctxt_lock
;
/*
* IPATH_STATUS_*,
...
...
drivers/infiniband/hw/ipath/ipath_keys.c
View file @
2a0d8366
...
...
@@ -132,6 +132,7 @@ int ipath_lkey_ok(struct ipath_qp *qp, struct ipath_sge *isge,
* (see ipath_get_dma_mr and ipath_dma.c).
*/
if
(
sge
->
lkey
==
0
)
{
/* always a kernel port, no locking needed */
struct
ipath_pd
*
pd
=
to_ipd
(
qp
->
ibqp
.
pd
);
if
(
pd
->
user
)
{
...
...
@@ -211,6 +212,7 @@ int ipath_rkey_ok(struct ipath_qp *qp, struct ipath_sge_state *ss,
* (see ipath_get_dma_mr and ipath_dma.c).
*/
if
(
rkey
==
0
)
{
/* always a kernel port, no locking needed */
struct
ipath_pd
*
pd
=
to_ipd
(
qp
->
ibqp
.
pd
);
if
(
pd
->
user
)
{
...
...
drivers/infiniband/hw/ipath/ipath_mad.c
View file @
2a0d8366
...
...
@@ -348,6 +348,7 @@ static int recv_subn_get_portinfo(struct ib_smp *smp,
*/
static
int
get_pkeys
(
struct
ipath_devdata
*
dd
,
u16
*
pkeys
)
{
/* always a kernel port, no locking needed */
struct
ipath_portdata
*
pd
=
dd
->
ipath_pd
[
0
];
memcpy
(
pkeys
,
pd
->
port_pkeys
,
sizeof
(
pd
->
port_pkeys
));
...
...
@@ -730,6 +731,7 @@ static int set_pkeys(struct ipath_devdata *dd, u16 *pkeys)
int
i
;
int
changed
=
0
;
/* always a kernel port, no locking needed */
pd
=
dd
->
ipath_pd
[
0
];
for
(
i
=
0
;
i
<
ARRAY_SIZE
(
pd
->
port_pkeys
);
i
++
)
{
...
...
drivers/infiniband/hw/ipath/ipath_qp.c
View file @
2a0d8366
...
...
@@ -745,6 +745,7 @@ struct ib_qp *ipath_create_qp(struct ib_pd *ibpd,
struct
ipath_swqe
*
swq
=
NULL
;
struct
ipath_ibdev
*
dev
;
size_t
sz
;
size_t
sg_list_sz
;
struct
ib_qp
*
ret
;
if
(
init_attr
->
create_flags
)
{
...
...
@@ -789,19 +790,31 @@ struct ib_qp *ipath_create_qp(struct ib_pd *ibpd,
goto
bail
;
}
sz
=
sizeof
(
*
qp
);
sg_list_sz
=
0
;
if
(
init_attr
->
srq
)
{
struct
ipath_srq
*
srq
=
to_isrq
(
init_attr
->
srq
);
sz
+=
sizeof
(
*
qp
->
r_sg_list
)
*
srq
->
rq
.
max_sge
;
}
else
sz
+=
sizeof
(
*
qp
->
r_sg_list
)
*
init_attr
->
cap
.
max_recv_sge
;
qp
=
kmalloc
(
sz
,
GFP_KERNEL
);
if
(
srq
->
rq
.
max_sge
>
1
)
sg_list_sz
=
sizeof
(
*
qp
->
r_sg_list
)
*
(
srq
->
rq
.
max_sge
-
1
);
}
else
if
(
init_attr
->
cap
.
max_recv_sge
>
1
)
sg_list_sz
=
sizeof
(
*
qp
->
r_sg_list
)
*
(
init_attr
->
cap
.
max_recv_sge
-
1
);
qp
=
kmalloc
(
sz
+
sg_list_sz
,
GFP_KERNEL
);
if
(
!
qp
)
{
ret
=
ERR_PTR
(
-
ENOMEM
);
goto
bail_swq
;
}
if
(
sg_list_sz
&&
(
init_attr
->
qp_type
==
IB_QPT_UD
||
init_attr
->
qp_type
==
IB_QPT_SMI
||
init_attr
->
qp_type
==
IB_QPT_GSI
))
{
qp
->
r_ud_sg_list
=
kmalloc
(
sg_list_sz
,
GFP_KERNEL
);
if
(
!
qp
->
r_ud_sg_list
)
{
ret
=
ERR_PTR
(
-
ENOMEM
);
goto
bail_qp
;
}
}
else
qp
->
r_ud_sg_list
=
NULL
;
if
(
init_attr
->
srq
)
{
sz
=
0
;
qp
->
r_rq
.
size
=
0
;
...
...
@@ -818,7 +831,7 @@ struct ib_qp *ipath_create_qp(struct ib_pd *ibpd,
qp
->
r_rq
.
size
*
sz
);
if
(
!
qp
->
r_rq
.
wq
)
{
ret
=
ERR_PTR
(
-
ENOMEM
);
goto
bail_
qp
;
goto
bail_
sg_list
;
}
}
...
...
@@ -848,7 +861,7 @@ struct ib_qp *ipath_create_qp(struct ib_pd *ibpd,
if
(
err
)
{
ret
=
ERR_PTR
(
err
);
vfree
(
qp
->
r_rq
.
wq
);
goto
bail_
qp
;
goto
bail_
sg_list
;
}
qp
->
ip
=
NULL
;
qp
->
s_tx
=
NULL
;
...
...
@@ -925,6 +938,8 @@ struct ib_qp *ipath_create_qp(struct ib_pd *ibpd,
vfree
(
qp
->
r_rq
.
wq
);
ipath_free_qp
(
&
dev
->
qp_table
,
qp
);
free_qpn
(
&
dev
->
qp_table
,
qp
->
ibqp
.
qp_num
);
bail_sg_list:
kfree
(
qp
->
r_ud_sg_list
);
bail_qp:
kfree
(
qp
);
bail_swq:
...
...
@@ -989,6 +1004,7 @@ int ipath_destroy_qp(struct ib_qp *ibqp)
kref_put
(
&
qp
->
ip
->
ref
,
ipath_release_mmap_info
);
else
vfree
(
qp
->
r_rq
.
wq
);
kfree
(
qp
->
r_ud_sg_list
);
vfree
(
qp
->
s_wq
);
kfree
(
qp
);
return
0
;
...
...
drivers/infiniband/hw/ipath/ipath_rc.c
View file @
2a0d8366
...
...
@@ -573,9 +573,8 @@ int ipath_make_rc_req(struct ipath_qp *qp)
ohdr
->
u
.
rc
.
reth
.
length
=
cpu_to_be32
(
qp
->
s_len
);
qp
->
s_state
=
OP
(
RDMA_READ_REQUEST
);
hwords
+=
sizeof
(
ohdr
->
u
.
rc
.
reth
)
/
sizeof
(
u32
);
bth2
=
qp
->
s_psn
++
&
IPATH_PSN_MASK
;
if
(
ipath_cmp24
(
qp
->
s_psn
,
qp
->
s_next_psn
)
>
0
)
qp
->
s_next_psn
=
qp
->
s_psn
;
bth2
=
qp
->
s_psn
&
IPATH_PSN_MASK
;
qp
->
s_psn
=
wqe
->
lpsn
+
1
;
ss
=
NULL
;
len
=
0
;
qp
->
s_cur
++
;
...
...
drivers/infiniband/hw/ipath/ipath_sdma.c
View file @
2a0d8366
...
...
@@ -698,10 +698,8 @@ int ipath_sdma_verbs_send(struct ipath_devdata *dd,
addr
=
dma_map_single
(
&
dd
->
pcidev
->
dev
,
tx
->
txreq
.
map_addr
,
tx
->
map_len
,
DMA_TO_DEVICE
);
if
(
dma_mapping_error
(
&
dd
->
pcidev
->
dev
,
addr
))
{
ret
=
-
EIO
;
goto
unlock
;
}
if
(
dma_mapping_error
(
&
dd
->
pcidev
->
dev
,
addr
))
goto
ioerr
;
dwoffset
=
tx
->
map_len
>>
2
;
make_sdma_desc
(
dd
,
sdmadesc
,
(
u64
)
addr
,
dwoffset
,
0
);
...
...
@@ -741,6 +739,8 @@ int ipath_sdma_verbs_send(struct ipath_devdata *dd,
dw
=
(
len
+
3
)
>>
2
;
addr
=
dma_map_single
(
&
dd
->
pcidev
->
dev
,
sge
->
vaddr
,
dw
<<
2
,
DMA_TO_DEVICE
);
if
(
dma_mapping_error
(
&
dd
->
pcidev
->
dev
,
addr
))
goto
unmap
;
make_sdma_desc
(
dd
,
sdmadesc
,
(
u64
)
addr
,
dw
,
dwoffset
);
/* SDmaUseLargeBuf has to be set in every descriptor */
if
(
tx
->
txreq
.
flags
&
IPATH_SDMA_TXREQ_F_USELARGEBUF
)
...
...
@@ -798,7 +798,18 @@ int ipath_sdma_verbs_send(struct ipath_devdata *dd,
list_add_tail
(
&
tx
->
txreq
.
list
,
&
dd
->
ipath_sdma_activelist
);
if
(
tx
->
txreq
.
flags
&
IPATH_SDMA_TXREQ_F_VL15
)
vl15_watchdog_enq
(
dd
);
goto
unlock
;
unmap:
while
(
tail
!=
dd
->
ipath_sdma_descq_tail
)
{
if
(
!
tail
)
tail
=
dd
->
ipath_sdma_descq_cnt
-
1
;
else
tail
--
;
unmap_desc
(
dd
,
tail
);
}
ioerr:
ret
=
-
EIO
;
unlock:
spin_unlock_irqrestore
(
&
dd
->
ipath_sdma_lock
,
flags
);
fail:
...
...
drivers/infiniband/hw/ipath/ipath_stats.c
View file @
2a0d8366
...
...
@@ -112,6 +112,14 @@ u64 ipath_snap_cntr(struct ipath_devdata *dd, ipath_creg creg)
dd
->
ipath_lastrpkts
=
val
;
}
val64
=
dd
->
ipath_rpkts
;
}
else
if
(
creg
==
dd
->
ipath_cregs
->
cr_ibsymbolerrcnt
)
{
if
(
dd
->
ibdeltainprog
)
val64
-=
val64
-
dd
->
ibsymsnap
;
val64
-=
dd
->
ibsymdelta
;
}
else
if
(
creg
==
dd
->
ipath_cregs
->
cr_iblinkerrrecovcnt
)
{
if
(
dd
->
ibdeltainprog
)
val64
-=
val64
-
dd
->
iblnkerrsnap
;
val64
-=
dd
->
iblnkerrdelta
;
}
else
val64
=
(
u64
)
val
;
...
...
drivers/infiniband/hw/ipath/ipath_ud.c
View file @
2a0d8366
...
...
@@ -70,8 +70,6 @@ static void ipath_ud_loopback(struct ipath_qp *sqp, struct ipath_swqe *swqe)
goto
done
;
}
rsge
.
sg_list
=
NULL
;
/*
* Check that the qkey matches (except for QP0, see 9.6.1.4.1).
* Qkeys with the high order bit set mean use the
...
...
@@ -115,21 +113,6 @@ static void ipath_ud_loopback(struct ipath_qp *sqp, struct ipath_swqe *swqe)
rq
=
&
qp
->
r_rq
;
}
if
(
rq
->
max_sge
>
1
)
{
/*
* XXX We could use GFP_KERNEL if ipath_do_send()
* was always called from the tasklet instead of
* from ipath_post_send().
*/
rsge
.
sg_list
=
kmalloc
((
rq
->
max_sge
-
1
)
*
sizeof
(
struct
ipath_sge
),
GFP_ATOMIC
);
if
(
!
rsge
.
sg_list
)
{
dev
->
n_pkt_drops
++
;
goto
drop
;
}
}
/*
* Get the next work request entry to find where to put the data.
* Note that it is safe to drop the lock after changing rq->tail
...
...
@@ -147,6 +130,7 @@ static void ipath_ud_loopback(struct ipath_qp *sqp, struct ipath_swqe *swqe)
goto
drop
;
}
wqe
=
get_rwqe_ptr
(
rq
,
tail
);
rsge
.
sg_list
=
qp
->
r_ud_sg_list
;
if
(
!
ipath_init_sge
(
qp
,
wqe
,
&
rlen
,
&
rsge
))
{
spin_unlock_irqrestore
(
&
rq
->
lock
,
flags
);
dev
->
n_pkt_drops
++
;
...
...
@@ -242,7 +226,6 @@ static void ipath_ud_loopback(struct ipath_qp *sqp, struct ipath_swqe *swqe)
ipath_cq_enter
(
to_icq
(
qp
->
ibqp
.
recv_cq
),
&
wc
,
swqe
->
wr
.
send_flags
&
IB_SEND_SOLICITED
);
drop:
kfree
(
rsge
.
sg_list
);
if
(
atomic_dec_and_test
(
&
qp
->
refcount
))
wake_up
(
&
qp
->
wait
);
done:
;
...
...
drivers/infiniband/hw/ipath/ipath_verbs.c
View file @
2a0d8366
...
...
@@ -1852,7 +1852,7 @@ unsigned ipath_get_npkeys(struct ipath_devdata *dd)
}
/**
* ipath_get_pkey - return the indexed PKEY from the port
0
PKEY table
* ipath_get_pkey - return the indexed PKEY from the port PKEY table
* @dd: the infinipath device
* @index: the PKEY index
*/
...
...
@@ -1860,6 +1860,7 @@ unsigned ipath_get_pkey(struct ipath_devdata *dd, unsigned index)
{
unsigned
ret
;
/* always a kernel port, no locking needed */
if
(
index
>=
ARRAY_SIZE
(
dd
->
ipath_pd
[
0
]
->
port_pkeys
))
ret
=
0
;
else
...
...
drivers/infiniband/hw/ipath/ipath_verbs.h
View file @
2a0d8366
...
...
@@ -431,6 +431,7 @@ struct ipath_qp {
u32
s_lsn
;
/* limit sequence number (credit) */
struct
ipath_swqe
*
s_wq
;
/* send work queue */
struct
ipath_swqe
*
s_wqe
;
struct
ipath_sge
*
r_ud_sg_list
;
struct
ipath_rq
r_rq
;
/* receive work queue */
struct
ipath_sge
r_sg_list
[
0
];
/* verified SGEs */
};
...
...
drivers/infiniband/hw/mlx4/cq.c
View file @
2a0d8366
...
...
@@ -222,7 +222,7 @@ struct ib_cq *mlx4_ib_create_cq(struct ib_device *ibdev, int entries, int vector
}
err
=
mlx4_cq_alloc
(
dev
->
dev
,
entries
,
&
cq
->
buf
.
mtt
,
uar
,
cq
->
db
.
dma
,
&
cq
->
mcq
,
0
);
cq
->
db
.
dma
,
&
cq
->
mcq
,
vector
,
0
);
if
(
err
)
goto
err_dbmap
;
...
...
@@ -325,15 +325,17 @@ static int mlx4_ib_get_outstanding_cqes(struct mlx4_ib_cq *cq)
static
void
mlx4_ib_cq_resize_copy_cqes
(
struct
mlx4_ib_cq
*
cq
)
{
struct
mlx4_cqe
*
cqe
;
struct
mlx4_cqe
*
cqe
,
*
new_cqe
;
int
i
;
i
=
cq
->
mcq
.
cons_index
;
cqe
=
get_cqe
(
cq
,
i
&
cq
->
ibcq
.
cqe
);
while
((
cqe
->
owner_sr_opcode
&
MLX4_CQE_OPCODE_MASK
)
!=
MLX4_CQE_OPCODE_RESIZE
)
{
memcpy
(
get_cqe_from_buf
(
&
cq
->
resize_buf
->
buf
,
(
i
+
1
)
&
cq
->
resize_buf
->
cqe
),
get_cqe
(
cq
,
i
&
cq
->
ibcq
.
cqe
),
sizeof
(
struct
mlx4_cqe
));
new_cqe
=
get_cqe_from_buf
(
&
cq
->
resize_buf
->
buf
,
(
i
+
1
)
&
cq
->
resize_buf
->
cqe
);
memcpy
(
new_cqe
,
get_cqe
(
cq
,
i
&
cq
->
ibcq
.
cqe
),
sizeof
(
struct
mlx4_cqe
));
new_cqe
->
owner_sr_opcode
=
(
cqe
->
owner_sr_opcode
&
~
MLX4_CQE_OWNER_MASK
)
|
(((
i
+
1
)
&
(
cq
->
resize_buf
->
cqe
+
1
))
?
MLX4_CQE_OWNER_MASK
:
0
);
cqe
=
get_cqe
(
cq
,
++
i
&
cq
->
ibcq
.
cqe
);
}
++
cq
->
mcq
.
cons_index
;
...
...
drivers/infiniband/hw/mlx4/main.c
View file @
2a0d8366
...
...
@@ -578,7 +578,7 @@ static void *mlx4_ib_add(struct mlx4_dev *dev)
mlx4_foreach_port
(
i
,
dev
,
MLX4_PORT_TYPE_IB
)
ibdev
->
num_ports
++
;
ibdev
->
ib_dev
.
phys_port_cnt
=
ibdev
->
num_ports
;
ibdev
->
ib_dev
.
num_comp_vectors
=
1
;
ibdev
->
ib_dev
.
num_comp_vectors
=
dev
->
caps
.
num_comp_vectors
;
ibdev
->
ib_dev
.
dma_device
=
&
dev
->
pdev
->
dev
;
ibdev
->
ib_dev
.
uverbs_abi_ver
=
MLX4_IB_UVERBS_ABI_VERSION
;
...
...
drivers/infiniband/hw/nes/nes.h
View file @
2a0d8366
...
...
@@ -137,14 +137,18 @@
#ifdef CONFIG_INFINIBAND_NES_DEBUG
#define nes_debug(level, fmt, args...) \
do { \
if (level & nes_debug_level) \
printk(KERN_ERR PFX "%s[%u]: " fmt, __func__, __LINE__, ##args)
printk(KERN_ERR PFX "%s[%u]: " fmt, __func__, __LINE__, ##args); \
} while (0)
#define assert(expr) \
if (!(expr)) { \
do { \
if (!(expr)) { \
printk(KERN_ERR PFX "Assertion failed! %s, %s, %s, line %d\n", \
#expr, __FILE__, __func__, __LINE__); \
}
} \
} while (0)
#define NES_EVENT_TIMEOUT 1200000
#else
...
...
drivers/infiniband/hw/nes/nes_cm.c
View file @
2a0d8366
This diff is collapsed.
Click to expand it.
drivers/infiniband/hw/nes/nes_cm.h
View file @
2a0d8366
...
...
@@ -76,6 +76,10 @@ enum nes_timer_type {
NES_TIMER_TYPE_CLOSE
,
};
#define NES_PASSIVE_STATE_INDICATED 0
#define NES_DO_NOT_SEND_RESET_EVENT 1
#define NES_SEND_RESET_EVENT 2
#define MAX_NES_IFS 4
#define SET_ACK 1
...
...
@@ -161,6 +165,8 @@ struct nes_timer_entry {
#define NES_CM_DEF_SEQ2 0x18ed5740
#define NES_CM_DEF_LOCAL_ID2 0xb807
#define MAX_CM_BUFFER 512
typedef
u32
nes_addr_t
;
...
...
@@ -254,8 +260,6 @@ struct nes_cm_listener {
/* per connection node and node state information */
struct
nes_cm_node
{
u32
hashkey
;
nes_addr_t
loc_addr
,
rem_addr
;
u16
loc_port
,
rem_port
;
...
...
@@ -292,7 +296,10 @@ struct nes_cm_node {
int
apbvt_set
;
int
accept_pend
;
int
freed
;
struct
list_head
timer_entry
;
struct
list_head
reset_entry
;
struct
nes_qp
*
nesqp
;
atomic_t
passive_state
;
};
/* structure for client or CM to fill when making CM api calls. */
...
...
@@ -350,7 +357,6 @@ struct nes_cm_core {
u32
mtu
;
u32
free_tx_pkt_max
;
u32
rx_pkt_posted
;
struct
sk_buff_head
tx_free_list
;
atomic_t
ht_node_cnt
;
struct
list_head
connected_nodes
;
/* struct list_head hashtable[NES_CM_HASHTABLE_SIZE]; */
...
...
@@ -390,7 +396,7 @@ struct nes_cm_ops {
struct
nes_cm_node
*
);
int
(
*
reject
)(
struct
nes_cm_core
*
,
struct
ietf_mpa_frame
*
,
struct
nes_cm_node
*
);
void
(
*
recv_pkt
)(
struct
nes_cm_core
*
,
struct
nes_vnic
*
,
int
(
*
recv_pkt
)(
struct
nes_cm_core
*
,
struct
nes_vnic
*
,
struct
sk_buff
*
);
int
(
*
destroy_cm_core
)(
struct
nes_cm_core
*
);
int
(
*
get
)(
struct
nes_cm_core
*
);
...
...
drivers/infiniband/hw/nes/nes_hw.c
View file @
2a0d8366
...
...
@@ -2700,9 +2700,15 @@ void nes_nic_ce_handler(struct nes_device *nesdev, struct nes_hw_nic_cq *cq)
pkt_type, (pkt_type & NES_PKT_TYPE_APBVT_MASK)); */
if
((
pkt_type
&
NES_PKT_TYPE_APBVT_MASK
)
==
NES_PKT_TYPE_APBVT_BITS
)
{
nes_cm_recv
(
rx_skb
,
nesvnic
->
netdev
);
}
else
{
if
((
cqe_misc
&
NES_NIC_CQE_TAG_VALID
)
&&
(
nesvnic
->
vlan_grp
!=
NULL
))
{
if
(
nes_cm_recv
(
rx_skb
,
nesvnic
->
netdev
))
rx_skb
=
NULL
;
}
if
(
rx_skb
==
NULL
)
goto
skip_rx_indicate0
;
if
((
cqe_misc
&
NES_NIC_CQE_TAG_VALID
)
&&
(
nesvnic
->
vlan_grp
!=
NULL
))
{
vlan_tag
=
(
u16
)(
le32_to_cpu
(
cq
->
cq_vbase
[
head
].
cqe_words
[
NES_NIC_CQE_TAG_PKT_TYPE_IDX
])
>>
16
);
...
...
@@ -2719,8 +2725,8 @@ void nes_nic_ce_handler(struct nes_device *nesdev, struct nes_hw_nic_cq *cq)
else
nes_netif_rx
(
rx_skb
);
}
}
skip_rx_indicate0:
nesvnic
->
netdev
->
last_rx
=
jiffies
;
/* nesvnic->netstats.rx_packets++; */
/* nesvnic->netstats.rx_bytes += rx_pkt_size; */
...
...
drivers/infiniband/hw/nes/nes_utils.c
View file @
2a0d8366
...
...
@@ -540,11 +540,14 @@ struct nes_cqp_request *nes_get_cqp_request(struct nes_device *nesdev)
if
(
!
list_empty
(
&
nesdev
->
cqp_avail_reqs
))
{
spin_lock_irqsave
(
&
nesdev
->
cqp
.
lock
,
flags
);
if
(
!
list_empty
(
&
nesdev
->
cqp_avail_reqs
))
{
cqp_request
=
list_entry
(
nesdev
->
cqp_avail_reqs
.
next
,
struct
nes_cqp_request
,
list
);
list_del_init
(
&
cqp_request
->
list
);
}
spin_unlock_irqrestore
(
&
nesdev
->
cqp
.
lock
,
flags
);
}
else
{
}
if
(
cqp_request
==
NULL
)
{
cqp_request
=
kzalloc
(
sizeof
(
struct
nes_cqp_request
),
GFP_KERNEL
);
if
(
cqp_request
)
{
cqp_request
->
dynamic
=
1
;
...
...
drivers/infiniband/hw/nes/nes_verbs.c
View file @
2a0d8366
...
...
@@ -1695,13 +1695,8 @@ static struct ib_cq *nes_create_cq(struct ib_device *ibdev, int entries,
/* use 4k pbl */
nes_debug
(
NES_DBG_CQ
,
"pbl_entries=%u, use a 4k PBL
\n
"
,
pbl_entries
);
if
(
nesadapter
->
free_4kpbl
==
0
)
{
if
(
cqp_request
->
dynamic
)
{
spin_unlock_irqrestore
(
&
nesadapter
->
pbl_lock
,
flags
);
kfree
(
cqp_request
);
}
else
{
list_add_tail
(
&
cqp_request
->
list
,
&
nesdev
->
cqp_avail_reqs
);
spin_unlock_irqrestore
(
&
nesadapter
->
pbl_lock
,
flags
);
}
nes_free_cqp_request
(
nesdev
,
cqp_request
);
if
(
!
context
)
pci_free_consistent
(
nesdev
->
pcidev
,
nescq
->
cq_mem_size
,
mem
,
nescq
->
hw_cq
.
cq_pbase
);
...
...
@@ -1717,13 +1712,8 @@ static struct ib_cq *nes_create_cq(struct ib_device *ibdev, int entries,
/* use 256 byte pbl */
nes_debug
(
NES_DBG_CQ
,
"pbl_entries=%u, use a 256 byte PBL
\n
"
,
pbl_entries
);
if
(
nesadapter
->
free_256pbl
==
0
)
{
if
(
cqp_request
->
dynamic
)
{
spin_unlock_irqrestore
(
&
nesadapter
->
pbl_lock
,
flags
);
kfree
(
cqp_request
);
}
else
{
list_add_tail
(
&
cqp_request
->
list
,
&
nesdev
->
cqp_avail_reqs
);
spin_unlock_irqrestore
(
&
nesadapter
->
pbl_lock
,
flags
);
}
nes_free_cqp_request
(
nesdev
,
cqp_request
);
if
(
!
context
)
pci_free_consistent
(
nesdev
->
pcidev
,
nescq
->
cq_mem_size
,
mem
,
nescq
->
hw_cq
.
cq_pbase
);
...
...
@@ -1928,13 +1918,8 @@ static int nes_reg_mr(struct nes_device *nesdev, struct nes_pd *nespd,
/* Two level PBL */
if
((
pbl_count
+
1
)
>
nesadapter
->
free_4kpbl
)
{
nes_debug
(
NES_DBG_MR
,
"Out of 4KB Pbls for two level request.
\n
"
);
if
(
cqp_request
->
dynamic
)
{
spin_unlock_irqrestore
(
&
nesadapter
->
pbl_lock
,
flags
);
kfree
(
cqp_request
);
}
else
{
list_add_tail
(
&
cqp_request
->
list
,
&
nesdev
->
cqp_avail_reqs
);
spin_unlock_irqrestore
(
&
nesadapter
->
pbl_lock
,
flags
);
}
nes_free_cqp_request
(
nesdev
,
cqp_request
);
return
-
ENOMEM
;
}
else
{
nesadapter
->
free_4kpbl
-=
pbl_count
+
1
;
...
...
@@ -1942,13 +1927,8 @@ static int nes_reg_mr(struct nes_device *nesdev, struct nes_pd *nespd,
}
else
if
(
residual_page_count
>
32
)
{
if
(
pbl_count
>
nesadapter
->
free_4kpbl
)
{
nes_debug
(
NES_DBG_MR
,
"Out of 4KB Pbls.
\n
"
);
if
(
cqp_request
->
dynamic
)
{
spin_unlock_irqrestore
(
&
nesadapter
->
pbl_lock
,
flags
);
kfree
(
cqp_request
);
}
else
{
list_add_tail
(
&
cqp_request
->
list
,
&
nesdev
->
cqp_avail_reqs
);
spin_unlock_irqrestore
(
&
nesadapter
->
pbl_lock
,
flags
);
}
nes_free_cqp_request
(
nesdev
,
cqp_request
);
return
-
ENOMEM
;
}
else
{
nesadapter
->
free_4kpbl
-=
pbl_count
;
...
...
@@ -1956,13 +1936,8 @@ static int nes_reg_mr(struct nes_device *nesdev, struct nes_pd *nespd,
}
else
{
if
(
pbl_count
>
nesadapter
->
free_256pbl
)
{
nes_debug
(
NES_DBG_MR
,
"Out of 256B Pbls.
\n
"
);
if
(
cqp_request
->
dynamic
)
{
spin_unlock_irqrestore
(
&
nesadapter
->
pbl_lock
,
flags
);
kfree
(
cqp_request
);
}
else
{
list_add_tail
(
&
cqp_request
->
list
,
&
nesdev
->
cqp_avail_reqs
);
spin_unlock_irqrestore
(
&
nesadapter
->
pbl_lock
,
flags
);
}
nes_free_cqp_request
(
nesdev
,
cqp_request
);
return
-
ENOMEM
;
}
else
{
nesadapter
->
free_256pbl
-=
pbl_count
;
...
...
drivers/infiniband/ulp/iser/iscsi_iser.h
View file @
2a0d8366
...
...
@@ -252,6 +252,9 @@ struct iser_conn {
wait_queue_head_t
wait
;
/* waitq for conn/disconn */
atomic_t
post_recv_buf_count
;
/* posted rx count */
atomic_t
post_send_buf_count
;
/* posted tx count */
atomic_t
unexpected_pdu_count
;
/* count of received *
* unexpected pdus *
* not yet retired */
char
name
[
ISER_OBJECT_NAME_SIZE
];
struct
iser_page_vec
*
page_vec
;
/* represents SG to fmr maps*
* maps serialized as tx is*/
...
...
drivers/infiniband/ulp/iser/iser_initiator.c
View file @
2a0d8366
...
...
@@ -183,14 +183,8 @@ static int iser_post_receive_control(struct iscsi_conn *conn)
struct
iser_regd_buf
*
regd_data
;
struct
iser_dto
*
recv_dto
=
NULL
;
struct
iser_device
*
device
=
iser_conn
->
ib_conn
->
device
;
int
rx_data_size
,
err
=
0
;
rx_desc
=
kmem_cache_alloc
(
ig
.
desc_cache
,
GFP_NOIO
);
if
(
rx_desc
==
NULL
)
{
iser_err
(
"Failed to alloc desc for post recv
\n
"
);
return
-
ENOMEM
;
}
rx_desc
->
type
=
ISCSI_RX
;
int
rx_data_size
,
err
;
int
posts
,
outstanding_unexp_pdus
;
/* for the login sequence we must support rx of upto 8K; login is done
* after conn create/bind (connect) and conn stop/bind (reconnect),
...
...
@@ -201,9 +195,26 @@ static int iser_post_receive_control(struct iscsi_conn *conn)
else
/* FIXME till user space sets conn->max_recv_dlength correctly */
rx_data_size
=
128
;
outstanding_unexp_pdus
=
atomic_xchg
(
&
iser_conn
->
ib_conn
->
unexpected_pdu_count
,
0
);
/*
* in addition to the response buffer, replace those consumed by
* unexpected pdus.
*/
for
(
posts
=
0
;
posts
<
1
+
outstanding_unexp_pdus
;
posts
++
)
{
rx_desc
=
kmem_cache_alloc
(
ig
.
desc_cache
,
GFP_NOIO
);
if
(
rx_desc
==
NULL
)
{
iser_err
(
"Failed to alloc desc for post recv %d
\n
"
,
posts
);
err
=
-
ENOMEM
;
goto
post_rx_cache_alloc_failure
;
}
rx_desc
->
type
=
ISCSI_RX
;
rx_desc
->
data
=
kmalloc
(
rx_data_size
,
GFP_NOIO
);
if
(
rx_desc
->
data
==
NULL
)
{
iser_err
(
"Failed to alloc data buf for post recv
\n
"
);
iser_err
(
"Failed to alloc data buf for post recv %d
\n
"
,
posts
);
err
=
-
ENOMEM
;
goto
post_rx_kmalloc_failure
;
}
...
...
@@ -233,14 +244,31 @@ static int iser_post_receive_control(struct iscsi_conn *conn)
iser_dto_add_regd_buff
(
recv_dto
,
regd_data
,
0
,
0
);
err
=
iser_post_recv
(
rx_desc
);
if
(
!
err
)
if
(
err
)
{
iser_err
(
"Failed iser_post_recv for post %d
\n
"
,
posts
);
goto
post_rx_post_recv_failure
;
}
}
/* all posts successful */
return
0
;
/* iser_post_recv failed */
post_rx_post_recv_failure:
iser_dto_buffs_release
(
recv_dto
);
kfree
(
rx_desc
->
data
);
post_rx_kmalloc_failure:
kmem_cache_free
(
ig
.
desc_cache
,
rx_desc
);
post_rx_cache_alloc_failure:
if
(
posts
>
0
)
{
/*
* response buffer posted, but did not replace all unexpected
* pdu recv bufs. Ignore error, retry occurs next send
*/
outstanding_unexp_pdus
-=
(
posts
-
1
);
err
=
0
;
}
atomic_add
(
outstanding_unexp_pdus
,
&
iser_conn
->
ib_conn
->
unexpected_pdu_count
);
return
err
;
}
...
...
@@ -274,8 +302,10 @@ int iser_conn_set_full_featured_mode(struct iscsi_conn *conn)
struct
iscsi_iser_conn
*
iser_conn
=
conn
->
dd_data
;
int
i
;
/* no need to keep it in a var, we are after login so if this should
* be negotiated, by now the result should be available here */
/*
* FIXME this value should be declared to the target during login with
* the MaxOutstandingUnexpectedPDUs key when supported
*/
int
initial_post_recv_bufs_num
=
ISER_MAX_RX_MISC_PDUS
;
iser_dbg
(
"Initially post: %d
\n
"
,
initial_post_recv_bufs_num
);
...
...
@@ -478,6 +508,7 @@ int iser_send_control(struct iscsi_conn *conn,
int
err
=
0
;
struct
iser_regd_buf
*
regd_buf
;
struct
iser_device
*
device
;
unsigned
char
opcode
;
if
(
!
iser_conn_state_comp
(
iser_conn
->
ib_conn
,
ISER_CONN_UP
))
{
iser_err
(
"Failed to send, conn: 0x%p is not up
\n
"
,
iser_conn
->
ib_conn
);
...
...
@@ -512,11 +543,16 @@ int iser_send_control(struct iscsi_conn *conn,
data_seg_len
);
}
opcode
=
task
->
hdr
->
opcode
&
ISCSI_OPCODE_MASK
;
/* post recv buffer for response if one is expected */
if
(
!
(
opcode
==
ISCSI_OP_NOOP_OUT
&&
task
->
hdr
->
itt
==
RESERVED_ITT
))
{
if
(
iser_post_receive_control
(
conn
)
!=
0
)
{
iser_err
(
"post_rcv_buff failed!
\n
"
);
err
=
-
ENOMEM
;
goto
send_control_error
;
}
}
err
=
iser_post_send
(
mdesc
);
if
(
!
err
)
...
...
@@ -586,6 +622,20 @@ void iser_rcv_completion(struct iser_desc *rx_desc,
* parallel to the execution of iser_conn_term. So the code that waits *
* for the posted rx bufs refcount to become zero handles everything */
atomic_dec
(
&
conn
->
ib_conn
->
post_recv_buf_count
);
/*
* if an unexpected PDU was received then the recv wr consumed must
* be replaced, this is done in the next send of a control-type PDU
*/
if
(
opcode
==
ISCSI_OP_NOOP_IN
&&
hdr
->
itt
==
RESERVED_ITT
)
{
/* nop-in with itt = 0xffffffff */
atomic_inc
(
&
conn
->
ib_conn
->
unexpected_pdu_count
);
}
else
if
(
opcode
==
ISCSI_OP_ASYNC_EVENT
)
{
/* asyncronous message */
atomic_inc
(
&
conn
->
ib_conn
->
unexpected_pdu_count
);
}
/* a reject PDU consumes the recv buf posted for the response */
}
void
iser_snd_completion
(
struct
iser_desc
*
tx_desc
)
...
...
drivers/infiniband/ulp/iser/iser_verbs.c
View file @
2a0d8366
...
...
@@ -498,6 +498,7 @@ void iser_conn_init(struct iser_conn *ib_conn)
init_waitqueue_head
(
&
ib_conn
->
wait
);
atomic_set
(
&
ib_conn
->
post_recv_buf_count
,
0
);
atomic_set
(
&
ib_conn
->
post_send_buf_count
,
0
);
atomic_set
(
&
ib_conn
->
unexpected_pdu_count
,
0
);
atomic_set
(
&
ib_conn
->
refcount
,
1
);
INIT_LIST_HEAD
(
&
ib_conn
->
conn_list
);
spin_lock_init
(
&
ib_conn
->
lock
);
...
...
drivers/net/mlx4/cq.c
View file @
2a0d8366
...
...
@@ -189,7 +189,7 @@ EXPORT_SYMBOL_GPL(mlx4_cq_resize);
int
mlx4_cq_alloc
(
struct
mlx4_dev
*
dev
,
int
nent
,
struct
mlx4_mtt
*
mtt
,
struct
mlx4_uar
*
uar
,
u64
db_rec
,
struct
mlx4_cq
*
cq
,
int
collapsed
)
unsigned
vector
,
int
collapsed
)
{
struct
mlx4_priv
*
priv
=
mlx4_priv
(
dev
);
struct
mlx4_cq_table
*
cq_table
=
&
priv
->
cq_table
;
...
...
@@ -198,6 +198,11 @@ int mlx4_cq_alloc(struct mlx4_dev *dev, int nent, struct mlx4_mtt *mtt,
u64
mtt_addr
;
int
err
;
if
(
vector
>=
dev
->
caps
.
num_comp_vectors
)
return
-
EINVAL
;
cq
->
vector
=
vector
;
cq
->
cqn
=
mlx4_bitmap_alloc
(
&
cq_table
->
bitmap
);
if
(
cq
->
cqn
==
-
1
)
return
-
ENOMEM
;
...
...
@@ -227,7 +232,7 @@ int mlx4_cq_alloc(struct mlx4_dev *dev, int nent, struct mlx4_mtt *mtt,
cq_context
->
flags
=
cpu_to_be32
(
!!
collapsed
<<
18
);
cq_context
->
logsize_usrpage
=
cpu_to_be32
((
ilog2
(
nent
)
<<
24
)
|
uar
->
index
);
cq_context
->
comp_eqn
=
priv
->
eq_table
.
eq
[
MLX4_EQ_COMP
].
eqn
;
cq_context
->
comp_eqn
=
priv
->
eq_table
.
eq
[
vector
].
eqn
;
cq_context
->
log_page_size
=
mtt
->
page_shift
-
MLX4_ICM_PAGE_SHIFT
;
mtt_addr
=
mlx4_mtt_addr
(
dev
,
mtt
);
...
...
@@ -276,7 +281,7 @@ void mlx4_cq_free(struct mlx4_dev *dev, struct mlx4_cq *cq)
if
(
err
)
mlx4_warn
(
dev
,
"HW2SW_CQ failed (%d) for CQN %06x
\n
"
,
err
,
cq
->
cqn
);
synchronize_irq
(
priv
->
eq_table
.
eq
[
MLX4_EQ_COMP
].
irq
);
synchronize_irq
(
priv
->
eq_table
.
eq
[
cq
->
vector
].
irq
);
spin_lock_irq
(
&
cq_table
->
lock
);
radix_tree_delete
(
&
cq_table
->
tree
,
cq
->
cqn
);
...
...
drivers/net/mlx4/en_cq.c
View file @
2a0d8366
...
...
@@ -51,10 +51,13 @@ int mlx4_en_create_cq(struct mlx4_en_priv *priv,
int
err
;
cq
->
size
=
entries
;
if
(
mode
==
RX
)
if
(
mode
==
RX
)
{
cq
->
buf_size
=
cq
->
size
*
sizeof
(
struct
mlx4_cqe
);
else
cq
->
vector
=
ring
%
mdev
->
dev
->
caps
.
num_comp_vectors
;
}
else
{
cq
->
buf_size
=
sizeof
(
struct
mlx4_cqe
);
cq
->
vector
=
0
;
}
cq
->
ring
=
ring
;
cq
->
is_tx
=
mode
;
...
...
@@ -86,7 +89,7 @@ int mlx4_en_activate_cq(struct mlx4_en_priv *priv, struct mlx4_en_cq *cq)
memset
(
cq
->
buf
,
0
,
cq
->
buf_size
);
err
=
mlx4_cq_alloc
(
mdev
->
dev
,
cq
->
size
,
&
cq
->
wqres
.
mtt
,
&
mdev
->
priv_uar
,
cq
->
wqres
.
db
.
dma
,
&
cq
->
mcq
,
cq
->
is_tx
);
cq
->
wqres
.
db
.
dma
,
&
cq
->
mcq
,
cq
->
vector
,
cq
->
is_tx
);
if
(
err
)
return
err
;
...
...
drivers/net/mlx4/en_main.c
View file @
2a0d8366
...
...
@@ -170,9 +170,9 @@ static void *mlx4_en_add(struct mlx4_dev *dev)
mlx4_info
(
mdev
,
"Using %d tx rings for port:%d
\n
"
,
mdev
->
profile
.
prof
[
i
].
tx_ring_num
,
i
);
if
(
!
mdev
->
profile
.
prof
[
i
].
rx_ring_num
)
{
mdev
->
profile
.
prof
[
i
].
rx_ring_num
=
1
;
mdev
->
profile
.
prof
[
i
].
rx_ring_num
=
dev
->
caps
.
num_comp_vectors
;
mlx4_info
(
mdev
,
"Defaulting to %d rx rings for port:%d
\n
"
,
1
,
i
);
mdev
->
profile
.
prof
[
i
].
rx_ring_num
,
i
);
}
else
mlx4_info
(
mdev
,
"Using %d rx rings for port:%d
\n
"
,
mdev
->
profile
.
prof
[
i
].
rx_ring_num
,
i
);
...
...
drivers/net/mlx4/eq.c
View file @
2a0d8366
...
...
@@ -243,10 +243,6 @@ static int mlx4_eq_int(struct mlx4_dev *dev, struct mlx4_eq *eq)
* least that often.
*/
if
(
unlikely
(
set_ci
>=
MLX4_NUM_SPARE_EQE
))
{
/*
* Conditional on hca_type is OK here because
* this is a rare case, not the fast path.
*/
eq_set_ci
(
eq
,
0
);
set_ci
=
0
;
}
...
...
@@ -266,7 +262,7 @@ static irqreturn_t mlx4_interrupt(int irq, void *dev_ptr)
writel
(
priv
->
eq_table
.
clr_mask
,
priv
->
eq_table
.
clr_int
);
for
(
i
=
0
;
i
<
MLX4_NUM_EQ
;
++
i
)
for
(
i
=
0
;
i
<
dev
->
caps
.
num_comp_vectors
+
1
;
++
i
)
work
|=
mlx4_eq_int
(
dev
,
&
priv
->
eq_table
.
eq
[
i
]);
return
IRQ_RETVAL
(
work
);
...
...
@@ -304,6 +300,17 @@ static int mlx4_HW2SW_EQ(struct mlx4_dev *dev, struct mlx4_cmd_mailbox *mailbox,
MLX4_CMD_TIME_CLASS_A
);
}
static
int
mlx4_num_eq_uar
(
struct
mlx4_dev
*
dev
)
{
/*
* Each UAR holds 4 EQ doorbells. To figure out how many UARs
* we need to map, take the difference of highest index and
* the lowest index we'll use and add 1.
*/
return
(
dev
->
caps
.
num_comp_vectors
+
1
+
dev
->
caps
.
reserved_eqs
)
/
4
-
dev
->
caps
.
reserved_eqs
/
4
+
1
;
}
static
void
__iomem
*
mlx4_get_eq_uar
(
struct
mlx4_dev
*
dev
,
struct
mlx4_eq
*
eq
)
{
struct
mlx4_priv
*
priv
=
mlx4_priv
(
dev
);
...
...
@@ -483,9 +490,11 @@ static void mlx4_free_irqs(struct mlx4_dev *dev)
if
(
eq_table
->
have_irq
)
free_irq
(
dev
->
pdev
->
irq
,
dev
);
for
(
i
=
0
;
i
<
MLX4_NUM_EQ
;
++
i
)
for
(
i
=
0
;
i
<
dev
->
caps
.
num_comp_vectors
+
1
;
++
i
)
if
(
eq_table
->
eq
[
i
].
have_irq
)
free_irq
(
eq_table
->
eq
[
i
].
irq
,
eq_table
->
eq
+
i
);
kfree
(
eq_table
->
irq_names
);
}
static
int
mlx4_map_clr_int
(
struct
mlx4_dev
*
dev
)
...
...
@@ -551,57 +560,93 @@ void mlx4_unmap_eq_icm(struct mlx4_dev *dev)
__free_page
(
priv
->
eq_table
.
icm_page
);
}
int
mlx4_alloc_eq_table
(
struct
mlx4_dev
*
dev
)
{
struct
mlx4_priv
*
priv
=
mlx4_priv
(
dev
);
priv
->
eq_table
.
eq
=
kcalloc
(
dev
->
caps
.
num_eqs
-
dev
->
caps
.
reserved_eqs
,
sizeof
*
priv
->
eq_table
.
eq
,
GFP_KERNEL
);
if
(
!
priv
->
eq_table
.
eq
)
return
-
ENOMEM
;
return
0
;
}
void
mlx4_free_eq_table
(
struct
mlx4_dev
*
dev
)
{
kfree
(
mlx4_priv
(
dev
)
->
eq_table
.
eq
);
}
int
mlx4_init_eq_table
(
struct
mlx4_dev
*
dev
)
{
struct
mlx4_priv
*
priv
=
mlx4_priv
(
dev
);
int
err
;
int
i
;
priv
->
eq_table
.
uar_map
=
kcalloc
(
sizeof
*
priv
->
eq_table
.
uar_map
,
mlx4_num_eq_uar
(
dev
),
GFP_KERNEL
);
if
(
!
priv
->
eq_table
.
uar_map
)
{
err
=
-
ENOMEM
;
goto
err_out_free
;
}
err
=
mlx4_bitmap_init
(
&
priv
->
eq_table
.
bitmap
,
dev
->
caps
.
num_eqs
,
dev
->
caps
.
num_eqs
-
1
,
dev
->
caps
.
reserved_eqs
,
0
);
if
(
err
)
return
err
;
goto
err_out_free
;
for
(
i
=
0
;
i
<
ARRAY_SIZE
(
priv
->
eq_table
.
uar_map
);
++
i
)
for
(
i
=
0
;
i
<
mlx4_num_eq_uar
(
dev
);
++
i
)
priv
->
eq_table
.
uar_map
[
i
]
=
NULL
;
err
=
mlx4_map_clr_int
(
dev
);
if
(
err
)
goto
err_out_
free
;
goto
err_out_
bitmap
;
priv
->
eq_table
.
clr_mask
=
swab32
(
1
<<
(
priv
->
eq_table
.
inta_pin
&
31
));
priv
->
eq_table
.
clr_int
=
priv
->
clr_base
+
(
priv
->
eq_table
.
inta_pin
<
32
?
4
:
0
);
priv
->
eq_table
.
irq_names
=
kmalloc
(
16
*
dev
->
caps
.
num_comp_vectors
,
GFP_KERNEL
);
if
(
!
priv
->
eq_table
.
irq_names
)
{
err
=
-
ENOMEM
;
goto
err_out_bitmap
;
}
for
(
i
=
0
;
i
<
dev
->
caps
.
num_comp_vectors
;
++
i
)
{
err
=
mlx4_create_eq
(
dev
,
dev
->
caps
.
num_cqs
+
MLX4_NUM_SPARE_EQE
,
(
dev
->
flags
&
MLX4_FLAG_MSI_X
)
?
MLX4_EQ_COMP
:
0
,
&
priv
->
eq_table
.
eq
[
MLX4_EQ_COMP
]);
(
dev
->
flags
&
MLX4_FLAG_MSI_X
)
?
i
:
0
,
&
priv
->
eq_table
.
eq
[
i
]);
if
(
err
)
goto
err_out_unmap
;
}
err
=
mlx4_create_eq
(
dev
,
MLX4_NUM_ASYNC_EQE
+
MLX4_NUM_SPARE_EQE
,
(
dev
->
flags
&
MLX4_FLAG_MSI_X
)
?
MLX4_EQ_ASYNC
:
0
,
&
priv
->
eq_table
.
eq
[
MLX4_EQ_ASYNC
]);
(
dev
->
flags
&
MLX4_FLAG_MSI_X
)
?
dev
->
caps
.
num_comp_vectors
:
0
,
&
priv
->
eq_table
.
eq
[
dev
->
caps
.
num_comp_vectors
]);
if
(
err
)
goto
err_out_comp
;
if
(
dev
->
flags
&
MLX4_FLAG_MSI_X
)
{
static
const
char
*
eq_name
[]
=
{
[
MLX4_EQ_COMP
]
=
DRV_NAME
" (comp)"
,
[
MLX4_EQ_ASYNC
]
=
DRV_NAME
" (async)"
};
static
const
char
async_eq_name
[]
=
"mlx4-async"
;
const
char
*
eq_name
;
for
(
i
=
0
;
i
<
dev
->
caps
.
num_comp_vectors
+
1
;
++
i
)
{
if
(
i
<
dev
->
caps
.
num_comp_vectors
)
{
snprintf
(
priv
->
eq_table
.
irq_names
+
i
*
16
,
16
,
"mlx4-comp-%d"
,
i
);
eq_name
=
priv
->
eq_table
.
irq_names
+
i
*
16
;
}
else
eq_name
=
async_eq_name
;
for
(
i
=
0
;
i
<
MLX4_NUM_EQ
;
++
i
)
{
err
=
request_irq
(
priv
->
eq_table
.
eq
[
i
].
irq
,
mlx4_msi_x_interrupt
,
0
,
eq_name
[
i
],
priv
->
eq_table
.
eq
+
i
);
mlx4_msi_x_interrupt
,
0
,
eq_name
,
priv
->
eq_table
.
eq
+
i
);
if
(
err
)
goto
err_out_async
;
priv
->
eq_table
.
eq
[
i
].
have_irq
=
1
;
}
}
else
{
err
=
request_irq
(
dev
->
pdev
->
irq
,
mlx4_interrupt
,
IRQF_SHARED
,
DRV_NAME
,
dev
);
...
...
@@ -612,28 +657,36 @@ int mlx4_init_eq_table(struct mlx4_dev *dev)
}
err
=
mlx4_MAP_EQ
(
dev
,
MLX4_ASYNC_EVENT_MASK
,
0
,
priv
->
eq_table
.
eq
[
MLX4_EQ_ASYNC
].
eqn
);
priv
->
eq_table
.
eq
[
dev
->
caps
.
num_comp_vectors
].
eqn
);
if
(
err
)
mlx4_warn
(
dev
,
"MAP_EQ for async EQ %d failed (%d)
\n
"
,
priv
->
eq_table
.
eq
[
MLX4_EQ_ASYNC
].
eqn
,
err
);
priv
->
eq_table
.
eq
[
dev
->
caps
.
num_comp_vectors
].
eqn
,
err
);
for
(
i
=
0
;
i
<
MLX4_NUM_EQ
;
++
i
)
for
(
i
=
0
;
i
<
dev
->
caps
.
num_comp_vectors
+
1
;
++
i
)
eq_set_ci
(
&
priv
->
eq_table
.
eq
[
i
],
1
);
return
0
;
err_out_async:
mlx4_free_eq
(
dev
,
&
priv
->
eq_table
.
eq
[
MLX4_EQ_ASYNC
]);
mlx4_free_eq
(
dev
,
&
priv
->
eq_table
.
eq
[
dev
->
caps
.
num_comp_vectors
]);
err_out_comp:
mlx4_free_eq
(
dev
,
&
priv
->
eq_table
.
eq
[
MLX4_EQ_COMP
])
;
i
=
dev
->
caps
.
num_comp_vectors
-
1
;
err_out_unmap:
while
(
i
>=
0
)
{
mlx4_free_eq
(
dev
,
&
priv
->
eq_table
.
eq
[
i
]);
--
i
;
}
mlx4_unmap_clr_int
(
dev
);
mlx4_free_irqs
(
dev
);
err_out_
free
:
err_out_
bitmap
:
mlx4_bitmap_cleanup
(
&
priv
->
eq_table
.
bitmap
);
err_out_free:
kfree
(
priv
->
eq_table
.
uar_map
);
return
err
;
}
...
...
@@ -643,18 +696,20 @@ void mlx4_cleanup_eq_table(struct mlx4_dev *dev)
int
i
;
mlx4_MAP_EQ
(
dev
,
MLX4_ASYNC_EVENT_MASK
,
1
,
priv
->
eq_table
.
eq
[
MLX4_EQ_ASYNC
].
eqn
);
priv
->
eq_table
.
eq
[
dev
->
caps
.
num_comp_vectors
].
eqn
);
mlx4_free_irqs
(
dev
);
for
(
i
=
0
;
i
<
MLX4_NUM_EQ
;
++
i
)
for
(
i
=
0
;
i
<
dev
->
caps
.
num_comp_vectors
+
1
;
++
i
)
mlx4_free_eq
(
dev
,
&
priv
->
eq_table
.
eq
[
i
]);
mlx4_unmap_clr_int
(
dev
);
for
(
i
=
0
;
i
<
ARRAY_SIZE
(
priv
->
eq_table
.
uar_map
);
++
i
)
for
(
i
=
0
;
i
<
mlx4_num_eq_uar
(
dev
);
++
i
)
if
(
priv
->
eq_table
.
uar_map
[
i
])
iounmap
(
priv
->
eq_table
.
uar_map
[
i
]);
mlx4_bitmap_cleanup
(
&
priv
->
eq_table
.
bitmap
);
kfree
(
priv
->
eq_table
.
uar_map
);
}
drivers/net/mlx4/main.c
View file @
2a0d8366
...
...
@@ -421,9 +421,7 @@ static int mlx4_init_cmpt_table(struct mlx4_dev *dev, u64 cmpt_base,
((
u64
)
(
MLX4_CMPT_TYPE_EQ
*
cmpt_entry_sz
)
<<
MLX4_CMPT_SHIFT
),
cmpt_entry_sz
,
roundup_pow_of_two
(
MLX4_NUM_EQ
+
dev
->
caps
.
reserved_eqs
),
MLX4_NUM_EQ
+
dev
->
caps
.
reserved_eqs
,
0
,
0
);
dev
->
caps
.
num_eqs
,
dev
->
caps
.
num_eqs
,
0
,
0
);
if
(
err
)
goto
err_cq
;
...
...
@@ -810,12 +808,12 @@ static int mlx4_setup_hca(struct mlx4_dev *dev)
if
(
dev
->
flags
&
MLX4_FLAG_MSI_X
)
{
mlx4_warn
(
dev
,
"NOP command failed to generate MSI-X "
"interrupt IRQ %d).
\n
"
,
priv
->
eq_table
.
eq
[
MLX4_EQ_ASYNC
].
irq
);
priv
->
eq_table
.
eq
[
dev
->
caps
.
num_comp_vectors
].
irq
);
mlx4_warn
(
dev
,
"Trying again without MSI-X.
\n
"
);
}
else
{
mlx4_err
(
dev
,
"NOP command failed to generate interrupt "
"(IRQ %d), aborting.
\n
"
,
priv
->
eq_table
.
eq
[
MLX4_EQ_ASYNC
].
irq
);
priv
->
eq_table
.
eq
[
dev
->
caps
.
num_comp_vectors
].
irq
);
mlx4_err
(
dev
,
"BIOS or ACPI interrupt routing problem?
\n
"
);
}
...
...
@@ -908,31 +906,50 @@ static int mlx4_setup_hca(struct mlx4_dev *dev)
static
void
mlx4_enable_msi_x
(
struct
mlx4_dev
*
dev
)
{
struct
mlx4_priv
*
priv
=
mlx4_priv
(
dev
);
struct
msix_entry
entries
[
MLX4_NUM_EQ
];
struct
msix_entry
*
entries
;
int
nreq
;
int
err
;
int
i
;
if
(
msi_x
)
{
for
(
i
=
0
;
i
<
MLX4_NUM_EQ
;
++
i
)
nreq
=
min
(
dev
->
caps
.
num_eqs
-
dev
->
caps
.
reserved_eqs
,
num_possible_cpus
()
+
1
);
entries
=
kcalloc
(
nreq
,
sizeof
*
entries
,
GFP_KERNEL
);
if
(
!
entries
)
goto
no_msi
;
for
(
i
=
0
;
i
<
nreq
;
++
i
)
entries
[
i
].
entry
=
i
;
err
=
pci_enable_msix
(
dev
->
pdev
,
entries
,
ARRAY_SIZE
(
entries
));
retry:
err
=
pci_enable_msix
(
dev
->
pdev
,
entries
,
nreq
);
if
(
err
)
{
if
(
err
>
0
)
mlx4_info
(
dev
,
"Only %d MSI-X vectors available, "
"not using MSI-X
\n
"
,
err
);
/* Try again if at least 2 vectors are available */
if
(
err
>
1
)
{
mlx4_info
(
dev
,
"Requested %d vectors, "
"but only %d MSI-X vectors available, "
"trying again
\n
"
,
nreq
,
err
);
nreq
=
err
;
goto
retry
;
}
goto
no_msi
;
}
for
(
i
=
0
;
i
<
MLX4_NUM_EQ
;
++
i
)
dev
->
caps
.
num_comp_vectors
=
nreq
-
1
;
for
(
i
=
0
;
i
<
nreq
;
++
i
)
priv
->
eq_table
.
eq
[
i
].
irq
=
entries
[
i
].
vector
;
dev
->
flags
|=
MLX4_FLAG_MSI_X
;
kfree
(
entries
);
return
;
}
no_msi:
for
(
i
=
0
;
i
<
MLX4_NUM_EQ
;
++
i
)
dev
->
caps
.
num_comp_vectors
=
1
;
for
(
i
=
0
;
i
<
2
;
++
i
)
priv
->
eq_table
.
eq
[
i
].
irq
=
dev
->
pdev
->
irq
;
}
...
...
@@ -1074,6 +1091,10 @@ static int __mlx4_init_one(struct pci_dev *pdev, const struct pci_device_id *id)
if
(
err
)
goto
err_cmd
;
err
=
mlx4_alloc_eq_table
(
dev
);
if
(
err
)
goto
err_close
;
mlx4_enable_msi_x
(
dev
);
err
=
mlx4_setup_hca
(
dev
);
...
...
@@ -1084,7 +1105,7 @@ static int __mlx4_init_one(struct pci_dev *pdev, const struct pci_device_id *id)
}
if
(
err
)
goto
err_
close
;
goto
err_
free_eq
;
for
(
port
=
1
;
port
<=
dev
->
caps
.
num_ports
;
port
++
)
{
err
=
mlx4_init_port_info
(
dev
,
port
);
...
...
@@ -1114,6 +1135,9 @@ static int __mlx4_init_one(struct pci_dev *pdev, const struct pci_device_id *id)
mlx4_cleanup_pd_table
(
dev
);
mlx4_cleanup_uar_table
(
dev
);
err_free_eq:
mlx4_free_eq_table
(
dev
);
err_close:
if
(
dev
->
flags
&
MLX4_FLAG_MSI_X
)
pci_disable_msix
(
pdev
);
...
...
@@ -1177,6 +1201,7 @@ static void mlx4_remove_one(struct pci_dev *pdev)
iounmap
(
priv
->
kar
);
mlx4_uar_free
(
dev
,
&
priv
->
driver_uar
);
mlx4_cleanup_uar_table
(
dev
);
mlx4_free_eq_table
(
dev
);
mlx4_close_hca
(
dev
);
mlx4_cmd_cleanup
(
dev
);
...
...
drivers/net/mlx4/mlx4.h
View file @
2a0d8366
...
...
@@ -62,12 +62,6 @@ enum {
MLX4_MTT_ENTRY_PER_SEG
=
8
};
enum
{
MLX4_EQ_ASYNC
,
MLX4_EQ_COMP
,
MLX4_NUM_EQ
};
enum
{
MLX4_NUM_PDS
=
1
<<
15
};
...
...
@@ -205,10 +199,11 @@ struct mlx4_cq_table {
struct
mlx4_eq_table
{
struct
mlx4_bitmap
bitmap
;
char
*
irq_names
;
void
__iomem
*
clr_int
;
void
__iomem
*
uar_map
[(
MLX4_NUM_EQ
+
6
)
/
4
]
;
void
__iomem
**
uar_map
;
u32
clr_mask
;
struct
mlx4_eq
eq
[
MLX4_NUM_EQ
]
;
struct
mlx4_eq
*
eq
;
u64
icm_virt
;
struct
page
*
icm_page
;
dma_addr_t
icm_dma
;
...
...
@@ -328,6 +323,9 @@ void mlx4_bitmap_cleanup(struct mlx4_bitmap *bitmap);
int
mlx4_reset
(
struct
mlx4_dev
*
dev
);
int
mlx4_alloc_eq_table
(
struct
mlx4_dev
*
dev
);
void
mlx4_free_eq_table
(
struct
mlx4_dev
*
dev
);
int
mlx4_init_pd_table
(
struct
mlx4_dev
*
dev
);
int
mlx4_init_uar_table
(
struct
mlx4_dev
*
dev
);
int
mlx4_init_mr_table
(
struct
mlx4_dev
*
dev
);
...
...
drivers/net/mlx4/profile.c
View file @
2a0d8366
...
...
@@ -107,7 +107,9 @@ u64 mlx4_make_profile(struct mlx4_dev *dev,
profile
[
MLX4_RES_AUXC
].
num
=
request
->
num_qp
;
profile
[
MLX4_RES_SRQ
].
num
=
request
->
num_srq
;
profile
[
MLX4_RES_CQ
].
num
=
request
->
num_cq
;
profile
[
MLX4_RES_EQ
].
num
=
MLX4_NUM_EQ
+
dev_cap
->
reserved_eqs
;
profile
[
MLX4_RES_EQ
].
num
=
min
(
dev_cap
->
max_eqs
,
dev_cap
->
reserved_eqs
+
num_possible_cpus
()
+
1
);
profile
[
MLX4_RES_DMPT
].
num
=
request
->
num_mpt
;
profile
[
MLX4_RES_CMPT
].
num
=
MLX4_NUM_CMPTS
;
profile
[
MLX4_RES_MTT
].
num
=
request
->
num_mtt
;
...
...
include/linux/mlx4/device.h
View file @
2a0d8366
...
...
@@ -206,6 +206,7 @@ struct mlx4_caps {
int
reserved_cqs
;
int
num_eqs
;
int
reserved_eqs
;
int
num_comp_vectors
;
int
num_mpts
;
int
num_mtt_segs
;
int
fmr_reserved_mtts
;
...
...
@@ -328,6 +329,7 @@ struct mlx4_cq {
int
arm_sn
;
int
cqn
;
unsigned
vector
;
atomic_t
refcount
;
struct
completion
free
;
...
...
@@ -437,7 +439,7 @@ void mlx4_free_hwq_res(struct mlx4_dev *mdev, struct mlx4_hwq_resources *wqres,
int
mlx4_cq_alloc
(
struct
mlx4_dev
*
dev
,
int
nent
,
struct
mlx4_mtt
*
mtt
,
struct
mlx4_uar
*
uar
,
u64
db_rec
,
struct
mlx4_cq
*
cq
,
int
collapsed
);
unsigned
vector
,
int
collapsed
);
void
mlx4_cq_free
(
struct
mlx4_dev
*
dev
,
struct
mlx4_cq
*
cq
);
int
mlx4_qp_reserve_range
(
struct
mlx4_dev
*
dev
,
int
cnt
,
int
align
,
int
*
base
);
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment