Commit f221dcd9 authored by David S. Miller's avatar David S. Miller

Merge branch 'net-smc-next'

Ursula Braun says:

====================
net/smc: patches for net-next

here are some patches for net/smc. Most important are
improvements for socket closing.
====================
Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
parents c02b7a91 2c9c1682
...@@ -147,7 +147,6 @@ static int smc_release(struct socket *sock) ...@@ -147,7 +147,6 @@ static int smc_release(struct socket *sock)
schedule_delayed_work(&smc->sock_put_work, schedule_delayed_work(&smc->sock_put_work,
SMC_CLOSE_SOCK_PUT_DELAY); SMC_CLOSE_SOCK_PUT_DELAY);
} }
sk->sk_prot->unhash(sk);
release_sock(sk); release_sock(sk);
sock_put(sk); sock_put(sk);
...@@ -451,6 +450,9 @@ static int smc_connect_rdma(struct smc_sock *smc) ...@@ -451,6 +450,9 @@ static int smc_connect_rdma(struct smc_sock *smc)
goto decline_rdma_unlock; goto decline_rdma_unlock;
} }
smc_close_init(smc);
smc_rx_init(smc);
if (local_contact == SMC_FIRST_CONTACT) { if (local_contact == SMC_FIRST_CONTACT) {
rc = smc_ib_ready_link(link); rc = smc_ib_ready_link(link);
if (rc) { if (rc) {
...@@ -477,7 +479,6 @@ static int smc_connect_rdma(struct smc_sock *smc) ...@@ -477,7 +479,6 @@ static int smc_connect_rdma(struct smc_sock *smc)
mutex_unlock(&smc_create_lgr_pending); mutex_unlock(&smc_create_lgr_pending);
smc_tx_init(smc); smc_tx_init(smc);
smc_rx_init(smc);
out_connected: out_connected:
smc_copy_sock_settings_to_clc(smc); smc_copy_sock_settings_to_clc(smc);
...@@ -637,7 +638,8 @@ struct sock *smc_accept_dequeue(struct sock *parent, ...@@ -637,7 +638,8 @@ struct sock *smc_accept_dequeue(struct sock *parent,
smc_accept_unlink(new_sk); smc_accept_unlink(new_sk);
if (new_sk->sk_state == SMC_CLOSED) { if (new_sk->sk_state == SMC_CLOSED) {
/* tbd in follow-on patch: close this sock */ new_sk->sk_prot->unhash(new_sk);
sock_put(new_sk);
continue; continue;
} }
if (new_sock) if (new_sock)
...@@ -657,8 +659,13 @@ void smc_close_non_accepted(struct sock *sk) ...@@ -657,8 +659,13 @@ void smc_close_non_accepted(struct sock *sk)
if (!sk->sk_lingertime) if (!sk->sk_lingertime)
/* wait for peer closing */ /* wait for peer closing */
sk->sk_lingertime = SMC_MAX_STREAM_WAIT_TIMEOUT; sk->sk_lingertime = SMC_MAX_STREAM_WAIT_TIMEOUT;
if (!smc->use_fallback) if (smc->use_fallback) {
sk->sk_state = SMC_CLOSED;
} else {
smc_close_active(smc); smc_close_active(smc);
sock_set_flag(sk, SOCK_DEAD);
sk->sk_shutdown |= SHUTDOWN_MASK;
}
if (smc->clcsock) { if (smc->clcsock) {
struct socket *tcp; struct socket *tcp;
...@@ -666,11 +673,9 @@ void smc_close_non_accepted(struct sock *sk) ...@@ -666,11 +673,9 @@ void smc_close_non_accepted(struct sock *sk)
smc->clcsock = NULL; smc->clcsock = NULL;
sock_release(tcp); sock_release(tcp);
} }
sock_set_flag(sk, SOCK_DEAD);
sk->sk_shutdown |= SHUTDOWN_MASK;
if (smc->use_fallback) { if (smc->use_fallback) {
schedule_delayed_work(&smc->sock_put_work, TCP_TIMEWAIT_LEN); schedule_delayed_work(&smc->sock_put_work, TCP_TIMEWAIT_LEN);
} else { } else if (sk->sk_state == SMC_CLOSED) {
smc_conn_free(&smc->conn); smc_conn_free(&smc->conn);
schedule_delayed_work(&smc->sock_put_work, schedule_delayed_work(&smc->sock_put_work,
SMC_CLOSE_SOCK_PUT_DELAY); SMC_CLOSE_SOCK_PUT_DELAY);
...@@ -800,6 +805,9 @@ static void smc_listen_work(struct work_struct *work) ...@@ -800,6 +805,9 @@ static void smc_listen_work(struct work_struct *work)
goto decline_rdma; goto decline_rdma;
} }
smc_close_init(new_smc);
smc_rx_init(new_smc);
rc = smc_clc_send_accept(new_smc, local_contact); rc = smc_clc_send_accept(new_smc, local_contact);
if (rc) if (rc)
goto out_err; goto out_err;
...@@ -839,7 +847,6 @@ static void smc_listen_work(struct work_struct *work) ...@@ -839,7 +847,6 @@ static void smc_listen_work(struct work_struct *work)
} }
smc_tx_init(new_smc); smc_tx_init(new_smc);
smc_rx_init(new_smc);
out_connected: out_connected:
sk_refcnt_debug_inc(newsmcsk); sk_refcnt_debug_inc(newsmcsk);
......
...@@ -164,6 +164,7 @@ struct smc_connection { ...@@ -164,6 +164,7 @@ struct smc_connection {
#ifndef KERNEL_HAS_ATOMIC64 #ifndef KERNEL_HAS_ATOMIC64
spinlock_t acurs_lock; /* protect cursors */ spinlock_t acurs_lock; /* protect cursors */
#endif #endif
struct work_struct close_work; /* peer sent some closing */
}; };
struct smc_sock { /* smc sock container */ struct smc_sock { /* smc sock container */
......
...@@ -217,8 +217,13 @@ static void smc_cdc_msg_recv_action(struct smc_sock *smc, ...@@ -217,8 +217,13 @@ static void smc_cdc_msg_recv_action(struct smc_sock *smc,
smc->sk.sk_err = ECONNRESET; smc->sk.sk_err = ECONNRESET;
conn->local_tx_ctrl.conn_state_flags.peer_conn_abort = 1; conn->local_tx_ctrl.conn_state_flags.peer_conn_abort = 1;
} }
if (smc_cdc_rxed_any_close_or_senddone(conn)) if (smc_cdc_rxed_any_close_or_senddone(conn)) {
smc_close_passive_received(smc); smc->sk.sk_shutdown |= RCV_SHUTDOWN;
if (smc->clcsock && smc->clcsock->sk)
smc->clcsock->sk->sk_shutdown |= RCV_SHUTDOWN;
sock_set_flag(&smc->sk, SOCK_DONE);
schedule_work(&conn->close_work);
}
/* piggy backed tx info */ /* piggy backed tx info */
/* trigger sndbuf consumer: RDMA write into peer RMBE and CDC */ /* trigger sndbuf consumer: RDMA write into peer RMBE and CDC */
...@@ -228,8 +233,6 @@ static void smc_cdc_msg_recv_action(struct smc_sock *smc, ...@@ -228,8 +233,6 @@ static void smc_cdc_msg_recv_action(struct smc_sock *smc,
smc_close_wake_tx_prepared(smc); smc_close_wake_tx_prepared(smc);
} }
/* subsequent patch: trigger socket release if connection closed */
/* socket connected but not accepted */ /* socket connected but not accepted */
if (!smc->sk.sk_socket) if (!smc->sk.sk_socket)
return; return;
......
...@@ -117,7 +117,6 @@ void smc_close_active_abort(struct smc_sock *smc) ...@@ -117,7 +117,6 @@ void smc_close_active_abort(struct smc_sock *smc)
struct smc_cdc_conn_state_flags *txflags = struct smc_cdc_conn_state_flags *txflags =
&smc->conn.local_tx_ctrl.conn_state_flags; &smc->conn.local_tx_ctrl.conn_state_flags;
bh_lock_sock(&smc->sk);
smc->sk.sk_err = ECONNABORTED; smc->sk.sk_err = ECONNABORTED;
if (smc->clcsock && smc->clcsock->sk) { if (smc->clcsock && smc->clcsock->sk) {
smc->clcsock->sk->sk_err = ECONNABORTED; smc->clcsock->sk->sk_err = ECONNABORTED;
...@@ -125,6 +124,7 @@ void smc_close_active_abort(struct smc_sock *smc) ...@@ -125,6 +124,7 @@ void smc_close_active_abort(struct smc_sock *smc)
} }
switch (smc->sk.sk_state) { switch (smc->sk.sk_state) {
case SMC_INIT: case SMC_INIT:
case SMC_ACTIVE:
smc->sk.sk_state = SMC_PEERABORTWAIT; smc->sk.sk_state = SMC_PEERABORTWAIT;
break; break;
case SMC_APPCLOSEWAIT1: case SMC_APPCLOSEWAIT1:
...@@ -161,10 +161,15 @@ void smc_close_active_abort(struct smc_sock *smc) ...@@ -161,10 +161,15 @@ void smc_close_active_abort(struct smc_sock *smc)
} }
sock_set_flag(&smc->sk, SOCK_DEAD); sock_set_flag(&smc->sk, SOCK_DEAD);
bh_unlock_sock(&smc->sk);
smc->sk.sk_state_change(&smc->sk); smc->sk.sk_state_change(&smc->sk);
} }
static inline bool smc_close_sent_any_close(struct smc_connection *conn)
{
return conn->local_tx_ctrl.conn_state_flags.peer_conn_abort ||
conn->local_tx_ctrl.conn_state_flags.peer_conn_closed;
}
int smc_close_active(struct smc_sock *smc) int smc_close_active(struct smc_sock *smc)
{ {
struct smc_cdc_conn_state_flags *txflags = struct smc_cdc_conn_state_flags *txflags =
...@@ -185,8 +190,7 @@ int smc_close_active(struct smc_sock *smc) ...@@ -185,8 +190,7 @@ int smc_close_active(struct smc_sock *smc)
case SMC_INIT: case SMC_INIT:
sk->sk_state = SMC_CLOSED; sk->sk_state = SMC_CLOSED;
if (smc->smc_listen_work.func) if (smc->smc_listen_work.func)
flush_work(&smc->smc_listen_work); cancel_work_sync(&smc->smc_listen_work);
sock_put(sk);
break; break;
case SMC_LISTEN: case SMC_LISTEN:
sk->sk_state = SMC_CLOSED; sk->sk_state = SMC_CLOSED;
...@@ -198,7 +202,7 @@ int smc_close_active(struct smc_sock *smc) ...@@ -198,7 +202,7 @@ int smc_close_active(struct smc_sock *smc)
} }
release_sock(sk); release_sock(sk);
smc_close_cleanup_listen(sk); smc_close_cleanup_listen(sk);
flush_work(&smc->tcp_listen_work); cancel_work_sync(&smc->smc_listen_work);
lock_sock(sk); lock_sock(sk);
break; break;
case SMC_ACTIVE: case SMC_ACTIVE:
...@@ -218,7 +222,7 @@ int smc_close_active(struct smc_sock *smc) ...@@ -218,7 +222,7 @@ int smc_close_active(struct smc_sock *smc)
case SMC_APPFINCLOSEWAIT: case SMC_APPFINCLOSEWAIT:
/* socket already shutdown wr or both (active close) */ /* socket already shutdown wr or both (active close) */
if (txflags->peer_done_writing && if (txflags->peer_done_writing &&
!txflags->peer_conn_closed) { !smc_close_sent_any_close(conn)) {
/* just shutdown wr done, send close request */ /* just shutdown wr done, send close request */
rc = smc_close_final(conn); rc = smc_close_final(conn);
} }
...@@ -248,6 +252,13 @@ int smc_close_active(struct smc_sock *smc) ...@@ -248,6 +252,13 @@ int smc_close_active(struct smc_sock *smc)
break; break;
case SMC_PEERCLOSEWAIT1: case SMC_PEERCLOSEWAIT1:
case SMC_PEERCLOSEWAIT2: case SMC_PEERCLOSEWAIT2:
if (txflags->peer_done_writing &&
!smc_close_sent_any_close(conn)) {
/* just shutdown wr done, send close request */
rc = smc_close_final(conn);
}
/* peer sending PeerConnectionClosed will cause transition */
break;
case SMC_PEERFINCLOSEWAIT: case SMC_PEERFINCLOSEWAIT:
/* peer sending PeerConnectionClosed will cause transition */ /* peer sending PeerConnectionClosed will cause transition */
break; break;
...@@ -285,7 +296,7 @@ static void smc_close_passive_abort_received(struct smc_sock *smc) ...@@ -285,7 +296,7 @@ static void smc_close_passive_abort_received(struct smc_sock *smc)
case SMC_PEERCLOSEWAIT1: case SMC_PEERCLOSEWAIT1:
case SMC_PEERCLOSEWAIT2: case SMC_PEERCLOSEWAIT2:
if (txflags->peer_done_writing && if (txflags->peer_done_writing &&
!txflags->peer_conn_closed) { !smc_close_sent_any_close(&smc->conn)) {
/* just shutdown, but not yet closed locally */ /* just shutdown, but not yet closed locally */
smc_close_abort(&smc->conn); smc_close_abort(&smc->conn);
sk->sk_state = SMC_PROCESSABORT; sk->sk_state = SMC_PROCESSABORT;
...@@ -306,22 +317,27 @@ static void smc_close_passive_abort_received(struct smc_sock *smc) ...@@ -306,22 +317,27 @@ static void smc_close_passive_abort_received(struct smc_sock *smc)
/* Some kind of closing has been received: peer_conn_closed, peer_conn_abort, /* Some kind of closing has been received: peer_conn_closed, peer_conn_abort,
* or peer_done_writing. * or peer_done_writing.
* Called under tasklet context.
*/ */
void smc_close_passive_received(struct smc_sock *smc) static void smc_close_passive_work(struct work_struct *work)
{ {
struct smc_cdc_conn_state_flags *rxflags = struct smc_connection *conn = container_of(work,
&smc->conn.local_rx_ctrl.conn_state_flags; struct smc_connection,
close_work);
struct smc_sock *smc = container_of(conn, struct smc_sock, conn);
struct smc_cdc_conn_state_flags *rxflags;
struct sock *sk = &smc->sk; struct sock *sk = &smc->sk;
int old_state; int old_state;
sk->sk_shutdown |= RCV_SHUTDOWN; lock_sock(&smc->sk);
if (smc->clcsock && smc->clcsock->sk)
smc->clcsock->sk->sk_shutdown |= RCV_SHUTDOWN;
sock_set_flag(&smc->sk, SOCK_DONE);
old_state = sk->sk_state; old_state = sk->sk_state;
if (!conn->alert_token_local) {
/* abnormal termination */
smc_close_active_abort(smc);
goto wakeup;
}
rxflags = &smc->conn.local_rx_ctrl.conn_state_flags;
if (rxflags->peer_conn_abort) { if (rxflags->peer_conn_abort) {
smc_close_passive_abort_received(smc); smc_close_passive_abort_received(smc);
goto wakeup; goto wakeup;
...@@ -331,7 +347,7 @@ void smc_close_passive_received(struct smc_sock *smc) ...@@ -331,7 +347,7 @@ void smc_close_passive_received(struct smc_sock *smc)
case SMC_INIT: case SMC_INIT:
if (atomic_read(&smc->conn.bytes_to_rcv) || if (atomic_read(&smc->conn.bytes_to_rcv) ||
(rxflags->peer_done_writing && (rxflags->peer_done_writing &&
!rxflags->peer_conn_closed)) !smc_cdc_rxed_any_close(conn)))
sk->sk_state = SMC_APPCLOSEWAIT1; sk->sk_state = SMC_APPCLOSEWAIT1;
else else
sk->sk_state = SMC_CLOSED; sk->sk_state = SMC_CLOSED;
...@@ -348,7 +364,7 @@ void smc_close_passive_received(struct smc_sock *smc) ...@@ -348,7 +364,7 @@ void smc_close_passive_received(struct smc_sock *smc)
if (!smc_cdc_rxed_any_close(&smc->conn)) if (!smc_cdc_rxed_any_close(&smc->conn))
break; break;
if (sock_flag(sk, SOCK_DEAD) && if (sock_flag(sk, SOCK_DEAD) &&
(sk->sk_shutdown == SHUTDOWN_MASK)) { smc_close_sent_any_close(conn)) {
/* smc_release has already been called locally */ /* smc_release has already been called locally */
sk->sk_state = SMC_CLOSED; sk->sk_state = SMC_CLOSED;
} else { } else {
...@@ -367,17 +383,19 @@ void smc_close_passive_received(struct smc_sock *smc) ...@@ -367,17 +383,19 @@ void smc_close_passive_received(struct smc_sock *smc)
} }
wakeup: wakeup:
if (old_state != sk->sk_state)
sk->sk_state_change(sk);
sk->sk_data_ready(sk); /* wakeup blocked rcvbuf consumers */ sk->sk_data_ready(sk); /* wakeup blocked rcvbuf consumers */
sk->sk_write_space(sk); /* wakeup blocked sndbuf producers */ sk->sk_write_space(sk); /* wakeup blocked sndbuf producers */
if ((sk->sk_state == SMC_CLOSED) && if (old_state != sk->sk_state) {
(sock_flag(sk, SOCK_DEAD) || (old_state == SMC_INIT))) { sk->sk_state_change(sk);
smc_conn_free(&smc->conn); if ((sk->sk_state == SMC_CLOSED) &&
schedule_delayed_work(&smc->sock_put_work, (sock_flag(sk, SOCK_DEAD) || !sk->sk_socket)) {
SMC_CLOSE_SOCK_PUT_DELAY); smc_conn_free(&smc->conn);
schedule_delayed_work(&smc->sock_put_work,
SMC_CLOSE_SOCK_PUT_DELAY);
}
} }
release_sock(&smc->sk);
} }
void smc_close_sock_put_work(struct work_struct *work) void smc_close_sock_put_work(struct work_struct *work)
...@@ -442,3 +460,9 @@ int smc_close_shutdown_write(struct smc_sock *smc) ...@@ -442,3 +460,9 @@ int smc_close_shutdown_write(struct smc_sock *smc)
sk->sk_state_change(&smc->sk); sk->sk_state_change(&smc->sk);
return rc; return rc;
} }
/* Initialize close properties on connection establishment. */
void smc_close_init(struct smc_sock *smc)
{
INIT_WORK(&smc->conn.close_work, smc_close_passive_work);
}
...@@ -21,8 +21,8 @@ ...@@ -21,8 +21,8 @@
void smc_close_wake_tx_prepared(struct smc_sock *smc); void smc_close_wake_tx_prepared(struct smc_sock *smc);
void smc_close_active_abort(struct smc_sock *smc); void smc_close_active_abort(struct smc_sock *smc);
int smc_close_active(struct smc_sock *smc); int smc_close_active(struct smc_sock *smc);
void smc_close_passive_received(struct smc_sock *smc);
void smc_close_sock_put_work(struct work_struct *work); void smc_close_sock_put_work(struct work_struct *work);
int smc_close_shutdown_write(struct smc_sock *smc); int smc_close_shutdown_write(struct smc_sock *smc);
void smc_close_init(struct smc_sock *smc);
#endif /* SMC_CLOSE_H */ #endif /* SMC_CLOSE_H */
...@@ -316,7 +316,7 @@ void smc_lgr_terminate(struct smc_link_group *lgr) ...@@ -316,7 +316,7 @@ void smc_lgr_terminate(struct smc_link_group *lgr)
smc = container_of(conn, struct smc_sock, conn); smc = container_of(conn, struct smc_sock, conn);
sock_hold(&smc->sk); sock_hold(&smc->sk);
__smc_lgr_unregister_conn(conn); __smc_lgr_unregister_conn(conn);
smc_close_active_abort(smc); schedule_work(&conn->close_work);
sock_put(&smc->sk); sock_put(&smc->sk);
node = rb_first(&lgr->conns_all); node = rb_first(&lgr->conns_all);
} }
......
...@@ -179,8 +179,6 @@ static void smc_ib_global_event_handler(struct ib_event_handler *handler, ...@@ -179,8 +179,6 @@ static void smc_ib_global_event_handler(struct ib_event_handler *handler,
u8 port_idx; u8 port_idx;
smcibdev = container_of(handler, struct smc_ib_device, event_handler); smcibdev = container_of(handler, struct smc_ib_device, event_handler);
if (!smc_pnet_find_ib(smcibdev->ibdev->name))
return;
switch (ibevent->event) { switch (ibevent->event) {
case IB_EVENT_PORT_ERR: case IB_EVENT_PORT_ERR:
...@@ -259,7 +257,6 @@ int smc_ib_create_queue_pair(struct smc_link *lnk) ...@@ -259,7 +257,6 @@ int smc_ib_create_queue_pair(struct smc_link *lnk)
.max_recv_wr = SMC_WR_BUF_CNT * 3, .max_recv_wr = SMC_WR_BUF_CNT * 3,
.max_send_sge = SMC_IB_MAX_SEND_SGE, .max_send_sge = SMC_IB_MAX_SEND_SGE,
.max_recv_sge = 1, .max_recv_sge = 1,
.max_inline_data = SMC_WR_TX_SIZE,
}, },
.sq_sig_type = IB_SIGNAL_REQ_WR, .sq_sig_type = IB_SIGNAL_REQ_WR,
.qp_type = IB_QPT_RC, .qp_type = IB_QPT_RC,
......
...@@ -219,7 +219,7 @@ static bool smc_pnetid_valid(const char *pnet_name, char *pnetid) ...@@ -219,7 +219,7 @@ static bool smc_pnetid_valid(const char *pnet_name, char *pnetid)
} }
/* Find an infiniband device by a given name. The device might not exist. */ /* Find an infiniband device by a given name. The device might not exist. */
struct smc_ib_device *smc_pnet_find_ib(char *ib_name) static struct smc_ib_device *smc_pnet_find_ib(char *ib_name)
{ {
struct smc_ib_device *ibdev; struct smc_ib_device *ibdev;
...@@ -523,8 +523,11 @@ void smc_pnet_find_roce_resource(struct sock *sk, ...@@ -523,8 +523,11 @@ void smc_pnet_find_roce_resource(struct sock *sk,
read_lock(&smc_pnettable.lock); read_lock(&smc_pnettable.lock);
list_for_each_entry(pnetelem, &smc_pnettable.pnetlist, list) { list_for_each_entry(pnetelem, &smc_pnettable.pnetlist, list) {
if (dst->dev == pnetelem->ndev) { if (dst->dev == pnetelem->ndev) {
*smcibdev = pnetelem->smcibdev; if (smc_ib_port_active(pnetelem->smcibdev,
*ibport = pnetelem->ib_port; pnetelem->ib_port)) {
*smcibdev = pnetelem->smcibdev;
*ibport = pnetelem->ib_port;
}
break; break;
} }
} }
......
...@@ -16,7 +16,6 @@ struct smc_ib_device; ...@@ -16,7 +16,6 @@ struct smc_ib_device;
int smc_pnet_init(void) __init; int smc_pnet_init(void) __init;
void smc_pnet_exit(void); void smc_pnet_exit(void);
int smc_pnet_remove_by_ibdev(struct smc_ib_device *ibdev); int smc_pnet_remove_by_ibdev(struct smc_ib_device *ibdev);
struct smc_ib_device *smc_pnet_find_ib(char *ib_name);
void smc_pnet_find_roce_resource(struct sock *sk, void smc_pnet_find_roce_resource(struct sock *sk,
struct smc_ib_device **smcibdev, u8 *ibport); struct smc_ib_device **smcibdev, u8 *ibport);
......
...@@ -36,11 +36,10 @@ static void smc_rx_data_ready(struct sock *sk) ...@@ -36,11 +36,10 @@ static void smc_rx_data_ready(struct sock *sk)
if (skwq_has_sleeper(wq)) if (skwq_has_sleeper(wq))
wake_up_interruptible_sync_poll(&wq->wait, POLLIN | POLLPRI | wake_up_interruptible_sync_poll(&wq->wait, POLLIN | POLLPRI |
POLLRDNORM | POLLRDBAND); POLLRDNORM | POLLRDBAND);
sk_wake_async(sk, SOCK_WAKE_WAITD, POLL_IN);
if ((sk->sk_shutdown == SHUTDOWN_MASK) || if ((sk->sk_shutdown == SHUTDOWN_MASK) ||
(sk->sk_state == SMC_CLOSED)) (sk->sk_state == SMC_CLOSED))
sk_wake_async(sk, SOCK_WAKE_WAITD, POLL_HUP); sk_wake_async(sk, SOCK_WAKE_WAITD, POLL_HUP);
else
sk_wake_async(sk, SOCK_WAKE_WAITD, POLL_IN);
rcu_read_unlock(); rcu_read_unlock();
} }
......
...@@ -431,9 +431,13 @@ static void smc_tx_work(struct work_struct *work) ...@@ -431,9 +431,13 @@ static void smc_tx_work(struct work_struct *work)
struct smc_connection, struct smc_connection,
tx_work); tx_work);
struct smc_sock *smc = container_of(conn, struct smc_sock, conn); struct smc_sock *smc = container_of(conn, struct smc_sock, conn);
int rc;
lock_sock(&smc->sk); lock_sock(&smc->sk);
smc_tx_sndbuf_nonempty(conn); rc = smc_tx_sndbuf_nonempty(conn);
if (!rc && conn->local_rx_ctrl.prod_flags.write_blocked &&
!atomic_read(&conn->bytes_to_rcv))
conn->local_rx_ctrl.prod_flags.write_blocked = 0;
release_sock(&smc->sk); release_sock(&smc->sk);
} }
......
...@@ -447,7 +447,7 @@ static void smc_wr_init_sge(struct smc_link *lnk) ...@@ -447,7 +447,7 @@ static void smc_wr_init_sge(struct smc_link *lnk)
lnk->wr_tx_ibs[i].num_sge = 1; lnk->wr_tx_ibs[i].num_sge = 1;
lnk->wr_tx_ibs[i].opcode = IB_WR_SEND; lnk->wr_tx_ibs[i].opcode = IB_WR_SEND;
lnk->wr_tx_ibs[i].send_flags = lnk->wr_tx_ibs[i].send_flags =
IB_SEND_SIGNALED | IB_SEND_SOLICITED | IB_SEND_INLINE; IB_SEND_SIGNALED | IB_SEND_SOLICITED;
} }
for (i = 0; i < lnk->wr_rx_cnt; i++) { for (i = 0; i < lnk->wr_rx_cnt; i++) {
lnk->wr_rx_sges[i].addr = lnk->wr_rx_sges[i].addr =
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment