Commit 796fff0c authored by David S. Miller's avatar David S. Miller

Merge branch 'smc-fixes'

Ursula Braun says:

====================
net/smc: fixes 2019-04-11

here are some fixes in different areas of the smc code for the net
tree.
====================
Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
parents 988dc4a9 f61bca58
...@@ -2084,12 +2084,6 @@ static inline bool skwq_has_sleeper(struct socket_wq *wq) ...@@ -2084,12 +2084,6 @@ static inline bool skwq_has_sleeper(struct socket_wq *wq)
* @p: poll_table * @p: poll_table
* *
* See the comments in the wq_has_sleeper function. * See the comments in the wq_has_sleeper function.
*
* Do not derive sock from filp->private_data here. An SMC socket establishes
* an internal TCP socket that is used in the fallback case. All socket
* operations on the SMC socket are then forwarded to the TCP socket. In case of
* poll, the filp->private_data pointer references the SMC socket because the
* TCP socket has no file assigned.
*/ */
static inline void sock_poll_wait(struct file *filp, struct socket *sock, static inline void sock_poll_wait(struct file *filp, struct socket *sock,
poll_table *p) poll_table *p)
......
...@@ -167,10 +167,9 @@ static int smc_release(struct socket *sock) ...@@ -167,10 +167,9 @@ static int smc_release(struct socket *sock)
if (sk->sk_state == SMC_CLOSED) { if (sk->sk_state == SMC_CLOSED) {
if (smc->clcsock) { if (smc->clcsock) {
mutex_lock(&smc->clcsock_release_lock); release_sock(sk);
sock_release(smc->clcsock); smc_clcsock_release(smc);
smc->clcsock = NULL; lock_sock(sk);
mutex_unlock(&smc->clcsock_release_lock);
} }
if (!smc->use_fallback) if (!smc->use_fallback)
smc_conn_free(&smc->conn); smc_conn_free(&smc->conn);
...@@ -446,10 +445,19 @@ static void smc_link_save_peer_info(struct smc_link *link, ...@@ -446,10 +445,19 @@ static void smc_link_save_peer_info(struct smc_link *link,
link->peer_mtu = clc->qp_mtu; link->peer_mtu = clc->qp_mtu;
} }
static void smc_switch_to_fallback(struct smc_sock *smc)
{
smc->use_fallback = true;
if (smc->sk.sk_socket && smc->sk.sk_socket->file) {
smc->clcsock->file = smc->sk.sk_socket->file;
smc->clcsock->file->private_data = smc->clcsock;
}
}
/* fall back during connect */ /* fall back during connect */
static int smc_connect_fallback(struct smc_sock *smc, int reason_code) static int smc_connect_fallback(struct smc_sock *smc, int reason_code)
{ {
smc->use_fallback = true; smc_switch_to_fallback(smc);
smc->fallback_rsn = reason_code; smc->fallback_rsn = reason_code;
smc_copy_sock_settings_to_clc(smc); smc_copy_sock_settings_to_clc(smc);
if (smc->sk.sk_state == SMC_INIT) if (smc->sk.sk_state == SMC_INIT)
...@@ -775,10 +783,14 @@ static void smc_connect_work(struct work_struct *work) ...@@ -775,10 +783,14 @@ static void smc_connect_work(struct work_struct *work)
smc->sk.sk_err = -rc; smc->sk.sk_err = -rc;
out: out:
if (smc->sk.sk_err) if (!sock_flag(&smc->sk, SOCK_DEAD)) {
smc->sk.sk_state_change(&smc->sk); if (smc->sk.sk_err) {
else smc->sk.sk_state_change(&smc->sk);
smc->sk.sk_write_space(&smc->sk); } else { /* allow polling before and after fallback decision */
smc->clcsock->sk->sk_write_space(smc->clcsock->sk);
smc->sk.sk_write_space(&smc->sk);
}
}
kfree(smc->connect_info); kfree(smc->connect_info);
smc->connect_info = NULL; smc->connect_info = NULL;
release_sock(&smc->sk); release_sock(&smc->sk);
...@@ -872,11 +884,11 @@ static int smc_clcsock_accept(struct smc_sock *lsmc, struct smc_sock **new_smc) ...@@ -872,11 +884,11 @@ static int smc_clcsock_accept(struct smc_sock *lsmc, struct smc_sock **new_smc)
if (rc < 0) if (rc < 0)
lsk->sk_err = -rc; lsk->sk_err = -rc;
if (rc < 0 || lsk->sk_state == SMC_CLOSED) { if (rc < 0 || lsk->sk_state == SMC_CLOSED) {
new_sk->sk_prot->unhash(new_sk);
if (new_clcsock) if (new_clcsock)
sock_release(new_clcsock); sock_release(new_clcsock);
new_sk->sk_state = SMC_CLOSED; new_sk->sk_state = SMC_CLOSED;
sock_set_flag(new_sk, SOCK_DEAD); sock_set_flag(new_sk, SOCK_DEAD);
new_sk->sk_prot->unhash(new_sk);
sock_put(new_sk); /* final */ sock_put(new_sk); /* final */
*new_smc = NULL; *new_smc = NULL;
goto out; goto out;
...@@ -927,16 +939,21 @@ struct sock *smc_accept_dequeue(struct sock *parent, ...@@ -927,16 +939,21 @@ struct sock *smc_accept_dequeue(struct sock *parent,
smc_accept_unlink(new_sk); smc_accept_unlink(new_sk);
if (new_sk->sk_state == SMC_CLOSED) { if (new_sk->sk_state == SMC_CLOSED) {
new_sk->sk_prot->unhash(new_sk);
if (isk->clcsock) { if (isk->clcsock) {
sock_release(isk->clcsock); sock_release(isk->clcsock);
isk->clcsock = NULL; isk->clcsock = NULL;
} }
new_sk->sk_prot->unhash(new_sk);
sock_put(new_sk); /* final */ sock_put(new_sk); /* final */
continue; continue;
} }
if (new_sock) if (new_sock) {
sock_graft(new_sk, new_sock); sock_graft(new_sk, new_sock);
if (isk->use_fallback) {
smc_sk(new_sk)->clcsock->file = new_sock->file;
isk->clcsock->file->private_data = isk->clcsock;
}
}
return new_sk; return new_sk;
} }
return NULL; return NULL;
...@@ -956,6 +973,7 @@ void smc_close_non_accepted(struct sock *sk) ...@@ -956,6 +973,7 @@ void smc_close_non_accepted(struct sock *sk)
sock_set_flag(sk, SOCK_DEAD); sock_set_flag(sk, SOCK_DEAD);
sk->sk_shutdown |= SHUTDOWN_MASK; sk->sk_shutdown |= SHUTDOWN_MASK;
} }
sk->sk_prot->unhash(sk);
if (smc->clcsock) { if (smc->clcsock) {
struct socket *tcp; struct socket *tcp;
...@@ -971,7 +989,6 @@ void smc_close_non_accepted(struct sock *sk) ...@@ -971,7 +989,6 @@ void smc_close_non_accepted(struct sock *sk)
smc_conn_free(&smc->conn); smc_conn_free(&smc->conn);
} }
release_sock(sk); release_sock(sk);
sk->sk_prot->unhash(sk);
sock_put(sk); /* final sock_put */ sock_put(sk); /* final sock_put */
} }
...@@ -1037,13 +1054,13 @@ static void smc_listen_out(struct smc_sock *new_smc) ...@@ -1037,13 +1054,13 @@ static void smc_listen_out(struct smc_sock *new_smc)
struct smc_sock *lsmc = new_smc->listen_smc; struct smc_sock *lsmc = new_smc->listen_smc;
struct sock *newsmcsk = &new_smc->sk; struct sock *newsmcsk = &new_smc->sk;
lock_sock_nested(&lsmc->sk, SINGLE_DEPTH_NESTING);
if (lsmc->sk.sk_state == SMC_LISTEN) { if (lsmc->sk.sk_state == SMC_LISTEN) {
lock_sock_nested(&lsmc->sk, SINGLE_DEPTH_NESTING);
smc_accept_enqueue(&lsmc->sk, newsmcsk); smc_accept_enqueue(&lsmc->sk, newsmcsk);
release_sock(&lsmc->sk);
} else { /* no longer listening */ } else { /* no longer listening */
smc_close_non_accepted(newsmcsk); smc_close_non_accepted(newsmcsk);
} }
release_sock(&lsmc->sk);
/* Wake up accept */ /* Wake up accept */
lsmc->sk.sk_data_ready(&lsmc->sk); lsmc->sk.sk_data_ready(&lsmc->sk);
...@@ -1087,7 +1104,7 @@ static void smc_listen_decline(struct smc_sock *new_smc, int reason_code, ...@@ -1087,7 +1104,7 @@ static void smc_listen_decline(struct smc_sock *new_smc, int reason_code,
return; return;
} }
smc_conn_free(&new_smc->conn); smc_conn_free(&new_smc->conn);
new_smc->use_fallback = true; smc_switch_to_fallback(new_smc);
new_smc->fallback_rsn = reason_code; new_smc->fallback_rsn = reason_code;
if (reason_code && reason_code != SMC_CLC_DECL_PEERDECL) { if (reason_code && reason_code != SMC_CLC_DECL_PEERDECL) {
if (smc_clc_send_decline(new_smc, reason_code) < 0) { if (smc_clc_send_decline(new_smc, reason_code) < 0) {
...@@ -1237,6 +1254,9 @@ static void smc_listen_work(struct work_struct *work) ...@@ -1237,6 +1254,9 @@ static void smc_listen_work(struct work_struct *work)
int rc = 0; int rc = 0;
u8 ibport; u8 ibport;
if (new_smc->listen_smc->sk.sk_state != SMC_LISTEN)
return smc_listen_out_err(new_smc);
if (new_smc->use_fallback) { if (new_smc->use_fallback) {
smc_listen_out_connected(new_smc); smc_listen_out_connected(new_smc);
return; return;
...@@ -1244,7 +1264,7 @@ static void smc_listen_work(struct work_struct *work) ...@@ -1244,7 +1264,7 @@ static void smc_listen_work(struct work_struct *work)
/* check if peer is smc capable */ /* check if peer is smc capable */
if (!tcp_sk(newclcsock->sk)->syn_smc) { if (!tcp_sk(newclcsock->sk)->syn_smc) {
new_smc->use_fallback = true; smc_switch_to_fallback(new_smc);
new_smc->fallback_rsn = SMC_CLC_DECL_PEERNOSMC; new_smc->fallback_rsn = SMC_CLC_DECL_PEERNOSMC;
smc_listen_out_connected(new_smc); smc_listen_out_connected(new_smc);
return; return;
...@@ -1501,7 +1521,7 @@ static int smc_sendmsg(struct socket *sock, struct msghdr *msg, size_t len) ...@@ -1501,7 +1521,7 @@ static int smc_sendmsg(struct socket *sock, struct msghdr *msg, size_t len)
if (msg->msg_flags & MSG_FASTOPEN) { if (msg->msg_flags & MSG_FASTOPEN) {
if (sk->sk_state == SMC_INIT) { if (sk->sk_state == SMC_INIT) {
smc->use_fallback = true; smc_switch_to_fallback(smc);
smc->fallback_rsn = SMC_CLC_DECL_OPTUNSUPP; smc->fallback_rsn = SMC_CLC_DECL_OPTUNSUPP;
} else { } else {
rc = -EINVAL; rc = -EINVAL;
...@@ -1703,7 +1723,7 @@ static int smc_setsockopt(struct socket *sock, int level, int optname, ...@@ -1703,7 +1723,7 @@ static int smc_setsockopt(struct socket *sock, int level, int optname,
case TCP_FASTOPEN_NO_COOKIE: case TCP_FASTOPEN_NO_COOKIE:
/* option not supported by SMC */ /* option not supported by SMC */
if (sk->sk_state == SMC_INIT) { if (sk->sk_state == SMC_INIT) {
smc->use_fallback = true; smc_switch_to_fallback(smc);
smc->fallback_rsn = SMC_CLC_DECL_OPTUNSUPP; smc->fallback_rsn = SMC_CLC_DECL_OPTUNSUPP;
} else { } else {
if (!smc->use_fallback) if (!smc->use_fallback)
......
...@@ -21,6 +21,22 @@ ...@@ -21,6 +21,22 @@
#define SMC_CLOSE_WAIT_LISTEN_CLCSOCK_TIME (5 * HZ) #define SMC_CLOSE_WAIT_LISTEN_CLCSOCK_TIME (5 * HZ)
/* release the clcsock that is assigned to the smc_sock */
void smc_clcsock_release(struct smc_sock *smc)
{
struct socket *tcp;
if (smc->listen_smc && current_work() != &smc->smc_listen_work)
cancel_work_sync(&smc->smc_listen_work);
mutex_lock(&smc->clcsock_release_lock);
if (smc->clcsock) {
tcp = smc->clcsock;
smc->clcsock = NULL;
sock_release(tcp);
}
mutex_unlock(&smc->clcsock_release_lock);
}
static void smc_close_cleanup_listen(struct sock *parent) static void smc_close_cleanup_listen(struct sock *parent)
{ {
struct sock *sk; struct sock *sk;
...@@ -321,6 +337,7 @@ static void smc_close_passive_work(struct work_struct *work) ...@@ -321,6 +337,7 @@ static void smc_close_passive_work(struct work_struct *work)
close_work); close_work);
struct smc_sock *smc = container_of(conn, struct smc_sock, conn); struct smc_sock *smc = container_of(conn, struct smc_sock, conn);
struct smc_cdc_conn_state_flags *rxflags; struct smc_cdc_conn_state_flags *rxflags;
bool release_clcsock = false;
struct sock *sk = &smc->sk; struct sock *sk = &smc->sk;
int old_state; int old_state;
...@@ -400,13 +417,13 @@ static void smc_close_passive_work(struct work_struct *work) ...@@ -400,13 +417,13 @@ static void smc_close_passive_work(struct work_struct *work)
if ((sk->sk_state == SMC_CLOSED) && if ((sk->sk_state == SMC_CLOSED) &&
(sock_flag(sk, SOCK_DEAD) || !sk->sk_socket)) { (sock_flag(sk, SOCK_DEAD) || !sk->sk_socket)) {
smc_conn_free(conn); smc_conn_free(conn);
if (smc->clcsock) { if (smc->clcsock)
sock_release(smc->clcsock); release_clcsock = true;
smc->clcsock = NULL;
}
} }
} }
release_sock(sk); release_sock(sk);
if (release_clcsock)
smc_clcsock_release(smc);
sock_put(sk); /* sock_hold done by schedulers of close_work */ sock_put(sk); /* sock_hold done by schedulers of close_work */
} }
......
...@@ -23,5 +23,6 @@ void smc_close_wake_tx_prepared(struct smc_sock *smc); ...@@ -23,5 +23,6 @@ void smc_close_wake_tx_prepared(struct smc_sock *smc);
int smc_close_active(struct smc_sock *smc); int smc_close_active(struct smc_sock *smc);
int smc_close_shutdown_write(struct smc_sock *smc); int smc_close_shutdown_write(struct smc_sock *smc);
void smc_close_init(struct smc_sock *smc); void smc_close_init(struct smc_sock *smc);
void smc_clcsock_release(struct smc_sock *smc);
#endif /* SMC_CLOSE_H */ #endif /* SMC_CLOSE_H */
...@@ -289,6 +289,11 @@ struct smcd_dev *smcd_alloc_dev(struct device *parent, const char *name, ...@@ -289,6 +289,11 @@ struct smcd_dev *smcd_alloc_dev(struct device *parent, const char *name,
INIT_LIST_HEAD(&smcd->vlan); INIT_LIST_HEAD(&smcd->vlan);
smcd->event_wq = alloc_ordered_workqueue("ism_evt_wq-%s)", smcd->event_wq = alloc_ordered_workqueue("ism_evt_wq-%s)",
WQ_MEM_RECLAIM, name); WQ_MEM_RECLAIM, name);
if (!smcd->event_wq) {
kfree(smcd->conn);
kfree(smcd);
return NULL;
}
return smcd; return smcd;
} }
EXPORT_SYMBOL_GPL(smcd_alloc_dev); EXPORT_SYMBOL_GPL(smcd_alloc_dev);
......
...@@ -603,7 +603,8 @@ static int smc_pnet_flush(struct sk_buff *skb, struct genl_info *info) ...@@ -603,7 +603,8 @@ static int smc_pnet_flush(struct sk_buff *skb, struct genl_info *info)
{ {
struct net *net = genl_info_net(info); struct net *net = genl_info_net(info);
return smc_pnet_remove_by_pnetid(net, NULL); smc_pnet_remove_by_pnetid(net, NULL);
return 0;
} }
/* SMC_PNETID generic netlink operation definition */ /* SMC_PNETID generic netlink operation definition */
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment