Commit a65878d6 authored by Eelco Chaudron's avatar Eelco Chaudron Committed by David S. Miller

net: openvswitch: fixes potential deadlock in dp cleanup code

The previous patch introduced a deadlock, this patch fixes it by making
sure the work is canceled without holding the global ovs lock. This is
done by moving the reorder processing one layer up to the netns level.

Fixes: eac87c41 ("net: openvswitch: reorder masks array based on usage")
Reported-by: syzbot+2c4ff3614695f75ce26c@syzkaller.appspotmail.com
Reported-by: syzbot+bad6507e5db05017b008@syzkaller.appspotmail.com
Reviewed-by: default avatarPaolo <pabeni@redhat.com>
Signed-off-by: default avatarEelco Chaudron <echaudro@redhat.com>
Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
parent dfd3d526
...@@ -1655,7 +1655,6 @@ static int ovs_dp_cmd_new(struct sk_buff *skb, struct genl_info *info) ...@@ -1655,7 +1655,6 @@ static int ovs_dp_cmd_new(struct sk_buff *skb, struct genl_info *info)
goto err_destroy_reply; goto err_destroy_reply;
ovs_dp_set_net(dp, sock_net(skb->sk)); ovs_dp_set_net(dp, sock_net(skb->sk));
INIT_DELAYED_WORK(&dp->masks_rebalance, ovs_dp_masks_rebalance);
/* Allocate table. */ /* Allocate table. */
err = ovs_flow_tbl_init(&dp->table); err = ovs_flow_tbl_init(&dp->table);
...@@ -1715,9 +1714,6 @@ static int ovs_dp_cmd_new(struct sk_buff *skb, struct genl_info *info) ...@@ -1715,9 +1714,6 @@ static int ovs_dp_cmd_new(struct sk_buff *skb, struct genl_info *info)
ovs_net = net_generic(ovs_dp_get_net(dp), ovs_net_id); ovs_net = net_generic(ovs_dp_get_net(dp), ovs_net_id);
list_add_tail_rcu(&dp->list_node, &ovs_net->dps); list_add_tail_rcu(&dp->list_node, &ovs_net->dps);
schedule_delayed_work(&dp->masks_rebalance,
msecs_to_jiffies(DP_MASKS_REBALANCE_INTERVAL));
ovs_unlock(); ovs_unlock();
ovs_notify(&dp_datapath_genl_family, reply, info); ovs_notify(&dp_datapath_genl_family, reply, info);
...@@ -1762,9 +1758,6 @@ static void __dp_destroy(struct datapath *dp) ...@@ -1762,9 +1758,6 @@ static void __dp_destroy(struct datapath *dp)
/* RCU destroy the flow table */ /* RCU destroy the flow table */
call_rcu(&dp->rcu, destroy_dp_rcu); call_rcu(&dp->rcu, destroy_dp_rcu);
/* Cancel remaining work. */
cancel_delayed_work_sync(&dp->masks_rebalance);
} }
static int ovs_dp_cmd_del(struct sk_buff *skb, struct genl_info *info) static int ovs_dp_cmd_del(struct sk_buff *skb, struct genl_info *info)
...@@ -2349,14 +2342,18 @@ static int ovs_vport_cmd_dump(struct sk_buff *skb, struct netlink_callback *cb) ...@@ -2349,14 +2342,18 @@ static int ovs_vport_cmd_dump(struct sk_buff *skb, struct netlink_callback *cb)
static void ovs_dp_masks_rebalance(struct work_struct *work) static void ovs_dp_masks_rebalance(struct work_struct *work)
{ {
struct datapath *dp = container_of(work, struct datapath, struct ovs_net *ovs_net = container_of(work, struct ovs_net,
masks_rebalance.work); masks_rebalance.work);
struct datapath *dp;
ovs_lock(); ovs_lock();
ovs_flow_masks_rebalance(&dp->table);
list_for_each_entry(dp, &ovs_net->dps, list_node)
ovs_flow_masks_rebalance(&dp->table);
ovs_unlock(); ovs_unlock();
schedule_delayed_work(&dp->masks_rebalance, schedule_delayed_work(&ovs_net->masks_rebalance,
msecs_to_jiffies(DP_MASKS_REBALANCE_INTERVAL)); msecs_to_jiffies(DP_MASKS_REBALANCE_INTERVAL));
} }
...@@ -2454,6 +2451,9 @@ static int __net_init ovs_init_net(struct net *net) ...@@ -2454,6 +2451,9 @@ static int __net_init ovs_init_net(struct net *net)
INIT_LIST_HEAD(&ovs_net->dps); INIT_LIST_HEAD(&ovs_net->dps);
INIT_WORK(&ovs_net->dp_notify_work, ovs_dp_notify_wq); INIT_WORK(&ovs_net->dp_notify_work, ovs_dp_notify_wq);
INIT_DELAYED_WORK(&ovs_net->masks_rebalance, ovs_dp_masks_rebalance);
schedule_delayed_work(&ovs_net->masks_rebalance,
msecs_to_jiffies(DP_MASKS_REBALANCE_INTERVAL));
return ovs_ct_init(net); return ovs_ct_init(net);
} }
...@@ -2508,6 +2508,7 @@ static void __net_exit ovs_exit_net(struct net *dnet) ...@@ -2508,6 +2508,7 @@ static void __net_exit ovs_exit_net(struct net *dnet)
ovs_unlock(); ovs_unlock();
cancel_delayed_work_sync(&ovs_net->masks_rebalance);
cancel_work_sync(&ovs_net->dp_notify_work); cancel_work_sync(&ovs_net->dp_notify_work);
} }
......
...@@ -84,9 +84,6 @@ struct datapath { ...@@ -84,9 +84,6 @@ struct datapath {
/* Switch meters. */ /* Switch meters. */
struct dp_meter_table meter_tbl; struct dp_meter_table meter_tbl;
/* re-balance flow masks timer */
struct delayed_work masks_rebalance;
}; };
/** /**
...@@ -135,6 +132,7 @@ struct dp_upcall_info { ...@@ -135,6 +132,7 @@ struct dp_upcall_info {
struct ovs_net { struct ovs_net {
struct list_head dps; struct list_head dps;
struct work_struct dp_notify_work; struct work_struct dp_notify_work;
struct delayed_work masks_rebalance;
#if IS_ENABLED(CONFIG_NETFILTER_CONNCOUNT) #if IS_ENABLED(CONFIG_NETFILTER_CONNCOUNT)
struct ovs_ct_limit_info *ct_limit_info; struct ovs_ct_limit_info *ct_limit_info;
#endif #endif
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment