Commit 1a4f1a02 authored by David S. Miller's avatar David S. Miller

Merge tag 'mlx5-updates-2019-08-22' of git://git.kernel.org/pub/scm/linux/kernel/git/saeed/linux

Saeed Mahameed says:

====================
mlx5-updates-2019-08-22

Misc updates for mlx5e net device driver

1) Maxim and Tariq add the support for LAG TX port affinity distribution
When VF LAG is enabled, VFs netdevs will round-robin the TX affinity
of their tx queues among the different LAG ports.
2) Aya adds the support for ip-in-ip RSS.
3) Marina adds the support for ip-in-ip TX TSO and checksum offloads.
4) Moshe adds a device internal drop counter to mlx5 ethtool stats.
====================
Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
parents 42aa15cf 25948b87
...@@ -922,6 +922,7 @@ static bool devx_is_general_cmd(void *in, struct mlx5_ib_dev *dev) ...@@ -922,6 +922,7 @@ static bool devx_is_general_cmd(void *in, struct mlx5_ib_dev *dev)
case MLX5_CMD_OP_QUERY_CONG_STATUS: case MLX5_CMD_OP_QUERY_CONG_STATUS:
case MLX5_CMD_OP_QUERY_CONG_PARAMS: case MLX5_CMD_OP_QUERY_CONG_PARAMS:
case MLX5_CMD_OP_QUERY_CONG_STATISTICS: case MLX5_CMD_OP_QUERY_CONG_STATISTICS:
case MLX5_CMD_OP_QUERY_LAG:
return true; return true;
default: default:
return false; return false;
......
...@@ -86,7 +86,7 @@ struct mlx5_core_srq *mlx5_cmd_get_srq(struct mlx5_ib_dev *dev, u32 srqn) ...@@ -86,7 +86,7 @@ struct mlx5_core_srq *mlx5_cmd_get_srq(struct mlx5_ib_dev *dev, u32 srqn)
xa_lock(&table->array); xa_lock(&table->array);
srq = xa_load(&table->array, srqn); srq = xa_load(&table->array, srqn);
if (srq) if (srq)
atomic_inc(&srq->common.refcount); refcount_inc(&srq->common.refcount);
xa_unlock(&table->array); xa_unlock(&table->array);
return srq; return srq;
...@@ -592,7 +592,7 @@ int mlx5_cmd_create_srq(struct mlx5_ib_dev *dev, struct mlx5_core_srq *srq, ...@@ -592,7 +592,7 @@ int mlx5_cmd_create_srq(struct mlx5_ib_dev *dev, struct mlx5_core_srq *srq,
if (err) if (err)
return err; return err;
atomic_set(&srq->common.refcount, 1); refcount_set(&srq->common.refcount, 1);
init_completion(&srq->common.free); init_completion(&srq->common.free);
err = xa_err(xa_store_irq(&table->array, srq->srqn, srq, GFP_KERNEL)); err = xa_err(xa_store_irq(&table->array, srq->srqn, srq, GFP_KERNEL));
...@@ -675,7 +675,7 @@ static int srq_event_notifier(struct notifier_block *nb, ...@@ -675,7 +675,7 @@ static int srq_event_notifier(struct notifier_block *nb,
xa_lock(&table->array); xa_lock(&table->array);
srq = xa_load(&table->array, srqn); srq = xa_load(&table->array, srqn);
if (srq) if (srq)
atomic_inc(&srq->common.refcount); refcount_inc(&srq->common.refcount);
xa_unlock(&table->array); xa_unlock(&table->array);
if (!srq) if (!srq)
......
...@@ -446,6 +446,8 @@ static int mlx5_internal_err_ret_value(struct mlx5_core_dev *dev, u16 op, ...@@ -446,6 +446,8 @@ static int mlx5_internal_err_ret_value(struct mlx5_core_dev *dev, u16 op,
case MLX5_CMD_OP_CREATE_UMEM: case MLX5_CMD_OP_CREATE_UMEM:
case MLX5_CMD_OP_DESTROY_UMEM: case MLX5_CMD_OP_DESTROY_UMEM:
case MLX5_CMD_OP_ALLOC_MEMIC: case MLX5_CMD_OP_ALLOC_MEMIC:
case MLX5_CMD_OP_MODIFY_XRQ:
case MLX5_CMD_OP_RELEASE_XRQ_ERROR:
*status = MLX5_DRIVER_STATUS_ABORTED; *status = MLX5_DRIVER_STATUS_ABORTED;
*synd = MLX5_DRIVER_SYND; *synd = MLX5_DRIVER_SYND;
return -EIO; return -EIO;
...@@ -637,6 +639,8 @@ const char *mlx5_command_str(int command) ...@@ -637,6 +639,8 @@ const char *mlx5_command_str(int command)
MLX5_COMMAND_STR_CASE(DESTROY_UCTX); MLX5_COMMAND_STR_CASE(DESTROY_UCTX);
MLX5_COMMAND_STR_CASE(CREATE_UMEM); MLX5_COMMAND_STR_CASE(CREATE_UMEM);
MLX5_COMMAND_STR_CASE(DESTROY_UMEM); MLX5_COMMAND_STR_CASE(DESTROY_UMEM);
MLX5_COMMAND_STR_CASE(RELEASE_XRQ_ERROR);
MLX5_COMMAND_STR_CASE(MODIFY_XRQ);
default: return "unknown command opcode"; default: return "unknown command opcode";
} }
} }
......
...@@ -546,7 +546,7 @@ static void mlx5_fw_tracer_save_trace(struct mlx5_fw_tracer *tracer, ...@@ -546,7 +546,7 @@ static void mlx5_fw_tracer_save_trace(struct mlx5_fw_tracer *tracer,
trace_data->timestamp = timestamp; trace_data->timestamp = timestamp;
trace_data->lost = lost; trace_data->lost = lost;
trace_data->event_id = event_id; trace_data->event_id = event_id;
strncpy(trace_data->msg, msg, TRACE_STR_MSG); strscpy_pad(trace_data->msg, msg, TRACE_STR_MSG);
tracer->st_arr.saved_traces_index = tracer->st_arr.saved_traces_index =
(tracer->st_arr.saved_traces_index + 1) & (SAVED_TRACES_NUM - 1); (tracer->st_arr.saved_traces_index + 1) & (SAVED_TRACES_NUM - 1);
......
...@@ -163,6 +163,14 @@ enum mlx5e_rq_group { ...@@ -163,6 +163,14 @@ enum mlx5e_rq_group {
#define MLX5E_NUM_RQ_GROUPS(g) (1 + MLX5E_RQ_GROUP_##g) #define MLX5E_NUM_RQ_GROUPS(g) (1 + MLX5E_RQ_GROUP_##g)
}; };
static inline u8 mlx5e_get_num_lag_ports(struct mlx5_core_dev *mdev)
{
if (mlx5_lag_is_lacp_owner(mdev))
return 1;
return clamp_t(u8, MLX5_CAP_GEN(mdev, num_lag_ports), 1, MLX5_MAX_PORTS);
}
static inline u16 mlx5_min_rx_wqes(int wq_type, u32 wq_size) static inline u16 mlx5_min_rx_wqes(int wq_type, u32 wq_size)
{ {
switch (wq_type) { switch (wq_type) {
...@@ -705,6 +713,7 @@ struct mlx5e_channel { ...@@ -705,6 +713,7 @@ struct mlx5e_channel {
struct net_device *netdev; struct net_device *netdev;
__be32 mkey_be; __be32 mkey_be;
u8 num_tc; u8 num_tc;
u8 lag_port;
/* XDP_REDIRECT */ /* XDP_REDIRECT */
struct mlx5e_xdpsq xdpsq; struct mlx5e_xdpsq xdpsq;
...@@ -818,7 +827,7 @@ struct mlx5e_priv { ...@@ -818,7 +827,7 @@ struct mlx5e_priv {
struct mlx5e_rq drop_rq; struct mlx5e_rq drop_rq;
struct mlx5e_channels channels; struct mlx5e_channels channels;
u32 tisn[MLX5E_MAX_NUM_TC]; u32 tisn[MLX5_MAX_PORTS][MLX5E_MAX_NUM_TC];
struct mlx5e_rqt indir_rqt; struct mlx5e_rqt indir_rqt;
struct mlx5e_tir indir_tir[MLX5E_NUM_INDIR_TIRS]; struct mlx5e_tir indir_tir[MLX5E_NUM_INDIR_TIRS];
struct mlx5e_tir inner_indir_tir[MLX5E_NUM_INDIR_TIRS]; struct mlx5e_tir inner_indir_tir[MLX5E_NUM_INDIR_TIRS];
...@@ -1056,12 +1065,6 @@ int mlx5e_modify_sq(struct mlx5_core_dev *mdev, u32 sqn, ...@@ -1056,12 +1065,6 @@ int mlx5e_modify_sq(struct mlx5_core_dev *mdev, u32 sqn,
void mlx5e_activate_txqsq(struct mlx5e_txqsq *sq); void mlx5e_activate_txqsq(struct mlx5e_txqsq *sq);
void mlx5e_tx_disable_queue(struct netdev_queue *txq); void mlx5e_tx_disable_queue(struct netdev_queue *txq);
static inline bool mlx5e_tunnel_inner_ft_supported(struct mlx5_core_dev *mdev)
{
return (MLX5_CAP_ETH(mdev, tunnel_stateless_gre) &&
MLX5_CAP_FLOWTABLE_NIC_RX(mdev, ft_field_support.inner_ip_version));
}
static inline bool mlx5_tx_swp_supported(struct mlx5_core_dev *mdev) static inline bool mlx5_tx_swp_supported(struct mlx5_core_dev *mdev)
{ {
return MLX5_CAP_ETH(mdev, swp) && return MLX5_CAP_ETH(mdev, swp) &&
...@@ -1107,6 +1110,7 @@ int mlx5e_create_tis(struct mlx5_core_dev *mdev, void *in, u32 *tisn); ...@@ -1107,6 +1110,7 @@ int mlx5e_create_tis(struct mlx5_core_dev *mdev, void *in, u32 *tisn);
void mlx5e_destroy_tis(struct mlx5_core_dev *mdev, u32 tisn); void mlx5e_destroy_tis(struct mlx5_core_dev *mdev, u32 tisn);
int mlx5e_create_tises(struct mlx5e_priv *priv); int mlx5e_create_tises(struct mlx5e_priv *priv);
void mlx5e_destroy_tises(struct mlx5e_priv *priv);
int mlx5e_update_nic_rx(struct mlx5e_priv *priv); int mlx5e_update_nic_rx(struct mlx5e_priv *priv);
void mlx5e_update_carrier(struct mlx5e_priv *priv); void mlx5e_update_carrier(struct mlx5e_priv *priv);
int mlx5e_close(struct net_device *netdev); int mlx5e_close(struct net_device *netdev);
......
...@@ -95,9 +95,15 @@ struct mlx5e_tirc_config { ...@@ -95,9 +95,15 @@ struct mlx5e_tirc_config {
enum mlx5e_tunnel_types { enum mlx5e_tunnel_types {
MLX5E_TT_IPV4_GRE, MLX5E_TT_IPV4_GRE,
MLX5E_TT_IPV6_GRE, MLX5E_TT_IPV6_GRE,
MLX5E_TT_IPV4_IPIP,
MLX5E_TT_IPV6_IPIP,
MLX5E_TT_IPV4_IPV6,
MLX5E_TT_IPV6_IPV6,
MLX5E_NUM_TUNNEL_TT, MLX5E_NUM_TUNNEL_TT,
}; };
bool mlx5e_tunnel_inner_ft_supported(struct mlx5_core_dev *mdev);
/* L3/L4 traffic type classifier */ /* L3/L4 traffic type classifier */
struct mlx5e_ttc_table { struct mlx5e_ttc_table {
struct mlx5e_flow_table ft; struct mlx5e_flow_table ft;
...@@ -232,5 +238,8 @@ void mlx5e_disable_cvlan_filter(struct mlx5e_priv *priv); ...@@ -232,5 +238,8 @@ void mlx5e_disable_cvlan_filter(struct mlx5e_priv *priv);
int mlx5e_create_flow_steering(struct mlx5e_priv *priv); int mlx5e_create_flow_steering(struct mlx5e_priv *priv);
void mlx5e_destroy_flow_steering(struct mlx5e_priv *priv); void mlx5e_destroy_flow_steering(struct mlx5e_priv *priv);
bool mlx5e_tunnel_proto_supported(struct mlx5_core_dev *mdev, u8 proto_type);
bool mlx5e_any_tunnel_proto_supported(struct mlx5_core_dev *mdev);
#endif /* __MLX5E_FLOW_STEER_H__ */ #endif /* __MLX5E_FLOW_STEER_H__ */
...@@ -1431,7 +1431,7 @@ static __u32 mlx5e_get_wol_supported(struct mlx5_core_dev *mdev) ...@@ -1431,7 +1431,7 @@ static __u32 mlx5e_get_wol_supported(struct mlx5_core_dev *mdev)
return ret; return ret;
} }
static __u32 mlx5e_refomrat_wol_mode_mlx5_to_linux(u8 mode) static __u32 mlx5e_reformat_wol_mode_mlx5_to_linux(u8 mode)
{ {
__u32 ret = 0; __u32 ret = 0;
...@@ -1459,7 +1459,7 @@ static __u32 mlx5e_refomrat_wol_mode_mlx5_to_linux(u8 mode) ...@@ -1459,7 +1459,7 @@ static __u32 mlx5e_refomrat_wol_mode_mlx5_to_linux(u8 mode)
return ret; return ret;
} }
static u8 mlx5e_refomrat_wol_mode_linux_to_mlx5(__u32 mode) static u8 mlx5e_reformat_wol_mode_linux_to_mlx5(__u32 mode)
{ {
u8 ret = 0; u8 ret = 0;
...@@ -1505,7 +1505,7 @@ static void mlx5e_get_wol(struct net_device *netdev, ...@@ -1505,7 +1505,7 @@ static void mlx5e_get_wol(struct net_device *netdev,
if (err) if (err)
return; return;
wol->wolopts = mlx5e_refomrat_wol_mode_mlx5_to_linux(mlx5_wol_mode); wol->wolopts = mlx5e_reformat_wol_mode_mlx5_to_linux(mlx5_wol_mode);
} }
static int mlx5e_set_wol(struct net_device *netdev, struct ethtool_wolinfo *wol) static int mlx5e_set_wol(struct net_device *netdev, struct ethtool_wolinfo *wol)
...@@ -1521,7 +1521,7 @@ static int mlx5e_set_wol(struct net_device *netdev, struct ethtool_wolinfo *wol) ...@@ -1521,7 +1521,7 @@ static int mlx5e_set_wol(struct net_device *netdev, struct ethtool_wolinfo *wol)
if (wol->wolopts & ~wol_supported) if (wol->wolopts & ~wol_supported)
return -EINVAL; return -EINVAL;
mlx5_wol_mode = mlx5e_refomrat_wol_mode_linux_to_mlx5(wol->wolopts); mlx5_wol_mode = mlx5e_reformat_wol_mode_linux_to_mlx5(wol->wolopts);
return mlx5_set_port_wol(mdev, mlx5_wol_mode); return mlx5_set_port_wol(mdev, mlx5_wol_mode);
} }
......
...@@ -747,8 +747,55 @@ static struct mlx5e_etype_proto ttc_tunnel_rules[] = { ...@@ -747,8 +747,55 @@ static struct mlx5e_etype_proto ttc_tunnel_rules[] = {
.etype = ETH_P_IPV6, .etype = ETH_P_IPV6,
.proto = IPPROTO_GRE, .proto = IPPROTO_GRE,
}, },
[MLX5E_TT_IPV4_IPIP] = {
.etype = ETH_P_IP,
.proto = IPPROTO_IPIP,
},
[MLX5E_TT_IPV6_IPIP] = {
.etype = ETH_P_IPV6,
.proto = IPPROTO_IPIP,
},
[MLX5E_TT_IPV4_IPV6] = {
.etype = ETH_P_IP,
.proto = IPPROTO_IPV6,
},
[MLX5E_TT_IPV6_IPV6] = {
.etype = ETH_P_IPV6,
.proto = IPPROTO_IPV6,
},
}; };
bool mlx5e_tunnel_proto_supported(struct mlx5_core_dev *mdev, u8 proto_type)
{
switch (proto_type) {
case IPPROTO_GRE:
return MLX5_CAP_ETH(mdev, tunnel_stateless_gre);
case IPPROTO_IPIP:
case IPPROTO_IPV6:
return MLX5_CAP_ETH(mdev, tunnel_stateless_ip_over_ip);
default:
return false;
}
}
bool mlx5e_any_tunnel_proto_supported(struct mlx5_core_dev *mdev)
{
int tt;
for (tt = 0; tt < MLX5E_NUM_TUNNEL_TT; tt++) {
if (mlx5e_tunnel_proto_supported(mdev, ttc_tunnel_rules[tt].proto))
return true;
}
return false;
}
bool mlx5e_tunnel_inner_ft_supported(struct mlx5_core_dev *mdev)
{
return (mlx5e_any_tunnel_proto_supported(mdev) &&
MLX5_CAP_FLOWTABLE_NIC_RX(mdev, ft_field_support.inner_ip_version));
}
static u8 mlx5e_etype_to_ipv(u16 ethertype) static u8 mlx5e_etype_to_ipv(u16 ethertype)
{ {
if (ethertype == ETH_P_IP) if (ethertype == ETH_P_IP)
...@@ -838,6 +885,9 @@ static int mlx5e_generate_ttc_table_rules(struct mlx5e_priv *priv, ...@@ -838,6 +885,9 @@ static int mlx5e_generate_ttc_table_rules(struct mlx5e_priv *priv,
dest.type = MLX5_FLOW_DESTINATION_TYPE_FLOW_TABLE; dest.type = MLX5_FLOW_DESTINATION_TYPE_FLOW_TABLE;
dest.ft = params->inner_ttc->ft.t; dest.ft = params->inner_ttc->ft.t;
for (tt = 0; tt < MLX5E_NUM_TUNNEL_TT; tt++) { for (tt = 0; tt < MLX5E_NUM_TUNNEL_TT; tt++) {
if (!mlx5e_tunnel_proto_supported(priv->mdev,
ttc_tunnel_rules[tt].proto))
continue;
rules[tt] = mlx5e_generate_ttc_rule(priv, ft, &dest, rules[tt] = mlx5e_generate_ttc_rule(priv, ft, &dest,
ttc_tunnel_rules[tt].etype, ttc_tunnel_rules[tt].etype,
ttc_tunnel_rules[tt].proto); ttc_tunnel_rules[tt].proto);
......
...@@ -1442,7 +1442,7 @@ int mlx5e_open_xdpsq(struct mlx5e_channel *c, struct mlx5e_params *params, ...@@ -1442,7 +1442,7 @@ int mlx5e_open_xdpsq(struct mlx5e_channel *c, struct mlx5e_params *params,
return err; return err;
csp.tis_lst_sz = 1; csp.tis_lst_sz = 1;
csp.tisn = c->priv->tisn[0]; /* tc = 0 */ csp.tisn = c->priv->tisn[c->lag_port][0]; /* tc = 0 */
csp.cqn = sq->cq.mcq.cqn; csp.cqn = sq->cq.mcq.cqn;
csp.wq_ctrl = &sq->wq_ctrl; csp.wq_ctrl = &sq->wq_ctrl;
csp.min_inline_mode = sq->min_inline_mode; csp.min_inline_mode = sq->min_inline_mode;
...@@ -1692,7 +1692,7 @@ static int mlx5e_open_sqs(struct mlx5e_channel *c, ...@@ -1692,7 +1692,7 @@ static int mlx5e_open_sqs(struct mlx5e_channel *c,
for (tc = 0; tc < params->num_tc; tc++) { for (tc = 0; tc < params->num_tc; tc++) {
int txq_ix = c->ix + tc * priv->max_nch; int txq_ix = c->ix + tc * priv->max_nch;
err = mlx5e_open_txqsq(c, c->priv->tisn[tc], txq_ix, err = mlx5e_open_txqsq(c, c->priv->tisn[c->lag_port][tc], txq_ix,
params, &cparam->sq, &c->sq[tc], tc); params, &cparam->sq, &c->sq[tc], tc);
if (err) if (err)
goto err_close_sqs; goto err_close_sqs;
...@@ -1926,6 +1926,13 @@ static void mlx5e_close_queues(struct mlx5e_channel *c) ...@@ -1926,6 +1926,13 @@ static void mlx5e_close_queues(struct mlx5e_channel *c)
mlx5e_close_cq(&c->icosq.cq); mlx5e_close_cq(&c->icosq.cq);
} }
static u8 mlx5e_enumerate_lag_port(struct mlx5_core_dev *mdev, int ix)
{
u16 port_aff_bias = mlx5_core_is_pf(mdev) ? 0 : MLX5_CAP_GEN(mdev, vhca_id);
return (ix + port_aff_bias) % mlx5e_get_num_lag_ports(mdev);
}
static int mlx5e_open_channel(struct mlx5e_priv *priv, int ix, static int mlx5e_open_channel(struct mlx5e_priv *priv, int ix,
struct mlx5e_params *params, struct mlx5e_params *params,
struct mlx5e_channel_param *cparam, struct mlx5e_channel_param *cparam,
...@@ -1960,6 +1967,7 @@ static int mlx5e_open_channel(struct mlx5e_priv *priv, int ix, ...@@ -1960,6 +1967,7 @@ static int mlx5e_open_channel(struct mlx5e_priv *priv, int ix,
c->xdp = !!params->xdp_prog; c->xdp = !!params->xdp_prog;
c->stats = &priv->channel_stats[ix].ch; c->stats = &priv->channel_stats[ix].ch;
c->irq_desc = irq_to_desc(irq); c->irq_desc = irq_to_desc(irq);
c->lag_port = mlx5e_enumerate_lag_port(priv->mdev, ix);
err = mlx5e_alloc_xps_cpumask(c, params); err = mlx5e_alloc_xps_cpumask(c, params);
if (err) if (err)
...@@ -3179,39 +3187,58 @@ void mlx5e_destroy_tis(struct mlx5_core_dev *mdev, u32 tisn) ...@@ -3179,39 +3187,58 @@ void mlx5e_destroy_tis(struct mlx5_core_dev *mdev, u32 tisn)
mlx5_core_destroy_tis(mdev, tisn); mlx5_core_destroy_tis(mdev, tisn);
} }
void mlx5e_destroy_tises(struct mlx5e_priv *priv)
{
int tc, i;
for (i = 0; i < mlx5e_get_num_lag_ports(priv->mdev); i++)
for (tc = 0; tc < priv->profile->max_tc; tc++)
mlx5e_destroy_tis(priv->mdev, priv->tisn[i][tc]);
}
static bool mlx5e_lag_should_assign_affinity(struct mlx5_core_dev *mdev)
{
return MLX5_CAP_GEN(mdev, lag_tx_port_affinity) && mlx5e_get_num_lag_ports(mdev) > 1;
}
int mlx5e_create_tises(struct mlx5e_priv *priv) int mlx5e_create_tises(struct mlx5e_priv *priv)
{ {
int tc, i;
int err; int err;
int tc;
for (tc = 0; tc < priv->profile->max_tc; tc++) { for (i = 0; i < mlx5e_get_num_lag_ports(priv->mdev); i++) {
u32 in[MLX5_ST_SZ_DW(create_tis_in)] = {}; for (tc = 0; tc < priv->profile->max_tc; tc++) {
void *tisc; u32 in[MLX5_ST_SZ_DW(create_tis_in)] = {};
void *tisc;
tisc = MLX5_ADDR_OF(create_tis_in, in, ctx); tisc = MLX5_ADDR_OF(create_tis_in, in, ctx);
MLX5_SET(tisc, tisc, prio, tc << 1); MLX5_SET(tisc, tisc, prio, tc << 1);
err = mlx5e_create_tis(priv->mdev, in, &priv->tisn[tc]); if (mlx5e_lag_should_assign_affinity(priv->mdev))
if (err) MLX5_SET(tisc, tisc, lag_tx_port_affinity, i + 1);
goto err_close_tises;
err = mlx5e_create_tis(priv->mdev, in, &priv->tisn[i][tc]);
if (err)
goto err_close_tises;
}
} }
return 0; return 0;
err_close_tises: err_close_tises:
for (tc--; tc >= 0; tc--) for (; i >= 0; i--) {
mlx5e_destroy_tis(priv->mdev, priv->tisn[tc]); for (tc--; tc >= 0; tc--)
mlx5e_destroy_tis(priv->mdev, priv->tisn[i][tc]);
tc = priv->profile->max_tc;
}
return err; return err;
} }
static void mlx5e_cleanup_nic_tx(struct mlx5e_priv *priv) static void mlx5e_cleanup_nic_tx(struct mlx5e_priv *priv)
{ {
int tc; mlx5e_destroy_tises(priv);
for (tc = 0; tc < priv->profile->max_tc; tc++)
mlx5e_destroy_tis(priv->mdev, priv->tisn[tc]);
} }
static void mlx5e_build_indir_tir_ctx_common(struct mlx5e_priv *priv, static void mlx5e_build_indir_tir_ctx_common(struct mlx5e_priv *priv,
...@@ -4216,6 +4243,8 @@ static netdev_features_t mlx5e_tunnel_features_check(struct mlx5e_priv *priv, ...@@ -4216,6 +4243,8 @@ static netdev_features_t mlx5e_tunnel_features_check(struct mlx5e_priv *priv,
switch (proto) { switch (proto) {
case IPPROTO_GRE: case IPPROTO_GRE:
case IPPROTO_IPIP:
case IPPROTO_IPV6:
return features; return features;
case IPPROTO_UDP: case IPPROTO_UDP:
udph = udp_hdr(skb); udph = udp_hdr(skb);
...@@ -4852,7 +4881,7 @@ static void mlx5e_build_nic_netdev(struct net_device *netdev) ...@@ -4852,7 +4881,7 @@ static void mlx5e_build_nic_netdev(struct net_device *netdev)
netdev->hw_features |= NETIF_F_HW_VLAN_STAG_TX; netdev->hw_features |= NETIF_F_HW_VLAN_STAG_TX;
if (mlx5_vxlan_allowed(mdev->vxlan) || mlx5_geneve_tx_allowed(mdev) || if (mlx5_vxlan_allowed(mdev->vxlan) || mlx5_geneve_tx_allowed(mdev) ||
MLX5_CAP_ETH(mdev, tunnel_stateless_gre)) { mlx5e_any_tunnel_proto_supported(mdev)) {
netdev->hw_enc_features |= NETIF_F_HW_CSUM; netdev->hw_enc_features |= NETIF_F_HW_CSUM;
netdev->hw_enc_features |= NETIF_F_TSO; netdev->hw_enc_features |= NETIF_F_TSO;
netdev->hw_enc_features |= NETIF_F_TSO6; netdev->hw_enc_features |= NETIF_F_TSO6;
...@@ -4867,7 +4896,7 @@ static void mlx5e_build_nic_netdev(struct net_device *netdev) ...@@ -4867,7 +4896,7 @@ static void mlx5e_build_nic_netdev(struct net_device *netdev)
netdev->gso_partial_features = NETIF_F_GSO_UDP_TUNNEL_CSUM; netdev->gso_partial_features = NETIF_F_GSO_UDP_TUNNEL_CSUM;
} }
if (MLX5_CAP_ETH(mdev, tunnel_stateless_gre)) { if (mlx5e_tunnel_proto_supported(mdev, IPPROTO_GRE)) {
netdev->hw_features |= NETIF_F_GSO_GRE | netdev->hw_features |= NETIF_F_GSO_GRE |
NETIF_F_GSO_GRE_CSUM; NETIF_F_GSO_GRE_CSUM;
netdev->hw_enc_features |= NETIF_F_GSO_GRE | netdev->hw_enc_features |= NETIF_F_GSO_GRE |
...@@ -4876,6 +4905,15 @@ static void mlx5e_build_nic_netdev(struct net_device *netdev) ...@@ -4876,6 +4905,15 @@ static void mlx5e_build_nic_netdev(struct net_device *netdev)
NETIF_F_GSO_GRE_CSUM; NETIF_F_GSO_GRE_CSUM;
} }
if (mlx5e_tunnel_proto_supported(mdev, IPPROTO_IPIP)) {
netdev->hw_features |= NETIF_F_GSO_IPXIP4 |
NETIF_F_GSO_IPXIP6;
netdev->hw_enc_features |= NETIF_F_GSO_IPXIP4 |
NETIF_F_GSO_IPXIP6;
netdev->gso_partial_features |= NETIF_F_GSO_IPXIP4 |
NETIF_F_GSO_IPXIP6;
}
netdev->hw_features |= NETIF_F_GSO_PARTIAL; netdev->hw_features |= NETIF_F_GSO_PARTIAL;
netdev->gso_partial_features |= NETIF_F_GSO_UDP_L4; netdev->gso_partial_features |= NETIF_F_GSO_UDP_L4;
netdev->hw_features |= NETIF_F_GSO_UDP_L4; netdev->hw_features |= NETIF_F_GSO_UDP_L4;
......
...@@ -1621,7 +1621,7 @@ static int mlx5e_init_rep_tx(struct mlx5e_priv *priv) ...@@ -1621,7 +1621,7 @@ static int mlx5e_init_rep_tx(struct mlx5e_priv *priv)
{ {
struct mlx5e_rep_priv *rpriv = priv->ppriv; struct mlx5e_rep_priv *rpriv = priv->ppriv;
struct mlx5_rep_uplink_priv *uplink_priv; struct mlx5_rep_uplink_priv *uplink_priv;
int tc, err; int err;
err = mlx5e_create_tises(priv); err = mlx5e_create_tises(priv);
if (err) { if (err) {
...@@ -1657,18 +1657,15 @@ static int mlx5e_init_rep_tx(struct mlx5e_priv *priv) ...@@ -1657,18 +1657,15 @@ static int mlx5e_init_rep_tx(struct mlx5e_priv *priv)
tc_esw_cleanup: tc_esw_cleanup:
mlx5e_tc_esw_cleanup(&uplink_priv->tc_ht); mlx5e_tc_esw_cleanup(&uplink_priv->tc_ht);
destroy_tises: destroy_tises:
for (tc = 0; tc < priv->profile->max_tc; tc++) mlx5e_destroy_tises(priv);
mlx5e_destroy_tis(priv->mdev, priv->tisn[tc]);
return err; return err;
} }
static void mlx5e_cleanup_rep_tx(struct mlx5e_priv *priv) static void mlx5e_cleanup_rep_tx(struct mlx5e_priv *priv)
{ {
struct mlx5e_rep_priv *rpriv = priv->ppriv; struct mlx5e_rep_priv *rpriv = priv->ppriv;
int tc;
for (tc = 0; tc < priv->profile->max_tc; tc++) mlx5e_destroy_tises(priv);
mlx5e_destroy_tis(priv->mdev, priv->tisn[tc]);
if (rpriv->rep->vport == MLX5_VPORT_UPLINK) { if (rpriv->rep->vport == MLX5_VPORT_UPLINK) {
/* clean indirect TC block notifications */ /* clean indirect TC block notifications */
......
...@@ -369,17 +369,27 @@ static void mlx5e_grp_q_update_stats(struct mlx5e_priv *priv) ...@@ -369,17 +369,27 @@ static void mlx5e_grp_q_update_stats(struct mlx5e_priv *priv)
} }
#define VNIC_ENV_OFF(c) MLX5_BYTE_OFF(query_vnic_env_out, c) #define VNIC_ENV_OFF(c) MLX5_BYTE_OFF(query_vnic_env_out, c)
static const struct counter_desc vnic_env_stats_desc[] = { static const struct counter_desc vnic_env_stats_steer_desc[] = {
{ "rx_steer_missed_packets", { "rx_steer_missed_packets",
VNIC_ENV_OFF(vport_env.nic_receive_steering_discard) }, VNIC_ENV_OFF(vport_env.nic_receive_steering_discard) },
}; };
#define NUM_VNIC_ENV_COUNTERS ARRAY_SIZE(vnic_env_stats_desc) static const struct counter_desc vnic_env_stats_dev_oob_desc[] = {
{ "dev_internal_queue_oob",
VNIC_ENV_OFF(vport_env.internal_rq_out_of_buffer) },
};
#define NUM_VNIC_ENV_STEER_COUNTERS(dev) \
(MLX5_CAP_GEN(dev, nic_receive_steering_discard) ? \
ARRAY_SIZE(vnic_env_stats_steer_desc) : 0)
#define NUM_VNIC_ENV_DEV_OOB_COUNTERS(dev) \
(MLX5_CAP_GEN(dev, vnic_env_int_rq_oob) ? \
ARRAY_SIZE(vnic_env_stats_dev_oob_desc) : 0)
static int mlx5e_grp_vnic_env_get_num_stats(struct mlx5e_priv *priv) static int mlx5e_grp_vnic_env_get_num_stats(struct mlx5e_priv *priv)
{ {
return MLX5_CAP_GEN(priv->mdev, nic_receive_steering_discard) ? return NUM_VNIC_ENV_STEER_COUNTERS(priv->mdev) +
NUM_VNIC_ENV_COUNTERS : 0; NUM_VNIC_ENV_DEV_OOB_COUNTERS(priv->mdev);
} }
static int mlx5e_grp_vnic_env_fill_strings(struct mlx5e_priv *priv, u8 *data, static int mlx5e_grp_vnic_env_fill_strings(struct mlx5e_priv *priv, u8 *data,
...@@ -387,12 +397,13 @@ static int mlx5e_grp_vnic_env_fill_strings(struct mlx5e_priv *priv, u8 *data, ...@@ -387,12 +397,13 @@ static int mlx5e_grp_vnic_env_fill_strings(struct mlx5e_priv *priv, u8 *data,
{ {
int i; int i;
if (!MLX5_CAP_GEN(priv->mdev, nic_receive_steering_discard)) for (i = 0; i < NUM_VNIC_ENV_STEER_COUNTERS(priv->mdev); i++)
return idx; strcpy(data + (idx++) * ETH_GSTRING_LEN,
vnic_env_stats_steer_desc[i].format);
for (i = 0; i < NUM_VNIC_ENV_COUNTERS; i++) for (i = 0; i < NUM_VNIC_ENV_DEV_OOB_COUNTERS(priv->mdev); i++)
strcpy(data + (idx++) * ETH_GSTRING_LEN, strcpy(data + (idx++) * ETH_GSTRING_LEN,
vnic_env_stats_desc[i].format); vnic_env_stats_dev_oob_desc[i].format);
return idx; return idx;
} }
...@@ -401,12 +412,13 @@ static int mlx5e_grp_vnic_env_fill_stats(struct mlx5e_priv *priv, u64 *data, ...@@ -401,12 +412,13 @@ static int mlx5e_grp_vnic_env_fill_stats(struct mlx5e_priv *priv, u64 *data,
{ {
int i; int i;
if (!MLX5_CAP_GEN(priv->mdev, nic_receive_steering_discard)) for (i = 0; i < NUM_VNIC_ENV_STEER_COUNTERS(priv->mdev); i++)
return idx;
for (i = 0; i < NUM_VNIC_ENV_COUNTERS; i++)
data[idx++] = MLX5E_READ_CTR64_BE(priv->stats.vnic.query_vnic_env_out, data[idx++] = MLX5E_READ_CTR64_BE(priv->stats.vnic.query_vnic_env_out,
vnic_env_stats_desc, i); vnic_env_stats_steer_desc, i);
for (i = 0; i < NUM_VNIC_ENV_DEV_OOB_COUNTERS(priv->mdev); i++)
data[idx++] = MLX5E_READ_CTR32_BE(priv->stats.vnic.query_vnic_env_out,
vnic_env_stats_dev_oob_desc, i);
return idx; return idx;
} }
......
...@@ -324,10 +324,13 @@ create_map_eq(struct mlx5_core_dev *dev, struct mlx5_eq *eq, ...@@ -324,10 +324,13 @@ create_map_eq(struct mlx5_core_dev *dev, struct mlx5_eq *eq,
/** /**
* mlx5_eq_enable - Enable EQ for receiving EQEs * mlx5_eq_enable - Enable EQ for receiving EQEs
* @dev - Device which owns the eq * @dev : Device which owns the eq
* @eq - EQ to enable * @eq : EQ to enable
* @nb - notifier call block * @nb : Notifier call block
* mlx5_eq_enable - must be called after EQ is created in device. *
* Must be called after EQ is created in device.
*
* @return: 0 if no error
*/ */
int mlx5_eq_enable(struct mlx5_core_dev *dev, struct mlx5_eq *eq, int mlx5_eq_enable(struct mlx5_core_dev *dev, struct mlx5_eq *eq,
struct notifier_block *nb) struct notifier_block *nb)
...@@ -344,11 +347,12 @@ int mlx5_eq_enable(struct mlx5_core_dev *dev, struct mlx5_eq *eq, ...@@ -344,11 +347,12 @@ int mlx5_eq_enable(struct mlx5_core_dev *dev, struct mlx5_eq *eq,
EXPORT_SYMBOL(mlx5_eq_enable); EXPORT_SYMBOL(mlx5_eq_enable);
/** /**
* mlx5_eq_disable - Enable EQ for receiving EQEs * mlx5_eq_disable - Disable EQ for receiving EQEs
* @dev - Device which owns the eq * @dev : Device which owns the eq
* @eq - EQ to disable * @eq : EQ to disable
* @nb - notifier call block * @nb : Notifier call block
* mlx5_eq_disable - must be called before EQ is destroyed. *
* Must be called before EQ is destroyed.
*/ */
void mlx5_eq_disable(struct mlx5_core_dev *dev, struct mlx5_eq *eq, void mlx5_eq_disable(struct mlx5_core_dev *dev, struct mlx5_eq *eq,
struct notifier_block *nb) struct notifier_block *nb)
......
...@@ -1413,7 +1413,7 @@ static int esw_vport_egress_config(struct mlx5_eswitch *esw, ...@@ -1413,7 +1413,7 @@ static int esw_vport_egress_config(struct mlx5_eswitch *esw,
static bool element_type_supported(struct mlx5_eswitch *esw, int type) static bool element_type_supported(struct mlx5_eswitch *esw, int type)
{ {
struct mlx5_core_dev *dev = esw->dev = esw->dev; const struct mlx5_core_dev *dev = esw->dev;
switch (type) { switch (type) {
case SCHEDULING_CONTEXT_ELEMENT_TYPE_TSAR: case SCHEDULING_CONTEXT_ELEMENT_TYPE_TSAR:
......
...@@ -182,7 +182,7 @@ static int mlx5_cmd_create_flow_table(struct mlx5_flow_root_namespace *ns, ...@@ -182,7 +182,7 @@ static int mlx5_cmd_create_flow_table(struct mlx5_flow_root_namespace *ns,
} else { } else {
MLX5_SET(create_flow_table_in, in, MLX5_SET(create_flow_table_in, in,
flow_table_context.table_miss_action, flow_table_context.table_miss_action,
ns->def_miss_action); ft->def_miss_action);
} }
break; break;
...@@ -262,7 +262,7 @@ static int mlx5_cmd_modify_flow_table(struct mlx5_flow_root_namespace *ns, ...@@ -262,7 +262,7 @@ static int mlx5_cmd_modify_flow_table(struct mlx5_flow_root_namespace *ns,
} else { } else {
MLX5_SET(modify_flow_table_in, in, MLX5_SET(modify_flow_table_in, in,
flow_table_context.table_miss_action, flow_table_context.table_miss_action,
ns->def_miss_action); ft->def_miss_action);
} }
} }
......
...@@ -60,7 +60,8 @@ ...@@ -60,7 +60,8 @@
ADD_PRIO(num_prios_val, 0, num_levels_val, {},\ ADD_PRIO(num_prios_val, 0, num_levels_val, {},\
__VA_ARGS__)\ __VA_ARGS__)\
#define ADD_NS(...) {.type = FS_TYPE_NAMESPACE,\ #define ADD_NS(def_miss_act, ...) {.type = FS_TYPE_NAMESPACE, \
.def_miss_action = def_miss_act,\
.children = (struct init_tree_node[]) {__VA_ARGS__},\ .children = (struct init_tree_node[]) {__VA_ARGS__},\
.ar_size = INIT_TREE_NODE_ARRAY_SIZE(__VA_ARGS__) \ .ar_size = INIT_TREE_NODE_ARRAY_SIZE(__VA_ARGS__) \
} }
...@@ -131,33 +132,41 @@ static struct init_tree_node { ...@@ -131,33 +132,41 @@ static struct init_tree_node {
int num_leaf_prios; int num_leaf_prios;
int prio; int prio;
int num_levels; int num_levels;
enum mlx5_flow_table_miss_action def_miss_action;
} root_fs = { } root_fs = {
.type = FS_TYPE_NAMESPACE, .type = FS_TYPE_NAMESPACE,
.ar_size = 7, .ar_size = 7,
.children = (struct init_tree_node[]) { .children = (struct init_tree_node[]){
ADD_PRIO(0, BY_PASS_MIN_LEVEL, 0, ADD_PRIO(0, BY_PASS_MIN_LEVEL, 0, FS_CHAINING_CAPS,
FS_CHAINING_CAPS, ADD_NS(MLX5_FLOW_TABLE_MISS_ACTION_DEF,
ADD_NS(ADD_MULTIPLE_PRIO(MLX5_BY_PASS_NUM_PRIOS, ADD_MULTIPLE_PRIO(MLX5_BY_PASS_NUM_PRIOS,
BY_PASS_PRIO_NUM_LEVELS))), BY_PASS_PRIO_NUM_LEVELS))),
ADD_PRIO(0, LAG_MIN_LEVEL, 0, ADD_PRIO(0, LAG_MIN_LEVEL, 0, FS_CHAINING_CAPS,
FS_CHAINING_CAPS, ADD_NS(MLX5_FLOW_TABLE_MISS_ACTION_DEF,
ADD_NS(ADD_MULTIPLE_PRIO(LAG_NUM_PRIOS, ADD_MULTIPLE_PRIO(LAG_NUM_PRIOS,
LAG_PRIO_NUM_LEVELS))), LAG_PRIO_NUM_LEVELS))),
ADD_PRIO(0, OFFLOADS_MIN_LEVEL, 0, {}, ADD_PRIO(0, OFFLOADS_MIN_LEVEL, 0, {},
ADD_NS(ADD_MULTIPLE_PRIO(OFFLOADS_NUM_PRIOS, OFFLOADS_MAX_FT))), ADD_NS(MLX5_FLOW_TABLE_MISS_ACTION_DEF,
ADD_PRIO(0, ETHTOOL_MIN_LEVEL, 0, ADD_MULTIPLE_PRIO(OFFLOADS_NUM_PRIOS,
FS_CHAINING_CAPS, OFFLOADS_MAX_FT))),
ADD_NS(ADD_MULTIPLE_PRIO(ETHTOOL_NUM_PRIOS, ADD_PRIO(0, ETHTOOL_MIN_LEVEL, 0, FS_CHAINING_CAPS,
ETHTOOL_PRIO_NUM_LEVELS))), ADD_NS(MLX5_FLOW_TABLE_MISS_ACTION_DEF,
ADD_PRIO(0, KERNEL_MIN_LEVEL, 0, {}, ADD_MULTIPLE_PRIO(ETHTOOL_NUM_PRIOS,
ADD_NS(ADD_MULTIPLE_PRIO(KERNEL_NIC_TC_NUM_PRIOS, KERNEL_NIC_TC_NUM_LEVELS), ETHTOOL_PRIO_NUM_LEVELS))),
ADD_MULTIPLE_PRIO(KERNEL_NIC_NUM_PRIOS, ADD_PRIO(0, KERNEL_MIN_LEVEL, 0, {},
KERNEL_NIC_PRIO_NUM_LEVELS))), ADD_NS(MLX5_FLOW_TABLE_MISS_ACTION_DEF,
ADD_PRIO(0, BY_PASS_MIN_LEVEL, 0, ADD_MULTIPLE_PRIO(KERNEL_NIC_TC_NUM_PRIOS,
FS_CHAINING_CAPS, KERNEL_NIC_TC_NUM_LEVELS),
ADD_NS(ADD_MULTIPLE_PRIO(LEFTOVERS_NUM_PRIOS, LEFTOVERS_NUM_LEVELS))), ADD_MULTIPLE_PRIO(KERNEL_NIC_NUM_PRIOS,
ADD_PRIO(0, ANCHOR_MIN_LEVEL, 0, {}, KERNEL_NIC_PRIO_NUM_LEVELS))),
ADD_NS(ADD_MULTIPLE_PRIO(ANCHOR_NUM_PRIOS, ANCHOR_NUM_LEVELS))), ADD_PRIO(0, BY_PASS_MIN_LEVEL, 0, FS_CHAINING_CAPS,
ADD_NS(MLX5_FLOW_TABLE_MISS_ACTION_DEF,
ADD_MULTIPLE_PRIO(LEFTOVERS_NUM_PRIOS,
LEFTOVERS_NUM_LEVELS))),
ADD_PRIO(0, ANCHOR_MIN_LEVEL, 0, {},
ADD_NS(MLX5_FLOW_TABLE_MISS_ACTION_DEF,
ADD_MULTIPLE_PRIO(ANCHOR_NUM_PRIOS,
ANCHOR_NUM_LEVELS))),
} }
}; };
...@@ -167,8 +176,29 @@ static struct init_tree_node egress_root_fs = { ...@@ -167,8 +176,29 @@ static struct init_tree_node egress_root_fs = {
.children = (struct init_tree_node[]) { .children = (struct init_tree_node[]) {
ADD_PRIO(0, MLX5_BY_PASS_NUM_PRIOS, 0, ADD_PRIO(0, MLX5_BY_PASS_NUM_PRIOS, 0,
FS_CHAINING_CAPS_EGRESS, FS_CHAINING_CAPS_EGRESS,
ADD_NS(ADD_MULTIPLE_PRIO(MLX5_BY_PASS_NUM_PRIOS, ADD_NS(MLX5_FLOW_TABLE_MISS_ACTION_DEF,
ADD_MULTIPLE_PRIO(MLX5_BY_PASS_NUM_PRIOS,
BY_PASS_PRIO_NUM_LEVELS))),
}
};
#define RDMA_RX_BYPASS_PRIO 0
#define RDMA_RX_KERNEL_PRIO 1
static struct init_tree_node rdma_rx_root_fs = {
.type = FS_TYPE_NAMESPACE,
.ar_size = 2,
.children = (struct init_tree_node[]) {
[RDMA_RX_BYPASS_PRIO] =
ADD_PRIO(0, MLX5_BY_PASS_NUM_REGULAR_PRIOS, 0,
FS_CHAINING_CAPS,
ADD_NS(MLX5_FLOW_TABLE_MISS_ACTION_DEF,
ADD_MULTIPLE_PRIO(MLX5_BY_PASS_NUM_REGULAR_PRIOS,
BY_PASS_PRIO_NUM_LEVELS))), BY_PASS_PRIO_NUM_LEVELS))),
[RDMA_RX_KERNEL_PRIO] =
ADD_PRIO(0, MLX5_BY_PASS_NUM_REGULAR_PRIOS + 1, 0,
FS_CHAINING_CAPS,
ADD_NS(MLX5_FLOW_TABLE_MISS_ACTION_SWITCH_DOMAIN,
ADD_MULTIPLE_PRIO(1, 1))),
} }
}; };
...@@ -1014,6 +1044,7 @@ static struct mlx5_flow_table *__mlx5_create_flow_table(struct mlx5_flow_namespa ...@@ -1014,6 +1044,7 @@ static struct mlx5_flow_table *__mlx5_create_flow_table(struct mlx5_flow_namespa
tree_init_node(&ft->node, del_hw_flow_table, del_sw_flow_table); tree_init_node(&ft->node, del_hw_flow_table, del_sw_flow_table);
log_table_sz = ft->max_fte ? ilog2(ft->max_fte) : 0; log_table_sz = ft->max_fte ? ilog2(ft->max_fte) : 0;
next_ft = find_next_chained_ft(fs_prio); next_ft = find_next_chained_ft(fs_prio);
ft->def_miss_action = ns->def_miss_action;
err = root->cmds->create_flow_table(root, ft, log_table_sz, next_ft); err = root->cmds->create_flow_table(root, ft, log_table_sz, next_ft);
if (err) if (err)
goto free_ft; goto free_ft;
...@@ -2056,16 +2087,18 @@ struct mlx5_flow_namespace *mlx5_get_flow_namespace(struct mlx5_core_dev *dev, ...@@ -2056,16 +2087,18 @@ struct mlx5_flow_namespace *mlx5_get_flow_namespace(struct mlx5_core_dev *dev,
if (steering->sniffer_tx_root_ns) if (steering->sniffer_tx_root_ns)
return &steering->sniffer_tx_root_ns->ns; return &steering->sniffer_tx_root_ns->ns;
return NULL; return NULL;
case MLX5_FLOW_NAMESPACE_RDMA_RX:
if (steering->rdma_rx_root_ns)
return &steering->rdma_rx_root_ns->ns;
return NULL;
default: default:
break; break;
} }
if (type == MLX5_FLOW_NAMESPACE_EGRESS) { if (type == MLX5_FLOW_NAMESPACE_EGRESS) {
root_ns = steering->egress_root_ns; root_ns = steering->egress_root_ns;
} else if (type == MLX5_FLOW_NAMESPACE_RDMA_RX) {
root_ns = steering->rdma_rx_root_ns;
prio = RDMA_RX_BYPASS_PRIO;
} else if (type == MLX5_FLOW_NAMESPACE_RDMA_RX_KERNEL) {
root_ns = steering->rdma_rx_root_ns;
prio = RDMA_RX_KERNEL_PRIO;
} else { /* Must be NIC RX */ } else { /* Must be NIC RX */
root_ns = steering->root_ns; root_ns = steering->root_ns;
prio = type; prio = type;
...@@ -2155,7 +2188,8 @@ static struct mlx5_flow_namespace *fs_init_namespace(struct mlx5_flow_namespace ...@@ -2155,7 +2188,8 @@ static struct mlx5_flow_namespace *fs_init_namespace(struct mlx5_flow_namespace
return ns; return ns;
} }
static struct mlx5_flow_namespace *fs_create_namespace(struct fs_prio *prio) static struct mlx5_flow_namespace *fs_create_namespace(struct fs_prio *prio,
int def_miss_act)
{ {
struct mlx5_flow_namespace *ns; struct mlx5_flow_namespace *ns;
...@@ -2164,6 +2198,7 @@ static struct mlx5_flow_namespace *fs_create_namespace(struct fs_prio *prio) ...@@ -2164,6 +2198,7 @@ static struct mlx5_flow_namespace *fs_create_namespace(struct fs_prio *prio)
return ERR_PTR(-ENOMEM); return ERR_PTR(-ENOMEM);
fs_init_namespace(ns); fs_init_namespace(ns);
ns->def_miss_action = def_miss_act;
tree_init_node(&ns->node, NULL, del_sw_ns); tree_init_node(&ns->node, NULL, del_sw_ns);
tree_add_node(&ns->node, &prio->node); tree_add_node(&ns->node, &prio->node);
list_add_tail(&ns->node.list, &prio->node.children); list_add_tail(&ns->node.list, &prio->node.children);
...@@ -2230,7 +2265,7 @@ static int init_root_tree_recursive(struct mlx5_flow_steering *steering, ...@@ -2230,7 +2265,7 @@ static int init_root_tree_recursive(struct mlx5_flow_steering *steering,
base = &fs_prio->node; base = &fs_prio->node;
} else if (init_node->type == FS_TYPE_NAMESPACE) { } else if (init_node->type == FS_TYPE_NAMESPACE) {
fs_get_obj(fs_prio, fs_parent_node); fs_get_obj(fs_prio, fs_parent_node);
fs_ns = fs_create_namespace(fs_prio); fs_ns = fs_create_namespace(fs_prio, init_node->def_miss_action);
if (IS_ERR(fs_ns)) if (IS_ERR(fs_ns))
return PTR_ERR(fs_ns); return PTR_ERR(fs_ns);
base = &fs_ns->node; base = &fs_ns->node;
...@@ -2494,18 +2529,25 @@ static int init_sniffer_rx_root_ns(struct mlx5_flow_steering *steering) ...@@ -2494,18 +2529,25 @@ static int init_sniffer_rx_root_ns(struct mlx5_flow_steering *steering)
static int init_rdma_rx_root_ns(struct mlx5_flow_steering *steering) static int init_rdma_rx_root_ns(struct mlx5_flow_steering *steering)
{ {
struct fs_prio *prio; int err;
steering->rdma_rx_root_ns = create_root_ns(steering, FS_FT_RDMA_RX); steering->rdma_rx_root_ns = create_root_ns(steering, FS_FT_RDMA_RX);
if (!steering->rdma_rx_root_ns) if (!steering->rdma_rx_root_ns)
return -ENOMEM; return -ENOMEM;
steering->rdma_rx_root_ns->def_miss_action = err = init_root_tree(steering, &rdma_rx_root_fs,
MLX5_FLOW_TABLE_MISS_ACTION_SWITCH_DOMAIN; &steering->rdma_rx_root_ns->ns.node);
if (err)
goto out_err;
/* Create single prio */ set_prio_attrs(steering->rdma_rx_root_ns);
prio = fs_create_prio(&steering->rdma_rx_root_ns->ns, 0, 1);
return PTR_ERR_OR_ZERO(prio); return 0;
out_err:
cleanup_root_ns(steering->rdma_rx_root_ns);
steering->rdma_rx_root_ns = NULL;
return err;
} }
static int init_fdb_root_ns(struct mlx5_flow_steering *steering) static int init_fdb_root_ns(struct mlx5_flow_steering *steering)
{ {
...@@ -2543,7 +2585,7 @@ static int init_fdb_root_ns(struct mlx5_flow_steering *steering) ...@@ -2543,7 +2585,7 @@ static int init_fdb_root_ns(struct mlx5_flow_steering *steering)
} }
for (chain = 0; chain <= FDB_MAX_CHAIN; chain++) { for (chain = 0; chain <= FDB_MAX_CHAIN; chain++) {
ns = fs_create_namespace(maj_prio); ns = fs_create_namespace(maj_prio, MLX5_FLOW_TABLE_MISS_ACTION_DEF);
if (IS_ERR(ns)) { if (IS_ERR(ns)) {
err = PTR_ERR(ns); err = PTR_ERR(ns);
goto out_err; goto out_err;
......
...@@ -145,6 +145,7 @@ struct mlx5_flow_table { ...@@ -145,6 +145,7 @@ struct mlx5_flow_table {
struct list_head fwd_rules; struct list_head fwd_rules;
u32 flags; u32 flags;
struct rhltable fgs_hash; struct rhltable fgs_hash;
enum mlx5_flow_table_miss_action def_miss_action;
}; };
struct mlx5_ft_underlay_qp { struct mlx5_ft_underlay_qp {
...@@ -191,6 +192,7 @@ struct fs_prio { ...@@ -191,6 +192,7 @@ struct fs_prio {
struct mlx5_flow_namespace { struct mlx5_flow_namespace {
/* parent == NULL => root ns */ /* parent == NULL => root ns */
struct fs_node node; struct fs_node node;
enum mlx5_flow_table_miss_action def_miss_action;
}; };
struct mlx5_flow_group_mask { struct mlx5_flow_group_mask {
...@@ -219,7 +221,6 @@ struct mlx5_flow_root_namespace { ...@@ -219,7 +221,6 @@ struct mlx5_flow_root_namespace {
struct mutex chain_lock; struct mutex chain_lock;
struct list_head underlay_qpns; struct list_head underlay_qpns;
const struct mlx5_flow_cmds *cmds; const struct mlx5_flow_cmds *cmds;
enum mlx5_flow_table_miss_action def_miss_action;
}; };
int mlx5_init_fc_stats(struct mlx5_core_dev *dev); int mlx5_init_fc_stats(struct mlx5_core_dev *dev);
......
...@@ -279,7 +279,7 @@ static int mlx5i_init_tx(struct mlx5e_priv *priv) ...@@ -279,7 +279,7 @@ static int mlx5i_init_tx(struct mlx5e_priv *priv)
return err; return err;
} }
err = mlx5i_create_tis(priv->mdev, ipriv->qp.qpn, &priv->tisn[0]); err = mlx5i_create_tis(priv->mdev, ipriv->qp.qpn, &priv->tisn[0][0]);
if (err) { if (err) {
mlx5_core_warn(priv->mdev, "create tis failed, %d\n", err); mlx5_core_warn(priv->mdev, "create tis failed, %d\n", err);
goto err_destroy_underlay_qp; goto err_destroy_underlay_qp;
...@@ -296,7 +296,7 @@ static void mlx5i_cleanup_tx(struct mlx5e_priv *priv) ...@@ -296,7 +296,7 @@ static void mlx5i_cleanup_tx(struct mlx5e_priv *priv)
{ {
struct mlx5i_priv *ipriv = priv->ppriv; struct mlx5i_priv *ipriv = priv->ppriv;
mlx5e_destroy_tis(priv->mdev, priv->tisn[0]); mlx5e_destroy_tis(priv->mdev, priv->tisn[0][0]);
mlx5i_destroy_underlay_qp(priv->mdev, &ipriv->qp); mlx5i_destroy_underlay_qp(priv->mdev, &ipriv->qp);
} }
......
...@@ -210,7 +210,7 @@ static int mlx5i_pkey_open(struct net_device *netdev) ...@@ -210,7 +210,7 @@ static int mlx5i_pkey_open(struct net_device *netdev)
goto err_unint_underlay_qp; goto err_unint_underlay_qp;
} }
err = mlx5i_create_tis(mdev, ipriv->qp.qpn, &epriv->tisn[0]); err = mlx5i_create_tis(mdev, ipriv->qp.qpn, &epriv->tisn[0][0]);
if (err) { if (err) {
mlx5_core_warn(mdev, "create child tis failed, %d\n", err); mlx5_core_warn(mdev, "create child tis failed, %d\n", err);
goto err_remove_rx_uderlay_qp; goto err_remove_rx_uderlay_qp;
...@@ -228,7 +228,7 @@ static int mlx5i_pkey_open(struct net_device *netdev) ...@@ -228,7 +228,7 @@ static int mlx5i_pkey_open(struct net_device *netdev)
return 0; return 0;
err_clear_state_opened_flag: err_clear_state_opened_flag:
mlx5e_destroy_tis(mdev, epriv->tisn[0]); mlx5e_destroy_tis(mdev, epriv->tisn[0][0]);
err_remove_rx_uderlay_qp: err_remove_rx_uderlay_qp:
mlx5_fs_remove_rx_underlay_qpn(mdev, ipriv->qp.qpn); mlx5_fs_remove_rx_underlay_qpn(mdev, ipriv->qp.qpn);
err_unint_underlay_qp: err_unint_underlay_qp:
...@@ -257,7 +257,7 @@ static int mlx5i_pkey_close(struct net_device *netdev) ...@@ -257,7 +257,7 @@ static int mlx5i_pkey_close(struct net_device *netdev)
mlx5i_uninit_underlay_qp(priv); mlx5i_uninit_underlay_qp(priv);
mlx5e_deactivate_priv_channels(priv); mlx5e_deactivate_priv_channels(priv);
mlx5e_close_channels(&priv->channels); mlx5e_close_channels(&priv->channels);
mlx5e_destroy_tis(mdev, priv->tisn[0]); mlx5e_destroy_tis(mdev, priv->tisn[0][0]);
unlock: unlock:
mutex_unlock(&priv->state_lock); mutex_unlock(&priv->state_lock);
return 0; return 0;
......
...@@ -2,6 +2,7 @@ ...@@ -2,6 +2,7 @@
// Copyright (c) 2019 Mellanox Technologies. // Copyright (c) 2019 Mellanox Technologies.
#include "mlx5_core.h" #include "mlx5_core.h"
#include "lib/mlx5.h"
int mlx5_create_encryption_key(struct mlx5_core_dev *mdev, int mlx5_create_encryption_key(struct mlx5_core_dev *mdev,
void *key, u32 sz_bytes, void *key, u32 sz_bytes,
......
...@@ -496,6 +496,12 @@ static int handle_hca_cap_odp(struct mlx5_core_dev *dev) ...@@ -496,6 +496,12 @@ static int handle_hca_cap_odp(struct mlx5_core_dev *dev)
ODP_CAP_SET_MAX(dev, xrc_odp_caps.write); ODP_CAP_SET_MAX(dev, xrc_odp_caps.write);
ODP_CAP_SET_MAX(dev, xrc_odp_caps.read); ODP_CAP_SET_MAX(dev, xrc_odp_caps.read);
ODP_CAP_SET_MAX(dev, xrc_odp_caps.atomic); ODP_CAP_SET_MAX(dev, xrc_odp_caps.atomic);
ODP_CAP_SET_MAX(dev, dc_odp_caps.srq_receive);
ODP_CAP_SET_MAX(dev, dc_odp_caps.send);
ODP_CAP_SET_MAX(dev, dc_odp_caps.receive);
ODP_CAP_SET_MAX(dev, dc_odp_caps.write);
ODP_CAP_SET_MAX(dev, dc_odp_caps.read);
ODP_CAP_SET_MAX(dev, dc_odp_caps.atomic);
if (do_set) if (do_set)
err = set_caps(dev, set_ctx, set_sz, err = set_caps(dev, set_ctx, set_sz,
......
...@@ -53,7 +53,7 @@ mlx5_get_rsc(struct mlx5_qp_table *table, u32 rsn) ...@@ -53,7 +53,7 @@ mlx5_get_rsc(struct mlx5_qp_table *table, u32 rsn)
common = radix_tree_lookup(&table->tree, rsn); common = radix_tree_lookup(&table->tree, rsn);
if (common) if (common)
atomic_inc(&common->refcount); refcount_inc(&common->refcount);
spin_unlock_irqrestore(&table->lock, flags); spin_unlock_irqrestore(&table->lock, flags);
...@@ -62,7 +62,7 @@ mlx5_get_rsc(struct mlx5_qp_table *table, u32 rsn) ...@@ -62,7 +62,7 @@ mlx5_get_rsc(struct mlx5_qp_table *table, u32 rsn)
void mlx5_core_put_rsc(struct mlx5_core_rsc_common *common) void mlx5_core_put_rsc(struct mlx5_core_rsc_common *common)
{ {
if (atomic_dec_and_test(&common->refcount)) if (refcount_dec_and_test(&common->refcount))
complete(&common->free); complete(&common->free);
} }
...@@ -162,7 +162,7 @@ static int rsc_event_notifier(struct notifier_block *nb, ...@@ -162,7 +162,7 @@ static int rsc_event_notifier(struct notifier_block *nb,
common = mlx5_get_rsc(table, rsn); common = mlx5_get_rsc(table, rsn);
if (!common) { if (!common) {
mlx5_core_warn(dev, "Async event for bogus resource 0x%x\n", rsn); mlx5_core_dbg(dev, "Async event for unknown resource 0x%x\n", rsn);
return NOTIFY_OK; return NOTIFY_OK;
} }
...@@ -209,7 +209,7 @@ static int create_resource_common(struct mlx5_core_dev *dev, ...@@ -209,7 +209,7 @@ static int create_resource_common(struct mlx5_core_dev *dev,
if (err) if (err)
return err; return err;
atomic_set(&qp->common.refcount, 1); refcount_set(&qp->common.refcount, 1);
init_completion(&qp->common.free); init_completion(&qp->common.free);
qp->pid = current->pid; qp->pid = current->pid;
......
...@@ -51,7 +51,7 @@ static int mlx5_rdma_enable_roce_steering(struct mlx5_core_dev *dev) ...@@ -51,7 +51,7 @@ static int mlx5_rdma_enable_roce_steering(struct mlx5_core_dev *dev)
return -ENOMEM; return -ENOMEM;
} }
ns = mlx5_get_flow_namespace(dev, MLX5_FLOW_NAMESPACE_RDMA_RX); ns = mlx5_get_flow_namespace(dev, MLX5_FLOW_NAMESPACE_RDMA_RX_KERNEL);
if (!ns) { if (!ns) {
mlx5_core_err(dev, "Failed to get RDMA RX namespace"); mlx5_core_err(dev, "Failed to get RDMA RX namespace");
err = -EOPNOTSUPP; err = -EOPNOTSUPP;
......
...@@ -47,6 +47,7 @@ ...@@ -47,6 +47,7 @@
#include <linux/interrupt.h> #include <linux/interrupt.h>
#include <linux/idr.h> #include <linux/idr.h>
#include <linux/notifier.h> #include <linux/notifier.h>
#include <linux/refcount.h>
#include <linux/mlx5/device.h> #include <linux/mlx5/device.h>
#include <linux/mlx5/doorbell.h> #include <linux/mlx5/doorbell.h>
...@@ -390,7 +391,7 @@ enum mlx5_res_type { ...@@ -390,7 +391,7 @@ enum mlx5_res_type {
struct mlx5_core_rsc_common { struct mlx5_core_rsc_common {
enum mlx5_res_type res; enum mlx5_res_type res;
atomic_t refcount; refcount_t refcount;
struct completion free; struct completion free;
}; };
......
...@@ -75,6 +75,7 @@ enum mlx5_flow_namespace_type { ...@@ -75,6 +75,7 @@ enum mlx5_flow_namespace_type {
MLX5_FLOW_NAMESPACE_SNIFFER_TX, MLX5_FLOW_NAMESPACE_SNIFFER_TX,
MLX5_FLOW_NAMESPACE_EGRESS, MLX5_FLOW_NAMESPACE_EGRESS,
MLX5_FLOW_NAMESPACE_RDMA_RX, MLX5_FLOW_NAMESPACE_RDMA_RX,
MLX5_FLOW_NAMESPACE_RDMA_RX_KERNEL,
}; };
enum { enum {
......
...@@ -172,6 +172,8 @@ enum { ...@@ -172,6 +172,8 @@ enum {
MLX5_CMD_OP_QUERY_XRQ_DC_PARAMS_ENTRY = 0x725, MLX5_CMD_OP_QUERY_XRQ_DC_PARAMS_ENTRY = 0x725,
MLX5_CMD_OP_SET_XRQ_DC_PARAMS_ENTRY = 0x726, MLX5_CMD_OP_SET_XRQ_DC_PARAMS_ENTRY = 0x726,
MLX5_CMD_OP_QUERY_XRQ_ERROR_PARAMS = 0x727, MLX5_CMD_OP_QUERY_XRQ_ERROR_PARAMS = 0x727,
MLX5_CMD_OP_RELEASE_XRQ_ERROR = 0x729,
MLX5_CMD_OP_MODIFY_XRQ = 0x72a,
MLX5_CMD_OP_QUERY_ESW_FUNCTIONS = 0x740, MLX5_CMD_OP_QUERY_ESW_FUNCTIONS = 0x740,
MLX5_CMD_OP_QUERY_VPORT_STATE = 0x750, MLX5_CMD_OP_QUERY_VPORT_STATE = 0x750,
MLX5_CMD_OP_MODIFY_VPORT_STATE = 0x751, MLX5_CMD_OP_MODIFY_VPORT_STATE = 0x751,
...@@ -806,7 +808,9 @@ struct mlx5_ifc_per_protocol_networking_offload_caps_bits { ...@@ -806,7 +808,9 @@ struct mlx5_ifc_per_protocol_networking_offload_caps_bits {
u8 swp_csum[0x1]; u8 swp_csum[0x1];
u8 swp_lso[0x1]; u8 swp_lso[0x1];
u8 cqe_checksum_full[0x1]; u8 cqe_checksum_full[0x1];
u8 reserved_at_24[0xc]; u8 reserved_at_24[0x5];
u8 tunnel_stateless_ip_over_ip[0x1];
u8 reserved_at_2a[0x6];
u8 max_vxlan_udp_ports[0x8]; u8 max_vxlan_udp_ports[0x8];
u8 reserved_at_38[0x6]; u8 reserved_at_38[0x6];
u8 max_geneve_opt_len[0x1]; u8 max_geneve_opt_len[0x1];
...@@ -944,7 +948,9 @@ struct mlx5_ifc_odp_cap_bits { ...@@ -944,7 +948,9 @@ struct mlx5_ifc_odp_cap_bits {
struct mlx5_ifc_odp_per_transport_service_cap_bits xrc_odp_caps; struct mlx5_ifc_odp_per_transport_service_cap_bits xrc_odp_caps;
u8 reserved_at_100[0x700]; struct mlx5_ifc_odp_per_transport_service_cap_bits dc_odp_caps;
u8 reserved_at_120[0x6E0];
}; };
struct mlx5_ifc_calc_op { struct mlx5_ifc_calc_op {
...@@ -1114,7 +1120,9 @@ struct mlx5_ifc_cmd_hca_cap_bits { ...@@ -1114,7 +1120,9 @@ struct mlx5_ifc_cmd_hca_cap_bits {
u8 cache_line_128byte[0x1]; u8 cache_line_128byte[0x1];
u8 reserved_at_165[0x4]; u8 reserved_at_165[0x4];
u8 rts2rts_qp_counters_set_id[0x1]; u8 rts2rts_qp_counters_set_id[0x1];
u8 reserved_at_16a[0x5]; u8 reserved_at_16a[0x2];
u8 vnic_env_int_rq_oob[0x1];
u8 reserved_at_16d[0x2];
u8 qcam_reg[0x1]; u8 qcam_reg[0x1];
u8 gid_table_size[0x10]; u8 gid_table_size[0x10];
...@@ -1243,7 +1251,9 @@ struct mlx5_ifc_cmd_hca_cap_bits { ...@@ -1243,7 +1251,9 @@ struct mlx5_ifc_cmd_hca_cap_bits {
u8 reserved_at_263[0x8]; u8 reserved_at_263[0x8];
u8 log_bf_reg_size[0x5]; u8 log_bf_reg_size[0x5];
u8 reserved_at_270[0xb]; u8 reserved_at_270[0x8];
u8 lag_tx_port_affinity[0x1];
u8 reserved_at_279[0x2];
u8 lag_master[0x1]; u8 lag_master[0x1];
u8 num_lag_ports[0x4]; u8 num_lag_ports[0x4];
...@@ -2770,7 +2780,11 @@ struct mlx5_ifc_vnic_diagnostic_statistics_bits { ...@@ -2770,7 +2780,11 @@ struct mlx5_ifc_vnic_diagnostic_statistics_bits {
u8 transmit_discard_vport_down[0x40]; u8 transmit_discard_vport_down[0x40];
u8 reserved_at_140[0xec0]; u8 reserved_at_140[0xa0];
u8 internal_rq_out_of_buffer[0x20];
u8 reserved_at_200[0xe00];
}; };
struct mlx5_ifc_traffic_counter_bits { struct mlx5_ifc_traffic_counter_bits {
...@@ -9594,8 +9608,6 @@ struct mlx5_ifc_query_lag_out_bits { ...@@ -9594,8 +9608,6 @@ struct mlx5_ifc_query_lag_out_bits {
u8 syndrome[0x20]; u8 syndrome[0x20];
u8 reserved_at_40[0x40];
struct mlx5_ifc_lagc_bits ctx; struct mlx5_ifc_lagc_bits ctx;
}; };
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment