Commit 221c14f3 authored by Saeed Mahameed's avatar Saeed Mahameed

net/mlx5: Resource tables, Use async events chain

Remove the explicit call to QP/SRQ resources events handlers on several FW
events and let resources logic register resources events notifiers via the
new API.
Signed-off-by: default avatarSaeed Mahameed <saeedm@mellanox.com>
parent 71edc69c
...@@ -324,7 +324,6 @@ static irqreturn_t mlx5_eq_async_int(int irq, void *eq_ptr) ...@@ -324,7 +324,6 @@ static irqreturn_t mlx5_eq_async_int(int irq, void *eq_ptr)
struct mlx5_eqe *eqe; struct mlx5_eqe *eqe;
int set_ci = 0; int set_ci = 0;
u32 cqn = -1; u32 cqn = -1;
u32 rsn;
u8 port; u8 port;
dev = eq->dev; dev = eq->dev;
...@@ -340,34 +339,6 @@ static irqreturn_t mlx5_eq_async_int(int irq, void *eq_ptr) ...@@ -340,34 +339,6 @@ static irqreturn_t mlx5_eq_async_int(int irq, void *eq_ptr)
mlx5_core_dbg(eq->dev, "eqn %d, eqe type %s\n", mlx5_core_dbg(eq->dev, "eqn %d, eqe type %s\n",
eq->eqn, eqe_type_str(eqe->type)); eq->eqn, eqe_type_str(eqe->type));
switch (eqe->type) { switch (eqe->type) {
case MLX5_EVENT_TYPE_DCT_DRAINED:
rsn = be32_to_cpu(eqe->data.dct.dctn) & 0xffffff;
rsn |= (MLX5_RES_DCT << MLX5_USER_INDEX_LEN);
mlx5_rsc_event(dev, rsn, eqe->type);
break;
case MLX5_EVENT_TYPE_PATH_MIG:
case MLX5_EVENT_TYPE_COMM_EST:
case MLX5_EVENT_TYPE_SQ_DRAINED:
case MLX5_EVENT_TYPE_SRQ_LAST_WQE:
case MLX5_EVENT_TYPE_WQ_CATAS_ERROR:
case MLX5_EVENT_TYPE_PATH_MIG_FAILED:
case MLX5_EVENT_TYPE_WQ_INVAL_REQ_ERROR:
case MLX5_EVENT_TYPE_WQ_ACCESS_ERROR:
rsn = be32_to_cpu(eqe->data.qp_srq.qp_srq_n) & 0xffffff;
rsn |= (eqe->data.qp_srq.type << MLX5_USER_INDEX_LEN);
mlx5_core_dbg(dev, "event %s(%d) arrived on resource 0x%x\n",
eqe_type_str(eqe->type), eqe->type, rsn);
mlx5_rsc_event(dev, rsn, eqe->type);
break;
case MLX5_EVENT_TYPE_SRQ_RQ_LIMIT:
case MLX5_EVENT_TYPE_SRQ_CATAS_ERROR:
rsn = be32_to_cpu(eqe->data.qp_srq.qp_srq_n) & 0xffffff;
mlx5_core_dbg(dev, "SRQ event %s(%d): srqn 0x%x\n",
eqe_type_str(eqe->type), eqe->type, rsn);
mlx5_srq_event(dev, rsn, eqe->type);
break;
case MLX5_EVENT_TYPE_PORT_CHANGE: case MLX5_EVENT_TYPE_PORT_CHANGE:
port = (eqe->data.port.port >> 4) & 0xf; port = (eqe->data.port.port >> 4) & 0xf;
switch (eqe->sub_type) { switch (eqe->sub_type) {
......
...@@ -38,11 +38,11 @@ ...@@ -38,11 +38,11 @@
#include <linux/mlx5/transobj.h> #include <linux/mlx5/transobj.h>
#include "mlx5_core.h" #include "mlx5_core.h"
#include "lib/eq.h"
static struct mlx5_core_rsc_common *mlx5_get_rsc(struct mlx5_core_dev *dev, static struct mlx5_core_rsc_common *
u32 rsn) mlx5_get_rsc(struct mlx5_qp_table *table, u32 rsn)
{ {
struct mlx5_qp_table *table = &dev->priv.qp_table;
struct mlx5_core_rsc_common *common; struct mlx5_core_rsc_common *common;
spin_lock(&table->lock); spin_lock(&table->lock);
...@@ -53,11 +53,6 @@ static struct mlx5_core_rsc_common *mlx5_get_rsc(struct mlx5_core_dev *dev, ...@@ -53,11 +53,6 @@ static struct mlx5_core_rsc_common *mlx5_get_rsc(struct mlx5_core_dev *dev,
spin_unlock(&table->lock); spin_unlock(&table->lock);
if (!common) {
mlx5_core_warn(dev, "Async event for bogus resource 0x%x\n",
rsn);
return NULL;
}
return common; return common;
} }
...@@ -120,14 +115,52 @@ static bool is_event_type_allowed(int rsc_type, int event_type) ...@@ -120,14 +115,52 @@ static bool is_event_type_allowed(int rsc_type, int event_type)
} }
} }
void mlx5_rsc_event(struct mlx5_core_dev *dev, u32 rsn, int event_type) static int rsc_event_notifier(struct notifier_block *nb,
unsigned long type, void *data)
{ {
struct mlx5_core_rsc_common *common = mlx5_get_rsc(dev, rsn); struct mlx5_core_rsc_common *common;
struct mlx5_qp_table *table;
struct mlx5_core_dev *dev;
struct mlx5_core_dct *dct; struct mlx5_core_dct *dct;
u8 event_type = (u8)type;
struct mlx5_core_qp *qp; struct mlx5_core_qp *qp;
struct mlx5_priv *priv;
struct mlx5_eqe *eqe;
u32 rsn;
switch (event_type) {
case MLX5_EVENT_TYPE_DCT_DRAINED:
eqe = data;
rsn = be32_to_cpu(eqe->data.dct.dctn) & 0xffffff;
rsn |= (MLX5_RES_DCT << MLX5_USER_INDEX_LEN);
break;
case MLX5_EVENT_TYPE_PATH_MIG:
case MLX5_EVENT_TYPE_COMM_EST:
case MLX5_EVENT_TYPE_SQ_DRAINED:
case MLX5_EVENT_TYPE_SRQ_LAST_WQE:
case MLX5_EVENT_TYPE_WQ_CATAS_ERROR:
case MLX5_EVENT_TYPE_PATH_MIG_FAILED:
case MLX5_EVENT_TYPE_WQ_INVAL_REQ_ERROR:
case MLX5_EVENT_TYPE_WQ_ACCESS_ERROR:
eqe = data;
rsn = be32_to_cpu(eqe->data.qp_srq.qp_srq_n) & 0xffffff;
rsn |= (eqe->data.qp_srq.type << MLX5_USER_INDEX_LEN);
break;
default:
return NOTIFY_DONE;
}
table = mlx5_nb_cof(nb, struct mlx5_qp_table, nb);
priv = container_of(table, struct mlx5_priv, qp_table);
dev = container_of(priv, struct mlx5_core_dev, priv);
if (!common) mlx5_core_dbg(dev, "event (%d) arrived on resource 0x%x\n", eqe->type, rsn);
return;
common = mlx5_get_rsc(table, rsn);
if (!common) {
mlx5_core_warn(dev, "Async event for bogus resource 0x%x\n", rsn);
return NOTIFY_OK;
}
if (!is_event_type_allowed((rsn >> MLX5_USER_INDEX_LEN), event_type)) { if (!is_event_type_allowed((rsn >> MLX5_USER_INDEX_LEN), event_type)) {
mlx5_core_warn(dev, "event 0x%.2x is not allowed on resource 0x%.8x\n", mlx5_core_warn(dev, "event 0x%.2x is not allowed on resource 0x%.8x\n",
...@@ -152,6 +185,8 @@ void mlx5_rsc_event(struct mlx5_core_dev *dev, u32 rsn, int event_type) ...@@ -152,6 +185,8 @@ void mlx5_rsc_event(struct mlx5_core_dev *dev, u32 rsn, int event_type)
} }
out: out:
mlx5_core_put_rsc(common); mlx5_core_put_rsc(common);
return NOTIFY_OK;
} }
static int create_resource_common(struct mlx5_core_dev *dev, static int create_resource_common(struct mlx5_core_dev *dev,
...@@ -487,10 +522,16 @@ void mlx5_init_qp_table(struct mlx5_core_dev *dev) ...@@ -487,10 +522,16 @@ void mlx5_init_qp_table(struct mlx5_core_dev *dev)
spin_lock_init(&table->lock); spin_lock_init(&table->lock);
INIT_RADIX_TREE(&table->tree, GFP_ATOMIC); INIT_RADIX_TREE(&table->tree, GFP_ATOMIC);
mlx5_qp_debugfs_init(dev); mlx5_qp_debugfs_init(dev);
MLX5_NB_INIT(&table->nb, rsc_event_notifier, NOTIFY_ANY);
mlx5_eq_notifier_register(dev, &table->nb);
} }
void mlx5_cleanup_qp_table(struct mlx5_core_dev *dev) void mlx5_cleanup_qp_table(struct mlx5_core_dev *dev)
{ {
struct mlx5_qp_table *table = &dev->priv.qp_table;
mlx5_eq_notifier_unregister(dev, &table->nb);
mlx5_qp_debugfs_cleanup(dev); mlx5_qp_debugfs_cleanup(dev);
} }
...@@ -676,8 +717,9 @@ struct mlx5_core_rsc_common *mlx5_core_res_hold(struct mlx5_core_dev *dev, ...@@ -676,8 +717,9 @@ struct mlx5_core_rsc_common *mlx5_core_res_hold(struct mlx5_core_dev *dev,
enum mlx5_res_type res_type) enum mlx5_res_type res_type)
{ {
u32 rsn = res_num | (res_type << MLX5_USER_INDEX_LEN); u32 rsn = res_num | (res_type << MLX5_USER_INDEX_LEN);
struct mlx5_qp_table *table = &dev->priv.qp_table;
return mlx5_get_rsc(dev, rsn); return mlx5_get_rsc(table, rsn);
} }
EXPORT_SYMBOL_GPL(mlx5_core_res_hold); EXPORT_SYMBOL_GPL(mlx5_core_res_hold);
......
...@@ -36,13 +36,25 @@ ...@@ -36,13 +36,25 @@
#include <linux/mlx5/cmd.h> #include <linux/mlx5/cmd.h>
#include <linux/mlx5/srq.h> #include <linux/mlx5/srq.h>
#include <rdma/ib_verbs.h> #include <rdma/ib_verbs.h>
#include "mlx5_core.h"
#include <linux/mlx5/transobj.h> #include <linux/mlx5/transobj.h>
#include "mlx5_core.h"
#include "lib/eq.h"
void mlx5_srq_event(struct mlx5_core_dev *dev, u32 srqn, int event_type) static int srq_event_notifier(struct mlx5_srq_table *table,
unsigned long type, void *data)
{ {
struct mlx5_srq_table *table = &dev->priv.srq_table; struct mlx5_core_dev *dev;
struct mlx5_core_srq *srq; struct mlx5_core_srq *srq;
struct mlx5_priv *priv;
struct mlx5_eqe *eqe;
u32 srqn;
priv = container_of(table, struct mlx5_priv, srq_table);
dev = container_of(priv, struct mlx5_core_dev, priv);
eqe = data;
srqn = be32_to_cpu(eqe->data.qp_srq.qp_srq_n) & 0xffffff;
mlx5_core_dbg(dev, "SRQ event (%d): srqn 0x%x\n", eqe->type, srqn);
spin_lock(&table->lock); spin_lock(&table->lock);
...@@ -54,13 +66,35 @@ void mlx5_srq_event(struct mlx5_core_dev *dev, u32 srqn, int event_type) ...@@ -54,13 +66,35 @@ void mlx5_srq_event(struct mlx5_core_dev *dev, u32 srqn, int event_type)
if (!srq) { if (!srq) {
mlx5_core_warn(dev, "Async event for bogus SRQ 0x%08x\n", srqn); mlx5_core_warn(dev, "Async event for bogus SRQ 0x%08x\n", srqn);
return; return NOTIFY_OK;
} }
srq->event(srq, event_type); srq->event(srq, eqe->type);
if (atomic_dec_and_test(&srq->refcount)) if (atomic_dec_and_test(&srq->refcount))
complete(&srq->free); complete(&srq->free);
return NOTIFY_OK;
}
static int catas_err_notifier(struct notifier_block *nb,
unsigned long type, void *data)
{
struct mlx5_srq_table *table;
table = mlx5_nb_cof(nb, struct mlx5_srq_table, catas_err_nb);
/* type == MLX5_EVENT_TYPE_SRQ_CATAS_ERROR */
return srq_event_notifier(table, type, data);
}
static int rq_limit_notifier(struct notifier_block *nb,
unsigned long type, void *data)
{
struct mlx5_srq_table *table;
table = mlx5_nb_cof(nb, struct mlx5_srq_table, rq_limit_nb);
/* type == MLX5_EVENT_TYPE_SRQ_RQ_LIMIT */
return srq_event_notifier(table, type, data);
} }
static int get_pas_size(struct mlx5_srq_attr *in) static int get_pas_size(struct mlx5_srq_attr *in)
...@@ -708,9 +742,18 @@ void mlx5_init_srq_table(struct mlx5_core_dev *dev) ...@@ -708,9 +742,18 @@ void mlx5_init_srq_table(struct mlx5_core_dev *dev)
memset(table, 0, sizeof(*table)); memset(table, 0, sizeof(*table));
spin_lock_init(&table->lock); spin_lock_init(&table->lock);
INIT_RADIX_TREE(&table->tree, GFP_ATOMIC); INIT_RADIX_TREE(&table->tree, GFP_ATOMIC);
MLX5_NB_INIT(&table->catas_err_nb, catas_err_notifier, SRQ_CATAS_ERROR);
mlx5_eq_notifier_register(dev, &table->catas_err_nb);
MLX5_NB_INIT(&table->rq_limit_nb, rq_limit_notifier, SRQ_RQ_LIMIT);
mlx5_eq_notifier_register(dev, &table->rq_limit_nb);
} }
void mlx5_cleanup_srq_table(struct mlx5_core_dev *dev) void mlx5_cleanup_srq_table(struct mlx5_core_dev *dev)
{ {
/* nothing */ struct mlx5_srq_table *table = &dev->priv.srq_table;
mlx5_eq_notifier_unregister(dev, &table->rq_limit_nb);
mlx5_eq_notifier_unregister(dev, &table->catas_err_nb);
} }
...@@ -464,6 +464,8 @@ struct mlx5_core_health { ...@@ -464,6 +464,8 @@ struct mlx5_core_health {
}; };
struct mlx5_qp_table { struct mlx5_qp_table {
struct mlx5_nb nb;
/* protect radix tree /* protect radix tree
*/ */
spinlock_t lock; spinlock_t lock;
...@@ -471,6 +473,8 @@ struct mlx5_qp_table { ...@@ -471,6 +473,8 @@ struct mlx5_qp_table {
}; };
struct mlx5_srq_table { struct mlx5_srq_table {
struct mlx5_nb catas_err_nb;
struct mlx5_nb rq_limit_nb;
/* protect radix tree /* protect radix tree
*/ */
spinlock_t lock; spinlock_t lock;
...@@ -978,8 +982,6 @@ void mlx5_unregister_debugfs(void); ...@@ -978,8 +982,6 @@ void mlx5_unregister_debugfs(void);
void mlx5_fill_page_array(struct mlx5_frag_buf *buf, __be64 *pas); void mlx5_fill_page_array(struct mlx5_frag_buf *buf, __be64 *pas);
void mlx5_fill_page_frag_array(struct mlx5_frag_buf *frag_buf, __be64 *pas); void mlx5_fill_page_frag_array(struct mlx5_frag_buf *frag_buf, __be64 *pas);
void mlx5_rsc_event(struct mlx5_core_dev *dev, u32 rsn, int event_type);
void mlx5_srq_event(struct mlx5_core_dev *dev, u32 srqn, int event_type);
struct mlx5_core_srq *mlx5_core_get_srq(struct mlx5_core_dev *dev, u32 srqn); struct mlx5_core_srq *mlx5_core_get_srq(struct mlx5_core_dev *dev, u32 srqn);
int mlx5_vector2eqn(struct mlx5_core_dev *dev, int vector, int *eqn, int mlx5_vector2eqn(struct mlx5_core_dev *dev, int vector, int *eqn,
unsigned int *irqn); unsigned int *irqn);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment