Commit c9813b0b authored by Yangyang Li's avatar Yangyang Li Committed by Leon Romanovsky

RDMA/hns: Support SRQ record doorbell

Compared with normal doorbell, using record doorbell can shorten the
process of ringing the doorbell and reduce the latency.

Add a flag HNS_ROCE_CAP_FLAG_SRQ_RECORD_DB to allow FW to
enable/disable SRQ record doorbell.

If the flag above is set, allocate the dma buffer for SRQ record
doorbell and write the buffer address into SRQC during SRQ creation.

For userspace SRQ, add a flag HNS_ROCE_RSP_SRQ_CAP_RECORD_DB to notify
userspace whether the SRQ record doorbell is enabled.
Signed-off-by: default avatarYangyang Li <liyangyang20@huawei.com>
Signed-off-by: default avatarJunxian Huang <huangjunxian6@hisilicon.com>
Link: https://lore.kernel.org/r/20230926130026.583088-1-huangjunxian6@hisilicon.comSigned-off-by: default avatarLeon Romanovsky <leon@kernel.org>
parent 8dc0fd2f
...@@ -146,6 +146,7 @@ enum { ...@@ -146,6 +146,7 @@ enum {
HNS_ROCE_CAP_FLAG_SDI_MODE = BIT(14), HNS_ROCE_CAP_FLAG_SDI_MODE = BIT(14),
HNS_ROCE_CAP_FLAG_STASH = BIT(17), HNS_ROCE_CAP_FLAG_STASH = BIT(17),
HNS_ROCE_CAP_FLAG_CQE_INLINE = BIT(19), HNS_ROCE_CAP_FLAG_CQE_INLINE = BIT(19),
HNS_ROCE_CAP_FLAG_SRQ_RECORD_DB = BIT(22),
}; };
#define HNS_ROCE_DB_TYPE_COUNT 2 #define HNS_ROCE_DB_TYPE_COUNT 2
...@@ -453,6 +454,8 @@ struct hns_roce_srq { ...@@ -453,6 +454,8 @@ struct hns_roce_srq {
spinlock_t lock; spinlock_t lock;
struct mutex mutex; struct mutex mutex;
void (*event)(struct hns_roce_srq *srq, enum hns_roce_event event); void (*event)(struct hns_roce_srq *srq, enum hns_roce_event event);
struct hns_roce_db rdb;
u32 cap_flags;
}; };
struct hns_roce_uar_table { struct hns_roce_uar_table {
......
...@@ -941,20 +941,23 @@ static void fill_wqe_idx(struct hns_roce_srq *srq, unsigned int wqe_idx) ...@@ -941,20 +941,23 @@ static void fill_wqe_idx(struct hns_roce_srq *srq, unsigned int wqe_idx)
idx_que->head++; idx_que->head++;
} }
static void update_srq_db(struct hns_roce_v2_db *db, struct hns_roce_srq *srq) static void update_srq_db(struct hns_roce_srq *srq)
{ {
hr_reg_write(db, DB_TAG, srq->srqn); struct hns_roce_dev *hr_dev = to_hr_dev(srq->ibsrq.device);
hr_reg_write(db, DB_CMD, HNS_ROCE_V2_SRQ_DB); struct hns_roce_v2_db db;
hr_reg_write(db, DB_PI, srq->idx_que.head);
hr_reg_write(&db, DB_TAG, srq->srqn);
hr_reg_write(&db, DB_CMD, HNS_ROCE_V2_SRQ_DB);
hr_reg_write(&db, DB_PI, srq->idx_que.head);
hns_roce_write64(hr_dev, (__le32 *)&db, srq->db_reg);
} }
static int hns_roce_v2_post_srq_recv(struct ib_srq *ibsrq, static int hns_roce_v2_post_srq_recv(struct ib_srq *ibsrq,
const struct ib_recv_wr *wr, const struct ib_recv_wr *wr,
const struct ib_recv_wr **bad_wr) const struct ib_recv_wr **bad_wr)
{ {
struct hns_roce_dev *hr_dev = to_hr_dev(ibsrq->device);
struct hns_roce_srq *srq = to_hr_srq(ibsrq); struct hns_roce_srq *srq = to_hr_srq(ibsrq);
struct hns_roce_v2_db srq_db;
unsigned long flags; unsigned long flags;
int ret = 0; int ret = 0;
u32 max_sge; u32 max_sge;
...@@ -985,9 +988,11 @@ static int hns_roce_v2_post_srq_recv(struct ib_srq *ibsrq, ...@@ -985,9 +988,11 @@ static int hns_roce_v2_post_srq_recv(struct ib_srq *ibsrq,
} }
if (likely(nreq)) { if (likely(nreq)) {
update_srq_db(&srq_db, srq); if (srq->cap_flags & HNS_ROCE_SRQ_CAP_RECORD_DB)
*srq->rdb.db_record = srq->idx_que.head &
hns_roce_write64(hr_dev, (__le32 *)&srq_db, srq->db_reg); V2_DB_PRODUCER_IDX_M;
else
update_srq_db(srq);
} }
spin_unlock_irqrestore(&srq->lock, flags); spin_unlock_irqrestore(&srq->lock, flags);
...@@ -5630,6 +5635,14 @@ static int hns_roce_v2_write_srqc(struct hns_roce_srq *srq, void *mb_buf) ...@@ -5630,6 +5635,14 @@ static int hns_roce_v2_write_srqc(struct hns_roce_srq *srq, void *mb_buf)
hr_reg_write(ctx, SRQC_WQE_BUF_PG_SZ, hr_reg_write(ctx, SRQC_WQE_BUF_PG_SZ,
to_hr_hw_page_shift(srq->buf_mtr.hem_cfg.buf_pg_shift)); to_hr_hw_page_shift(srq->buf_mtr.hem_cfg.buf_pg_shift));
if (srq->cap_flags & HNS_ROCE_SRQ_CAP_RECORD_DB) {
hr_reg_enable(ctx, SRQC_DB_RECORD_EN);
hr_reg_write(ctx, SRQC_DB_RECORD_ADDR_L,
lower_32_bits(srq->rdb.dma) >> 1);
hr_reg_write(ctx, SRQC_DB_RECORD_ADDR_H,
upper_32_bits(srq->rdb.dma));
}
return hns_roce_v2_write_srqc_index_queue(srq, ctx); return hns_roce_v2_write_srqc_index_queue(srq, ctx);
} }
......
...@@ -5,6 +5,7 @@ ...@@ -5,6 +5,7 @@
#include <linux/pci.h> #include <linux/pci.h>
#include <rdma/ib_umem.h> #include <rdma/ib_umem.h>
#include <rdma/uverbs_ioctl.h>
#include "hns_roce_device.h" #include "hns_roce_device.h"
#include "hns_roce_cmd.h" #include "hns_roce_cmd.h"
#include "hns_roce_hem.h" #include "hns_roce_hem.h"
...@@ -387,6 +388,79 @@ static void free_srq_buf(struct hns_roce_dev *hr_dev, struct hns_roce_srq *srq) ...@@ -387,6 +388,79 @@ static void free_srq_buf(struct hns_roce_dev *hr_dev, struct hns_roce_srq *srq)
free_srq_idx(hr_dev, srq); free_srq_idx(hr_dev, srq);
} }
static int get_srq_ucmd(struct hns_roce_srq *srq, struct ib_udata *udata,
struct hns_roce_ib_create_srq *ucmd)
{
struct ib_device *ibdev = srq->ibsrq.device;
int ret;
ret = ib_copy_from_udata(ucmd, udata, min(udata->inlen, sizeof(*ucmd)));
if (ret) {
ibdev_err(ibdev, "failed to copy SRQ udata, ret = %d.\n", ret);
return ret;
}
return 0;
}
static void free_srq_db(struct hns_roce_dev *hr_dev, struct hns_roce_srq *srq,
struct ib_udata *udata)
{
struct hns_roce_ucontext *uctx;
if (!(srq->cap_flags & HNS_ROCE_SRQ_CAP_RECORD_DB))
return;
srq->cap_flags &= ~HNS_ROCE_SRQ_CAP_RECORD_DB;
if (udata) {
uctx = rdma_udata_to_drv_context(udata,
struct hns_roce_ucontext,
ibucontext);
hns_roce_db_unmap_user(uctx, &srq->rdb);
} else {
hns_roce_free_db(hr_dev, &srq->rdb);
}
}
static int alloc_srq_db(struct hns_roce_dev *hr_dev, struct hns_roce_srq *srq,
struct ib_udata *udata,
struct hns_roce_ib_create_srq_resp *resp)
{
struct hns_roce_ib_create_srq ucmd = {};
struct hns_roce_ucontext *uctx;
int ret;
if (udata) {
ret = get_srq_ucmd(srq, udata, &ucmd);
if (ret)
return ret;
if ((hr_dev->caps.flags & HNS_ROCE_CAP_FLAG_SRQ_RECORD_DB) &&
(ucmd.req_cap_flags & HNS_ROCE_SRQ_CAP_RECORD_DB)) {
uctx = rdma_udata_to_drv_context(udata,
struct hns_roce_ucontext, ibucontext);
ret = hns_roce_db_map_user(uctx, ucmd.db_addr,
&srq->rdb);
if (ret)
return ret;
srq->cap_flags |= HNS_ROCE_RSP_SRQ_CAP_RECORD_DB;
}
} else {
if (hr_dev->caps.flags & HNS_ROCE_CAP_FLAG_SRQ_RECORD_DB) {
ret = hns_roce_alloc_db(hr_dev, &srq->rdb, 1);
if (ret)
return ret;
*srq->rdb.db_record = 0;
srq->cap_flags |= HNS_ROCE_RSP_SRQ_CAP_RECORD_DB;
}
srq->db_reg = hr_dev->reg_base + SRQ_DB_REG;
}
return 0;
}
int hns_roce_create_srq(struct ib_srq *ib_srq, int hns_roce_create_srq(struct ib_srq *ib_srq,
struct ib_srq_init_attr *init_attr, struct ib_srq_init_attr *init_attr,
struct ib_udata *udata) struct ib_udata *udata)
...@@ -407,15 +481,20 @@ int hns_roce_create_srq(struct ib_srq *ib_srq, ...@@ -407,15 +481,20 @@ int hns_roce_create_srq(struct ib_srq *ib_srq,
if (ret) if (ret)
return ret; return ret;
ret = alloc_srqn(hr_dev, srq); ret = alloc_srq_db(hr_dev, srq, udata, &resp);
if (ret) if (ret)
goto err_srq_buf; goto err_srq_buf;
ret = alloc_srqn(hr_dev, srq);
if (ret)
goto err_srq_db;
ret = alloc_srqc(hr_dev, srq); ret = alloc_srqc(hr_dev, srq);
if (ret) if (ret)
goto err_srqn; goto err_srqn;
if (udata) { if (udata) {
resp.cap_flags = srq->cap_flags;
resp.srqn = srq->srqn; resp.srqn = srq->srqn;
if (ib_copy_to_udata(udata, &resp, if (ib_copy_to_udata(udata, &resp,
min(udata->outlen, sizeof(resp)))) { min(udata->outlen, sizeof(resp)))) {
...@@ -424,7 +503,6 @@ int hns_roce_create_srq(struct ib_srq *ib_srq, ...@@ -424,7 +503,6 @@ int hns_roce_create_srq(struct ib_srq *ib_srq,
} }
} }
srq->db_reg = hr_dev->reg_base + SRQ_DB_REG;
srq->event = hns_roce_ib_srq_event; srq->event = hns_roce_ib_srq_event;
refcount_set(&srq->refcount, 1); refcount_set(&srq->refcount, 1);
init_completion(&srq->free); init_completion(&srq->free);
...@@ -435,6 +513,8 @@ int hns_roce_create_srq(struct ib_srq *ib_srq, ...@@ -435,6 +513,8 @@ int hns_roce_create_srq(struct ib_srq *ib_srq,
free_srqc(hr_dev, srq); free_srqc(hr_dev, srq);
err_srqn: err_srqn:
free_srqn(hr_dev, srq); free_srqn(hr_dev, srq);
err_srq_db:
free_srq_db(hr_dev, srq, udata);
err_srq_buf: err_srq_buf:
free_srq_buf(hr_dev, srq); free_srq_buf(hr_dev, srq);
...@@ -448,6 +528,7 @@ int hns_roce_destroy_srq(struct ib_srq *ibsrq, struct ib_udata *udata) ...@@ -448,6 +528,7 @@ int hns_roce_destroy_srq(struct ib_srq *ibsrq, struct ib_udata *udata)
free_srqc(hr_dev, srq); free_srqc(hr_dev, srq);
free_srqn(hr_dev, srq); free_srqn(hr_dev, srq);
free_srq_db(hr_dev, srq, udata);
free_srq_buf(hr_dev, srq); free_srq_buf(hr_dev, srq);
return 0; return 0;
} }
......
...@@ -52,15 +52,25 @@ struct hns_roce_ib_create_cq_resp { ...@@ -52,15 +52,25 @@ struct hns_roce_ib_create_cq_resp {
__aligned_u64 cap_flags; __aligned_u64 cap_flags;
}; };
enum hns_roce_srq_cap_flags {
HNS_ROCE_SRQ_CAP_RECORD_DB = 1 << 0,
};
enum hns_roce_srq_cap_flags_resp {
HNS_ROCE_RSP_SRQ_CAP_RECORD_DB = 1 << 0,
};
struct hns_roce_ib_create_srq { struct hns_roce_ib_create_srq {
__aligned_u64 buf_addr; __aligned_u64 buf_addr;
__aligned_u64 db_addr; __aligned_u64 db_addr;
__aligned_u64 que_addr; __aligned_u64 que_addr;
__u32 req_cap_flags; /* Use enum hns_roce_srq_cap_flags */
__u32 reserved;
}; };
struct hns_roce_ib_create_srq_resp { struct hns_roce_ib_create_srq_resp {
__u32 srqn; __u32 srqn;
__u32 reserved; __u32 cap_flags; /* Use enum hns_roce_srq_cap_flags */
}; };
struct hns_roce_ib_create_qp { struct hns_roce_ib_create_qp {
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment