Commit 0705107f authored by Dave Jiang's avatar Dave Jiang Committed by Vinod Koul

dmaengine: idxd: move submission to sbitmap_queue

Kill the percpu-rwsem for work submission in favor of an sbitmap_queue.
Signed-off-by: default avatarDave Jiang <dave.jiang@intel.com>
Reviewed-by: default avatarTony Luck <tony.luck@intel.com>
Reviewed-by: default avatarDan Williams <dan.j.williams@intel.com>
Link: https://lore.kernel.org/r/159225446631.68253.8860709181621260997.stgit@djiang5-desk3.ch.intel.comSigned-off-by: default avatarVinod Koul <vkoul@kernel.org>
parent 77522b21
...@@ -285,8 +285,8 @@ config INTEL_IDMA64 ...@@ -285,8 +285,8 @@ config INTEL_IDMA64
config INTEL_IDXD config INTEL_IDXD
tristate "Intel Data Accelerators support" tristate "Intel Data Accelerators support"
depends on PCI && X86_64 depends on PCI && X86_64
depends on SBITMAP
select DMA_ENGINE select DMA_ENGINE
select SBITMAP
help help
Enable support for the Intel(R) data accelerators present Enable support for the Intel(R) data accelerators present
in Intel Xeon CPU. in Intel Xeon CPU.
......
...@@ -160,16 +160,14 @@ static int alloc_descs(struct idxd_wq *wq, int num) ...@@ -160,16 +160,14 @@ static int alloc_descs(struct idxd_wq *wq, int num)
int idxd_wq_alloc_resources(struct idxd_wq *wq) int idxd_wq_alloc_resources(struct idxd_wq *wq)
{ {
struct idxd_device *idxd = wq->idxd; struct idxd_device *idxd = wq->idxd;
struct idxd_group *group = wq->group;
struct device *dev = &idxd->pdev->dev; struct device *dev = &idxd->pdev->dev;
int rc, num_descs, i; int rc, num_descs, i;
if (wq->type != IDXD_WQT_KERNEL) if (wq->type != IDXD_WQT_KERNEL)
return 0; return 0;
num_descs = wq->size + wq->num_descs = wq->size;
idxd->hw.gen_cap.max_descs_per_engine * group->num_engines; num_descs = wq->size;
wq->num_descs = num_descs;
rc = alloc_hw_descs(wq, num_descs); rc = alloc_hw_descs(wq, num_descs);
if (rc < 0) if (rc < 0)
...@@ -187,7 +185,7 @@ int idxd_wq_alloc_resources(struct idxd_wq *wq) ...@@ -187,7 +185,7 @@ int idxd_wq_alloc_resources(struct idxd_wq *wq)
if (rc < 0) if (rc < 0)
goto fail_alloc_descs; goto fail_alloc_descs;
rc = sbitmap_init_node(&wq->sbmap, num_descs, -1, GFP_KERNEL, rc = sbitmap_queue_init_node(&wq->sbq, num_descs, -1, false, GFP_KERNEL,
dev_to_node(dev)); dev_to_node(dev));
if (rc < 0) if (rc < 0)
goto fail_sbitmap_init; goto fail_sbitmap_init;
...@@ -201,7 +199,7 @@ int idxd_wq_alloc_resources(struct idxd_wq *wq) ...@@ -201,7 +199,7 @@ int idxd_wq_alloc_resources(struct idxd_wq *wq)
sizeof(struct dsa_completion_record) * i; sizeof(struct dsa_completion_record) * i;
desc->id = i; desc->id = i;
desc->wq = wq; desc->wq = wq;
desc->cpu = -1;
dma_async_tx_descriptor_init(&desc->txd, &wq->dma_chan); dma_async_tx_descriptor_init(&desc->txd, &wq->dma_chan);
desc->txd.tx_submit = idxd_dma_tx_submit; desc->txd.tx_submit = idxd_dma_tx_submit;
} }
...@@ -227,7 +225,7 @@ void idxd_wq_free_resources(struct idxd_wq *wq) ...@@ -227,7 +225,7 @@ void idxd_wq_free_resources(struct idxd_wq *wq)
free_hw_descs(wq); free_hw_descs(wq);
free_descs(wq); free_descs(wq);
dma_free_coherent(dev, wq->compls_size, wq->compls, wq->compls_addr); dma_free_coherent(dev, wq->compls_size, wq->compls, wq->compls_addr);
sbitmap_free(&wq->sbmap); sbitmap_queue_free(&wq->sbq);
} }
int idxd_wq_enable(struct idxd_wq *wq) int idxd_wq_enable(struct idxd_wq *wq)
......
...@@ -104,7 +104,6 @@ struct idxd_wq { ...@@ -104,7 +104,6 @@ struct idxd_wq {
enum idxd_wq_state state; enum idxd_wq_state state;
unsigned long flags; unsigned long flags;
union wqcfg wqcfg; union wqcfg wqcfg;
atomic_t dq_count; /* dedicated queue flow control */
u32 vec_ptr; /* interrupt steering */ u32 vec_ptr; /* interrupt steering */
struct dsa_hw_desc **hw_descs; struct dsa_hw_desc **hw_descs;
int num_descs; int num_descs;
...@@ -112,10 +111,8 @@ struct idxd_wq { ...@@ -112,10 +111,8 @@ struct idxd_wq {
dma_addr_t compls_addr; dma_addr_t compls_addr;
int compls_size; int compls_size;
struct idxd_desc **descs; struct idxd_desc **descs;
struct sbitmap sbmap; struct sbitmap_queue sbq;
struct dma_chan dma_chan; struct dma_chan dma_chan;
struct percpu_rw_semaphore submit_lock;
wait_queue_head_t submit_waitq;
char name[WQ_NAME_SIZE + 1]; char name[WQ_NAME_SIZE + 1];
}; };
...@@ -201,6 +198,7 @@ struct idxd_desc { ...@@ -201,6 +198,7 @@ struct idxd_desc {
struct llist_node llnode; struct llist_node llnode;
struct list_head list; struct list_head list;
int id; int id;
int cpu;
struct idxd_wq *wq; struct idxd_wq *wq;
}; };
......
...@@ -141,17 +141,6 @@ static int idxd_setup_interrupts(struct idxd_device *idxd) ...@@ -141,17 +141,6 @@ static int idxd_setup_interrupts(struct idxd_device *idxd)
return rc; return rc;
} }
static void idxd_wqs_free_lock(struct idxd_device *idxd)
{
int i;
for (i = 0; i < idxd->max_wqs; i++) {
struct idxd_wq *wq = &idxd->wqs[i];
percpu_free_rwsem(&wq->submit_lock);
}
}
static int idxd_setup_internals(struct idxd_device *idxd) static int idxd_setup_internals(struct idxd_device *idxd)
{ {
struct device *dev = &idxd->pdev->dev; struct device *dev = &idxd->pdev->dev;
...@@ -181,19 +170,11 @@ static int idxd_setup_internals(struct idxd_device *idxd) ...@@ -181,19 +170,11 @@ static int idxd_setup_internals(struct idxd_device *idxd)
for (i = 0; i < idxd->max_wqs; i++) { for (i = 0; i < idxd->max_wqs; i++) {
struct idxd_wq *wq = &idxd->wqs[i]; struct idxd_wq *wq = &idxd->wqs[i];
int rc;
wq->id = i; wq->id = i;
wq->idxd = idxd; wq->idxd = idxd;
mutex_init(&wq->wq_lock); mutex_init(&wq->wq_lock);
atomic_set(&wq->dq_count, 0);
init_waitqueue_head(&wq->submit_waitq);
wq->idxd_cdev.minor = -1; wq->idxd_cdev.minor = -1;
rc = percpu_init_rwsem(&wq->submit_lock);
if (rc < 0) {
idxd_wqs_free_lock(idxd);
return rc;
}
} }
for (i = 0; i < idxd->max_engines; i++) { for (i = 0; i < idxd->max_engines; i++) {
...@@ -462,7 +443,6 @@ static void idxd_remove(struct pci_dev *pdev) ...@@ -462,7 +443,6 @@ static void idxd_remove(struct pci_dev *pdev)
dev_dbg(&pdev->dev, "%s called\n", __func__); dev_dbg(&pdev->dev, "%s called\n", __func__);
idxd_cleanup_sysfs(idxd); idxd_cleanup_sysfs(idxd);
idxd_shutdown(pdev); idxd_shutdown(pdev);
idxd_wqs_free_lock(idxd);
mutex_lock(&idxd_idr_lock); mutex_lock(&idxd_idr_lock);
idr_remove(&idxd_idrs[idxd->type], idxd->id); idr_remove(&idxd_idrs[idxd->type], idxd->id);
mutex_unlock(&idxd_idr_lock); mutex_unlock(&idxd_idr_lock);
......
...@@ -8,61 +8,61 @@ ...@@ -8,61 +8,61 @@
#include "idxd.h" #include "idxd.h"
#include "registers.h" #include "registers.h"
struct idxd_desc *idxd_alloc_desc(struct idxd_wq *wq, enum idxd_op_type optype) static struct idxd_desc *__get_desc(struct idxd_wq *wq, int idx, int cpu)
{ {
struct idxd_desc *desc; struct idxd_desc *desc;
int idx;
desc = wq->descs[idx];
memset(desc->hw, 0, sizeof(struct dsa_hw_desc));
memset(desc->completion, 0, sizeof(struct dsa_completion_record));
desc->cpu = cpu;
return desc;
}
struct idxd_desc *idxd_alloc_desc(struct idxd_wq *wq, enum idxd_op_type optype)
{
int cpu, idx;
struct idxd_device *idxd = wq->idxd; struct idxd_device *idxd = wq->idxd;
DEFINE_SBQ_WAIT(wait);
struct sbq_wait_state *ws;
struct sbitmap_queue *sbq;
if (idxd->state != IDXD_DEV_ENABLED) if (idxd->state != IDXD_DEV_ENABLED)
return ERR_PTR(-EIO); return ERR_PTR(-EIO);
if (optype == IDXD_OP_BLOCK) sbq = &wq->sbq;
percpu_down_read(&wq->submit_lock); idx = sbitmap_queue_get(sbq, &cpu);
else if (!percpu_down_read_trylock(&wq->submit_lock)) if (idx < 0) {
return ERR_PTR(-EBUSY); if (optype == IDXD_OP_NONBLOCK)
if (!atomic_add_unless(&wq->dq_count, 1, wq->size)) {
int rc;
if (optype == IDXD_OP_NONBLOCK) {
percpu_up_read(&wq->submit_lock);
return ERR_PTR(-EAGAIN); return ERR_PTR(-EAGAIN);
} else {
return __get_desc(wq, idx, cpu);
} }
percpu_up_read(&wq->submit_lock); ws = &sbq->ws[0];
percpu_down_write(&wq->submit_lock); for (;;) {
rc = wait_event_interruptible(wq->submit_waitq, sbitmap_prepare_to_wait(sbq, ws, &wait, TASK_INTERRUPTIBLE);
atomic_add_unless(&wq->dq_count, if (signal_pending_state(TASK_INTERRUPTIBLE, current))
1, wq->size) || break;
idxd->state != IDXD_DEV_ENABLED); idx = sbitmap_queue_get(sbq, &cpu);
percpu_up_write(&wq->submit_lock); if (idx > 0)
if (rc < 0) break;
return ERR_PTR(-EINTR); schedule();
if (idxd->state != IDXD_DEV_ENABLED)
return ERR_PTR(-EIO);
} else {
percpu_up_read(&wq->submit_lock);
} }
idx = sbitmap_get(&wq->sbmap, 0, false); sbitmap_finish_wait(sbq, ws, &wait);
if (idx < 0) { if (idx < 0)
atomic_dec(&wq->dq_count);
return ERR_PTR(-EAGAIN); return ERR_PTR(-EAGAIN);
}
desc = wq->descs[idx]; return __get_desc(wq, idx, cpu);
memset(desc->hw, 0, sizeof(struct dsa_hw_desc));
memset(desc->completion, 0, sizeof(struct dsa_completion_record));
return desc;
} }
void idxd_free_desc(struct idxd_wq *wq, struct idxd_desc *desc) void idxd_free_desc(struct idxd_wq *wq, struct idxd_desc *desc)
{ {
atomic_dec(&wq->dq_count); int cpu = desc->cpu;
sbitmap_clear_bit(&wq->sbmap, desc->id); desc->cpu = -1;
wake_up(&wq->submit_waitq); sbitmap_queue_clear(&wq->sbq, desc->id, cpu);
} }
int idxd_submit_desc(struct idxd_wq *wq, struct idxd_desc *desc) int idxd_submit_desc(struct idxd_wq *wq, struct idxd_desc *desc)
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment