Commit fea40451 authored by LABBE Corentin's avatar LABBE Corentin Committed by Herbert Xu

crypto: qce - dma_map_sg can handle chained SG

The qce driver use two dma_map_sg path according to SG are chained
or not.
Since dma_map_sg can handle both case, clean the code with all
references to sg chained.

Thus removing qce_mapsg, qce_unmapsg and qce_countsg functions.
Signed-off-by: default avatarLABBE Corentin <clabbe.montjoie@gmail.com>
Signed-off-by: default avatarHerbert Xu <herbert@gondor.apana.org.au>
parent 166db195
...@@ -44,10 +44,8 @@ static void qce_ablkcipher_done(void *data) ...@@ -44,10 +44,8 @@ static void qce_ablkcipher_done(void *data)
error); error);
if (diff_dst) if (diff_dst)
qce_unmapsg(qce->dev, rctx->src_sg, rctx->src_nents, dir_src, dma_unmap_sg(qce->dev, rctx->src_sg, rctx->src_nents, dir_src);
rctx->dst_chained); dma_unmap_sg(qce->dev, rctx->dst_sg, rctx->dst_nents, dir_dst);
qce_unmapsg(qce->dev, rctx->dst_sg, rctx->dst_nents, dir_dst,
rctx->dst_chained);
sg_free_table(&rctx->dst_tbl); sg_free_table(&rctx->dst_tbl);
...@@ -80,15 +78,11 @@ qce_ablkcipher_async_req_handle(struct crypto_async_request *async_req) ...@@ -80,15 +78,11 @@ qce_ablkcipher_async_req_handle(struct crypto_async_request *async_req)
dir_src = diff_dst ? DMA_TO_DEVICE : DMA_BIDIRECTIONAL; dir_src = diff_dst ? DMA_TO_DEVICE : DMA_BIDIRECTIONAL;
dir_dst = diff_dst ? DMA_FROM_DEVICE : DMA_BIDIRECTIONAL; dir_dst = diff_dst ? DMA_FROM_DEVICE : DMA_BIDIRECTIONAL;
rctx->src_nents = qce_countsg(req->src, req->nbytes, rctx->src_nents = sg_nents_for_len(req->src, req->nbytes);
&rctx->src_chained); if (diff_dst)
if (diff_dst) { rctx->dst_nents = sg_nents_for_len(req->dst, req->nbytes);
rctx->dst_nents = qce_countsg(req->dst, req->nbytes, else
&rctx->dst_chained);
} else {
rctx->dst_nents = rctx->src_nents; rctx->dst_nents = rctx->src_nents;
rctx->dst_chained = rctx->src_chained;
}
rctx->dst_nents += 1; rctx->dst_nents += 1;
...@@ -116,14 +110,12 @@ qce_ablkcipher_async_req_handle(struct crypto_async_request *async_req) ...@@ -116,14 +110,12 @@ qce_ablkcipher_async_req_handle(struct crypto_async_request *async_req)
sg_mark_end(sg); sg_mark_end(sg);
rctx->dst_sg = rctx->dst_tbl.sgl; rctx->dst_sg = rctx->dst_tbl.sgl;
ret = qce_mapsg(qce->dev, rctx->dst_sg, rctx->dst_nents, dir_dst, ret = dma_map_sg(qce->dev, rctx->dst_sg, rctx->dst_nents, dir_dst);
rctx->dst_chained);
if (ret < 0) if (ret < 0)
goto error_free; goto error_free;
if (diff_dst) { if (diff_dst) {
ret = qce_mapsg(qce->dev, req->src, rctx->src_nents, dir_src, ret = dma_map_sg(qce->dev, req->src, rctx->src_nents, dir_src);
rctx->src_chained);
if (ret < 0) if (ret < 0)
goto error_unmap_dst; goto error_unmap_dst;
rctx->src_sg = req->src; rctx->src_sg = req->src;
...@@ -149,11 +141,9 @@ qce_ablkcipher_async_req_handle(struct crypto_async_request *async_req) ...@@ -149,11 +141,9 @@ qce_ablkcipher_async_req_handle(struct crypto_async_request *async_req)
qce_dma_terminate_all(&qce->dma); qce_dma_terminate_all(&qce->dma);
error_unmap_src: error_unmap_src:
if (diff_dst) if (diff_dst)
qce_unmapsg(qce->dev, req->src, rctx->src_nents, dir_src, dma_unmap_sg(qce->dev, req->src, rctx->src_nents, dir_src);
rctx->src_chained);
error_unmap_dst: error_unmap_dst:
qce_unmapsg(qce->dev, rctx->dst_sg, rctx->dst_nents, dir_dst, dma_unmap_sg(qce->dev, rctx->dst_sg, rctx->dst_nents, dir_dst);
rctx->dst_chained);
error_free: error_free:
sg_free_table(&rctx->dst_tbl); sg_free_table(&rctx->dst_tbl);
return ret; return ret;
......
...@@ -32,8 +32,6 @@ struct qce_cipher_ctx { ...@@ -32,8 +32,6 @@ struct qce_cipher_ctx {
* @ivsize: IV size * @ivsize: IV size
* @src_nents: source entries * @src_nents: source entries
* @dst_nents: destination entries * @dst_nents: destination entries
* @src_chained: is source chained
* @dst_chained: is destination chained
* @result_sg: scatterlist used for result buffer * @result_sg: scatterlist used for result buffer
* @dst_tbl: destination sg table * @dst_tbl: destination sg table
* @dst_sg: destination sg pointer table beginning * @dst_sg: destination sg pointer table beginning
...@@ -47,8 +45,6 @@ struct qce_cipher_reqctx { ...@@ -47,8 +45,6 @@ struct qce_cipher_reqctx {
unsigned int ivsize; unsigned int ivsize;
int src_nents; int src_nents;
int dst_nents; int dst_nents;
bool src_chained;
bool dst_chained;
struct scatterlist result_sg; struct scatterlist result_sg;
struct sg_table dst_tbl; struct sg_table dst_tbl;
struct scatterlist *dst_sg; struct scatterlist *dst_sg;
......
...@@ -54,58 +54,6 @@ void qce_dma_release(struct qce_dma_data *dma) ...@@ -54,58 +54,6 @@ void qce_dma_release(struct qce_dma_data *dma)
kfree(dma->result_buf); kfree(dma->result_buf);
} }
int qce_mapsg(struct device *dev, struct scatterlist *sg, int nents,
enum dma_data_direction dir, bool chained)
{
int err;
if (chained) {
while (sg) {
err = dma_map_sg(dev, sg, 1, dir);
if (!err)
return -EFAULT;
sg = sg_next(sg);
}
} else {
err = dma_map_sg(dev, sg, nents, dir);
if (!err)
return -EFAULT;
}
return nents;
}
void qce_unmapsg(struct device *dev, struct scatterlist *sg, int nents,
enum dma_data_direction dir, bool chained)
{
if (chained)
while (sg) {
dma_unmap_sg(dev, sg, 1, dir);
sg = sg_next(sg);
}
else
dma_unmap_sg(dev, sg, nents, dir);
}
int qce_countsg(struct scatterlist *sglist, int nbytes, bool *chained)
{
struct scatterlist *sg = sglist;
int nents = 0;
if (chained)
*chained = false;
while (nbytes > 0 && sg) {
nents++;
nbytes -= sg->length;
if (!sg_is_last(sg) && (sg + 1)->length == 0 && chained)
*chained = true;
sg = sg_next(sg);
}
return nents;
}
struct scatterlist * struct scatterlist *
qce_sgtable_add(struct sg_table *sgt, struct scatterlist *new_sgl) qce_sgtable_add(struct sg_table *sgt, struct scatterlist *new_sgl)
{ {
......
...@@ -49,11 +49,6 @@ int qce_dma_prep_sgs(struct qce_dma_data *dma, struct scatterlist *sg_in, ...@@ -49,11 +49,6 @@ int qce_dma_prep_sgs(struct qce_dma_data *dma, struct scatterlist *sg_in,
dma_async_tx_callback cb, void *cb_param); dma_async_tx_callback cb, void *cb_param);
void qce_dma_issue_pending(struct qce_dma_data *dma); void qce_dma_issue_pending(struct qce_dma_data *dma);
int qce_dma_terminate_all(struct qce_dma_data *dma); int qce_dma_terminate_all(struct qce_dma_data *dma);
int qce_countsg(struct scatterlist *sg_list, int nbytes, bool *chained);
void qce_unmapsg(struct device *dev, struct scatterlist *sg, int nents,
enum dma_data_direction dir, bool chained);
int qce_mapsg(struct device *dev, struct scatterlist *sg, int nents,
enum dma_data_direction dir, bool chained);
struct scatterlist * struct scatterlist *
qce_sgtable_add(struct sg_table *sgt, struct scatterlist *sg_add); qce_sgtable_add(struct sg_table *sgt, struct scatterlist *sg_add);
......
...@@ -51,9 +51,8 @@ static void qce_ahash_done(void *data) ...@@ -51,9 +51,8 @@ static void qce_ahash_done(void *data)
if (error) if (error)
dev_dbg(qce->dev, "ahash dma termination error (%d)\n", error); dev_dbg(qce->dev, "ahash dma termination error (%d)\n", error);
qce_unmapsg(qce->dev, req->src, rctx->src_nents, DMA_TO_DEVICE, dma_unmap_sg(qce->dev, req->src, rctx->src_nents, DMA_TO_DEVICE);
rctx->src_chained); dma_unmap_sg(qce->dev, &rctx->result_sg, 1, DMA_FROM_DEVICE);
qce_unmapsg(qce->dev, &rctx->result_sg, 1, DMA_FROM_DEVICE, 0);
memcpy(rctx->digest, result->auth_iv, digestsize); memcpy(rctx->digest, result->auth_iv, digestsize);
if (req->result) if (req->result)
...@@ -92,16 +91,14 @@ static int qce_ahash_async_req_handle(struct crypto_async_request *async_req) ...@@ -92,16 +91,14 @@ static int qce_ahash_async_req_handle(struct crypto_async_request *async_req)
rctx->authklen = AES_KEYSIZE_128; rctx->authklen = AES_KEYSIZE_128;
} }
rctx->src_nents = qce_countsg(req->src, req->nbytes, rctx->src_nents = sg_nents_for_len(req->src, req->nbytes);
&rctx->src_chained); ret = dma_map_sg(qce->dev, req->src, rctx->src_nents, DMA_TO_DEVICE);
ret = qce_mapsg(qce->dev, req->src, rctx->src_nents, DMA_TO_DEVICE,
rctx->src_chained);
if (ret < 0) if (ret < 0)
return ret; return ret;
sg_init_one(&rctx->result_sg, qce->dma.result_buf, QCE_RESULT_BUF_SZ); sg_init_one(&rctx->result_sg, qce->dma.result_buf, QCE_RESULT_BUF_SZ);
ret = qce_mapsg(qce->dev, &rctx->result_sg, 1, DMA_FROM_DEVICE, 0); ret = dma_map_sg(qce->dev, &rctx->result_sg, 1, DMA_FROM_DEVICE);
if (ret < 0) if (ret < 0)
goto error_unmap_src; goto error_unmap_src;
...@@ -121,10 +118,9 @@ static int qce_ahash_async_req_handle(struct crypto_async_request *async_req) ...@@ -121,10 +118,9 @@ static int qce_ahash_async_req_handle(struct crypto_async_request *async_req)
error_terminate: error_terminate:
qce_dma_terminate_all(&qce->dma); qce_dma_terminate_all(&qce->dma);
error_unmap_dst: error_unmap_dst:
qce_unmapsg(qce->dev, &rctx->result_sg, 1, DMA_FROM_DEVICE, 0); dma_unmap_sg(qce->dev, &rctx->result_sg, 1, DMA_FROM_DEVICE);
error_unmap_src: error_unmap_src:
qce_unmapsg(qce->dev, req->src, rctx->src_nents, DMA_TO_DEVICE, dma_unmap_sg(qce->dev, req->src, rctx->src_nents, DMA_TO_DEVICE);
rctx->src_chained);
return ret; return ret;
} }
......
...@@ -36,7 +36,6 @@ struct qce_sha_ctx { ...@@ -36,7 +36,6 @@ struct qce_sha_ctx {
* @flags: operation flags * @flags: operation flags
* @src_orig: original request sg list * @src_orig: original request sg list
* @nbytes_orig: original request number of bytes * @nbytes_orig: original request number of bytes
* @src_chained: is source scatterlist chained
* @src_nents: source number of entries * @src_nents: source number of entries
* @byte_count: byte count * @byte_count: byte count
* @count: save count in states during update, import and export * @count: save count in states during update, import and export
...@@ -55,7 +54,6 @@ struct qce_sha_reqctx { ...@@ -55,7 +54,6 @@ struct qce_sha_reqctx {
unsigned long flags; unsigned long flags;
struct scatterlist *src_orig; struct scatterlist *src_orig;
unsigned int nbytes_orig; unsigned int nbytes_orig;
bool src_chained;
int src_nents; int src_nents;
__be32 byte_count[2]; __be32 byte_count[2];
u64 count; u64 count;
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment