crypto: caam - change return code in caam_jr_enqueue function
Based on commit 6b80ea389a
("crypto: change transient busy return code to -ENOSPC"),
change the return code of caam_jr_enqueue function to -EINPROGRESS, in
case of success, -ENOSPC in case the CAAM is busy (has no space left
in job ring queue), -EIO if it cannot map the caller's descriptor.
Update, also, the cases for resource-freeing for each algorithm type.
This is done for later use, on backlogging support in CAAM.
Signed-off-by: Iuliana Prodan <iuliana.prodan@nxp.com>
Reviewed-by: Horia Geanta <horia.geanta@nxp.com>
Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
This commit is contained in:
parent
d53e44fe98
commit
4d370a1036
@ -1398,9 +1398,7 @@ static inline int chachapoly_crypt(struct aead_request *req, bool encrypt)
|
||||
1);
|
||||
|
||||
ret = caam_jr_enqueue(jrdev, desc, aead_crypt_done, req);
|
||||
if (!ret) {
|
||||
ret = -EINPROGRESS;
|
||||
} else {
|
||||
if (ret != -EINPROGRESS) {
|
||||
aead_unmap(jrdev, edesc, req);
|
||||
kfree(edesc);
|
||||
}
|
||||
@ -1443,9 +1441,7 @@ static inline int aead_crypt(struct aead_request *req, bool encrypt)
|
||||
|
||||
desc = edesc->hw_desc;
|
||||
ret = caam_jr_enqueue(jrdev, desc, aead_crypt_done, req);
|
||||
if (!ret) {
|
||||
ret = -EINPROGRESS;
|
||||
} else {
|
||||
if (ret != -EINPROGRESS) {
|
||||
aead_unmap(jrdev, edesc, req);
|
||||
kfree(edesc);
|
||||
}
|
||||
@ -1488,9 +1484,7 @@ static inline int gcm_crypt(struct aead_request *req, bool encrypt)
|
||||
|
||||
desc = edesc->hw_desc;
|
||||
ret = caam_jr_enqueue(jrdev, desc, aead_crypt_done, req);
|
||||
if (!ret) {
|
||||
ret = -EINPROGRESS;
|
||||
} else {
|
||||
if (ret != -EINPROGRESS) {
|
||||
aead_unmap(jrdev, edesc, req);
|
||||
kfree(edesc);
|
||||
}
|
||||
@ -1706,9 +1700,7 @@ static inline int skcipher_crypt(struct skcipher_request *req, bool encrypt)
|
||||
desc = edesc->hw_desc;
|
||||
ret = caam_jr_enqueue(jrdev, desc, skcipher_crypt_done, req);
|
||||
|
||||
if (!ret) {
|
||||
ret = -EINPROGRESS;
|
||||
} else {
|
||||
if (ret != -EINPROGRESS) {
|
||||
skcipher_unmap(jrdev, edesc, req);
|
||||
kfree(edesc);
|
||||
}
|
||||
|
@ -395,7 +395,7 @@ static int hash_digest_key(struct caam_hash_ctx *ctx, u32 *keylen, u8 *key,
|
||||
init_completion(&result.completion);
|
||||
|
||||
ret = caam_jr_enqueue(jrdev, desc, split_key_done, &result);
|
||||
if (!ret) {
|
||||
if (ret == -EINPROGRESS) {
|
||||
/* in progress */
|
||||
wait_for_completion(&result.completion);
|
||||
ret = result.err;
|
||||
@ -828,10 +828,8 @@ static int ahash_update_ctx(struct ahash_request *req)
|
||||
desc_bytes(desc), 1);
|
||||
|
||||
ret = caam_jr_enqueue(jrdev, desc, ahash_done_bi, req);
|
||||
if (ret)
|
||||
if (ret != -EINPROGRESS)
|
||||
goto unmap_ctx;
|
||||
|
||||
ret = -EINPROGRESS;
|
||||
} else if (*next_buflen) {
|
||||
scatterwalk_map_and_copy(buf + *buflen, req->src, 0,
|
||||
req->nbytes, 0);
|
||||
@ -903,10 +901,9 @@ static int ahash_final_ctx(struct ahash_request *req)
|
||||
1);
|
||||
|
||||
ret = caam_jr_enqueue(jrdev, desc, ahash_done_ctx_src, req);
|
||||
if (ret)
|
||||
goto unmap_ctx;
|
||||
if (ret == -EINPROGRESS)
|
||||
return ret;
|
||||
|
||||
return -EINPROGRESS;
|
||||
unmap_ctx:
|
||||
ahash_unmap_ctx(jrdev, edesc, req, digestsize, DMA_BIDIRECTIONAL);
|
||||
kfree(edesc);
|
||||
@ -980,10 +977,9 @@ static int ahash_finup_ctx(struct ahash_request *req)
|
||||
1);
|
||||
|
||||
ret = caam_jr_enqueue(jrdev, desc, ahash_done_ctx_src, req);
|
||||
if (ret)
|
||||
goto unmap_ctx;
|
||||
if (ret == -EINPROGRESS)
|
||||
return ret;
|
||||
|
||||
return -EINPROGRESS;
|
||||
unmap_ctx:
|
||||
ahash_unmap_ctx(jrdev, edesc, req, digestsize, DMA_BIDIRECTIONAL);
|
||||
kfree(edesc);
|
||||
@ -1053,9 +1049,7 @@ static int ahash_digest(struct ahash_request *req)
|
||||
1);
|
||||
|
||||
ret = caam_jr_enqueue(jrdev, desc, ahash_done, req);
|
||||
if (!ret) {
|
||||
ret = -EINPROGRESS;
|
||||
} else {
|
||||
if (ret != -EINPROGRESS) {
|
||||
ahash_unmap_ctx(jrdev, edesc, req, digestsize, DMA_FROM_DEVICE);
|
||||
kfree(edesc);
|
||||
}
|
||||
@ -1105,9 +1099,7 @@ static int ahash_final_no_ctx(struct ahash_request *req)
|
||||
1);
|
||||
|
||||
ret = caam_jr_enqueue(jrdev, desc, ahash_done, req);
|
||||
if (!ret) {
|
||||
ret = -EINPROGRESS;
|
||||
} else {
|
||||
if (ret != -EINPROGRESS) {
|
||||
ahash_unmap_ctx(jrdev, edesc, req, digestsize, DMA_FROM_DEVICE);
|
||||
kfree(edesc);
|
||||
}
|
||||
@ -1218,10 +1210,9 @@ static int ahash_update_no_ctx(struct ahash_request *req)
|
||||
desc_bytes(desc), 1);
|
||||
|
||||
ret = caam_jr_enqueue(jrdev, desc, ahash_done_ctx_dst, req);
|
||||
if (ret)
|
||||
if (ret != -EINPROGRESS)
|
||||
goto unmap_ctx;
|
||||
|
||||
ret = -EINPROGRESS;
|
||||
state->update = ahash_update_ctx;
|
||||
state->finup = ahash_finup_ctx;
|
||||
state->final = ahash_final_ctx;
|
||||
@ -1310,9 +1301,7 @@ static int ahash_finup_no_ctx(struct ahash_request *req)
|
||||
1);
|
||||
|
||||
ret = caam_jr_enqueue(jrdev, desc, ahash_done, req);
|
||||
if (!ret) {
|
||||
ret = -EINPROGRESS;
|
||||
} else {
|
||||
if (ret != -EINPROGRESS) {
|
||||
ahash_unmap_ctx(jrdev, edesc, req, digestsize, DMA_FROM_DEVICE);
|
||||
kfree(edesc);
|
||||
}
|
||||
@ -1406,10 +1395,9 @@ static int ahash_update_first(struct ahash_request *req)
|
||||
desc_bytes(desc), 1);
|
||||
|
||||
ret = caam_jr_enqueue(jrdev, desc, ahash_done_ctx_dst, req);
|
||||
if (ret)
|
||||
if (ret != -EINPROGRESS)
|
||||
goto unmap_ctx;
|
||||
|
||||
ret = -EINPROGRESS;
|
||||
state->update = ahash_update_ctx;
|
||||
state->finup = ahash_finup_ctx;
|
||||
state->final = ahash_final_ctx;
|
||||
|
@ -634,8 +634,8 @@ static int caam_rsa_enc(struct akcipher_request *req)
|
||||
init_rsa_pub_desc(edesc->hw_desc, &edesc->pdb.pub);
|
||||
|
||||
ret = caam_jr_enqueue(jrdev, edesc->hw_desc, rsa_pub_done, req);
|
||||
if (!ret)
|
||||
return -EINPROGRESS;
|
||||
if (ret == -EINPROGRESS)
|
||||
return ret;
|
||||
|
||||
rsa_pub_unmap(jrdev, edesc, req);
|
||||
|
||||
@ -667,8 +667,8 @@ static int caam_rsa_dec_priv_f1(struct akcipher_request *req)
|
||||
init_rsa_priv_f1_desc(edesc->hw_desc, &edesc->pdb.priv_f1);
|
||||
|
||||
ret = caam_jr_enqueue(jrdev, edesc->hw_desc, rsa_priv_f_done, req);
|
||||
if (!ret)
|
||||
return -EINPROGRESS;
|
||||
if (ret == -EINPROGRESS)
|
||||
return ret;
|
||||
|
||||
rsa_priv_f1_unmap(jrdev, edesc, req);
|
||||
|
||||
@ -700,8 +700,8 @@ static int caam_rsa_dec_priv_f2(struct akcipher_request *req)
|
||||
init_rsa_priv_f2_desc(edesc->hw_desc, &edesc->pdb.priv_f2);
|
||||
|
||||
ret = caam_jr_enqueue(jrdev, edesc->hw_desc, rsa_priv_f_done, req);
|
||||
if (!ret)
|
||||
return -EINPROGRESS;
|
||||
if (ret == -EINPROGRESS)
|
||||
return ret;
|
||||
|
||||
rsa_priv_f2_unmap(jrdev, edesc, req);
|
||||
|
||||
@ -733,8 +733,8 @@ static int caam_rsa_dec_priv_f3(struct akcipher_request *req)
|
||||
init_rsa_priv_f3_desc(edesc->hw_desc, &edesc->pdb.priv_f3);
|
||||
|
||||
ret = caam_jr_enqueue(jrdev, edesc->hw_desc, rsa_priv_f_done, req);
|
||||
if (!ret)
|
||||
return -EINPROGRESS;
|
||||
if (ret == -EINPROGRESS)
|
||||
return ret;
|
||||
|
||||
rsa_priv_f3_unmap(jrdev, edesc, req);
|
||||
|
||||
|
@ -133,7 +133,7 @@ static inline int submit_job(struct caam_rng_ctx *ctx, int to_current)
|
||||
dev_dbg(jrdev, "submitting job %d\n", !(to_current ^ ctx->current_buf));
|
||||
init_completion(&bd->filled);
|
||||
err = caam_jr_enqueue(jrdev, desc, rng_done, ctx);
|
||||
if (err)
|
||||
if (err != -EINPROGRESS)
|
||||
complete(&bd->filled); /* don't wait on failed job*/
|
||||
else
|
||||
atomic_inc(&bd->empty); /* note if pending */
|
||||
@ -153,7 +153,7 @@ static int caam_read(struct hwrng *rng, void *data, size_t max, bool wait)
|
||||
if (atomic_read(&bd->empty) == BUF_EMPTY) {
|
||||
err = submit_job(ctx, 1);
|
||||
/* if can't submit job, can't even wait */
|
||||
if (err)
|
||||
if (err != -EINPROGRESS)
|
||||
return 0;
|
||||
}
|
||||
/* no immediate data, so exit if not waiting */
|
||||
|
@ -324,8 +324,8 @@ void caam_jr_free(struct device *rdev)
|
||||
EXPORT_SYMBOL(caam_jr_free);
|
||||
|
||||
/**
|
||||
* caam_jr_enqueue() - Enqueue a job descriptor head. Returns 0 if OK,
|
||||
* -EBUSY if the queue is full, -EIO if it cannot map the caller's
|
||||
* caam_jr_enqueue() - Enqueue a job descriptor head. Returns -EINPROGRESS
|
||||
* if OK, -ENOSPC if the queue is full, -EIO if it cannot map the caller's
|
||||
* descriptor.
|
||||
* @dev: device of the job ring to be used. This device should have
|
||||
* been assigned prior by caam_jr_register().
|
||||
@ -377,7 +377,7 @@ int caam_jr_enqueue(struct device *dev, u32 *desc,
|
||||
CIRC_SPACE(head, tail, JOBR_DEPTH) <= 0) {
|
||||
spin_unlock_bh(&jrp->inplock);
|
||||
dma_unmap_single(dev, desc_dma, desc_size, DMA_TO_DEVICE);
|
||||
return -EBUSY;
|
||||
return -ENOSPC;
|
||||
}
|
||||
|
||||
head_entry = &jrp->entinfo[head];
|
||||
@ -414,7 +414,7 @@ int caam_jr_enqueue(struct device *dev, u32 *desc,
|
||||
|
||||
spin_unlock_bh(&jrp->inplock);
|
||||
|
||||
return 0;
|
||||
return -EINPROGRESS;
|
||||
}
|
||||
EXPORT_SYMBOL(caam_jr_enqueue);
|
||||
|
||||
|
@ -108,7 +108,7 @@ int gen_split_key(struct device *jrdev, u8 *key_out,
|
||||
init_completion(&result.completion);
|
||||
|
||||
ret = caam_jr_enqueue(jrdev, desc, split_key_done, &result);
|
||||
if (!ret) {
|
||||
if (ret == -EINPROGRESS) {
|
||||
/* in progress */
|
||||
wait_for_completion(&result.completion);
|
||||
ret = result.err;
|
||||
|
Loading…
Reference in New Issue
Block a user