crypto: msm: disable aead alg related apis temporarily

Just temporarily disable aead alg related apis for 4.4 upgrade.
4.4 upstream cryto driver has a lot of changes on aead related APIS
and data structures, and hw crypto driver can not be compiled on 4.4.

Signed-off-by: Zhen Kong <zkong@codeaurora.org>
This commit is contained in:
Zhen Kong 2016-02-26 18:00:10 -08:00 committed by David Keitel
parent 1db07b7a79
commit 97a7e8fcc4
3 changed files with 35 additions and 18 deletions

View file

@ -207,7 +207,7 @@ static int count_sg(struct scatterlist *sg, int nbytes)
{
int i;
for (i = 0; nbytes > 0; i++, sg = scatterwalk_sg_next(sg))
for (i = 0; nbytes > 0; i++, sg = sg_next(sg))
nbytes -= sg->length;
return i;
}
@ -219,7 +219,7 @@ static int qce_dma_map_sg(struct device *dev, struct scatterlist *sg, int nents,
for (i = 0; i < nents; ++i) {
dma_map_sg(dev, sg, 1, direction);
sg = scatterwalk_sg_next(sg);
sg = sg_next(sg);
}
return nents;
@ -232,7 +232,7 @@ static int qce_dma_unmap_sg(struct device *dev, struct scatterlist *sg,
for (i = 0; i < nents; ++i) {
dma_unmap_sg(dev, sg, 1, direction);
sg = scatterwalk_sg_next(sg);
sg = sg_next(sg);
}
return nents;
@ -2035,6 +2035,7 @@ static int _qce_unlock_other_pipes(struct qce_device *pce_dev, int req_info)
static inline void qce_free_req_info(struct qce_device *pce_dev, int req_info,
bool is_complete);
#if 0
static int _aead_complete(struct qce_device *pce_dev, int req_info)
{
struct aead_request *areq;
@ -2124,6 +2125,7 @@ static int _aead_complete(struct qce_device *pce_dev, int req_info)
}
return 0;
};
#endif
static int _sha_complete(struct qce_device *pce_dev, int req_info)
{
@ -2450,7 +2452,7 @@ static int _qce_sps_add_sg_data(struct qce_device *pce_dev,
addr += data_cnt;
len -= data_cnt;
}
sg_src = scatterwalk_sg_next(sg_src);
sg_src = sg_next(sg_src);
}
return 0;
}
@ -2936,9 +2938,11 @@ static void _qce_req_complete(struct qce_device *pce_dev, unsigned int req_info)
case QCE_XFER_HASHING:
_sha_complete(pce_dev, req_info);
break;
#if 0
case QCE_XFER_AEAD:
_aead_complete(pce_dev, req_info);
break;
#endif
case QCE_XFER_F8:
_f8_complete(pce_dev, req_info);
break;
@ -4569,7 +4573,7 @@ static int select_mode(struct qce_device *pce_dev,
return 0;
}
#if 0
static int _qce_aead_ccm_req(void *handle, struct qce_req *q_req)
{
struct qce_device *pce_dev = (struct qce_device *) handle;
@ -4782,6 +4786,7 @@ bad:
return rc;
}
#endif
static int _qce_suspend(void *handle)
{
@ -4844,6 +4849,7 @@ static int _qce_resume(void *handle)
struct qce_pm_table qce_pm_table = {_qce_suspend, _qce_resume};
EXPORT_SYMBOL(qce_pm_table);
#if 0
int qce_aead_req(void *handle, struct qce_req *q_req)
{
struct qce_device *pce_dev = (struct qce_device *)handle;
@ -5053,6 +5059,7 @@ bad:
return rc;
}
EXPORT_SYMBOL(qce_aead_req);
#endif
int qce_ablk_cipher_req(void *handle, struct qce_req *c_req)
{

View file

@ -691,7 +691,7 @@ static int qcrypto_count_sg(struct scatterlist *sg, int nbytes)
{
int i;
for (i = 0; nbytes > 0 && sg != NULL; i++, sg = scatterwalk_sg_next(sg))
for (i = 0; nbytes > 0 && sg != NULL; i++, sg = sg_next(sg))
nbytes -= sg->length;
return i;
@ -708,7 +708,7 @@ static size_t qcrypto_sg_copy_from_buffer(struct scatterlist *sgl,
buf += len;
buflen -= len;
offset += len;
sgl = scatterwalk_sg_next(sgl);
sgl = sg_next(sgl);
}
return offset;
@ -725,7 +725,7 @@ static size_t qcrypto_sg_copy_to_buffer(struct scatterlist *sgl,
buf += len;
buflen -= len;
offset += len;
sgl = scatterwalk_sg_next(sgl);
sgl = sg_next(sgl);
}
return offset;
@ -887,6 +887,7 @@ static int _qcrypto_cra_aes_ablkcipher_init(struct crypto_tfm *tfm)
return _qcrypto_cra_ablkcipher_init(tfm);
};
#if 0
static int _qcrypto_cra_aead_sha1_init(struct crypto_tfm *tfm)
{
int rc;
@ -930,6 +931,7 @@ static int _qcrypto_cra_aead_rfc4309_ccm_init(struct crypto_tfm *tfm)
ctx->auth_alg = QCE_HASH_AES_CMAC;
return rc;
}
#endif
static void _qcrypto_cra_ablkcipher_exit(struct crypto_tfm *tfm)
{
@ -1606,7 +1608,7 @@ static void _qce_ablk_cipher_complete(void *cookie, unsigned char *icb,
req_done(pqcrypto_req_control);
};
#if 0
static void _qce_aead_complete(void *cookie, unsigned char *icv,
unsigned char *iv, int ret)
{
@ -1728,6 +1730,7 @@ static void _qce_aead_complete(void *cookie, unsigned char *icv,
pqcrypto_req_control->arsp->res = ret;
req_done(pqcrypto_req_control);
}
#endif
static int aead_ccm_set_msg_len(u8 *block, unsigned int msglen, int csize)
{
@ -1942,6 +1945,7 @@ static int _qcrypto_process_ahash(struct crypto_engine *pengine,
return ret;
}
#if 0
static int _qcrypto_process_aead(struct crypto_engine *pengine,
struct qcrypto_req_control *pqcrypto_req_control)
{
@ -2145,6 +2149,7 @@ static int _qcrypto_process_aead(struct crypto_engine *pengine,
return ret;
}
#endif
static struct crypto_engine *_qcrypto_static_assign_engine(
struct crypto_priv *cp)
@ -2284,9 +2289,11 @@ again:
case CRYPTO_ALG_TYPE_AHASH:
ret = _qcrypto_process_ahash(pengine, pqcrypto_req_control);
break;
#if 0
case CRYPTO_ALG_TYPE_AEAD:
ret = _qcrypto_process_aead(pengine, pqcrypto_req_control);
break;
#endif
default:
ret = -EINVAL;
};
@ -2853,7 +2860,7 @@ static int _qcrypto_dec_aes_xts(struct ablkcipher_request *req)
return _qcrypto_queue_req(cp, ctx->pengine, &req->base);
};
#if 0
static int _qcrypto_aead_decrypt_aes_ccm(struct aead_request *req)
{
struct qcrypto_cipher_req_ctx *rctx;
@ -3259,6 +3266,7 @@ static int _qcrypto_aead_givencrypt_3des_cbc(struct aead_givcrypt_request *req)
pstat->aead_sha256_3des_enc++;
return _qcrypto_queue_req(cp, ctx->pengine, &areq->base);
}
#endif
static int _sha_init(struct ahash_request *req)
{
@ -3523,7 +3531,7 @@ static int _sha_update(struct ahash_request *req, uint32_t sha_block_size)
if ((len + sg_last->length) > nbytes)
break;
len += sg_last->length;
sg_last = scatterwalk_sg_next(sg_last);
sg_last = sg_next(sg_last);
}
if (rctx->trailing_buf_len) {
if (cp->ce_support.aligned_only) {
@ -4463,7 +4471,7 @@ static struct crypto_alg _qcrypto_ablk_cipher_xts_algo = {
},
},
};
#if 0
static struct crypto_alg _qcrypto_aead_sha1_hmac_algos[] = {
{
.cra_name = "authenc(hmac(sha1),cbc(aes))",
@ -4670,7 +4678,7 @@ static struct crypto_alg _qcrypto_aead_rfc4309_ccm_algo = {
}
}
};
#endif
static int _qcrypto_probe(struct platform_device *pdev)
{
@ -4891,6 +4899,7 @@ static int _qcrypto_probe(struct platform_device *pdev)
}
/* register crypto aead (hmac-sha1) algorithms the device supports */
#if 0
if (cp->ce_support.sha1_hmac_20 || cp->ce_support.sha1_hmac
|| cp->ce_support.sha_hmac) {
for (i = 0; i < ARRAY_SIZE(_qcrypto_aead_sha1_hmac_algos);
@ -4966,7 +4975,7 @@ static int _qcrypto_probe(struct platform_device *pdev)
}
}
}
#endif
if ((cp->ce_support.sha_hmac) || (cp->platform_support.sha_hmac)) {
/* register crypto hmac algorithms the device supports */
for (i = 0; i < ARRAY_SIZE(_qcrypto_sha_hmac_algos); i++) {
@ -5009,6 +5018,7 @@ static int _qcrypto_probe(struct platform_device *pdev)
* Register crypto cipher (aes-ccm) algorithms the
* device supports
*/
#if 0
if (cp->ce_support.aes_ccm) {
struct qcrypto_alg *q_alg;
@ -5070,7 +5080,7 @@ static int _qcrypto_probe(struct platform_device *pdev)
q_alg->cipher_alg.cra_driver_name);
}
}
#endif
mutex_unlock(&cp->engine_lock);

View file

@ -27,16 +27,16 @@
int qcrypto_cipher_set_device(struct ablkcipher_request *req, unsigned int dev);
int qcrypto_ahash_set_device(struct ahash_request *req, unsigned int dev);
int qcrypto_aead_set_device(struct aead_request *req, unsigned int dev);
/*int qcrypto_aead_set_device(struct aead_request *req, unsigned int dev);*/
int qcrypto_cipher_set_flag(struct ablkcipher_request *req, unsigned int flags);
int qcrypto_ahash_set_flag(struct ahash_request *req, unsigned int flags);
int qcrypto_aead_set_flag(struct aead_request *req, unsigned int flags);
/*int qcrypto_aead_set_flag(struct aead_request *req, unsigned int flags);*/
int qcrypto_cipher_clear_flag(struct ablkcipher_request *req,
unsigned int flags);
int qcrypto_ahash_clear_flag(struct ahash_request *req, unsigned int flags);
int qcrypto_aead_clear_flag(struct aead_request *req, unsigned int flags);
/*int qcrypto_aead_clear_flag(struct aead_request *req, unsigned int flags);*/
struct crypto_engine_entry {
u32 hw_instance;