/linux-4.19.296/crypto/ |
D | authencesn.c | 112 unsigned int cryptlen = req->cryptlen; in crypto_authenc_esn_genicv_tail() local 118 scatterwalk_map_and_copy(tmp + 1, dst, assoclen + cryptlen, 4, 0); in crypto_authenc_esn_genicv_tail() 121 scatterwalk_map_and_copy(hash, dst, assoclen + cryptlen, authsize, 1); in crypto_authenc_esn_genicv_tail() 146 unsigned int cryptlen = req->cryptlen; in crypto_authenc_esn_genicv() local 156 scatterwalk_map_and_copy(tmp + 1, dst, assoclen + cryptlen, 4, 1); in crypto_authenc_esn_genicv() 162 ahash_request_set_crypt(ahreq, dst, hash, assoclen + cryptlen); in crypto_authenc_esn_genicv() 205 unsigned int cryptlen = req->cryptlen; in crypto_authenc_esn_encrypt() local 225 skcipher_request_set_crypt(skreq, src, dst, cryptlen, req->iv); in crypto_authenc_esn_encrypt() 246 unsigned int cryptlen = req->cryptlen - authsize; in crypto_authenc_esn_decrypt_tail() local 257 scatterwalk_map_and_copy(tmp + 1, dst, assoclen + cryptlen, 4, 0); in crypto_authenc_esn_decrypt_tail() [all …]
|
D | lrw.c | 182 rctx->left -= subreq->cryptlen; in post_crypt() 215 unsigned cryptlen; in pre_crypt() local 224 cryptlen = subreq->cryptlen; in pre_crypt() 225 more = rctx->left > cryptlen; in pre_crypt() 227 cryptlen = rctx->left; in pre_crypt() 230 cryptlen, req->iv); in pre_crypt() 259 cryptlen, NULL); in pre_crypt() 296 subreq->cryptlen = LRW_BUFFER_SIZE; in init_crypt() 297 if (req->cryptlen > LRW_BUFFER_SIZE) { in init_crypt() 298 unsigned int n = min(req->cryptlen, (unsigned int)PAGE_SIZE); in init_crypt() [all …]
|
D | xts.c | 127 rctx->left -= subreq->cryptlen; in post_crypt() 158 unsigned cryptlen; in pre_crypt() local 164 cryptlen = subreq->cryptlen; in pre_crypt() 166 more = rctx->left > cryptlen; in pre_crypt() 168 cryptlen = rctx->left; in pre_crypt() 171 cryptlen, NULL); in pre_crypt() 193 cryptlen, NULL); in pre_crypt() 231 subreq->cryptlen = XTS_BUFFER_SIZE; in init_crypt() 232 if (req->cryptlen > XTS_BUFFER_SIZE) { in init_crypt() 233 unsigned int n = min(req->cryptlen, (unsigned int)PAGE_SIZE); in init_crypt() [all …]
|
D | aegis128.c | 311 u64 assoclen, u64 cryptlen) in crypto_aegis128_final() argument 314 u64 cryptbits = cryptlen * 8; in crypto_aegis128_final() 357 unsigned int cryptlen, in crypto_aegis128_crypt() argument 367 crypto_aegis128_final(&state, tag_xor, req->assoclen, cryptlen); in crypto_aegis128_crypt() 380 unsigned int cryptlen = req->cryptlen; in crypto_aegis128_encrypt() local 382 crypto_aegis128_crypt(req, &tag, cryptlen, &ops); in crypto_aegis128_encrypt() 384 scatterwalk_map_and_copy(tag.bytes, req->dst, req->assoclen + cryptlen, in crypto_aegis128_encrypt() 400 unsigned int cryptlen = req->cryptlen - authsize; in crypto_aegis128_decrypt() local 402 scatterwalk_map_and_copy(tag.bytes, req->src, req->assoclen + cryptlen, in crypto_aegis128_decrypt() 405 crypto_aegis128_crypt(req, &tag, cryptlen, &ops); in crypto_aegis128_decrypt()
|
D | chacha20poly1305.c | 47 __le64 cryptlen; member 67 unsigned int cryptlen; member 109 req->assoclen + rctx->cryptlen, in poly_verify_tag() 121 req->assoclen + rctx->cryptlen, in poly_copy_tag() 139 if (rctx->cryptlen == 0) in chacha_decrypt() 157 rctx->cryptlen, creq->iv); in chacha_decrypt() 170 if (rctx->cryptlen == req->cryptlen) /* encrypting */ in poly_tail_continue() 193 len = cpu_to_le64(rctx->cryptlen); in poly_tail() 194 memcpy(&preq->tail.cryptlen, &len, sizeof(len)); in poly_tail() 223 padlen = (bs - (rctx->cryptlen % bs)) % bs; in poly_cipherpad() [all …]
|
D | aegis256.c | 324 u64 assoclen, u64 cryptlen) in crypto_aegis256_final() argument 327 u64 cryptbits = cryptlen * 8; in crypto_aegis256_final() 372 unsigned int cryptlen, in crypto_aegis256_crypt() argument 382 crypto_aegis256_final(&state, tag_xor, req->assoclen, cryptlen); in crypto_aegis256_crypt() 395 unsigned int cryptlen = req->cryptlen; in crypto_aegis256_encrypt() local 397 crypto_aegis256_crypt(req, &tag, cryptlen, &ops); in crypto_aegis256_encrypt() 399 scatterwalk_map_and_copy(tag.bytes, req->dst, req->assoclen + cryptlen, in crypto_aegis256_encrypt() 415 unsigned int cryptlen = req->cryptlen - authsize; in crypto_aegis256_decrypt() local 417 scatterwalk_map_and_copy(tag.bytes, req->src, req->assoclen + cryptlen, in crypto_aegis256_decrypt() 420 crypto_aegis256_crypt(req, &tag, cryptlen, &ops); in crypto_aegis256_decrypt()
|
D | morus640.c | 386 u64 assoclen, u64 cryptlen) in crypto_morus640_final() argument 393 tmp.words[2] = lower_32_bits(cryptlen * 8); in crypto_morus640_final() 394 tmp.words[3] = upper_32_bits(cryptlen * 8); in crypto_morus640_final() 427 unsigned int cryptlen, in crypto_morus640_crypt() argument 437 crypto_morus640_final(&state, tag_xor, req->assoclen, cryptlen); in crypto_morus640_crypt() 451 unsigned int cryptlen = req->cryptlen; in crypto_morus640_encrypt() local 453 crypto_morus640_crypt(req, &tag, cryptlen, &ops); in crypto_morus640_encrypt() 457 req->assoclen + cryptlen, authsize, 1); in crypto_morus640_encrypt() 473 unsigned int cryptlen = req->cryptlen - authsize; in crypto_morus640_decrypt() local 476 req->assoclen + cryptlen, authsize, 0); in crypto_morus640_decrypt() [all …]
|
D | ccm.c | 138 unsigned int cryptlen) in format_input() argument 156 return set_msg_len(info + 16 - l, cryptlen, l); in format_input() 179 unsigned int cryptlen) in crypto_ccm_auth() argument 192 err = format_input(odata, req, cryptlen); in crypto_ccm_auth() 228 cryptlen += ilen; in crypto_ccm_auth() 231 ahash_request_set_crypt(ahreq, plain, pctx->odata, cryptlen); in crypto_ccm_auth() 246 req->assoclen + req->cryptlen, in crypto_ccm_encrypt_done() 302 unsigned int cryptlen = req->cryptlen; in crypto_ccm_encrypt() local 311 err = crypto_ccm_auth(req, sg_next(pctx->src), cryptlen); in crypto_ccm_encrypt() 322 skcipher_request_set_crypt(skreq, pctx->src, dst, cryptlen + 16, iv); in crypto_ccm_encrypt() [all …]
|
D | aegis128l.c | 374 u64 assoclen, u64 cryptlen) in crypto_aegis128l_final() argument 377 u64 cryptbits = cryptlen * 8; in crypto_aegis128l_final() 421 unsigned int cryptlen, in crypto_aegis128l_crypt() argument 431 crypto_aegis128l_final(&state, tag_xor, req->assoclen, cryptlen); in crypto_aegis128l_crypt() 444 unsigned int cryptlen = req->cryptlen; in crypto_aegis128l_encrypt() local 446 crypto_aegis128l_crypt(req, &tag, cryptlen, &ops); in crypto_aegis128l_encrypt() 448 scatterwalk_map_and_copy(tag.bytes, req->dst, req->assoclen + cryptlen, in crypto_aegis128l_encrypt() 464 unsigned int cryptlen = req->cryptlen - authsize; in crypto_aegis128l_decrypt() local 466 scatterwalk_map_and_copy(tag.bytes, req->src, req->assoclen + cryptlen, in crypto_aegis128l_decrypt() 469 crypto_aegis128l_crypt(req, &tag, cryptlen, &ops); in crypto_aegis128l_decrypt()
|
D | echainiv.c | 42 if (req->cryptlen < ivsize) in echainiv_encrypt() 56 req->assoclen + req->cryptlen, in echainiv_encrypt() 67 req->cryptlen, info); in echainiv_encrypt() 99 if (req->cryptlen < ivsize) in echainiv_decrypt() 109 req->cryptlen - ivsize, req->iv); in echainiv_decrypt()
|
D | morus1280.c | 387 u64 assoclen, u64 cryptlen) in crypto_morus1280_final() argument 393 tmp.words[1] = cryptlen * 8; in crypto_morus1280_final() 435 unsigned int cryptlen, in crypto_morus1280_crypt() argument 445 crypto_morus1280_final(&state, tag_xor, req->assoclen, cryptlen); in crypto_morus1280_crypt() 459 unsigned int cryptlen = req->cryptlen; in crypto_morus1280_encrypt() local 461 crypto_morus1280_crypt(req, &tag, cryptlen, &ops); in crypto_morus1280_encrypt() 465 req->assoclen + cryptlen, authsize, 1); in crypto_morus1280_encrypt() 481 unsigned int cryptlen = req->cryptlen - authsize; in crypto_morus1280_decrypt() local 484 req->assoclen + cryptlen, authsize, 0); in crypto_morus1280_decrypt() 487 crypto_morus1280_crypt(req, &tag, cryptlen, &ops); in crypto_morus1280_decrypt()
|
D | gcm.c | 62 unsigned int cryptlen; member 200 unsigned int cryptlen) in crypto_gcm_init_crypt() argument 212 cryptlen + sizeof(pctx->auth_tag), in crypto_gcm_init_crypt() 253 lengths.b = cpu_to_be64(gctx->cryptlen * 8); in gcm_hash_len() 314 remain = gcm_remain(gctx->cryptlen); in gcm_hash_crypt_continue() 343 if (gctx->cryptlen) in gcm_hash_assoc_remain_continue() 345 gctx->src, gctx->cryptlen, flags) ?: in gcm_hash_assoc_remain_continue() 441 req->assoclen + req->cryptlen, in gcm_enc_copy_hash() 452 gctx->cryptlen = req->cryptlen; in gcm_encrypt_continue() 480 crypto_gcm_init_crypt(req, req->cryptlen); in crypto_gcm_encrypt() [all …]
|
D | seqiv.c | 66 if (req->cryptlen < ivsize) in seqiv_aead_encrypt() 82 req->assoclen + req->cryptlen, in seqiv_aead_encrypt() 105 req->cryptlen - ivsize, info); in seqiv_aead_encrypt() 126 if (req->cryptlen < ivsize + crypto_aead_authsize(geniv)) in seqiv_aead_decrypt() 136 req->cryptlen - ivsize, req->iv); in seqiv_aead_decrypt()
|
D | authenc.c | 140 req->assoclen + req->cryptlen, in authenc_geniv_ahash_done() 164 req->assoclen + req->cryptlen); in crypto_authenc_genicv() 172 scatterwalk_map_and_copy(hash, req->dst, req->assoclen + req->cryptlen, in crypto_authenc_genicv() 215 unsigned int cryptlen = req->cryptlen; in crypto_authenc_encrypt() local 235 skcipher_request_set_crypt(skreq, src, dst, cryptlen, req->iv); in crypto_authenc_encrypt() 274 req->cryptlen - authsize, req->iv); in crypto_authenc_decrypt_tail() 311 req->assoclen + req->cryptlen - authsize); in crypto_authenc_decrypt()
|
D | cts.c | 114 lastn = req->cryptlen - offset; in cts_cbc_encrypt() 154 unsigned int nbytes = req->cryptlen; in crypto_cts_encrypt() 194 lastn = req->cryptlen - offset; in cts_cbc_decrypt() 246 unsigned int nbytes = req->cryptlen; in crypto_cts_decrypt()
|
D | skcipher.c | 458 walk->total = req->cryptlen; in skcipher_walk_skcipher() 558 walk->total = req->cryptlen; in skcipher_walk_aead() 567 walk->total = req->cryptlen; in skcipher_walk_aead_encrypt() 578 walk->total = req->cryptlen - crypto_aead_authsize(tfm); in skcipher_walk_aead_decrypt() 639 return crypt(&desc, req->dst, req->src, req->cryptlen); in skcipher_crypt_blkcipher() 735 ablkcipher_request_set_crypt(subreq, req->src, req->dst, req->cryptlen, in skcipher_crypt_ablkcipher()
|
D | drbg.c | 1828 u32 cryptlen = min3(inlen, outlen, (u32)DRBG_OUTSCRATCHLEN); in drbg_kcapi_sym_ctr() local 1832 cryptlen, drbg->V); in drbg_kcapi_sym_ctr() 1840 memcpy(outbuf, drbg->outscratchpad, cryptlen); in drbg_kcapi_sym_ctr() 1841 memzero_explicit(drbg->outscratchpad, cryptlen); in drbg_kcapi_sym_ctr() 1843 outlen -= cryptlen; in drbg_kcapi_sym_ctr() 1844 outbuf += cryptlen; in drbg_kcapi_sym_ctr()
|
D | pcrypt.c | 171 req->cryptlen, req->iv); in pcrypt_aead_encrypt() 215 req->cryptlen, req->iv); in pcrypt_aead_decrypt()
|
D | ctr.c | 309 req->cryptlen, iv); in crypto_rfc3686_crypt()
|
D | cryptd.c | 495 skcipher_request_set_crypt(subreq, req->src, req->dst, req->cryptlen, in cryptd_skcipher_encrypt() 523 skcipher_request_set_crypt(subreq, req->src, req->dst, req->cryptlen, in cryptd_skcipher_decrypt()
|
/linux-4.19.296/include/crypto/ |
D | aead.h | 90 unsigned int cryptlen; member 369 if (req->cryptlen < crypto_aead_authsize(aead)) in crypto_aead_decrypt() 512 unsigned int cryptlen, u8 *iv) in aead_request_set_crypt() argument 516 req->cryptlen = cryptlen; in aead_request_set_crypt()
|
D | skcipher.h | 30 unsigned int cryptlen; member 606 unsigned int cryptlen, void *iv) in skcipher_request_set_crypt() argument 610 req->cryptlen = cryptlen; in skcipher_request_set_crypt()
|
D | morus640_glue.h | 38 void (*final)(void *state, void *tag_xor, u64 assoclen, u64 cryptlen);
|
D | morus1280_glue.h | 38 void (*final)(void *state, void *tag_xor, u64 assoclen, u64 cryptlen);
|