/linux-4.19.296/drivers/misc/ |
D | vmw_balloon.c | 221 void (*add_page)(struct vmballoon *b, int idx, struct page *p); 222 int (*lock)(struct vmballoon *b, unsigned int num_pages, 224 int (*unlock)(struct vmballoon *b, unsigned int num_pages, 279 static bool vmballoon_send_start(struct vmballoon *b, unsigned long req_caps) in vmballoon_send_start() argument 284 STATS_INC(b->stats.start); in vmballoon_send_start() 290 b->capabilities = capabilities; in vmballoon_send_start() 294 b->capabilities = VMW_BALLOON_BASIC_CMDS; in vmballoon_send_start() 306 if ((b->capabilities & VMW_BALLOON_BATCHED_2M_CMDS) && in vmballoon_send_start() 307 (b->capabilities & VMW_BALLOON_BATCHED_CMDS)) in vmballoon_send_start() 308 b->supported_page_sizes = 2; in vmballoon_send_start() [all …]
|
/linux-4.19.296/include/linux/ |
D | jhash.h | 35 #define __jhash_mix(a, b, c) \ argument 37 a -= c; a ^= rol32(c, 4); c += b; \ 38 b -= a; b ^= rol32(a, 6); a += c; \ 39 c -= b; c ^= rol32(b, 8); b += a; \ 40 a -= c; a ^= rol32(c, 16); c += b; \ 41 b -= a; b ^= rol32(a, 19); a += c; \ 42 c -= b; c ^= rol32(b, 4); b += a; \ 46 #define __jhash_final(a, b, c) \ argument 48 c ^= b; c -= rol32(b, 14); \ 50 b ^= a; b -= rol32(a, 25); \ [all …]
|
D | overflow.h | 56 #define check_add_overflow(a, b, d) ({ \ argument 58 typeof(b) __b = (b); \ 65 #define check_sub_overflow(a, b, d) ({ \ argument 67 typeof(b) __b = (b); \ 74 #define check_mul_overflow(a, b, d) ({ \ argument 76 typeof(b) __b = (b); \ 87 #define __unsigned_add_overflow(a, b, d) ({ \ argument 89 typeof(b) __b = (b); \ 96 #define __unsigned_sub_overflow(a, b, d) ({ \ argument 98 typeof(b) __b = (b); \ [all …]
|
D | siphash.h | 33 u64 siphash_2u64(const u64 a, const u64 b, const siphash_key_t *key); 34 u64 siphash_3u64(const u64 a, const u64 b, const u64 c, 36 u64 siphash_4u64(const u64 a, const u64 b, const u64 c, const u64 d, 39 u64 siphash_3u32(const u32 a, const u32 b, const u32 c, 42 static inline u64 siphash_2u32(const u32 a, const u32 b, in siphash_2u32() argument 45 return siphash_1u64((u64)b << 32 | a, key); in siphash_2u32() 47 static inline u64 siphash_4u32(const u32 a, const u32 b, const u32 c, in siphash_4u32() argument 50 return siphash_2u64((u64)b << 32 | a, (u64)d << 32 | c, key); in siphash_4u32() 100 u32 hsiphash_2u32(const u32 a, const u32 b, const hsiphash_key_t *key); 101 u32 hsiphash_3u32(const u32 a, const u32 b, const u32 c, [all …]
|
D | jiffies.h | 105 #define time_after(a,b) \ argument 107 typecheck(unsigned long, b) && \ 108 ((long)((b) - (a)) < 0)) 109 #define time_before(a,b) time_after(b,a) argument 111 #define time_after_eq(a,b) \ argument 113 typecheck(unsigned long, b) && \ 114 ((long)((a) - (b)) >= 0)) 115 #define time_before_eq(a,b) time_after_eq(b,a) argument 120 #define time_in_range(a,b,c) \ argument 121 (time_after_eq(a,b) && \ [all …]
|
/linux-4.19.296/crypto/ |
D | memneq.c | 68 __crypto_memneq_generic(const void *a, const void *b, size_t size) in __crypto_memneq_generic() argument 74 neq |= *(unsigned long *)a ^ *(unsigned long *)b; in __crypto_memneq_generic() 77 b += sizeof(unsigned long); in __crypto_memneq_generic() 82 neq |= *(unsigned char *)a ^ *(unsigned char *)b; in __crypto_memneq_generic() 85 b += 1; in __crypto_memneq_generic() 92 static inline unsigned long __crypto_memneq_16(const void *a, const void *b) in __crypto_memneq_16() argument 98 neq |= *(unsigned long *)(a) ^ *(unsigned long *)(b); in __crypto_memneq_16() 100 neq |= *(unsigned long *)(a+8) ^ *(unsigned long *)(b+8); in __crypto_memneq_16() 103 neq |= *(unsigned int *)(a) ^ *(unsigned int *)(b); in __crypto_memneq_16() 105 neq |= *(unsigned int *)(a+4) ^ *(unsigned int *)(b+4); in __crypto_memneq_16() [all …]
|
D | des_generic.c | 617 #define DES_PC2(a, b, c, d) (T4(d) | T3(c) | T2(b) | T1(a)) argument 634 unsigned long a, b, c, d, w; in des_ekey() local 639 b = k[6]; b &= 0x0e; b <<= 4; b |= k[2] & 0x1e; b = pc1[b]; in des_ekey() 642 pe[15 * 2 + 0] = DES_PC2(a, b, c, d); d = rs[d]; in des_ekey() 643 pe[14 * 2 + 0] = DES_PC2(d, a, b, c); c = rs[c]; b = rs[b]; in des_ekey() 644 pe[13 * 2 + 0] = DES_PC2(b, c, d, a); a = rs[a]; d = rs[d]; in des_ekey() 645 pe[12 * 2 + 0] = DES_PC2(d, a, b, c); c = rs[c]; b = rs[b]; in des_ekey() 646 pe[11 * 2 + 0] = DES_PC2(b, c, d, a); a = rs[a]; d = rs[d]; in des_ekey() 647 pe[10 * 2 + 0] = DES_PC2(d, a, b, c); c = rs[c]; b = rs[b]; in des_ekey() 648 pe[ 9 * 2 + 0] = DES_PC2(b, c, d, a); a = rs[a]; d = rs[d]; in des_ekey() [all …]
|
D | md5.c | 45 u32 a, b, c, d; in md5_transform() local 48 b = hash[1]; in md5_transform() 52 MD5STEP(F1, a, b, c, d, in[0] + 0xd76aa478, 7); in md5_transform() 53 MD5STEP(F1, d, a, b, c, in[1] + 0xe8c7b756, 12); in md5_transform() 54 MD5STEP(F1, c, d, a, b, in[2] + 0x242070db, 17); in md5_transform() 55 MD5STEP(F1, b, c, d, a, in[3] + 0xc1bdceee, 22); in md5_transform() 56 MD5STEP(F1, a, b, c, d, in[4] + 0xf57c0faf, 7); in md5_transform() 57 MD5STEP(F1, d, a, b, c, in[5] + 0x4787c62a, 12); in md5_transform() 58 MD5STEP(F1, c, d, a, b, in[6] + 0xa8304613, 17); in md5_transform() 59 MD5STEP(F1, b, c, d, a, in[7] + 0xfd469501, 22); in md5_transform() [all …]
|
D | md4.c | 63 #define ROUND1(a,b,c,d,k,s) (a = lshift(a + F(b,c,d) + k, s)) argument 64 #define ROUND2(a,b,c,d,k,s) (a = lshift(a + G(b,c,d) + k + (u32)0x5A827999,s)) argument 65 #define ROUND3(a,b,c,d,k,s) (a = lshift(a + H(b,c,d) + k + (u32)0x6ED9EBA1,s)) argument 69 u32 a, b, c, d; in md4_transform() local 72 b = hash[1]; in md4_transform() 76 ROUND1(a, b, c, d, in[0], 3); in md4_transform() 77 ROUND1(d, a, b, c, in[1], 7); in md4_transform() 78 ROUND1(c, d, a, b, in[2], 11); in md4_transform() 79 ROUND1(b, c, d, a, in[3], 19); in md4_transform() 80 ROUND1(a, b, c, d, in[4], 3); in md4_transform() [all …]
|
D | sha256_generic.c | 73 u32 a, b, c, d, e, f, g, h, t1, t2; in sha256_transform() local 86 a=state[0]; b=state[1]; c=state[2]; d=state[3]; in sha256_transform() 91 t2 = e0(a) + Maj(a,b,c); d+=t1; h=t1+t2; in sha256_transform() 93 t2 = e0(h) + Maj(h,a,b); c+=t1; g=t1+t2; in sha256_transform() 95 t2 = e0(g) + Maj(g,h,a); b+=t1; f=t1+t2; in sha256_transform() 96 t1 = e + e1(b) + Ch(b,c,d) + 0xe9b5dba5 + W[ 3]; in sha256_transform() 98 t1 = d + e1(a) + Ch(a,b,c) + 0x3956c25b + W[ 4]; in sha256_transform() 100 t1 = c + e1(h) + Ch(h,a,b) + 0x59f111f1 + W[ 5]; in sha256_transform() 102 t1 = b + e1(g) + Ch(g,h,a) + 0x923f82a4 + W[ 6]; in sha256_transform() 103 t2 = e0(c) + Maj(c,d,e); f+=t1; b=t1+t2; in sha256_transform() [all …]
|
/linux-4.19.296/lib/ |
D | gcd.c | 22 unsigned long gcd(unsigned long a, unsigned long b) in gcd() argument 24 unsigned long r = a | b; in gcd() 26 if (!a || !b) in gcd() 29 b >>= __ffs(b); in gcd() 30 if (b == 1) in gcd() 37 if (a == b) in gcd() 40 if (a < b) in gcd() 41 swap(a, b); in gcd() 42 a -= b; in gcd() 49 unsigned long gcd(unsigned long a, unsigned long b) in gcd() argument [all …]
|
D | decompress_unxz.c | 177 static bool memeq(const void *a, const void *b, size_t size) in memeq() argument 180 const uint8_t *y = b; in memeq() 194 uint8_t *b = buf; in memzero() local 195 uint8_t *e = b + size; in memzero() 197 while (b != e) in memzero() 198 *b++ = '\0'; in memzero() 257 struct xz_buf b; in unxz() local 278 b.out = out; in unxz() 279 b.out_size = (size_t)-1; in unxz() 281 b.out_size = XZ_IOBUF_SIZE; in unxz() [all …]
|
D | sha256.c | 50 u32 a, b, c, d, e, f, g, h, t1, t2; in sha256_transform() local 63 a = state[0]; b = state[1]; c = state[2]; d = state[3]; in sha256_transform() 68 t2 = e0(a) + Maj(a, b, c); d += t1; h = t1 + t2; in sha256_transform() 70 t2 = e0(h) + Maj(h, a, b); c += t1; g = t1 + t2; in sha256_transform() 72 t2 = e0(g) + Maj(g, h, a); b += t1; f = t1 + t2; in sha256_transform() 73 t1 = e + e1(b) + Ch(b, c, d) + 0xe9b5dba5 + W[3]; in sha256_transform() 75 t1 = d + e1(a) + Ch(a, b, c) + 0x3956c25b + W[4]; in sha256_transform() 77 t1 = c + e1(h) + Ch(h, a, b) + 0x59f111f1 + W[5]; in sha256_transform() 79 t1 = b + e1(g) + Ch(g, h, a) + 0x923f82a4 + W[6]; in sha256_transform() 80 t2 = e0(c) + Maj(c, d, e); f += t1; b = t1 + t2; in sha256_transform() [all …]
|
D | siphash.c | 28 u64 b = ((u64)(len)) << 56; \ 35 v3 ^= b; \ 38 v0 ^= b; \ 62 b |= le64_to_cpu((__force __le64)(load_unaligned_zeropad(data) & in __siphash_aligned() 66 case 7: b |= ((u64)end[6]) << 48; in __siphash_aligned() 67 case 6: b |= ((u64)end[5]) << 40; in __siphash_aligned() 68 case 5: b |= ((u64)end[4]) << 32; in __siphash_aligned() 69 case 4: b |= le32_to_cpup(data); break; in __siphash_aligned() 70 case 3: b |= ((u64)end[2]) << 16; in __siphash_aligned() 71 case 2: b |= le16_to_cpup(data); break; in __siphash_aligned() [all …]
|
D | list_sort.c | 19 struct list_head *b), in merge() argument 20 struct list_head *a, struct list_head *b) in merge() 24 while (a && b) { in merge() 26 if ((*cmp)(priv, a, b) <= 0) { in merge() 30 tail->next = b; in merge() 31 b = b->next; in merge() 35 tail->next = a?:b; in merge() 48 struct list_head *b), in merge_and_restore_back_links() argument 50 struct list_head *a, struct list_head *b) in merge_and_restore_back_links() 55 while (a && b) { in merge_and_restore_back_links() [all …]
|
D | uuid.c | 51 static void __uuid_gen_common(__u8 b[16]) in __uuid_gen_common() 53 prandom_bytes(b, 16); in __uuid_gen_common() 55 b[8] = (b[8] & 0x3F) | 0x80; in __uuid_gen_common() 60 __uuid_gen_common(lu->b); in guid_gen() 62 lu->b[7] = (lu->b[7] & 0x0F) | 0x40; in guid_gen() 68 __uuid_gen_common(bu->b); in uuid_gen() 70 bu->b[6] = (bu->b[6] & 0x0F) | 0x40; in uuid_gen() 103 static int __uuid_parse(const char *uuid, __u8 b[16], const u8 ei[16]) in __uuid_parse() 115 b[ei[i]] = (hi << 4) | lo; in __uuid_parse() 123 return __uuid_parse(uuid, u->b, guid_index); in guid_parse() [all …]
|
/linux-4.19.296/fs/ntfs/ |
D | mst.c | 42 int post_read_mst_fixup(NTFS_RECORD *b, const u32 size) in post_read_mst_fixup() argument 48 usa_ofs = le16_to_cpu(b->usa_ofs); in post_read_mst_fixup() 50 usa_count = le16_to_cpu(b->usa_count) - 1; in post_read_mst_fixup() 58 usa_pos = (u16*)b + usa_ofs/sizeof(u16); in post_read_mst_fixup() 70 data_pos = (u16*)b + NTFS_BLOCK_SIZE/sizeof(u16) - 1; in post_read_mst_fixup() 81 b->magic = magic_BAAD; in post_read_mst_fixup() 87 usa_count = le16_to_cpu(b->usa_count) - 1; in post_read_mst_fixup() 88 data_pos = (u16*)b + NTFS_BLOCK_SIZE/sizeof(u16) - 1; in post_read_mst_fixup() 123 int pre_write_mst_fixup(NTFS_RECORD *b, const u32 size) in pre_write_mst_fixup() argument 130 if (!b || ntfs_is_baad_record(b->magic) || in pre_write_mst_fixup() [all …]
|
/linux-4.19.296/drivers/media/common/videobuf2/ |
D | videobuf2-v4l2.c | 56 static int __verify_planes_array(struct vb2_buffer *vb, const struct v4l2_buffer *b) in __verify_planes_array() argument 58 if (!V4L2_TYPE_IS_MULTIPLANAR(b->type)) in __verify_planes_array() 62 if (b->m.planes == NULL) { in __verify_planes_array() 67 if (b->length < vb->num_planes || b->length > VB2_MAX_PLANES) { in __verify_planes_array() 69 vb->num_planes, b->length); in __verify_planes_array() 85 static int __verify_length(struct vb2_buffer *vb, const struct v4l2_buffer *b) in __verify_length() argument 91 if (!V4L2_TYPE_IS_OUTPUT(b->type)) in __verify_length() 94 if (V4L2_TYPE_IS_MULTIPLANAR(b->type)) { in __verify_length() 96 length = (b->memory == VB2_MEMORY_USERPTR || in __verify_length() 97 b->memory == VB2_MEMORY_DMABUF) in __verify_length() [all …]
|
/linux-4.19.296/lib/xz/ |
D | xz_dec_stream.c | 157 static bool fill_temp(struct xz_dec *s, struct xz_buf *b) in fill_temp() argument 160 b->in_size - b->in_pos, s->temp.size - s->temp.pos); in fill_temp() 162 memcpy(s->temp.buf + s->temp.pos, b->in + b->in_pos, copy_size); in fill_temp() 163 b->in_pos += copy_size; in fill_temp() 218 static enum xz_ret dec_block(struct xz_dec *s, struct xz_buf *b) in dec_block() argument 222 s->in_start = b->in_pos; in dec_block() 223 s->out_start = b->out_pos; in dec_block() 227 ret = xz_dec_bcj_run(s->bcj, s->lzma2, b); in dec_block() 230 ret = xz_dec_lzma2_run(s->lzma2, b); in dec_block() 232 s->block.compressed += b->in_pos - s->in_start; in dec_block() [all …]
|
D | xz_dec_bcj.c | 83 static inline int bcj_x86_test_msbyte(uint8_t b) in bcj_x86_test_msbyte() argument 85 return b == 0x00 || b == 0xFF; in bcj_x86_test_msbyte() 101 uint8_t b; in bcj_x86() local 117 b = buf[i + 4 - mask_to_bit_num[prev_mask]]; in bcj_x86() 119 || bcj_x86_test_msbyte(b)) { in bcj_x86() 137 b = (uint8_t)(dest >> (24 - j)); in bcj_x86() 138 if (!bcj_x86_test_msbyte(b)) in bcj_x86() 399 static void bcj_flush(struct xz_dec_bcj *s, struct xz_buf *b) in bcj_flush() argument 403 copy_size = min_t(size_t, s->temp.filtered, b->out_size - b->out_pos); in bcj_flush() 404 memcpy(b->out + b->out_pos, s->temp.buf, copy_size); in bcj_flush() [all …]
|
/linux-4.19.296/include/asm-generic/bitops/ |
D | sched.h | 13 static inline int sched_find_first_bit(const unsigned long *b) in sched_find_first_bit() argument 16 if (b[0]) in sched_find_first_bit() 17 return __ffs(b[0]); in sched_find_first_bit() 18 return __ffs(b[1]) + 64; in sched_find_first_bit() 20 if (b[0]) in sched_find_first_bit() 21 return __ffs(b[0]); in sched_find_first_bit() 22 if (b[1]) in sched_find_first_bit() 23 return __ffs(b[1]) + 32; in sched_find_first_bit() 24 if (b[2]) in sched_find_first_bit() 25 return __ffs(b[2]) + 64; in sched_find_first_bit() [all …]
|
/linux-4.19.296/include/crypto/ |
D | gf128mul.h | 163 void gf128mul_lle(be128 *a, const be128 *b); 165 void gf128mul_bbe(be128 *a, const be128 *b); 185 u64 b = be64_to_cpu(x->b); in gf128mul_x_lle() local 189 u64 _tt = gf128mul_mask_from_bit(b, 0) & ((u64)0xe1 << 56); in gf128mul_x_lle() 191 r->b = cpu_to_be64((b >> 1) | (a << 63)); in gf128mul_x_lle() 198 u64 b = be64_to_cpu(x->b); in gf128mul_x_bbe() local 203 r->a = cpu_to_be64((a << 1) | (b >> 63)); in gf128mul_x_bbe() 204 r->b = cpu_to_be64((b << 1) ^ _tt); in gf128mul_x_bbe() 211 u64 b = le64_to_cpu(x->b); in gf128mul_x_ble() local 216 r->a = cpu_to_le64((a << 1) | (b >> 63)); in gf128mul_x_ble() [all …]
|
/linux-4.19.296/lib/842/ |
D | 842_compress.c | 118 #define get_input_data(p, o, b) \ argument 119 be##b##_to_cpu(get_unaligned((__be##b *)((p)->in + (o)))) 121 #define init_hashtable_nodes(p, b) do { \ argument 123 hash_init((p)->htable##b); \ 124 for (_i = 0; _i < ARRAY_SIZE((p)->node##b); _i++) { \ 125 (p)->node##b[_i].index = _i; \ 126 (p)->node##b[_i].data = 0; \ 127 INIT_HLIST_NODE(&(p)->node##b[_i].node); \ 131 #define find_index(p, b, n) ({ \ argument 132 struct sw842_hlist_node##b *_n; \ [all …]
|
/linux-4.19.296/include/linux/mtd/ |
D | ftl.h | 63 #define BLOCK_FREE(b) ((b) == 0xffffffff) argument 64 #define BLOCK_DELETED(b) (((b) == 0) || ((b) == 0xfffffffe)) argument 66 #define BLOCK_TYPE(b) ((b) & 0x7f) argument 67 #define BLOCK_ADDRESS(b) ((b) & ~0x7f) argument 68 #define BLOCK_NUMBER(b) ((b) >> 9) argument
|
/linux-4.19.296/fs/ext4/ |
D | hash.c | 19 __u32 a = in[0], b = in[1], c = in[2], d = in[3]; in TEA_transform() local 24 b0 += ((b1 << 4)+a) ^ (b1+sum) ^ ((b1 >> 5)+b); in TEA_transform() 43 #define ROUND(f, a, b, c, d, x, s) \ argument 44 (a += f(b, c, d) + x, a = rol32(a, s)) 54 __u32 a = buf[0], b = buf[1], c = buf[2], d = buf[3]; in half_md4_transform() local 57 ROUND(F, a, b, c, d, in[0] + K1, 3); in half_md4_transform() 58 ROUND(F, d, a, b, c, in[1] + K1, 7); in half_md4_transform() 59 ROUND(F, c, d, a, b, in[2] + K1, 11); in half_md4_transform() 60 ROUND(F, b, c, d, a, in[3] + K1, 19); in half_md4_transform() 61 ROUND(F, a, b, c, d, in[4] + K1, 3); in half_md4_transform() [all …]
|