/linux-4.19.296/block/ |
D | blk-stat.c | 55 int bucket; in blk_stat_add() local 67 bucket = cb->bucket_fn(rq); in blk_stat_add() 68 if (bucket < 0) in blk_stat_add() 71 stat = &get_cpu_ptr(cb->cpu_stat)[bucket]; in blk_stat_add() 81 unsigned int bucket; in blk_stat_timer_fn() local 84 for (bucket = 0; bucket < cb->buckets; bucket++) in blk_stat_timer_fn() 85 blk_rq_stat_init(&cb->stat[bucket]); in blk_stat_timer_fn() 91 for (bucket = 0; bucket < cb->buckets; bucket++) { in blk_stat_timer_fn() 92 blk_rq_stat_sum(&cb->stat[bucket], &cpu_stat[bucket]); in blk_stat_timer_fn() 93 blk_rq_stat_init(&cpu_stat[bucket]); in blk_stat_timer_fn() [all …]
|
D | blk-mq-debugfs.c | 40 int bucket; in queue_poll_stat_show() local 42 for (bucket = 0; bucket < BLK_MQ_POLL_STATS_BKTS/2; bucket++) { in queue_poll_stat_show() 43 seq_printf(m, "read (%d Bytes): ", 1 << (9+bucket)); in queue_poll_stat_show() 44 print_stat(m, &q->poll_stat[2*bucket]); in queue_poll_stat_show() 47 seq_printf(m, "write (%d Bytes): ", 1 << (9+bucket)); in queue_poll_stat_show() 48 print_stat(m, &q->poll_stat[2*bucket+1]); in queue_poll_stat_show()
|
D | blk-mq.c | 46 int ddir, bytes, bucket; in blk_mq_poll_stats_bkt() local 51 bucket = ddir + 2*(ilog2(bytes) - 9); in blk_mq_poll_stats_bkt() 53 if (bucket < 0) in blk_mq_poll_stats_bkt() 55 else if (bucket >= BLK_MQ_POLL_STATS_BKTS) in blk_mq_poll_stats_bkt() 58 return bucket; in blk_mq_poll_stats_bkt() 3063 int bucket; in blk_mq_poll_stats_fn() local 3065 for (bucket = 0; bucket < BLK_MQ_POLL_STATS_BKTS; bucket++) { in blk_mq_poll_stats_fn() 3066 if (cb->stat[bucket].nr_samples) in blk_mq_poll_stats_fn() 3067 q->poll_stat[bucket] = cb->stat[bucket]; in blk_mq_poll_stats_fn() 3076 int bucket; in blk_mq_poll_nsecs() local [all …]
|
D | blk-throttle.c | 2063 struct latency_bucket *bucket; in throtl_update_latency_buckets() local 2066 bucket = per_cpu_ptr(td->latency_buckets[rw], in throtl_update_latency_buckets() 2068 tmp->total_latency += bucket[i].total_latency; in throtl_update_latency_buckets() 2069 tmp->samples += bucket[i].samples; in throtl_update_latency_buckets() 2070 bucket[i].total_latency = 0; in throtl_update_latency_buckets() 2071 bucket[i].samples = 0; in throtl_update_latency_buckets() 2307 int bucket; in blk_throtl_bio_endio() local 2310 bucket = request_bucket_index(bio_issue_size(&bio->bi_issue)); in blk_throtl_bio_endio() 2311 threshold = tg->td->avg_buckets[rw][bucket].latency + in blk_throtl_bio_endio()
|
/linux-4.19.296/fs/dlm/ |
D | debug_fs.c | 370 unsigned bucket; member 429 unsigned bucket, entry; in table_seq_start() local 432 bucket = n >> 32; in table_seq_start() 435 if (bucket >= ls->ls_rsbtbl_size) in table_seq_start() 452 tree = toss ? &ls->ls_rsbtbl[bucket].toss : &ls->ls_rsbtbl[bucket].keep; in table_seq_start() 454 spin_lock(&ls->ls_rsbtbl[bucket].lock); in table_seq_start() 461 ri->bucket = bucket; in table_seq_start() 462 spin_unlock(&ls->ls_rsbtbl[bucket].lock); in table_seq_start() 467 spin_unlock(&ls->ls_rsbtbl[bucket].lock); in table_seq_start() 477 bucket++; in table_seq_start() [all …]
|
D | dir.c | 203 uint32_t hash, bucket; in find_rsb_root() local 207 bucket = hash & (ls->ls_rsbtbl_size - 1); in find_rsb_root() 209 spin_lock(&ls->ls_rsbtbl[bucket].lock); in find_rsb_root() 210 rv = dlm_search_rsb_tree(&ls->ls_rsbtbl[bucket].keep, name, len, &r); in find_rsb_root() 212 rv = dlm_search_rsb_tree(&ls->ls_rsbtbl[bucket].toss, in find_rsb_root() 214 spin_unlock(&ls->ls_rsbtbl[bucket].lock); in find_rsb_root()
|
/linux-4.19.296/drivers/cpuidle/governors/ |
D | menu.c | 128 unsigned int bucket; member 145 int bucket = 0; in which_bucket() local 154 bucket = BUCKETS/2; in which_bucket() 157 return bucket; in which_bucket() 159 return bucket + 1; in which_bucket() 161 return bucket + 2; in which_bucket() 163 return bucket + 3; in which_bucket() 165 return bucket + 4; in which_bucket() 166 return bucket + 5; in which_bucket() 311 data->bucket = which_bucket(data->next_timer_us, nr_iowaiters); in menu_select() [all …]
|
/linux-4.19.296/fs/nfs/ |
D | pnfs_nfs.c | 75 struct pnfs_commit_bucket *bucket; in pnfs_generic_clear_request_commit() local 77 bucket = list_first_entry(&req->wb_list, in pnfs_generic_clear_request_commit() 80 freeme = bucket->wlseg; in pnfs_generic_clear_request_commit() 81 bucket->wlseg = NULL; in pnfs_generic_clear_request_commit() 90 pnfs_generic_scan_ds_commit_list(struct pnfs_commit_bucket *bucket, in pnfs_generic_scan_ds_commit_list() argument 94 struct list_head *src = &bucket->written; in pnfs_generic_scan_ds_commit_list() 95 struct list_head *dst = &bucket->committing; in pnfs_generic_scan_ds_commit_list() 103 if (bucket->clseg == NULL) in pnfs_generic_scan_ds_commit_list() 104 bucket->clseg = pnfs_get_lseg(bucket->wlseg); in pnfs_generic_scan_ds_commit_list() 106 pnfs_put_lseg(bucket->wlseg); in pnfs_generic_scan_ds_commit_list() [all …]
|
/linux-4.19.296/fs/ocfs2/ |
D | xattr.c | 135 struct ocfs2_xattr_bucket *bucket; member 289 struct ocfs2_xattr_bucket *bucket, 311 struct ocfs2_xattr_bucket *bucket, 332 struct ocfs2_xattr_bucket *bucket; in ocfs2_xattr_bucket_new() local 337 bucket = kzalloc(sizeof(struct ocfs2_xattr_bucket), GFP_NOFS); in ocfs2_xattr_bucket_new() 338 if (bucket) { in ocfs2_xattr_bucket_new() 339 bucket->bu_inode = inode; in ocfs2_xattr_bucket_new() 340 bucket->bu_blocks = blks; in ocfs2_xattr_bucket_new() 343 return bucket; in ocfs2_xattr_bucket_new() 346 static void ocfs2_xattr_bucket_relse(struct ocfs2_xattr_bucket *bucket) in ocfs2_xattr_bucket_relse() argument [all …]
|
/linux-4.19.296/lib/ |
D | stackdepot.c | 186 static inline struct stack_record *find_stack(struct stack_record *bucket, in find_stack() argument 192 for (found = bucket; found; found = found->next) { in find_stack() 226 struct stack_record *found = NULL, **bucket; in depot_save_stack() local 235 bucket = &stack_table[hash & STACK_HASH_MASK]; in depot_save_stack() 242 found = find_stack(smp_load_acquire(bucket), trace->entries, in depot_save_stack() 271 found = find_stack(*bucket, trace->entries, trace->nr_entries, hash); in depot_save_stack() 277 new->next = *bucket; in depot_save_stack() 282 smp_store_release(bucket, new); in depot_save_stack()
|
D | rhashtable.c | 38 struct rhash_head __rcu *bucket; member 131 INIT_RHT_NULLS_HEAD(ntbl[i].bucket); in nested_table_alloc() 1208 return &ntbl[subhash].bucket; in rht_bucket_nested() 1238 return &ntbl[hash].bucket; in rht_bucket_nested_insert()
|
/linux-4.19.296/include/trace/events/ |
D | bcache.h | 68 __field(size_t, bucket ) 72 __entry->bucket = PTR_BUCKET_NR(b->c, &b->key, 0); 75 TP_printk("bucket %zu", __entry->bucket) 246 __field(size_t, bucket ) 252 __entry->bucket = PTR_BUCKET_NR(b->c, &b->key, 0); 257 TP_printk("bucket %zu", __entry->bucket) 348 __field(size_t, bucket ) 353 __entry->bucket = PTR_BUCKET_NR(b->c, &b->key, 0); 357 TP_printk("bucket %zu keys %u", __entry->bucket, __entry->keys) 407 TP_PROTO(struct cache *ca, size_t bucket), [all …]
|
/linux-4.19.296/include/net/ |
D | transp_v6.h | 48 __u16 srcp, __u16 destp, int rqueue, int bucket); 51 __u16 destp, int bucket) in ip6_dgram_sock_seq_show() argument 54 bucket); in ip6_dgram_sock_seq_show()
|
D | llc.h | 75 u32 bucket = hash_32(ifindex, LLC_SK_DEV_HASH_BITS); in llc_sk_dev_hash() local 77 return &sap->sk_dev_hash[bucket]; in llc_sk_dev_hash()
|
D | raw.h | 50 int bucket; member
|
D | ping.h | 51 int bucket; member
|
/linux-4.19.296/fs/xfs/libxfs/ |
D | xfs_ag.c | 220 int bucket; in xfs_agflblock_init() local 229 for (bucket = 0; bucket < xfs_agfl_size(mp); bucket++) in xfs_agflblock_init() 230 agfl_bno[bucket] = cpu_to_be32(NULLAGBLOCK); in xfs_agflblock_init() 240 int bucket; in xfs_agiblock_init() local 258 for (bucket = 0; bucket < XFS_AGI_UNLINKED_BUCKETS; bucket++) in xfs_agiblock_init() 259 agi->agi_unlinked[bucket] = cpu_to_be32(NULLAGINO); in xfs_agiblock_init()
|
/linux-4.19.296/drivers/misc/vmw_vmci/ |
D | vmci_doorbell.c | 128 u32 bucket = VMCI_DOORBELL_HASH(idx); in dbell_index_table_find() local 131 hlist_for_each_entry(dbell, &vmci_doorbell_it.entries[bucket], in dbell_index_table_find() 147 u32 bucket; in dbell_index_table_add() local 195 bucket = VMCI_DOORBELL_HASH(entry->idx); in dbell_index_table_add() 196 hlist_add_head(&entry->node, &vmci_doorbell_it.entries[bucket]); in dbell_index_table_add() 360 u32 bucket = VMCI_DOORBELL_HASH(notify_idx); in dbell_fire_entries() local 365 hlist_for_each_entry(dbell, &vmci_doorbell_it.entries[bucket], node) { in dbell_fire_entries()
|
/linux-4.19.296/fs/btrfs/ |
D | compression.c | 750 struct bucket_item *bucket; member 763 kfree(workspace->bucket); in free_heuristic_ws() 780 ws->bucket = kcalloc(BUCKET_SIZE, sizeof(*ws->bucket), GFP_KERNEL); in alloc_heuristic_ws() 781 if (!ws->bucket) in alloc_heuristic_ws() 1269 for (i = 0; i < BUCKET_SIZE && ws->bucket[i].count > 0; i++) { in shannon_entropy() 1270 p = ws->bucket[i].count; in shannon_entropy() 1402 struct bucket_item *bucket = ws->bucket; in byte_core_set_size() local 1405 radix_sort(ws->bucket, ws->bucket_b, BUCKET_SIZE); in byte_core_set_size() 1408 coreset_sum += bucket[i].count; in byte_core_set_size() 1413 for (; i < BYTE_CORE_SET_HIGH && bucket[i].count > 0; i++) { in byte_core_set_size() [all …]
|
/linux-4.19.296/fs/fscache/ |
D | cookie.c | 192 unsigned int bucket; in fscache_hash_cookie() local 194 bucket = candidate->key_hash & (ARRAY_SIZE(fscache_cookie_hash) - 1); in fscache_hash_cookie() 195 h = &fscache_cookie_hash[bucket]; in fscache_hash_cookie() 835 unsigned int bucket; in fscache_unhash_cookie() local 837 bucket = cookie->key_hash & (ARRAY_SIZE(fscache_cookie_hash) - 1); in fscache_unhash_cookie() 838 h = &fscache_cookie_hash[bucket]; in fscache_unhash_cookie()
|
/linux-4.19.296/include/linux/ |
D | atalk.h | 149 int bucket; member
|
/linux-4.19.296/fs/ocfs2/dlm/ |
D | dlmdomain.c | 175 struct hlist_head *bucket; in __dlm_insert_lockres() local 179 bucket = dlm_lockres_hash(dlm, res->lockname.hash); in __dlm_insert_lockres() 184 hlist_add_head(&res->hash_node, bucket); in __dlm_insert_lockres() 195 struct hlist_head *bucket; in __dlm_lookup_lockres_full() local 202 bucket = dlm_lockres_hash(dlm, hash); in __dlm_lookup_lockres_full() 204 hlist_for_each_entry(res, bucket, hash_node) { in __dlm_lookup_lockres_full() 426 struct hlist_head *bucket; in dlm_migrate_all_locks() local 436 bucket = dlm_lockres_hash(dlm, i); in dlm_migrate_all_locks() 437 iter = bucket->first; in dlm_migrate_all_locks()
|
D | dlmmaster.c | 330 struct hlist_head *bucket; in __dlm_insert_mle() local 334 bucket = dlm_master_hash(dlm, mle->mnamehash); in __dlm_insert_mle() 335 hlist_add_head(&mle->master_hash_node, bucket); in __dlm_insert_mle() 344 struct hlist_head *bucket; in dlm_find_mle() local 350 bucket = dlm_master_hash(dlm, hash); in dlm_find_mle() 351 hlist_for_each_entry(tmpmle, bucket, master_hash_node) { in dlm_find_mle() 3351 struct hlist_head *bucket; in dlm_clean_master_list() local 3362 bucket = dlm_master_hash(dlm, i); in dlm_clean_master_list() 3363 hlist_for_each_entry_safe(mle, tmp, bucket, master_hash_node) { in dlm_clean_master_list() 3551 struct hlist_head *bucket; in dlm_force_free_mles() local [all …]
|
/linux-4.19.296/fs/xfs/ |
D | xfs_log_recover.c | 2018 struct list_head *bucket; in xlog_recover_buffer_pass1() local 2033 bucket = XLOG_BUF_CANCEL_BUCKET(log, buf_f->blf_blkno); in xlog_recover_buffer_pass1() 2034 list_for_each_entry(bcp, bucket, bc_list) { in xlog_recover_buffer_pass1() 2047 list_add_tail(&bcp->bc_list, bucket); in xlog_recover_buffer_pass1() 2065 struct list_head *bucket; in xlog_peek_buffer_cancelled() local 2074 bucket = XLOG_BUF_CANCEL_BUCKET(log, blkno); in xlog_peek_buffer_cancelled() 2075 list_for_each_entry(bcp, bucket, bc_list) { in xlog_peek_buffer_cancelled() 5013 int bucket) in xlog_recover_clear_agi_bucket() argument 5030 agi->agi_unlinked[bucket] = cpu_to_be32(NULLAGINO); in xlog_recover_clear_agi_bucket() 5032 (sizeof(xfs_agino_t) * bucket); in xlog_recover_clear_agi_bucket() [all …]
|
/linux-4.19.296/fs/omfs/ |
D | dir.c | 28 int bucket = omfs_hash(name, namelen, nbuckets); in omfs_get_bucket() local 30 *ofs = OMFS_DIR_START + bucket * 8; in omfs_get_bucket()
|