Home
last modified time | relevance | path

Searched refs:bucket (Results 1 – 25 of 38) sorted by relevance

12

/linux-4.19.296/block/
Dblk-stat.c55 int bucket; in blk_stat_add() local
67 bucket = cb->bucket_fn(rq); in blk_stat_add()
68 if (bucket < 0) in blk_stat_add()
71 stat = &get_cpu_ptr(cb->cpu_stat)[bucket]; in blk_stat_add()
81 unsigned int bucket; in blk_stat_timer_fn() local
84 for (bucket = 0; bucket < cb->buckets; bucket++) in blk_stat_timer_fn()
85 blk_rq_stat_init(&cb->stat[bucket]); in blk_stat_timer_fn()
91 for (bucket = 0; bucket < cb->buckets; bucket++) { in blk_stat_timer_fn()
92 blk_rq_stat_sum(&cb->stat[bucket], &cpu_stat[bucket]); in blk_stat_timer_fn()
93 blk_rq_stat_init(&cpu_stat[bucket]); in blk_stat_timer_fn()
[all …]
Dblk-mq-debugfs.c40 int bucket; in queue_poll_stat_show() local
42 for (bucket = 0; bucket < BLK_MQ_POLL_STATS_BKTS/2; bucket++) { in queue_poll_stat_show()
43 seq_printf(m, "read (%d Bytes): ", 1 << (9+bucket)); in queue_poll_stat_show()
44 print_stat(m, &q->poll_stat[2*bucket]); in queue_poll_stat_show()
47 seq_printf(m, "write (%d Bytes): ", 1 << (9+bucket)); in queue_poll_stat_show()
48 print_stat(m, &q->poll_stat[2*bucket+1]); in queue_poll_stat_show()
Dblk-mq.c46 int ddir, bytes, bucket; in blk_mq_poll_stats_bkt() local
51 bucket = ddir + 2*(ilog2(bytes) - 9); in blk_mq_poll_stats_bkt()
53 if (bucket < 0) in blk_mq_poll_stats_bkt()
55 else if (bucket >= BLK_MQ_POLL_STATS_BKTS) in blk_mq_poll_stats_bkt()
58 return bucket; in blk_mq_poll_stats_bkt()
3063 int bucket; in blk_mq_poll_stats_fn() local
3065 for (bucket = 0; bucket < BLK_MQ_POLL_STATS_BKTS; bucket++) { in blk_mq_poll_stats_fn()
3066 if (cb->stat[bucket].nr_samples) in blk_mq_poll_stats_fn()
3067 q->poll_stat[bucket] = cb->stat[bucket]; in blk_mq_poll_stats_fn()
3076 int bucket; in blk_mq_poll_nsecs() local
[all …]
Dblk-throttle.c2063 struct latency_bucket *bucket; in throtl_update_latency_buckets() local
2066 bucket = per_cpu_ptr(td->latency_buckets[rw], in throtl_update_latency_buckets()
2068 tmp->total_latency += bucket[i].total_latency; in throtl_update_latency_buckets()
2069 tmp->samples += bucket[i].samples; in throtl_update_latency_buckets()
2070 bucket[i].total_latency = 0; in throtl_update_latency_buckets()
2071 bucket[i].samples = 0; in throtl_update_latency_buckets()
2307 int bucket; in blk_throtl_bio_endio() local
2310 bucket = request_bucket_index(bio_issue_size(&bio->bi_issue)); in blk_throtl_bio_endio()
2311 threshold = tg->td->avg_buckets[rw][bucket].latency + in blk_throtl_bio_endio()
/linux-4.19.296/fs/dlm/
Ddebug_fs.c370 unsigned bucket; member
429 unsigned bucket, entry; in table_seq_start() local
432 bucket = n >> 32; in table_seq_start()
435 if (bucket >= ls->ls_rsbtbl_size) in table_seq_start()
452 tree = toss ? &ls->ls_rsbtbl[bucket].toss : &ls->ls_rsbtbl[bucket].keep; in table_seq_start()
454 spin_lock(&ls->ls_rsbtbl[bucket].lock); in table_seq_start()
461 ri->bucket = bucket; in table_seq_start()
462 spin_unlock(&ls->ls_rsbtbl[bucket].lock); in table_seq_start()
467 spin_unlock(&ls->ls_rsbtbl[bucket].lock); in table_seq_start()
477 bucket++; in table_seq_start()
[all …]
Ddir.c203 uint32_t hash, bucket; in find_rsb_root() local
207 bucket = hash & (ls->ls_rsbtbl_size - 1); in find_rsb_root()
209 spin_lock(&ls->ls_rsbtbl[bucket].lock); in find_rsb_root()
210 rv = dlm_search_rsb_tree(&ls->ls_rsbtbl[bucket].keep, name, len, &r); in find_rsb_root()
212 rv = dlm_search_rsb_tree(&ls->ls_rsbtbl[bucket].toss, in find_rsb_root()
214 spin_unlock(&ls->ls_rsbtbl[bucket].lock); in find_rsb_root()
/linux-4.19.296/drivers/cpuidle/governors/
Dmenu.c128 unsigned int bucket; member
145 int bucket = 0; in which_bucket() local
154 bucket = BUCKETS/2; in which_bucket()
157 return bucket; in which_bucket()
159 return bucket + 1; in which_bucket()
161 return bucket + 2; in which_bucket()
163 return bucket + 3; in which_bucket()
165 return bucket + 4; in which_bucket()
166 return bucket + 5; in which_bucket()
311 data->bucket = which_bucket(data->next_timer_us, nr_iowaiters); in menu_select()
[all …]
/linux-4.19.296/fs/nfs/
Dpnfs_nfs.c75 struct pnfs_commit_bucket *bucket; in pnfs_generic_clear_request_commit() local
77 bucket = list_first_entry(&req->wb_list, in pnfs_generic_clear_request_commit()
80 freeme = bucket->wlseg; in pnfs_generic_clear_request_commit()
81 bucket->wlseg = NULL; in pnfs_generic_clear_request_commit()
90 pnfs_generic_scan_ds_commit_list(struct pnfs_commit_bucket *bucket, in pnfs_generic_scan_ds_commit_list() argument
94 struct list_head *src = &bucket->written; in pnfs_generic_scan_ds_commit_list()
95 struct list_head *dst = &bucket->committing; in pnfs_generic_scan_ds_commit_list()
103 if (bucket->clseg == NULL) in pnfs_generic_scan_ds_commit_list()
104 bucket->clseg = pnfs_get_lseg(bucket->wlseg); in pnfs_generic_scan_ds_commit_list()
106 pnfs_put_lseg(bucket->wlseg); in pnfs_generic_scan_ds_commit_list()
[all …]
/linux-4.19.296/fs/ocfs2/
Dxattr.c135 struct ocfs2_xattr_bucket *bucket; member
289 struct ocfs2_xattr_bucket *bucket,
311 struct ocfs2_xattr_bucket *bucket,
332 struct ocfs2_xattr_bucket *bucket; in ocfs2_xattr_bucket_new() local
337 bucket = kzalloc(sizeof(struct ocfs2_xattr_bucket), GFP_NOFS); in ocfs2_xattr_bucket_new()
338 if (bucket) { in ocfs2_xattr_bucket_new()
339 bucket->bu_inode = inode; in ocfs2_xattr_bucket_new()
340 bucket->bu_blocks = blks; in ocfs2_xattr_bucket_new()
343 return bucket; in ocfs2_xattr_bucket_new()
346 static void ocfs2_xattr_bucket_relse(struct ocfs2_xattr_bucket *bucket) in ocfs2_xattr_bucket_relse() argument
[all …]
/linux-4.19.296/lib/
Dstackdepot.c186 static inline struct stack_record *find_stack(struct stack_record *bucket, in find_stack() argument
192 for (found = bucket; found; found = found->next) { in find_stack()
226 struct stack_record *found = NULL, **bucket; in depot_save_stack() local
235 bucket = &stack_table[hash & STACK_HASH_MASK]; in depot_save_stack()
242 found = find_stack(smp_load_acquire(bucket), trace->entries, in depot_save_stack()
271 found = find_stack(*bucket, trace->entries, trace->nr_entries, hash); in depot_save_stack()
277 new->next = *bucket; in depot_save_stack()
282 smp_store_release(bucket, new); in depot_save_stack()
Drhashtable.c38 struct rhash_head __rcu *bucket; member
131 INIT_RHT_NULLS_HEAD(ntbl[i].bucket); in nested_table_alloc()
1208 return &ntbl[subhash].bucket; in rht_bucket_nested()
1238 return &ntbl[hash].bucket; in rht_bucket_nested_insert()
/linux-4.19.296/include/trace/events/
Dbcache.h68 __field(size_t, bucket )
72 __entry->bucket = PTR_BUCKET_NR(b->c, &b->key, 0);
75 TP_printk("bucket %zu", __entry->bucket)
246 __field(size_t, bucket )
252 __entry->bucket = PTR_BUCKET_NR(b->c, &b->key, 0);
257 TP_printk("bucket %zu", __entry->bucket)
348 __field(size_t, bucket )
353 __entry->bucket = PTR_BUCKET_NR(b->c, &b->key, 0);
357 TP_printk("bucket %zu keys %u", __entry->bucket, __entry->keys)
407 TP_PROTO(struct cache *ca, size_t bucket),
[all …]
/linux-4.19.296/include/net/
Dtransp_v6.h48 __u16 srcp, __u16 destp, int rqueue, int bucket);
51 __u16 destp, int bucket) in ip6_dgram_sock_seq_show() argument
54 bucket); in ip6_dgram_sock_seq_show()
Dllc.h75 u32 bucket = hash_32(ifindex, LLC_SK_DEV_HASH_BITS); in llc_sk_dev_hash() local
77 return &sap->sk_dev_hash[bucket]; in llc_sk_dev_hash()
Draw.h50 int bucket; member
Dping.h51 int bucket; member
/linux-4.19.296/fs/xfs/libxfs/
Dxfs_ag.c220 int bucket; in xfs_agflblock_init() local
229 for (bucket = 0; bucket < xfs_agfl_size(mp); bucket++) in xfs_agflblock_init()
230 agfl_bno[bucket] = cpu_to_be32(NULLAGBLOCK); in xfs_agflblock_init()
240 int bucket; in xfs_agiblock_init() local
258 for (bucket = 0; bucket < XFS_AGI_UNLINKED_BUCKETS; bucket++) in xfs_agiblock_init()
259 agi->agi_unlinked[bucket] = cpu_to_be32(NULLAGINO); in xfs_agiblock_init()
/linux-4.19.296/drivers/misc/vmw_vmci/
Dvmci_doorbell.c128 u32 bucket = VMCI_DOORBELL_HASH(idx); in dbell_index_table_find() local
131 hlist_for_each_entry(dbell, &vmci_doorbell_it.entries[bucket], in dbell_index_table_find()
147 u32 bucket; in dbell_index_table_add() local
195 bucket = VMCI_DOORBELL_HASH(entry->idx); in dbell_index_table_add()
196 hlist_add_head(&entry->node, &vmci_doorbell_it.entries[bucket]); in dbell_index_table_add()
360 u32 bucket = VMCI_DOORBELL_HASH(notify_idx); in dbell_fire_entries() local
365 hlist_for_each_entry(dbell, &vmci_doorbell_it.entries[bucket], node) { in dbell_fire_entries()
/linux-4.19.296/fs/btrfs/
Dcompression.c750 struct bucket_item *bucket; member
763 kfree(workspace->bucket); in free_heuristic_ws()
780 ws->bucket = kcalloc(BUCKET_SIZE, sizeof(*ws->bucket), GFP_KERNEL); in alloc_heuristic_ws()
781 if (!ws->bucket) in alloc_heuristic_ws()
1269 for (i = 0; i < BUCKET_SIZE && ws->bucket[i].count > 0; i++) { in shannon_entropy()
1270 p = ws->bucket[i].count; in shannon_entropy()
1402 struct bucket_item *bucket = ws->bucket; in byte_core_set_size() local
1405 radix_sort(ws->bucket, ws->bucket_b, BUCKET_SIZE); in byte_core_set_size()
1408 coreset_sum += bucket[i].count; in byte_core_set_size()
1413 for (; i < BYTE_CORE_SET_HIGH && bucket[i].count > 0; i++) { in byte_core_set_size()
[all …]
/linux-4.19.296/fs/fscache/
Dcookie.c192 unsigned int bucket; in fscache_hash_cookie() local
194 bucket = candidate->key_hash & (ARRAY_SIZE(fscache_cookie_hash) - 1); in fscache_hash_cookie()
195 h = &fscache_cookie_hash[bucket]; in fscache_hash_cookie()
835 unsigned int bucket; in fscache_unhash_cookie() local
837 bucket = cookie->key_hash & (ARRAY_SIZE(fscache_cookie_hash) - 1); in fscache_unhash_cookie()
838 h = &fscache_cookie_hash[bucket]; in fscache_unhash_cookie()
/linux-4.19.296/include/linux/
Datalk.h149 int bucket; member
/linux-4.19.296/fs/ocfs2/dlm/
Ddlmdomain.c175 struct hlist_head *bucket; in __dlm_insert_lockres() local
179 bucket = dlm_lockres_hash(dlm, res->lockname.hash); in __dlm_insert_lockres()
184 hlist_add_head(&res->hash_node, bucket); in __dlm_insert_lockres()
195 struct hlist_head *bucket; in __dlm_lookup_lockres_full() local
202 bucket = dlm_lockres_hash(dlm, hash); in __dlm_lookup_lockres_full()
204 hlist_for_each_entry(res, bucket, hash_node) { in __dlm_lookup_lockres_full()
426 struct hlist_head *bucket; in dlm_migrate_all_locks() local
436 bucket = dlm_lockres_hash(dlm, i); in dlm_migrate_all_locks()
437 iter = bucket->first; in dlm_migrate_all_locks()
Ddlmmaster.c330 struct hlist_head *bucket; in __dlm_insert_mle() local
334 bucket = dlm_master_hash(dlm, mle->mnamehash); in __dlm_insert_mle()
335 hlist_add_head(&mle->master_hash_node, bucket); in __dlm_insert_mle()
344 struct hlist_head *bucket; in dlm_find_mle() local
350 bucket = dlm_master_hash(dlm, hash); in dlm_find_mle()
351 hlist_for_each_entry(tmpmle, bucket, master_hash_node) { in dlm_find_mle()
3351 struct hlist_head *bucket; in dlm_clean_master_list() local
3362 bucket = dlm_master_hash(dlm, i); in dlm_clean_master_list()
3363 hlist_for_each_entry_safe(mle, tmp, bucket, master_hash_node) { in dlm_clean_master_list()
3551 struct hlist_head *bucket; in dlm_force_free_mles() local
[all …]
/linux-4.19.296/fs/xfs/
Dxfs_log_recover.c2018 struct list_head *bucket; in xlog_recover_buffer_pass1() local
2033 bucket = XLOG_BUF_CANCEL_BUCKET(log, buf_f->blf_blkno); in xlog_recover_buffer_pass1()
2034 list_for_each_entry(bcp, bucket, bc_list) { in xlog_recover_buffer_pass1()
2047 list_add_tail(&bcp->bc_list, bucket); in xlog_recover_buffer_pass1()
2065 struct list_head *bucket; in xlog_peek_buffer_cancelled() local
2074 bucket = XLOG_BUF_CANCEL_BUCKET(log, blkno); in xlog_peek_buffer_cancelled()
2075 list_for_each_entry(bcp, bucket, bc_list) { in xlog_peek_buffer_cancelled()
5013 int bucket) in xlog_recover_clear_agi_bucket() argument
5030 agi->agi_unlinked[bucket] = cpu_to_be32(NULLAGINO); in xlog_recover_clear_agi_bucket()
5032 (sizeof(xfs_agino_t) * bucket); in xlog_recover_clear_agi_bucket()
[all …]
/linux-4.19.296/fs/omfs/
Ddir.c28 int bucket = omfs_hash(name, namelen, nbuckets); in omfs_get_bucket() local
30 *ofs = OMFS_DIR_START + bucket * 8; in omfs_get_bucket()

12