/linux-4.19.296/include/linux/ |
D | cpumask.h | 33 #define cpumask_pr_args(maskp) nr_cpu_ids, cpumask_bits(maskp) 36 #define nr_cpu_ids 1U macro 38 extern unsigned int nr_cpu_ids; 44 #define nr_cpumask_bits nr_cpu_ids 243 (cpu) < nr_cpu_ids;) 255 (cpu) < nr_cpu_ids;) 291 (cpu) < nr_cpu_ids;) 895 nr_cpu_ids); in cpumap_print_to_pagebuf()
|
D | backing-dev-defs.h | 46 #define WB_STAT_BATCH (8*(1+ilog2(nr_cpu_ids)))
|
D | backing-dev.h | 101 return nr_cpu_ids * WB_STAT_BATCH; in wb_stat_error()
|
/linux-4.19.296/drivers/base/ |
D | cpu.c | 256 if (total_cpus && nr_cpu_ids < total_cpus) { in print_cpus_offline() 260 if (nr_cpu_ids == total_cpus-1) in print_cpus_offline() 261 n += snprintf(&buf[n], len - n, "%u", nr_cpu_ids); in print_cpus_offline() 264 nr_cpu_ids, total_cpus-1); in print_cpus_offline() 400 if (cpu < nr_cpu_ids && cpu_possible(cpu)) in get_cpu_device()
|
/linux-4.19.296/block/ |
D | blk-mq-cpumap.c | 27 if (ret < nr_cpu_ids) in get_first_sibling()
|
D | blk-mq.c | 1332 if (cpu >= nr_cpu_ids) in blk_mq_first_mapped_cpu() 1355 if (next_cpu >= nr_cpu_ids) in blk_mq_hctx_next_cpu() 2219 hctx->ctxs = kmalloc_array_node(nr_cpu_ids, sizeof(void *), in blk_mq_init_hctx() 2224 if (sbitmap_init_node(&hctx->ctx_map, nr_cpu_ids, ilog2(8), in blk_mq_init_hctx() 2605 q->queue_hw_ctx = kcalloc_node(nr_cpu_ids, sizeof(*(q->queue_hw_ctx)), in blk_mq_init_allocated_queue() 2619 q->nr_queues = nr_cpu_ids; in blk_mq_init_allocated_queue() 2822 if (set->nr_hw_queues > nr_cpu_ids) in blk_mq_alloc_tag_set() 2823 set->nr_hw_queues = nr_cpu_ids; in blk_mq_alloc_tag_set() 2825 set->tags = kcalloc_node(nr_cpu_ids, sizeof(struct blk_mq_tags *), in blk_mq_alloc_tag_set() 2831 set->mq_map = kcalloc_node(nr_cpu_ids, sizeof(*set->mq_map), in blk_mq_alloc_tag_set() [all …]
|
/linux-4.19.296/lib/ |
D | flex_proportions.c | 170 #define PROP_BATCH (8*(1+ilog2(nr_cpu_ids))) 208 if (val < (nr_cpu_ids * PROP_BATCH)) in fprop_reflect_period_percpu()
|
D | cpu_rmap.c | 39 obj_offset = ALIGN(offsetof(struct cpu_rmap, near[nr_cpu_ids]), in alloc_cpu_rmap()
|
/linux-4.19.296/drivers/xen/ |
D | cpu_hotplug.c | 57 if (cpu >= nr_cpu_ids || !cpu_possible(cpu)) in vcpu_hotplug()
|
D | evtchn.c | 442 if (unlikely(selected_cpu >= nr_cpu_ids)) in evtchn_bind_interdom_next_vcpu()
|
/linux-4.19.296/include/linux/netfilter/ |
D | x_tables.h | 435 if (nr_cpu_ids > 1) in xt_get_this_cpu_counter() 444 if (nr_cpu_ids > 1) in xt_get_per_cpu_counter()
|
/linux-4.19.296/drivers/cpuidle/ |
D | dt_idle_states.c | 119 cpu < nr_cpu_ids; cpu = cpumask_next(cpu, cpumask)) { in idle_state_valid()
|
/linux-4.19.296/drivers/irqchip/ |
D | irq-hip04.c | 157 if (cpu >= NR_HIP04_CPU_IF || cpu >= nr_cpu_ids) in hip04_irq_set_affinity()
|
D | irq-bcm6345-l1.c | 211 if (new_cpu >= nr_cpu_ids) in bcm6345_l1_set_affinity()
|
D | irq-gic-v3.c | 702 while (cpu < nr_cpu_ids) { in gic_compute_target_list() 706 if (next_cpu >= nr_cpu_ids) in gic_compute_target_list() 787 if (cpu >= nr_cpu_ids) in gic_set_affinity()
|
D | irq-gic.c | 335 if (cpu >= NR_GIC_CPU_IF || cpu >= nr_cpu_ids) in gic_set_affinity() 798 if (unlikely(nr_cpu_ids == 1)) { in gic_raise_softirq()
|
D | irq-gic-v3-its.c | 1170 if (cpu >= nr_cpu_ids) in its_set_affinity() 1951 its->collections = kcalloc(nr_cpu_ids, sizeof(*its->collections), in its_alloc_collections() 1956 for (i = 0; i < nr_cpu_ids; i++) in its_alloc_collections() 2501 if (cpu >= nr_cpu_ids) { in its_irq_domain_activate() 3373 entries = roundup_pow_of_two(nr_cpu_ids); in its_init_vpe_domain()
|
/linux-4.19.296/drivers/bus/ |
D | arm-cci.c | 269 for (cpu = 0; cpu < nr_cpu_ids; cpu++) { in cci_disable_port_by_cpu()
|
/linux-4.19.296/drivers/cpufreq/ |
D | speedstep-centrino.c | 452 if (good_cpu >= nr_cpu_ids) { in centrino_target()
|
D | cpufreq.c | 225 if (WARN_ON(cpu >= nr_cpu_ids)) in cpufreq_cpu_get()
|
/linux-4.19.296/drivers/pci/ |
D | pci-driver.c | 353 cpu = nr_cpu_ids; in pci_call_probe() 357 if (cpu < nr_cpu_ids) in pci_call_probe()
|
/linux-4.19.296/fs/ |
D | seq_file.c | 1088 for (*cpu = cpumask_next(*cpu, cpu_possible_mask); *cpu < nr_cpu_ids; in seq_hlist_next_percpu()
|
/linux-4.19.296/fs/pstore/ |
D | ram.c | 804 ? nr_cpu_ids in ramoops_probe()
|
/linux-4.19.296/fs/btrfs/ |
D | disk-io.c | 2637 (1 + ilog2(nr_cpu_ids)); in open_ctree() 2888 fs_info->dirty_metadata_batch = nodesize * (1 + ilog2(nr_cpu_ids)); in open_ctree() 2889 fs_info->delalloc_batch = sectorsize * 512 * (1 + ilog2(nr_cpu_ids)); in open_ctree()
|
/linux-4.19.296/fs/ext4/ |
D | ext4.h | 2914 #define EXT4_FREECLUSTERS_WATERMARK (4 * (percpu_counter_batch * nr_cpu_ids))
|