/linux-4.19.296/drivers/misc/sgi-xp/ |
D | xpc_main.c | 177 struct xpc_partition *part = from_timer(part, t, disengage_timer); in xpc_timeout_partition_disengage() local 179 DBUG_ON(time_is_after_jiffies(part->disengage_timeout)); in xpc_timeout_partition_disengage() 181 (void)xpc_partition_disengaged(part); in xpc_timeout_partition_disengage() 183 DBUG_ON(part->disengage_timeout != 0); in xpc_timeout_partition_disengage() 184 DBUG_ON(xpc_arch_ops.partition_engaged(XPC_PARTID(part))); in xpc_timeout_partition_disengage() 226 struct xpc_partition *part; in xpc_check_remote_hb() local 238 part = &xpc_partitions[partid]; in xpc_check_remote_hb() 240 if (part->act_state == XPC_P_AS_INACTIVE || in xpc_check_remote_hb() 241 part->act_state == XPC_P_AS_DEACTIVATING) { in xpc_check_remote_hb() 245 ret = xpc_arch_ops.get_remote_heartbeat(part); in xpc_check_remote_hb() [all …]
|
D | xpc_partition.c | 270 xpc_partition_disengaged(struct xpc_partition *part) in xpc_partition_disengaged() argument 272 short partid = XPC_PARTID(part); in xpc_partition_disengaged() 276 if (part->disengage_timeout) { in xpc_partition_disengaged() 278 if (time_is_after_jiffies(part->disengage_timeout)) { in xpc_partition_disengaged() 294 part->disengage_timeout = 0; in xpc_partition_disengaged() 298 del_singleshot_timer_sync(&part->disengage_timer); in xpc_partition_disengaged() 300 DBUG_ON(part->act_state != XPC_P_AS_DEACTIVATING && in xpc_partition_disengaged() 301 part->act_state != XPC_P_AS_INACTIVE); in xpc_partition_disengaged() 302 if (part->act_state != XPC_P_AS_INACTIVE) in xpc_partition_disengaged() 303 xpc_wakeup_channel_mgr(part); in xpc_partition_disengaged() [all …]
|
D | xpc_uv.c | 373 struct xpc_partition *part; in xpc_process_activate_IRQ_rcvd_uv() local 380 part = &xpc_partitions[partid]; in xpc_process_activate_IRQ_rcvd_uv() 382 if (part->sn.uv.act_state_req == 0) in xpc_process_activate_IRQ_rcvd_uv() 388 act_state_req = part->sn.uv.act_state_req; in xpc_process_activate_IRQ_rcvd_uv() 389 part->sn.uv.act_state_req = 0; in xpc_process_activate_IRQ_rcvd_uv() 393 if (part->act_state == XPC_P_AS_INACTIVE) in xpc_process_activate_IRQ_rcvd_uv() 394 xpc_activate_partition(part); in xpc_process_activate_IRQ_rcvd_uv() 395 else if (part->act_state == XPC_P_AS_DEACTIVATING) in xpc_process_activate_IRQ_rcvd_uv() 396 XPC_DEACTIVATE_PARTITION(part, xpReactivating); in xpc_process_activate_IRQ_rcvd_uv() 399 if (part->act_state == XPC_P_AS_INACTIVE) in xpc_process_activate_IRQ_rcvd_uv() [all …]
|
D | xpc_channel.c | 82 struct xpc_partition *part = &xpc_partitions[ch->partid]; in xpc_process_disconnect() local 101 if (part->act_state == XPC_P_AS_DEACTIVATING) { in xpc_process_disconnect() 157 atomic_dec(&part->nchannels_active); in xpc_process_disconnect() 168 if (part->act_state != XPC_P_AS_DEACTIVATING) { in xpc_process_disconnect() 170 spin_lock(&part->chctl_lock); in xpc_process_disconnect() 171 part->chctl.flags[ch->number] |= in xpc_process_disconnect() 173 spin_unlock(&part->chctl_lock); in xpc_process_disconnect() 183 xpc_process_openclose_chctl_flags(struct xpc_partition *part, int ch_number, in xpc_process_openclose_chctl_flags() argument 188 &part->remote_openclose_args[ch_number]; in xpc_process_openclose_chctl_flags() 189 struct xpc_channel *ch = &part->channels[ch_number]; in xpc_process_openclose_chctl_flags() [all …]
|
D | xpc_sn2.c | 268 xpc_check_for_sent_chctl_flags_sn2(struct xpc_partition *part) in xpc_check_for_sent_chctl_flags_sn2() argument 273 chctl.all_flags = xpc_receive_IRQ_amo_sn2(part->sn.sn2. in xpc_check_for_sent_chctl_flags_sn2() 278 spin_lock_irqsave(&part->chctl_lock, irq_flags); in xpc_check_for_sent_chctl_flags_sn2() 279 part->chctl.all_flags |= chctl.all_flags; in xpc_check_for_sent_chctl_flags_sn2() 280 spin_unlock_irqrestore(&part->chctl_lock, irq_flags); in xpc_check_for_sent_chctl_flags_sn2() 283 "0x%llx\n", XPC_PARTID(part), chctl.all_flags); in xpc_check_for_sent_chctl_flags_sn2() 285 xpc_wakeup_channel_mgr(part); in xpc_check_for_sent_chctl_flags_sn2() 308 struct xpc_partition *part = &xpc_partitions[partid]; in xpc_handle_notify_IRQ_sn2() local 312 if (xpc_part_ref(part)) { in xpc_handle_notify_IRQ_sn2() 313 xpc_check_for_sent_chctl_flags_sn2(part); in xpc_handle_notify_IRQ_sn2() [all …]
|
D | xpc.h | 937 xpc_wakeup_channel_mgr(struct xpc_partition *part) in xpc_wakeup_channel_mgr() argument 939 if (atomic_inc_return(&part->channel_mgr_requests) == 1) in xpc_wakeup_channel_mgr() 940 wake_up(&part->channel_mgr_wq); in xpc_wakeup_channel_mgr() 971 xpc_part_deref(struct xpc_partition *part) in xpc_part_deref() argument 973 s32 refs = atomic_dec_return(&part->references); in xpc_part_deref() 976 if (refs == 0 && part->setup_state == XPC_P_SS_WTEARDOWN) in xpc_part_deref() 977 wake_up(&part->teardown_wq); in xpc_part_deref() 981 xpc_part_ref(struct xpc_partition *part) in xpc_part_ref() argument 985 atomic_inc(&part->references); in xpc_part_ref() 986 setup = (part->setup_state == XPC_P_SS_SETUP); in xpc_part_ref() [all …]
|
/linux-4.19.296/drivers/irqchip/ |
D | irq-partition-percpu.c | 36 static bool partition_check_cpu(struct partition_desc *part, in partition_check_cpu() argument 39 return cpumask_test_cpu(cpu, &part->parts[hwirq].mask); in partition_check_cpu() 44 struct partition_desc *part = irq_data_get_irq_chip_data(d); in partition_irq_mask() local 45 struct irq_chip *chip = irq_desc_get_chip(part->chained_desc); in partition_irq_mask() 46 struct irq_data *data = irq_desc_get_irq_data(part->chained_desc); in partition_irq_mask() 48 if (partition_check_cpu(part, smp_processor_id(), d->hwirq) && in partition_irq_mask() 55 struct partition_desc *part = irq_data_get_irq_chip_data(d); in partition_irq_unmask() local 56 struct irq_chip *chip = irq_desc_get_chip(part->chained_desc); in partition_irq_unmask() 57 struct irq_data *data = irq_desc_get_irq_data(part->chained_desc); in partition_irq_unmask() 59 if (partition_check_cpu(part, smp_processor_id(), d->hwirq) && in partition_irq_unmask() [all …]
|
/linux-4.19.296/include/linux/ |
D | genhd.h | 26 #define part_to_dev(part) (&((part)->__dev)) argument 156 struct hd_struct __rcu *part[]; member 216 static inline struct gendisk *part_to_disk(struct hd_struct *part) in part_to_disk() argument 218 if (likely(part)) { in part_to_disk() 219 if (part->partno) in part_to_disk() 220 return dev_to_disk(part_to_dev(part)->parent); in part_to_disk() 222 return dev_to_disk(part_to_dev(part)); in part_to_disk() 245 static inline dev_t part_devt(struct hd_struct *part) in part_devt() argument 247 return part_to_dev(part)->devt; in part_devt() 253 static inline void disk_put_part(struct hd_struct *part) in disk_put_part() argument [all …]
|
D | bcm963xx_nvram.h | 51 #define BCM963XX_NVRAM_NAND_PART_OFFSET(nvram, part) \ argument 52 bcm963xx_nvram_nand_part_offset(nvram, BCM963XX_NVRAM_NAND_PART_ ##part) 56 enum bcm963xx_nvram_nand_part part) in bcm963xx_nvram_nand_part_offset() argument 58 return nvram->nand_part_offset[part] * SZ_1K; in bcm963xx_nvram_nand_part_offset() 61 #define BCM963XX_NVRAM_NAND_PART_SIZE(nvram, part) \ argument 62 bcm963xx_nvram_nand_part_size(nvram, BCM963XX_NVRAM_NAND_PART_ ##part) 66 enum bcm963xx_nvram_nand_part part) in bcm963xx_nvram_nand_part_size() argument 68 return nvram->nand_part_size[part] * SZ_1K; in bcm963xx_nvram_nand_part_size()
|
/linux-4.19.296/fs/orangefs/ |
D | dir.c | 17 struct orangefs_dir_part *part; member 134 struct orangefs_dir_part *part, *new; in parse_readdir() local 138 part = od->part; in parse_readdir() 139 while (part) { in parse_readdir() 141 if (part->next) in parse_readdir() 142 part = part->next; in parse_readdir() 151 if (!od->part) in parse_readdir() 152 od->part = new; in parse_readdir() 154 part->next = new; in parse_readdir() 189 static int fill_from_part(struct orangefs_dir_part *part, in fill_from_part() argument [all …]
|
/linux-4.19.296/block/partitions/ |
D | mac.c | 16 extern void note_bootable_part(dev_t dev, int part, int goodness); 41 struct mac_partition *part; in mac_partition() local 59 if (partoffset + sizeof(*part) > datasize) in mac_partition() 61 part = (struct mac_partition *) (data + partoffset); in mac_partition() 62 if (be16_to_cpu(part->signature) != MAC_PARTITION_MAGIC) { in mac_partition() 66 blocks_in_map = be32_to_cpu(part->map_count); in mac_partition() 82 part = (struct mac_partition *) (data + pos%512); in mac_partition() 83 if (be16_to_cpu(part->signature) != MAC_PARTITION_MAGIC) in mac_partition() 86 be32_to_cpu(part->start_block) * (secsize/512), in mac_partition() 87 be32_to_cpu(part->block_count) * (secsize/512)); in mac_partition() [all …]
|
D | atari.c | 59 if (!VALID_PARTITION(&rs->part[0], hd_size) && in atari_partition() 60 !VALID_PARTITION(&rs->part[1], hd_size) && in atari_partition() 61 !VALID_PARTITION(&rs->part[2], hd_size) && in atari_partition() 62 !VALID_PARTITION(&rs->part[3], hd_size)) { in atari_partition() 72 pi = &rs->part[0]; in atari_partition() 74 for (slot = 1; pi < &rs->part[4] && slot < state->limit; slot++, pi++) { in atari_partition() 103 if (!(xrs->part[0].flg & 1)) { in atari_partition() 110 partsect + be32_to_cpu(xrs->part[0].st), in atari_partition() 111 be32_to_cpu(xrs->part[0].siz)); in atari_partition() 113 if (!(xrs->part[1].flg & 1)) { in atari_partition() [all …]
|
D | acorn.c | 71 struct riscix_part part[8]; member 92 int part; in riscix_partition() local 97 for (part = 0; part < 8; part++) { in riscix_partition() 98 if (rr->part[part].one && in riscix_partition() 99 memcmp(rr->part[part].name, "All\0", 4)) { in riscix_partition() 101 le32_to_cpu(rr->part[part].start), in riscix_partition() 102 le32_to_cpu(rr->part[part].length)); in riscix_partition() 104 strlcat(state->pp_buf, rr->part[part].name, PAGE_SIZE); in riscix_partition()
|
D | amiga.c | 36 int blk, part, res = 0; in amiga_partition() local 85 for (part = 1; blk>0 && part<=16; part++, put_dev_sector(sect)) { in amiga_partition()
|
/linux-4.19.296/lib/ |
D | flex_array.c | 168 struct flex_array_part *part = fa->parts[part_nr]; in __fa_get_part() local 169 if (!part) { in __fa_get_part() 170 part = kmalloc(sizeof(struct flex_array_part), flags); in __fa_get_part() 171 if (!part) in __fa_get_part() 174 memset(part, FLEX_ARRAY_FREE, in __fa_get_part() 176 fa->parts[part_nr] = part; in __fa_get_part() 178 return part; in __fa_get_part() 202 struct flex_array_part *part; in flex_array_put() local 210 part = (struct flex_array_part *)&fa->parts[0]; in flex_array_put() 213 part = __fa_get_part(fa, part_nr, flags); in flex_array_put() [all …]
|
D | list_sort.c | 106 struct list_head *part[MAX_LIST_LENGTH_BITS+1]; /* sorted partial lists in list_sort() local 115 memset(part, 0, sizeof(part)); in list_sort() 125 for (lev = 0; part[lev]; lev++) { in list_sort() 126 cur = merge(priv, cmp, part[lev], cur); in list_sort() 127 part[lev] = NULL; in list_sort() 130 if (unlikely(lev >= ARRAY_SIZE(part)-1)) { in list_sort() 136 part[lev] = cur; in list_sort() 140 if (part[lev]) in list_sort() 141 list = merge(priv, cmp, part[lev], list); in list_sort() 143 merge_and_restore_back_links(priv, cmp, head, part[max_lev], list); in list_sort()
|
/linux-4.19.296/drivers/misc/ |
D | sram.c | 42 struct sram_partition *part; in sram_read() local 44 part = container_of(attr, struct sram_partition, battr); in sram_read() 46 mutex_lock(&part->lock); in sram_read() 47 memcpy_fromio(buf, part->base + pos, count); in sram_read() 48 mutex_unlock(&part->lock); in sram_read() 57 struct sram_partition *part; in sram_write() local 59 part = container_of(attr, struct sram_partition, battr); in sram_write() 61 mutex_lock(&part->lock); in sram_write() 62 memcpy_toio(part->base + pos, buf, count); in sram_write() 63 mutex_unlock(&part->lock); in sram_write() [all …]
|
D | sram-exec.c | 31 struct sram_partition *part) in sram_check_protect_exec() argument 33 unsigned long base = (unsigned long)part->base; in sram_check_protect_exec() 45 int sram_add_protect_exec(struct sram_partition *part) in sram_add_protect_exec() argument 48 list_add_tail(&part->list, &exec_pool_list); in sram_add_protect_exec() 84 struct sram_partition *part = NULL, *p; in sram_exec_copy() local 92 part = p; in sram_exec_copy() 96 if (!part) in sram_exec_copy() 102 base = (unsigned long)part->base; in sram_exec_copy() 105 mutex_lock(&part->lock); in sram_exec_copy() 115 mutex_unlock(&part->lock); in sram_exec_copy()
|
D | sram.h | 43 struct sram_partition *part); 44 int sram_add_protect_exec(struct sram_partition *part); 48 struct sram_partition *part) in sram_check_protect_exec() argument 53 static inline int sram_add_protect_exec(struct sram_partition *part) in sram_add_protect_exec() argument
|
/linux-4.19.296/block/ |
D | genhd.c | 48 void part_inc_in_flight(struct request_queue *q, struct hd_struct *part, int rw) in part_inc_in_flight() argument 53 atomic_inc(&part->in_flight[rw]); in part_inc_in_flight() 54 if (part->partno) in part_inc_in_flight() 55 atomic_inc(&part_to_disk(part)->part0.in_flight[rw]); in part_inc_in_flight() 58 void part_dec_in_flight(struct request_queue *q, struct hd_struct *part, int rw) in part_dec_in_flight() argument 63 atomic_dec(&part->in_flight[rw]); in part_dec_in_flight() 64 if (part->partno) in part_dec_in_flight() 65 atomic_dec(&part_to_disk(part)->part0.in_flight[rw]); in part_dec_in_flight() 68 void part_in_flight(struct request_queue *q, struct hd_struct *part, in part_in_flight() argument 72 blk_mq_in_flight(q, part, inflight); in part_in_flight() [all …]
|
D | partition-generic.c | 237 struct hd_struct *part = dev_to_part(dev); in part_uevent() local 239 add_uevent_var(env, "PARTN=%u", part->partno); in part_uevent() 240 if (part->info && part->info->volname[0]) in part_uevent() 241 add_uevent_var(env, "PARTNAME=%s", part->info->volname); in part_uevent() 254 struct hd_struct *part = container_of(to_rcu_work(work), struct hd_struct, in delete_partition_work_fn() local 257 part->start_sect = 0; in delete_partition_work_fn() 258 part->nr_sects = 0; in delete_partition_work_fn() 259 part_stat_set_all(part, 0); in delete_partition_work_fn() 260 put_device(part_to_dev(part)); in delete_partition_work_fn() 265 struct hd_struct *part = container_of(ref, struct hd_struct, ref); in __delete_partition() local [all …]
|
D | ioctl.c | 17 struct hd_struct *part, *lpart; in blkpg_ioctl() local 57 while ((part = disk_part_iter_next(&piter))) { in blkpg_ioctl() 58 if (!(start + length <= part->start_sect || in blkpg_ioctl() 59 start >= part->start_sect + part->nr_sects)) { in blkpg_ioctl() 68 part = add_partition(disk, partno, start, length, in blkpg_ioctl() 71 return PTR_ERR_OR_ZERO(part); in blkpg_ioctl() 73 part = disk_get_part(disk, partno); in blkpg_ioctl() 74 if (!part) in blkpg_ioctl() 77 bdevp = bdget(part_devt(part)); in blkpg_ioctl() 78 disk_put_part(part); in blkpg_ioctl() [all …]
|
/linux-4.19.296/drivers/char/ |
D | generic_nvram.c | 85 int part, offset; in nvram_ioctl() local 89 if (copy_from_user(&part, (void __user*)arg, sizeof(part)) != 0) in nvram_ioctl() 91 if (part < pmac_nvram_OF || part > pmac_nvram_NR) in nvram_ioctl() 93 offset = pmac_get_partition(part); in nvram_ioctl()
|
/linux-4.19.296/fs/hfs/ |
D | part_tbl.c | 82 (HFS_SB(sb)->part < 0 || HFS_SB(sb)->part == i)) { in hfs_part_find() 98 (HFS_SB(sb)->part < 0 || HFS_SB(sb)->part == i)) { in hfs_part_find()
|
/linux-4.19.296/fs/hfsplus/ |
D | part_tbl.c | 81 (sbi->part < 0 || sbi->part == i)) { in hfs_parse_old_pmap() 102 (sbi->part < 0 || sbi->part == i)) { in hfs_parse_new_pmap()
|