Home
last modified time | relevance | path

Searched refs:part (Results 1 – 25 of 88) sorted by relevance

1234

/linux-4.19.296/drivers/misc/sgi-xp/
Dxpc_main.c177 struct xpc_partition *part = from_timer(part, t, disengage_timer); in xpc_timeout_partition_disengage() local
179 DBUG_ON(time_is_after_jiffies(part->disengage_timeout)); in xpc_timeout_partition_disengage()
181 (void)xpc_partition_disengaged(part); in xpc_timeout_partition_disengage()
183 DBUG_ON(part->disengage_timeout != 0); in xpc_timeout_partition_disengage()
184 DBUG_ON(xpc_arch_ops.partition_engaged(XPC_PARTID(part))); in xpc_timeout_partition_disengage()
226 struct xpc_partition *part; in xpc_check_remote_hb() local
238 part = &xpc_partitions[partid]; in xpc_check_remote_hb()
240 if (part->act_state == XPC_P_AS_INACTIVE || in xpc_check_remote_hb()
241 part->act_state == XPC_P_AS_DEACTIVATING) { in xpc_check_remote_hb()
245 ret = xpc_arch_ops.get_remote_heartbeat(part); in xpc_check_remote_hb()
[all …]
Dxpc_partition.c270 xpc_partition_disengaged(struct xpc_partition *part) in xpc_partition_disengaged() argument
272 short partid = XPC_PARTID(part); in xpc_partition_disengaged()
276 if (part->disengage_timeout) { in xpc_partition_disengaged()
278 if (time_is_after_jiffies(part->disengage_timeout)) { in xpc_partition_disengaged()
294 part->disengage_timeout = 0; in xpc_partition_disengaged()
298 del_singleshot_timer_sync(&part->disengage_timer); in xpc_partition_disengaged()
300 DBUG_ON(part->act_state != XPC_P_AS_DEACTIVATING && in xpc_partition_disengaged()
301 part->act_state != XPC_P_AS_INACTIVE); in xpc_partition_disengaged()
302 if (part->act_state != XPC_P_AS_INACTIVE) in xpc_partition_disengaged()
303 xpc_wakeup_channel_mgr(part); in xpc_partition_disengaged()
[all …]
Dxpc_uv.c373 struct xpc_partition *part; in xpc_process_activate_IRQ_rcvd_uv() local
380 part = &xpc_partitions[partid]; in xpc_process_activate_IRQ_rcvd_uv()
382 if (part->sn.uv.act_state_req == 0) in xpc_process_activate_IRQ_rcvd_uv()
388 act_state_req = part->sn.uv.act_state_req; in xpc_process_activate_IRQ_rcvd_uv()
389 part->sn.uv.act_state_req = 0; in xpc_process_activate_IRQ_rcvd_uv()
393 if (part->act_state == XPC_P_AS_INACTIVE) in xpc_process_activate_IRQ_rcvd_uv()
394 xpc_activate_partition(part); in xpc_process_activate_IRQ_rcvd_uv()
395 else if (part->act_state == XPC_P_AS_DEACTIVATING) in xpc_process_activate_IRQ_rcvd_uv()
396 XPC_DEACTIVATE_PARTITION(part, xpReactivating); in xpc_process_activate_IRQ_rcvd_uv()
399 if (part->act_state == XPC_P_AS_INACTIVE) in xpc_process_activate_IRQ_rcvd_uv()
[all …]
Dxpc_channel.c82 struct xpc_partition *part = &xpc_partitions[ch->partid]; in xpc_process_disconnect() local
101 if (part->act_state == XPC_P_AS_DEACTIVATING) { in xpc_process_disconnect()
157 atomic_dec(&part->nchannels_active); in xpc_process_disconnect()
168 if (part->act_state != XPC_P_AS_DEACTIVATING) { in xpc_process_disconnect()
170 spin_lock(&part->chctl_lock); in xpc_process_disconnect()
171 part->chctl.flags[ch->number] |= in xpc_process_disconnect()
173 spin_unlock(&part->chctl_lock); in xpc_process_disconnect()
183 xpc_process_openclose_chctl_flags(struct xpc_partition *part, int ch_number, in xpc_process_openclose_chctl_flags() argument
188 &part->remote_openclose_args[ch_number]; in xpc_process_openclose_chctl_flags()
189 struct xpc_channel *ch = &part->channels[ch_number]; in xpc_process_openclose_chctl_flags()
[all …]
Dxpc_sn2.c268 xpc_check_for_sent_chctl_flags_sn2(struct xpc_partition *part) in xpc_check_for_sent_chctl_flags_sn2() argument
273 chctl.all_flags = xpc_receive_IRQ_amo_sn2(part->sn.sn2. in xpc_check_for_sent_chctl_flags_sn2()
278 spin_lock_irqsave(&part->chctl_lock, irq_flags); in xpc_check_for_sent_chctl_flags_sn2()
279 part->chctl.all_flags |= chctl.all_flags; in xpc_check_for_sent_chctl_flags_sn2()
280 spin_unlock_irqrestore(&part->chctl_lock, irq_flags); in xpc_check_for_sent_chctl_flags_sn2()
283 "0x%llx\n", XPC_PARTID(part), chctl.all_flags); in xpc_check_for_sent_chctl_flags_sn2()
285 xpc_wakeup_channel_mgr(part); in xpc_check_for_sent_chctl_flags_sn2()
308 struct xpc_partition *part = &xpc_partitions[partid]; in xpc_handle_notify_IRQ_sn2() local
312 if (xpc_part_ref(part)) { in xpc_handle_notify_IRQ_sn2()
313 xpc_check_for_sent_chctl_flags_sn2(part); in xpc_handle_notify_IRQ_sn2()
[all …]
Dxpc.h937 xpc_wakeup_channel_mgr(struct xpc_partition *part) in xpc_wakeup_channel_mgr() argument
939 if (atomic_inc_return(&part->channel_mgr_requests) == 1) in xpc_wakeup_channel_mgr()
940 wake_up(&part->channel_mgr_wq); in xpc_wakeup_channel_mgr()
971 xpc_part_deref(struct xpc_partition *part) in xpc_part_deref() argument
973 s32 refs = atomic_dec_return(&part->references); in xpc_part_deref()
976 if (refs == 0 && part->setup_state == XPC_P_SS_WTEARDOWN) in xpc_part_deref()
977 wake_up(&part->teardown_wq); in xpc_part_deref()
981 xpc_part_ref(struct xpc_partition *part) in xpc_part_ref() argument
985 atomic_inc(&part->references); in xpc_part_ref()
986 setup = (part->setup_state == XPC_P_SS_SETUP); in xpc_part_ref()
[all …]
/linux-4.19.296/drivers/irqchip/
Dirq-partition-percpu.c36 static bool partition_check_cpu(struct partition_desc *part, in partition_check_cpu() argument
39 return cpumask_test_cpu(cpu, &part->parts[hwirq].mask); in partition_check_cpu()
44 struct partition_desc *part = irq_data_get_irq_chip_data(d); in partition_irq_mask() local
45 struct irq_chip *chip = irq_desc_get_chip(part->chained_desc); in partition_irq_mask()
46 struct irq_data *data = irq_desc_get_irq_data(part->chained_desc); in partition_irq_mask()
48 if (partition_check_cpu(part, smp_processor_id(), d->hwirq) && in partition_irq_mask()
55 struct partition_desc *part = irq_data_get_irq_chip_data(d); in partition_irq_unmask() local
56 struct irq_chip *chip = irq_desc_get_chip(part->chained_desc); in partition_irq_unmask()
57 struct irq_data *data = irq_desc_get_irq_data(part->chained_desc); in partition_irq_unmask()
59 if (partition_check_cpu(part, smp_processor_id(), d->hwirq) && in partition_irq_unmask()
[all …]
/linux-4.19.296/include/linux/
Dgenhd.h26 #define part_to_dev(part) (&((part)->__dev)) argument
156 struct hd_struct __rcu *part[]; member
216 static inline struct gendisk *part_to_disk(struct hd_struct *part) in part_to_disk() argument
218 if (likely(part)) { in part_to_disk()
219 if (part->partno) in part_to_disk()
220 return dev_to_disk(part_to_dev(part)->parent); in part_to_disk()
222 return dev_to_disk(part_to_dev(part)); in part_to_disk()
245 static inline dev_t part_devt(struct hd_struct *part) in part_devt() argument
247 return part_to_dev(part)->devt; in part_devt()
253 static inline void disk_put_part(struct hd_struct *part) in disk_put_part() argument
[all …]
Dbcm963xx_nvram.h51 #define BCM963XX_NVRAM_NAND_PART_OFFSET(nvram, part) \ argument
52 bcm963xx_nvram_nand_part_offset(nvram, BCM963XX_NVRAM_NAND_PART_ ##part)
56 enum bcm963xx_nvram_nand_part part) in bcm963xx_nvram_nand_part_offset() argument
58 return nvram->nand_part_offset[part] * SZ_1K; in bcm963xx_nvram_nand_part_offset()
61 #define BCM963XX_NVRAM_NAND_PART_SIZE(nvram, part) \ argument
62 bcm963xx_nvram_nand_part_size(nvram, BCM963XX_NVRAM_NAND_PART_ ##part)
66 enum bcm963xx_nvram_nand_part part) in bcm963xx_nvram_nand_part_size() argument
68 return nvram->nand_part_size[part] * SZ_1K; in bcm963xx_nvram_nand_part_size()
/linux-4.19.296/fs/orangefs/
Ddir.c17 struct orangefs_dir_part *part; member
134 struct orangefs_dir_part *part, *new; in parse_readdir() local
138 part = od->part; in parse_readdir()
139 while (part) { in parse_readdir()
141 if (part->next) in parse_readdir()
142 part = part->next; in parse_readdir()
151 if (!od->part) in parse_readdir()
152 od->part = new; in parse_readdir()
154 part->next = new; in parse_readdir()
189 static int fill_from_part(struct orangefs_dir_part *part, in fill_from_part() argument
[all …]
/linux-4.19.296/block/partitions/
Dmac.c16 extern void note_bootable_part(dev_t dev, int part, int goodness);
41 struct mac_partition *part; in mac_partition() local
59 if (partoffset + sizeof(*part) > datasize) in mac_partition()
61 part = (struct mac_partition *) (data + partoffset); in mac_partition()
62 if (be16_to_cpu(part->signature) != MAC_PARTITION_MAGIC) { in mac_partition()
66 blocks_in_map = be32_to_cpu(part->map_count); in mac_partition()
82 part = (struct mac_partition *) (data + pos%512); in mac_partition()
83 if (be16_to_cpu(part->signature) != MAC_PARTITION_MAGIC) in mac_partition()
86 be32_to_cpu(part->start_block) * (secsize/512), in mac_partition()
87 be32_to_cpu(part->block_count) * (secsize/512)); in mac_partition()
[all …]
Datari.c59 if (!VALID_PARTITION(&rs->part[0], hd_size) && in atari_partition()
60 !VALID_PARTITION(&rs->part[1], hd_size) && in atari_partition()
61 !VALID_PARTITION(&rs->part[2], hd_size) && in atari_partition()
62 !VALID_PARTITION(&rs->part[3], hd_size)) { in atari_partition()
72 pi = &rs->part[0]; in atari_partition()
74 for (slot = 1; pi < &rs->part[4] && slot < state->limit; slot++, pi++) { in atari_partition()
103 if (!(xrs->part[0].flg & 1)) { in atari_partition()
110 partsect + be32_to_cpu(xrs->part[0].st), in atari_partition()
111 be32_to_cpu(xrs->part[0].siz)); in atari_partition()
113 if (!(xrs->part[1].flg & 1)) { in atari_partition()
[all …]
Dacorn.c71 struct riscix_part part[8]; member
92 int part; in riscix_partition() local
97 for (part = 0; part < 8; part++) { in riscix_partition()
98 if (rr->part[part].one && in riscix_partition()
99 memcmp(rr->part[part].name, "All\0", 4)) { in riscix_partition()
101 le32_to_cpu(rr->part[part].start), in riscix_partition()
102 le32_to_cpu(rr->part[part].length)); in riscix_partition()
104 strlcat(state->pp_buf, rr->part[part].name, PAGE_SIZE); in riscix_partition()
Damiga.c36 int blk, part, res = 0; in amiga_partition() local
85 for (part = 1; blk>0 && part<=16; part++, put_dev_sector(sect)) { in amiga_partition()
/linux-4.19.296/lib/
Dflex_array.c168 struct flex_array_part *part = fa->parts[part_nr]; in __fa_get_part() local
169 if (!part) { in __fa_get_part()
170 part = kmalloc(sizeof(struct flex_array_part), flags); in __fa_get_part()
171 if (!part) in __fa_get_part()
174 memset(part, FLEX_ARRAY_FREE, in __fa_get_part()
176 fa->parts[part_nr] = part; in __fa_get_part()
178 return part; in __fa_get_part()
202 struct flex_array_part *part; in flex_array_put() local
210 part = (struct flex_array_part *)&fa->parts[0]; in flex_array_put()
213 part = __fa_get_part(fa, part_nr, flags); in flex_array_put()
[all …]
Dlist_sort.c106 struct list_head *part[MAX_LIST_LENGTH_BITS+1]; /* sorted partial lists in list_sort() local
115 memset(part, 0, sizeof(part)); in list_sort()
125 for (lev = 0; part[lev]; lev++) { in list_sort()
126 cur = merge(priv, cmp, part[lev], cur); in list_sort()
127 part[lev] = NULL; in list_sort()
130 if (unlikely(lev >= ARRAY_SIZE(part)-1)) { in list_sort()
136 part[lev] = cur; in list_sort()
140 if (part[lev]) in list_sort()
141 list = merge(priv, cmp, part[lev], list); in list_sort()
143 merge_and_restore_back_links(priv, cmp, head, part[max_lev], list); in list_sort()
/linux-4.19.296/drivers/misc/
Dsram.c42 struct sram_partition *part; in sram_read() local
44 part = container_of(attr, struct sram_partition, battr); in sram_read()
46 mutex_lock(&part->lock); in sram_read()
47 memcpy_fromio(buf, part->base + pos, count); in sram_read()
48 mutex_unlock(&part->lock); in sram_read()
57 struct sram_partition *part; in sram_write() local
59 part = container_of(attr, struct sram_partition, battr); in sram_write()
61 mutex_lock(&part->lock); in sram_write()
62 memcpy_toio(part->base + pos, buf, count); in sram_write()
63 mutex_unlock(&part->lock); in sram_write()
[all …]
Dsram-exec.c31 struct sram_partition *part) in sram_check_protect_exec() argument
33 unsigned long base = (unsigned long)part->base; in sram_check_protect_exec()
45 int sram_add_protect_exec(struct sram_partition *part) in sram_add_protect_exec() argument
48 list_add_tail(&part->list, &exec_pool_list); in sram_add_protect_exec()
84 struct sram_partition *part = NULL, *p; in sram_exec_copy() local
92 part = p; in sram_exec_copy()
96 if (!part) in sram_exec_copy()
102 base = (unsigned long)part->base; in sram_exec_copy()
105 mutex_lock(&part->lock); in sram_exec_copy()
115 mutex_unlock(&part->lock); in sram_exec_copy()
Dsram.h43 struct sram_partition *part);
44 int sram_add_protect_exec(struct sram_partition *part);
48 struct sram_partition *part) in sram_check_protect_exec() argument
53 static inline int sram_add_protect_exec(struct sram_partition *part) in sram_add_protect_exec() argument
/linux-4.19.296/block/
Dgenhd.c48 void part_inc_in_flight(struct request_queue *q, struct hd_struct *part, int rw) in part_inc_in_flight() argument
53 atomic_inc(&part->in_flight[rw]); in part_inc_in_flight()
54 if (part->partno) in part_inc_in_flight()
55 atomic_inc(&part_to_disk(part)->part0.in_flight[rw]); in part_inc_in_flight()
58 void part_dec_in_flight(struct request_queue *q, struct hd_struct *part, int rw) in part_dec_in_flight() argument
63 atomic_dec(&part->in_flight[rw]); in part_dec_in_flight()
64 if (part->partno) in part_dec_in_flight()
65 atomic_dec(&part_to_disk(part)->part0.in_flight[rw]); in part_dec_in_flight()
68 void part_in_flight(struct request_queue *q, struct hd_struct *part, in part_in_flight() argument
72 blk_mq_in_flight(q, part, inflight); in part_in_flight()
[all …]
Dpartition-generic.c237 struct hd_struct *part = dev_to_part(dev); in part_uevent() local
239 add_uevent_var(env, "PARTN=%u", part->partno); in part_uevent()
240 if (part->info && part->info->volname[0]) in part_uevent()
241 add_uevent_var(env, "PARTNAME=%s", part->info->volname); in part_uevent()
254 struct hd_struct *part = container_of(to_rcu_work(work), struct hd_struct, in delete_partition_work_fn() local
257 part->start_sect = 0; in delete_partition_work_fn()
258 part->nr_sects = 0; in delete_partition_work_fn()
259 part_stat_set_all(part, 0); in delete_partition_work_fn()
260 put_device(part_to_dev(part)); in delete_partition_work_fn()
265 struct hd_struct *part = container_of(ref, struct hd_struct, ref); in __delete_partition() local
[all …]
Dioctl.c17 struct hd_struct *part, *lpart; in blkpg_ioctl() local
57 while ((part = disk_part_iter_next(&piter))) { in blkpg_ioctl()
58 if (!(start + length <= part->start_sect || in blkpg_ioctl()
59 start >= part->start_sect + part->nr_sects)) { in blkpg_ioctl()
68 part = add_partition(disk, partno, start, length, in blkpg_ioctl()
71 return PTR_ERR_OR_ZERO(part); in blkpg_ioctl()
73 part = disk_get_part(disk, partno); in blkpg_ioctl()
74 if (!part) in blkpg_ioctl()
77 bdevp = bdget(part_devt(part)); in blkpg_ioctl()
78 disk_put_part(part); in blkpg_ioctl()
[all …]
/linux-4.19.296/drivers/char/
Dgeneric_nvram.c85 int part, offset; in nvram_ioctl() local
89 if (copy_from_user(&part, (void __user*)arg, sizeof(part)) != 0) in nvram_ioctl()
91 if (part < pmac_nvram_OF || part > pmac_nvram_NR) in nvram_ioctl()
93 offset = pmac_get_partition(part); in nvram_ioctl()
/linux-4.19.296/fs/hfs/
Dpart_tbl.c82 (HFS_SB(sb)->part < 0 || HFS_SB(sb)->part == i)) { in hfs_part_find()
98 (HFS_SB(sb)->part < 0 || HFS_SB(sb)->part == i)) { in hfs_part_find()
/linux-4.19.296/fs/hfsplus/
Dpart_tbl.c81 (sbi->part < 0 || sbi->part == i)) { in hfs_parse_old_pmap()
102 (sbi->part < 0 || sbi->part == i)) { in hfs_parse_new_pmap()

1234