/linux-4.19.296/fs/xfs/libxfs/ |
D | xfs_btree.c | 61 struct xfs_btree_cur *cur, in __xfs_btree_check_lblock() argument 66 struct xfs_mount *mp = cur->bc_mp; in __xfs_btree_check_lblock() 67 xfs_btnum_t btnum = cur->bc_btnum; in __xfs_btree_check_lblock() 85 cur->bc_ops->get_maxrecs(cur, level)) in __xfs_btree_check_lblock() 88 !xfs_btree_check_lptr(cur, be64_to_cpu(block->bb_u.l.bb_leftsib), in __xfs_btree_check_lblock() 92 !xfs_btree_check_lptr(cur, be64_to_cpu(block->bb_u.l.bb_rightsib), in __xfs_btree_check_lblock() 102 struct xfs_btree_cur *cur, in xfs_btree_check_lblock() argument 107 struct xfs_mount *mp = cur->bc_mp; in xfs_btree_check_lblock() 110 fa = __xfs_btree_check_lblock(cur, block, level, bp); in xfs_btree_check_lblock() 127 struct xfs_btree_cur *cur, in __xfs_btree_check_sblock() argument [all …]
|
D | xfs_refcount.c | 47 struct xfs_btree_cur *cur, in xfs_refcount_lookup_le() argument 51 trace_xfs_refcount_lookup(cur->bc_mp, cur->bc_private.a.agno, bno, in xfs_refcount_lookup_le() 53 cur->bc_rec.rc.rc_startblock = bno; in xfs_refcount_lookup_le() 54 cur->bc_rec.rc.rc_blockcount = 0; in xfs_refcount_lookup_le() 55 return xfs_btree_lookup(cur, XFS_LOOKUP_LE, stat); in xfs_refcount_lookup_le() 64 struct xfs_btree_cur *cur, in xfs_refcount_lookup_ge() argument 68 trace_xfs_refcount_lookup(cur->bc_mp, cur->bc_private.a.agno, bno, in xfs_refcount_lookup_ge() 70 cur->bc_rec.rc.rc_startblock = bno; in xfs_refcount_lookup_ge() 71 cur->bc_rec.rc.rc_blockcount = 0; in xfs_refcount_lookup_ge() 72 return xfs_btree_lookup(cur, XFS_LOOKUP_GE, stat); in xfs_refcount_lookup_ge() [all …]
|
D | xfs_ialloc_btree.c | 28 struct xfs_btree_cur *cur, in xfs_inobt_get_minrecs() argument 31 return cur->bc_mp->m_inobt_mnr[level != 0]; in xfs_inobt_get_minrecs() 36 struct xfs_btree_cur *cur) in xfs_inobt_dup_cursor() argument 38 return xfs_inobt_init_cursor(cur->bc_mp, cur->bc_tp, in xfs_inobt_dup_cursor() 39 cur->bc_private.a.agbp, cur->bc_private.a.agno, in xfs_inobt_dup_cursor() 40 cur->bc_btnum); in xfs_inobt_dup_cursor() 45 struct xfs_btree_cur *cur, in xfs_inobt_set_root() argument 49 struct xfs_buf *agbp = cur->bc_private.a.agbp; in xfs_inobt_set_root() 54 xfs_ialloc_log_agi(cur->bc_tp, agbp, XFS_AGI_ROOT | XFS_AGI_LEVEL); in xfs_inobt_set_root() 59 struct xfs_btree_cur *cur, in xfs_finobt_set_root() argument [all …]
|
D | xfs_iext_tree.c | 144 static inline struct xfs_iext_rec *cur_rec(struct xfs_iext_cursor *cur) in cur_rec() argument 146 return &cur->leaf->recs[cur->pos]; in cur_rec() 150 struct xfs_iext_cursor *cur) in xfs_iext_valid() argument 152 if (!cur->leaf) in xfs_iext_valid() 154 if (cur->pos < 0 || cur->pos >= xfs_iext_max_recs(ifp)) in xfs_iext_valid() 156 if (xfs_iext_rec_is_empty(cur_rec(cur))) in xfs_iext_valid() 203 struct xfs_iext_cursor *cur) in xfs_iext_first() argument 205 cur->pos = 0; in xfs_iext_first() 206 cur->leaf = xfs_iext_find_first_leaf(ifp); in xfs_iext_first() 212 struct xfs_iext_cursor *cur) in xfs_iext_last() argument [all …]
|
D | xfs_alloc_btree.c | 26 struct xfs_btree_cur *cur) in xfs_allocbt_dup_cursor() argument 28 return xfs_allocbt_init_cursor(cur->bc_mp, cur->bc_tp, in xfs_allocbt_dup_cursor() 29 cur->bc_private.a.agbp, cur->bc_private.a.agno, in xfs_allocbt_dup_cursor() 30 cur->bc_btnum); in xfs_allocbt_dup_cursor() 35 struct xfs_btree_cur *cur, in xfs_allocbt_set_root() argument 39 struct xfs_buf *agbp = cur->bc_private.a.agbp; in xfs_allocbt_set_root() 42 int btnum = cur->bc_btnum; in xfs_allocbt_set_root() 43 struct xfs_perag *pag = xfs_perag_get(cur->bc_mp, seqno); in xfs_allocbt_set_root() 52 xfs_alloc_log_agf(cur->bc_tp, agbp, XFS_AGF_ROOTS | XFS_AGF_LEVELS); in xfs_allocbt_set_root() 57 struct xfs_btree_cur *cur, in xfs_allocbt_alloc_block() argument [all …]
|
D | xfs_refcount_btree.c | 27 struct xfs_btree_cur *cur) in xfs_refcountbt_dup_cursor() argument 29 return xfs_refcountbt_init_cursor(cur->bc_mp, cur->bc_tp, in xfs_refcountbt_dup_cursor() 30 cur->bc_private.a.agbp, cur->bc_private.a.agno); in xfs_refcountbt_dup_cursor() 35 struct xfs_btree_cur *cur, in xfs_refcountbt_set_root() argument 39 struct xfs_buf *agbp = cur->bc_private.a.agbp; in xfs_refcountbt_set_root() 42 struct xfs_perag *pag = xfs_perag_get(cur->bc_mp, seqno); in xfs_refcountbt_set_root() 51 xfs_alloc_log_agf(cur->bc_tp, agbp, in xfs_refcountbt_set_root() 57 struct xfs_btree_cur *cur, in xfs_refcountbt_alloc_block() argument 62 struct xfs_buf *agbp = cur->bc_private.a.agbp; in xfs_refcountbt_alloc_block() 68 args.tp = cur->bc_tp; in xfs_refcountbt_alloc_block() [all …]
|
D | xfs_rmap_btree.c | 55 struct xfs_btree_cur *cur) in xfs_rmapbt_dup_cursor() argument 57 return xfs_rmapbt_init_cursor(cur->bc_mp, cur->bc_tp, in xfs_rmapbt_dup_cursor() 58 cur->bc_private.a.agbp, cur->bc_private.a.agno); in xfs_rmapbt_dup_cursor() 63 struct xfs_btree_cur *cur, in xfs_rmapbt_set_root() argument 67 struct xfs_buf *agbp = cur->bc_private.a.agbp; in xfs_rmapbt_set_root() 70 int btnum = cur->bc_btnum; in xfs_rmapbt_set_root() 71 struct xfs_perag *pag = xfs_perag_get(cur->bc_mp, seqno); in xfs_rmapbt_set_root() 80 xfs_alloc_log_agf(cur->bc_tp, agbp, XFS_AGF_ROOTS | XFS_AGF_LEVELS); in xfs_rmapbt_set_root() 85 struct xfs_btree_cur *cur, in xfs_rmapbt_alloc_block() argument 90 struct xfs_buf *agbp = cur->bc_private.a.agbp; in xfs_rmapbt_alloc_block() [all …]
|
D | xfs_rmap.c | 38 struct xfs_btree_cur *cur, in xfs_rmap_lookup_le() argument 46 cur->bc_rec.r.rm_startblock = bno; in xfs_rmap_lookup_le() 47 cur->bc_rec.r.rm_blockcount = len; in xfs_rmap_lookup_le() 48 cur->bc_rec.r.rm_owner = owner; in xfs_rmap_lookup_le() 49 cur->bc_rec.r.rm_offset = offset; in xfs_rmap_lookup_le() 50 cur->bc_rec.r.rm_flags = flags; in xfs_rmap_lookup_le() 51 return xfs_btree_lookup(cur, XFS_LOOKUP_LE, stat); in xfs_rmap_lookup_le() 60 struct xfs_btree_cur *cur, in xfs_rmap_lookup_eq() argument 68 cur->bc_rec.r.rm_startblock = bno; in xfs_rmap_lookup_eq() 69 cur->bc_rec.r.rm_blockcount = len; in xfs_rmap_lookup_eq() [all …]
|
D | xfs_bmap_btree.c | 167 struct xfs_btree_cur *cur) in xfs_bmbt_dup_cursor() argument 171 new = xfs_bmbt_init_cursor(cur->bc_mp, cur->bc_tp, in xfs_bmbt_dup_cursor() 172 cur->bc_private.b.ip, cur->bc_private.b.whichfork); in xfs_bmbt_dup_cursor() 178 new->bc_private.b.flags = cur->bc_private.b.flags; in xfs_bmbt_dup_cursor() 199 struct xfs_btree_cur *cur, in xfs_bmbt_alloc_block() argument 208 args.tp = cur->bc_tp; in xfs_bmbt_alloc_block() 209 args.mp = cur->bc_mp; in xfs_bmbt_alloc_block() 210 args.fsbno = cur->bc_tp->t_firstblock; in xfs_bmbt_alloc_block() 211 xfs_rmap_ino_bmbt_owner(&args.oinfo, cur->bc_private.b.ip->i_ino, in xfs_bmbt_alloc_block() 212 cur->bc_private.b.whichfork); in xfs_bmbt_alloc_block() [all …]
|
D | xfs_btree.h | 88 #define XFS_BTREE_STATS_INC(cur, stat) \ argument 89 XFS_STATS_INC_OFF((cur)->bc_mp, (cur)->bc_statoff + __XBTS_ ## stat) 90 #define XFS_BTREE_STATS_ADD(cur, stat, val) \ argument 91 XFS_STATS_ADD_OFF((cur)->bc_mp, (cur)->bc_statoff + __XBTS_ ## stat, val) 106 void (*set_root)(struct xfs_btree_cur *cur, 110 int (*alloc_block)(struct xfs_btree_cur *cur, 114 int (*free_block)(struct xfs_btree_cur *cur, struct xfs_buf *bp); 117 void (*update_lastrec)(struct xfs_btree_cur *cur, 123 int (*get_minrecs)(struct xfs_btree_cur *cur, int level); 124 int (*get_maxrecs)(struct xfs_btree_cur *cur, int level); [all …]
|
D | xfs_ialloc.c | 53 struct xfs_btree_cur *cur, /* btree cursor */ in xfs_inobt_lookup() argument 58 cur->bc_rec.i.ir_startino = ino; in xfs_inobt_lookup() 59 cur->bc_rec.i.ir_holemask = 0; in xfs_inobt_lookup() 60 cur->bc_rec.i.ir_count = 0; in xfs_inobt_lookup() 61 cur->bc_rec.i.ir_freecount = 0; in xfs_inobt_lookup() 62 cur->bc_rec.i.ir_free = 0; in xfs_inobt_lookup() 63 return xfs_btree_lookup(cur, dir, stat); in xfs_inobt_lookup() 72 struct xfs_btree_cur *cur, /* btree cursor */ in xfs_inobt_update() argument 78 if (xfs_sb_version_hassparseinodes(&cur->bc_mp->m_sb)) { in xfs_inobt_update() 87 return xfs_btree_update(cur, &rec); in xfs_inobt_update() [all …]
|
D | xfs_inode_fork.h | 104 void xfs_iext_insert(struct xfs_inode *, struct xfs_iext_cursor *cur, 112 struct xfs_iext_cursor *cur, 116 struct xfs_iext_cursor *cur, 119 struct xfs_iext_cursor *cur, 122 struct xfs_iext_cursor *cur, 131 struct xfs_iext_cursor *cur, struct xfs_bmbt_irec *gotp) in xfs_iext_next_extent() argument 133 xfs_iext_next(ifp, cur); in xfs_iext_next_extent() 134 return xfs_iext_get_extent(ifp, cur, gotp); in xfs_iext_next_extent() 138 struct xfs_iext_cursor *cur, struct xfs_bmbt_irec *gotp) in xfs_iext_prev_extent() argument 140 xfs_iext_prev(ifp, cur); in xfs_iext_prev_extent() [all …]
|
/linux-4.19.296/fs/xfs/scrub/ |
D | btree.c | 34 struct xfs_btree_cur *cur, in __xchk_btree_process_error() argument 55 if (cur->bc_flags & XFS_BTREE_ROOT_IN_INODE) in __xchk_btree_process_error() 56 trace_xchk_ifork_btree_op_error(sc, cur, level, in __xchk_btree_process_error() 59 trace_xchk_btree_op_error(sc, cur, level, in __xchk_btree_process_error() 69 struct xfs_btree_cur *cur, in xchk_btree_process_error() argument 73 return __xchk_btree_process_error(sc, cur, level, error, in xchk_btree_process_error() 80 struct xfs_btree_cur *cur, in xchk_btree_xref_process_error() argument 84 return __xchk_btree_process_error(sc, cur, level, error, in xchk_btree_xref_process_error() 92 struct xfs_btree_cur *cur, in __xchk_btree_set_corrupt() argument 99 if (cur->bc_flags & XFS_BTREE_ROOT_IN_INODE) in __xchk_btree_set_corrupt() [all …]
|
D | trace.c | 26 struct xfs_btree_cur *cur, in xchk_btree_cur_fsbno() argument 29 if (level < cur->bc_nlevels && cur->bc_bufs[level]) in xchk_btree_cur_fsbno() 30 return XFS_DADDR_TO_FSB(cur->bc_mp, cur->bc_bufs[level]->b_bn); in xchk_btree_cur_fsbno() 31 else if (level == cur->bc_nlevels - 1 && in xchk_btree_cur_fsbno() 32 cur->bc_flags & XFS_BTREE_LONG_PTRS) in xchk_btree_cur_fsbno() 33 return XFS_INO_TO_FSB(cur->bc_mp, cur->bc_private.b.ip->i_ino); in xchk_btree_cur_fsbno() 34 else if (!(cur->bc_flags & XFS_BTREE_LONG_PTRS)) in xchk_btree_cur_fsbno() 35 return XFS_AGB_TO_FSB(cur->bc_mp, cur->bc_private.a.agno, 0); in xchk_btree_cur_fsbno()
|
D | ialloc.c | 105 struct xfs_mount *mp = bs->cur->bc_mp; in xchk_iallocbt_chunk() 106 xfs_agnumber_t agno = bs->cur->bc_private.a.agno; in xchk_iallocbt_chunk() 113 xchk_btree_set_corrupt(bs->sc, bs->cur, 0); in xchk_iallocbt_chunk() 140 struct xfs_mount *mp = bs->cur->bc_mp; in xchk_iallocbt_check_cluster_freemask() 153 xchk_btree_set_corrupt(bs->sc, bs->cur, 0); in xchk_iallocbt_check_cluster_freemask() 159 error = xfs_icache_inode_is_allocated(mp, bs->cur->bc_tp, in xchk_iallocbt_check_cluster_freemask() 178 xchk_btree_set_corrupt(bs->sc, bs->cur, 0); in xchk_iallocbt_check_cluster_freemask() 191 struct xfs_mount *mp = bs->cur->bc_mp; in xchk_iallocbt_check_freemask() 213 fsino = XFS_AGINO_TO_INO(mp, bs->cur->bc_private.a.agno, agino); in xchk_iallocbt_check_freemask() 226 xchk_btree_set_corrupt(bs->sc, bs->cur, 0); in xchk_iallocbt_check_freemask() [all …]
|
/linux-4.19.296/drivers/char/ |
D | bsr.c | 168 struct bsr_dev *cur, *n; in bsr_cleanup_devs() local 170 list_for_each_entry_safe(cur, n, &bsr_devs, bsr_list) { in bsr_cleanup_devs() 171 if (cur->bsr_device) { in bsr_cleanup_devs() 172 cdev_del(&cur->bsr_cdev); in bsr_cleanup_devs() 173 device_del(cur->bsr_device); in bsr_cleanup_devs() 175 list_del(&cur->bsr_list); in bsr_cleanup_devs() 176 kfree(cur); in bsr_cleanup_devs() 200 struct bsr_dev *cur = kzalloc(sizeof(struct bsr_dev), in bsr_add_node() local 205 if (!cur) { in bsr_add_node() 214 kfree(cur); in bsr_add_node() [all …]
|
/linux-4.19.296/lib/zstd/ |
D | zstd_opt.h | 440 U32 cur, match_num, last_pos, litlen, price; in ZSTD_compressBlock_opt_generic() local 457 cur = 0; in ZSTD_compressBlock_opt_generic() 482 cur = 0; in ZSTD_compressBlock_opt_generic() 515 for (cur = 1; cur <= last_pos; cur++) { in ZSTD_compressBlock_opt_generic() 516 inr = ip + cur; in ZSTD_compressBlock_opt_generic() 518 if (opt[cur - 1].mlen == 1) { in ZSTD_compressBlock_opt_generic() 519 litlen = opt[cur - 1].litlen + 1; in ZSTD_compressBlock_opt_generic() 520 if (cur > litlen) { in ZSTD_compressBlock_opt_generic() 521 price = opt[cur - litlen].price + ZSTD_getLiteralPrice(seqStorePtr, litlen, inr - litlen); in ZSTD_compressBlock_opt_generic() 526 price = opt[cur - 1].price + ZSTD_getLiteralPrice(seqStorePtr, litlen, inr - 1); in ZSTD_compressBlock_opt_generic() [all …]
|
/linux-4.19.296/include/linux/ |
D | iversion.h | 176 u64 cur, old, new; in inode_maybe_inc_iversion() local 189 cur = inode_peek_iversion_raw(inode); in inode_maybe_inc_iversion() 192 if (!force && !(cur & I_VERSION_QUERIED)) in inode_maybe_inc_iversion() 196 new = (cur & ~I_VERSION_QUERIED) + I_VERSION_INCREMENT; in inode_maybe_inc_iversion() 198 old = atomic64_cmpxchg(&inode->i_version, cur, new); in inode_maybe_inc_iversion() 199 if (likely(old == cur)) in inode_maybe_inc_iversion() 201 cur = old; in inode_maybe_inc_iversion() 283 u64 cur, old, new; in inode_query_iversion() local 285 cur = inode_peek_iversion_raw(inode); in inode_query_iversion() 288 if (cur & I_VERSION_QUERIED) { in inode_query_iversion() [all …]
|
/linux-4.19.296/drivers/char/agp/ |
D | isoch.c | 25 struct agp_3_5_dev *cur, *n = list_entry(new, struct agp_3_5_dev, list); in agp_3_5_dev_list_insert() local 29 cur = list_entry(pos, struct agp_3_5_dev, list); in agp_3_5_dev_list_insert() 30 if (cur->maxbw > n->maxbw) in agp_3_5_dev_list_insert() 38 struct agp_3_5_dev *cur; in agp_3_5_dev_list_sort() local 46 cur = list_entry(pos, struct agp_3_5_dev, list); in agp_3_5_dev_list_sort() 47 dev = cur->dev; in agp_3_5_dev_list_sort() 49 pci_read_config_dword(dev, cur->capndx+AGPNISTAT, &nistat); in agp_3_5_dev_list_sort() 50 cur->maxbw = (nistat >> 16) & 0xff; in agp_3_5_dev_list_sort() 82 struct agp_3_5_dev *cur; in agp_3_5_isochronous_node_enable() local 138 cur = list_entry(pos, struct agp_3_5_dev, list); in agp_3_5_isochronous_node_enable() [all …]
|
/linux-4.19.296/drivers/misc/ocxl/ |
D | pasid.c | 15 struct id_range *cur; in dump_list() local 18 list_for_each_entry(cur, head, list) { in dump_list() 19 pr_debug("Range %d->%d\n", cur->start, cur->end); in dump_list() 28 struct id_range *cur, *new; in range_alloc() local 37 list_for_each_entry(cur, head, list) { in range_alloc() 38 if ((cur->start - last_end) > size) in range_alloc() 40 last_end = cur->end; in range_alloc() 41 pos = &cur->list; in range_alloc() 65 struct id_range *cur, *tmp; in range_free() local 67 list_for_each_entry_safe(cur, tmp, head, list) { in range_free() [all …]
|
/linux-4.19.296/drivers/clk/rockchip/ |
D | clk-pll.c | 165 struct rockchip_pll_rate_table cur; in rockchip_rk3036_pll_recalc_rate() local 168 rockchip_rk3036_pll_get_params(pll, &cur); in rockchip_rk3036_pll_recalc_rate() 170 rate64 *= cur.fbdiv; in rockchip_rk3036_pll_recalc_rate() 171 do_div(rate64, cur.refdiv); in rockchip_rk3036_pll_recalc_rate() 173 if (cur.dsmpd == 0) { in rockchip_rk3036_pll_recalc_rate() 175 u64 frac_rate64 = prate * cur.frac; in rockchip_rk3036_pll_recalc_rate() 177 do_div(frac_rate64, cur.refdiv); in rockchip_rk3036_pll_recalc_rate() 181 do_div(rate64, cur.postdiv1); in rockchip_rk3036_pll_recalc_rate() 182 do_div(rate64, cur.postdiv2); in rockchip_rk3036_pll_recalc_rate() 192 struct rockchip_pll_rate_table cur; in rockchip_rk3036_pll_set_params() local [all …]
|
/linux-4.19.296/lib/ |
D | errseq.c | 60 errseq_t cur, old; in errseq_set() local 89 cur = new; in errseq_set() 94 cur = cmpxchg(eseq, old, new); in errseq_set() 100 if (likely(cur == old || cur == new)) in errseq_set() 104 old = cur; in errseq_set() 106 return cur; in errseq_set() 146 errseq_t cur = READ_ONCE(*eseq); in errseq_check() local 148 if (likely(cur == since)) in errseq_check() 150 return -(cur & MAX_ERRNO); in errseq_check()
|
D | test_list_sort.c | 73 struct list_head *cur; in list_sort_test() local 99 for (cur = head.next; cur->next != &head; cur = cur->next) { in list_sort_test() 103 if (cur->next->prev != cur) { in list_sort_test() 108 cmp_result = cmp(NULL, cur, cur->next); in list_sort_test() 114 el = container_of(cur, struct debug_el, list); in list_sort_test() 115 el1 = container_of(cur->next, struct debug_el, list); in list_sort_test() 128 if (head.prev != cur) { in list_sort_test()
|
/linux-4.19.296/drivers/misc/vmw_vmci/ |
D | vmci_event.c | 57 struct vmci_subscription *cur, *p2; in vmci_event_exit() local 58 list_for_each_entry_safe(cur, p2, &subscriber_array[e], node) { in vmci_event_exit() 66 list_del(&cur->node); in vmci_event_exit() 67 kfree(cur); in vmci_event_exit() 80 struct vmci_subscription *cur; in event_find() local 81 list_for_each_entry(cur, &subscriber_array[e], node) { in event_find() 82 if (cur->id == sub_id) in event_find() 83 return cur; in event_find() 95 struct vmci_subscription *cur; in event_deliver() local 100 list_for_each_entry_rcu(cur, subscriber_list, node) { in event_deliver() [all …]
|
/linux-4.19.296/fs/xfs/ |
D | xfs_trans_ail.c | 155 struct xfs_ail_cursor *cur) in xfs_trans_ail_cursor_init() argument 157 cur->item = NULL; in xfs_trans_ail_cursor_init() 158 list_add_tail(&cur->list, &ailp->ail_cursors); in xfs_trans_ail_cursor_init() 168 struct xfs_ail_cursor *cur) in xfs_trans_ail_cursor_next() argument 170 struct xfs_log_item *lip = cur->item; in xfs_trans_ail_cursor_next() 175 cur->item = xfs_ail_next(ailp, lip); in xfs_trans_ail_cursor_next() 185 struct xfs_ail_cursor *cur) in xfs_trans_ail_cursor_done() argument 187 cur->item = NULL; in xfs_trans_ail_cursor_done() 188 list_del_init(&cur->list); in xfs_trans_ail_cursor_done() 204 struct xfs_ail_cursor *cur; in xfs_trans_ail_cursor_clear() local [all …]
|