Home
last modified time | relevance | path

Searched refs:cur (Results 1 – 25 of 205) sorted by relevance

123456789

/linux-4.19.296/fs/xfs/libxfs/
Dxfs_btree.c61 struct xfs_btree_cur *cur, in __xfs_btree_check_lblock() argument
66 struct xfs_mount *mp = cur->bc_mp; in __xfs_btree_check_lblock()
67 xfs_btnum_t btnum = cur->bc_btnum; in __xfs_btree_check_lblock()
85 cur->bc_ops->get_maxrecs(cur, level)) in __xfs_btree_check_lblock()
88 !xfs_btree_check_lptr(cur, be64_to_cpu(block->bb_u.l.bb_leftsib), in __xfs_btree_check_lblock()
92 !xfs_btree_check_lptr(cur, be64_to_cpu(block->bb_u.l.bb_rightsib), in __xfs_btree_check_lblock()
102 struct xfs_btree_cur *cur, in xfs_btree_check_lblock() argument
107 struct xfs_mount *mp = cur->bc_mp; in xfs_btree_check_lblock()
110 fa = __xfs_btree_check_lblock(cur, block, level, bp); in xfs_btree_check_lblock()
127 struct xfs_btree_cur *cur, in __xfs_btree_check_sblock() argument
[all …]
Dxfs_refcount.c47 struct xfs_btree_cur *cur, in xfs_refcount_lookup_le() argument
51 trace_xfs_refcount_lookup(cur->bc_mp, cur->bc_private.a.agno, bno, in xfs_refcount_lookup_le()
53 cur->bc_rec.rc.rc_startblock = bno; in xfs_refcount_lookup_le()
54 cur->bc_rec.rc.rc_blockcount = 0; in xfs_refcount_lookup_le()
55 return xfs_btree_lookup(cur, XFS_LOOKUP_LE, stat); in xfs_refcount_lookup_le()
64 struct xfs_btree_cur *cur, in xfs_refcount_lookup_ge() argument
68 trace_xfs_refcount_lookup(cur->bc_mp, cur->bc_private.a.agno, bno, in xfs_refcount_lookup_ge()
70 cur->bc_rec.rc.rc_startblock = bno; in xfs_refcount_lookup_ge()
71 cur->bc_rec.rc.rc_blockcount = 0; in xfs_refcount_lookup_ge()
72 return xfs_btree_lookup(cur, XFS_LOOKUP_GE, stat); in xfs_refcount_lookup_ge()
[all …]
Dxfs_ialloc_btree.c28 struct xfs_btree_cur *cur, in xfs_inobt_get_minrecs() argument
31 return cur->bc_mp->m_inobt_mnr[level != 0]; in xfs_inobt_get_minrecs()
36 struct xfs_btree_cur *cur) in xfs_inobt_dup_cursor() argument
38 return xfs_inobt_init_cursor(cur->bc_mp, cur->bc_tp, in xfs_inobt_dup_cursor()
39 cur->bc_private.a.agbp, cur->bc_private.a.agno, in xfs_inobt_dup_cursor()
40 cur->bc_btnum); in xfs_inobt_dup_cursor()
45 struct xfs_btree_cur *cur, in xfs_inobt_set_root() argument
49 struct xfs_buf *agbp = cur->bc_private.a.agbp; in xfs_inobt_set_root()
54 xfs_ialloc_log_agi(cur->bc_tp, agbp, XFS_AGI_ROOT | XFS_AGI_LEVEL); in xfs_inobt_set_root()
59 struct xfs_btree_cur *cur, in xfs_finobt_set_root() argument
[all …]
Dxfs_iext_tree.c144 static inline struct xfs_iext_rec *cur_rec(struct xfs_iext_cursor *cur) in cur_rec() argument
146 return &cur->leaf->recs[cur->pos]; in cur_rec()
150 struct xfs_iext_cursor *cur) in xfs_iext_valid() argument
152 if (!cur->leaf) in xfs_iext_valid()
154 if (cur->pos < 0 || cur->pos >= xfs_iext_max_recs(ifp)) in xfs_iext_valid()
156 if (xfs_iext_rec_is_empty(cur_rec(cur))) in xfs_iext_valid()
203 struct xfs_iext_cursor *cur) in xfs_iext_first() argument
205 cur->pos = 0; in xfs_iext_first()
206 cur->leaf = xfs_iext_find_first_leaf(ifp); in xfs_iext_first()
212 struct xfs_iext_cursor *cur) in xfs_iext_last() argument
[all …]
Dxfs_alloc_btree.c26 struct xfs_btree_cur *cur) in xfs_allocbt_dup_cursor() argument
28 return xfs_allocbt_init_cursor(cur->bc_mp, cur->bc_tp, in xfs_allocbt_dup_cursor()
29 cur->bc_private.a.agbp, cur->bc_private.a.agno, in xfs_allocbt_dup_cursor()
30 cur->bc_btnum); in xfs_allocbt_dup_cursor()
35 struct xfs_btree_cur *cur, in xfs_allocbt_set_root() argument
39 struct xfs_buf *agbp = cur->bc_private.a.agbp; in xfs_allocbt_set_root()
42 int btnum = cur->bc_btnum; in xfs_allocbt_set_root()
43 struct xfs_perag *pag = xfs_perag_get(cur->bc_mp, seqno); in xfs_allocbt_set_root()
52 xfs_alloc_log_agf(cur->bc_tp, agbp, XFS_AGF_ROOTS | XFS_AGF_LEVELS); in xfs_allocbt_set_root()
57 struct xfs_btree_cur *cur, in xfs_allocbt_alloc_block() argument
[all …]
Dxfs_refcount_btree.c27 struct xfs_btree_cur *cur) in xfs_refcountbt_dup_cursor() argument
29 return xfs_refcountbt_init_cursor(cur->bc_mp, cur->bc_tp, in xfs_refcountbt_dup_cursor()
30 cur->bc_private.a.agbp, cur->bc_private.a.agno); in xfs_refcountbt_dup_cursor()
35 struct xfs_btree_cur *cur, in xfs_refcountbt_set_root() argument
39 struct xfs_buf *agbp = cur->bc_private.a.agbp; in xfs_refcountbt_set_root()
42 struct xfs_perag *pag = xfs_perag_get(cur->bc_mp, seqno); in xfs_refcountbt_set_root()
51 xfs_alloc_log_agf(cur->bc_tp, agbp, in xfs_refcountbt_set_root()
57 struct xfs_btree_cur *cur, in xfs_refcountbt_alloc_block() argument
62 struct xfs_buf *agbp = cur->bc_private.a.agbp; in xfs_refcountbt_alloc_block()
68 args.tp = cur->bc_tp; in xfs_refcountbt_alloc_block()
[all …]
Dxfs_rmap_btree.c55 struct xfs_btree_cur *cur) in xfs_rmapbt_dup_cursor() argument
57 return xfs_rmapbt_init_cursor(cur->bc_mp, cur->bc_tp, in xfs_rmapbt_dup_cursor()
58 cur->bc_private.a.agbp, cur->bc_private.a.agno); in xfs_rmapbt_dup_cursor()
63 struct xfs_btree_cur *cur, in xfs_rmapbt_set_root() argument
67 struct xfs_buf *agbp = cur->bc_private.a.agbp; in xfs_rmapbt_set_root()
70 int btnum = cur->bc_btnum; in xfs_rmapbt_set_root()
71 struct xfs_perag *pag = xfs_perag_get(cur->bc_mp, seqno); in xfs_rmapbt_set_root()
80 xfs_alloc_log_agf(cur->bc_tp, agbp, XFS_AGF_ROOTS | XFS_AGF_LEVELS); in xfs_rmapbt_set_root()
85 struct xfs_btree_cur *cur, in xfs_rmapbt_alloc_block() argument
90 struct xfs_buf *agbp = cur->bc_private.a.agbp; in xfs_rmapbt_alloc_block()
[all …]
Dxfs_rmap.c38 struct xfs_btree_cur *cur, in xfs_rmap_lookup_le() argument
46 cur->bc_rec.r.rm_startblock = bno; in xfs_rmap_lookup_le()
47 cur->bc_rec.r.rm_blockcount = len; in xfs_rmap_lookup_le()
48 cur->bc_rec.r.rm_owner = owner; in xfs_rmap_lookup_le()
49 cur->bc_rec.r.rm_offset = offset; in xfs_rmap_lookup_le()
50 cur->bc_rec.r.rm_flags = flags; in xfs_rmap_lookup_le()
51 return xfs_btree_lookup(cur, XFS_LOOKUP_LE, stat); in xfs_rmap_lookup_le()
60 struct xfs_btree_cur *cur, in xfs_rmap_lookup_eq() argument
68 cur->bc_rec.r.rm_startblock = bno; in xfs_rmap_lookup_eq()
69 cur->bc_rec.r.rm_blockcount = len; in xfs_rmap_lookup_eq()
[all …]
Dxfs_bmap_btree.c167 struct xfs_btree_cur *cur) in xfs_bmbt_dup_cursor() argument
171 new = xfs_bmbt_init_cursor(cur->bc_mp, cur->bc_tp, in xfs_bmbt_dup_cursor()
172 cur->bc_private.b.ip, cur->bc_private.b.whichfork); in xfs_bmbt_dup_cursor()
178 new->bc_private.b.flags = cur->bc_private.b.flags; in xfs_bmbt_dup_cursor()
199 struct xfs_btree_cur *cur, in xfs_bmbt_alloc_block() argument
208 args.tp = cur->bc_tp; in xfs_bmbt_alloc_block()
209 args.mp = cur->bc_mp; in xfs_bmbt_alloc_block()
210 args.fsbno = cur->bc_tp->t_firstblock; in xfs_bmbt_alloc_block()
211 xfs_rmap_ino_bmbt_owner(&args.oinfo, cur->bc_private.b.ip->i_ino, in xfs_bmbt_alloc_block()
212 cur->bc_private.b.whichfork); in xfs_bmbt_alloc_block()
[all …]
Dxfs_btree.h88 #define XFS_BTREE_STATS_INC(cur, stat) \ argument
89 XFS_STATS_INC_OFF((cur)->bc_mp, (cur)->bc_statoff + __XBTS_ ## stat)
90 #define XFS_BTREE_STATS_ADD(cur, stat, val) \ argument
91 XFS_STATS_ADD_OFF((cur)->bc_mp, (cur)->bc_statoff + __XBTS_ ## stat, val)
106 void (*set_root)(struct xfs_btree_cur *cur,
110 int (*alloc_block)(struct xfs_btree_cur *cur,
114 int (*free_block)(struct xfs_btree_cur *cur, struct xfs_buf *bp);
117 void (*update_lastrec)(struct xfs_btree_cur *cur,
123 int (*get_minrecs)(struct xfs_btree_cur *cur, int level);
124 int (*get_maxrecs)(struct xfs_btree_cur *cur, int level);
[all …]
Dxfs_ialloc.c53 struct xfs_btree_cur *cur, /* btree cursor */ in xfs_inobt_lookup() argument
58 cur->bc_rec.i.ir_startino = ino; in xfs_inobt_lookup()
59 cur->bc_rec.i.ir_holemask = 0; in xfs_inobt_lookup()
60 cur->bc_rec.i.ir_count = 0; in xfs_inobt_lookup()
61 cur->bc_rec.i.ir_freecount = 0; in xfs_inobt_lookup()
62 cur->bc_rec.i.ir_free = 0; in xfs_inobt_lookup()
63 return xfs_btree_lookup(cur, dir, stat); in xfs_inobt_lookup()
72 struct xfs_btree_cur *cur, /* btree cursor */ in xfs_inobt_update() argument
78 if (xfs_sb_version_hassparseinodes(&cur->bc_mp->m_sb)) { in xfs_inobt_update()
87 return xfs_btree_update(cur, &rec); in xfs_inobt_update()
[all …]
Dxfs_inode_fork.h104 void xfs_iext_insert(struct xfs_inode *, struct xfs_iext_cursor *cur,
112 struct xfs_iext_cursor *cur,
116 struct xfs_iext_cursor *cur,
119 struct xfs_iext_cursor *cur,
122 struct xfs_iext_cursor *cur,
131 struct xfs_iext_cursor *cur, struct xfs_bmbt_irec *gotp) in xfs_iext_next_extent() argument
133 xfs_iext_next(ifp, cur); in xfs_iext_next_extent()
134 return xfs_iext_get_extent(ifp, cur, gotp); in xfs_iext_next_extent()
138 struct xfs_iext_cursor *cur, struct xfs_bmbt_irec *gotp) in xfs_iext_prev_extent() argument
140 xfs_iext_prev(ifp, cur); in xfs_iext_prev_extent()
[all …]
/linux-4.19.296/fs/xfs/scrub/
Dbtree.c34 struct xfs_btree_cur *cur, in __xchk_btree_process_error() argument
55 if (cur->bc_flags & XFS_BTREE_ROOT_IN_INODE) in __xchk_btree_process_error()
56 trace_xchk_ifork_btree_op_error(sc, cur, level, in __xchk_btree_process_error()
59 trace_xchk_btree_op_error(sc, cur, level, in __xchk_btree_process_error()
69 struct xfs_btree_cur *cur, in xchk_btree_process_error() argument
73 return __xchk_btree_process_error(sc, cur, level, error, in xchk_btree_process_error()
80 struct xfs_btree_cur *cur, in xchk_btree_xref_process_error() argument
84 return __xchk_btree_process_error(sc, cur, level, error, in xchk_btree_xref_process_error()
92 struct xfs_btree_cur *cur, in __xchk_btree_set_corrupt() argument
99 if (cur->bc_flags & XFS_BTREE_ROOT_IN_INODE) in __xchk_btree_set_corrupt()
[all …]
Dtrace.c26 struct xfs_btree_cur *cur, in xchk_btree_cur_fsbno() argument
29 if (level < cur->bc_nlevels && cur->bc_bufs[level]) in xchk_btree_cur_fsbno()
30 return XFS_DADDR_TO_FSB(cur->bc_mp, cur->bc_bufs[level]->b_bn); in xchk_btree_cur_fsbno()
31 else if (level == cur->bc_nlevels - 1 && in xchk_btree_cur_fsbno()
32 cur->bc_flags & XFS_BTREE_LONG_PTRS) in xchk_btree_cur_fsbno()
33 return XFS_INO_TO_FSB(cur->bc_mp, cur->bc_private.b.ip->i_ino); in xchk_btree_cur_fsbno()
34 else if (!(cur->bc_flags & XFS_BTREE_LONG_PTRS)) in xchk_btree_cur_fsbno()
35 return XFS_AGB_TO_FSB(cur->bc_mp, cur->bc_private.a.agno, 0); in xchk_btree_cur_fsbno()
Dialloc.c105 struct xfs_mount *mp = bs->cur->bc_mp; in xchk_iallocbt_chunk()
106 xfs_agnumber_t agno = bs->cur->bc_private.a.agno; in xchk_iallocbt_chunk()
113 xchk_btree_set_corrupt(bs->sc, bs->cur, 0); in xchk_iallocbt_chunk()
140 struct xfs_mount *mp = bs->cur->bc_mp; in xchk_iallocbt_check_cluster_freemask()
153 xchk_btree_set_corrupt(bs->sc, bs->cur, 0); in xchk_iallocbt_check_cluster_freemask()
159 error = xfs_icache_inode_is_allocated(mp, bs->cur->bc_tp, in xchk_iallocbt_check_cluster_freemask()
178 xchk_btree_set_corrupt(bs->sc, bs->cur, 0); in xchk_iallocbt_check_cluster_freemask()
191 struct xfs_mount *mp = bs->cur->bc_mp; in xchk_iallocbt_check_freemask()
213 fsino = XFS_AGINO_TO_INO(mp, bs->cur->bc_private.a.agno, agino); in xchk_iallocbt_check_freemask()
226 xchk_btree_set_corrupt(bs->sc, bs->cur, 0); in xchk_iallocbt_check_freemask()
[all …]
/linux-4.19.296/drivers/char/
Dbsr.c168 struct bsr_dev *cur, *n; in bsr_cleanup_devs() local
170 list_for_each_entry_safe(cur, n, &bsr_devs, bsr_list) { in bsr_cleanup_devs()
171 if (cur->bsr_device) { in bsr_cleanup_devs()
172 cdev_del(&cur->bsr_cdev); in bsr_cleanup_devs()
173 device_del(cur->bsr_device); in bsr_cleanup_devs()
175 list_del(&cur->bsr_list); in bsr_cleanup_devs()
176 kfree(cur); in bsr_cleanup_devs()
200 struct bsr_dev *cur = kzalloc(sizeof(struct bsr_dev), in bsr_add_node() local
205 if (!cur) { in bsr_add_node()
214 kfree(cur); in bsr_add_node()
[all …]
/linux-4.19.296/lib/zstd/
Dzstd_opt.h440 U32 cur, match_num, last_pos, litlen, price; in ZSTD_compressBlock_opt_generic() local
457 cur = 0; in ZSTD_compressBlock_opt_generic()
482 cur = 0; in ZSTD_compressBlock_opt_generic()
515 for (cur = 1; cur <= last_pos; cur++) { in ZSTD_compressBlock_opt_generic()
516 inr = ip + cur; in ZSTD_compressBlock_opt_generic()
518 if (opt[cur - 1].mlen == 1) { in ZSTD_compressBlock_opt_generic()
519 litlen = opt[cur - 1].litlen + 1; in ZSTD_compressBlock_opt_generic()
520 if (cur > litlen) { in ZSTD_compressBlock_opt_generic()
521 price = opt[cur - litlen].price + ZSTD_getLiteralPrice(seqStorePtr, litlen, inr - litlen); in ZSTD_compressBlock_opt_generic()
526 price = opt[cur - 1].price + ZSTD_getLiteralPrice(seqStorePtr, litlen, inr - 1); in ZSTD_compressBlock_opt_generic()
[all …]
/linux-4.19.296/include/linux/
Diversion.h176 u64 cur, old, new; in inode_maybe_inc_iversion() local
189 cur = inode_peek_iversion_raw(inode); in inode_maybe_inc_iversion()
192 if (!force && !(cur & I_VERSION_QUERIED)) in inode_maybe_inc_iversion()
196 new = (cur & ~I_VERSION_QUERIED) + I_VERSION_INCREMENT; in inode_maybe_inc_iversion()
198 old = atomic64_cmpxchg(&inode->i_version, cur, new); in inode_maybe_inc_iversion()
199 if (likely(old == cur)) in inode_maybe_inc_iversion()
201 cur = old; in inode_maybe_inc_iversion()
283 u64 cur, old, new; in inode_query_iversion() local
285 cur = inode_peek_iversion_raw(inode); in inode_query_iversion()
288 if (cur & I_VERSION_QUERIED) { in inode_query_iversion()
[all …]
/linux-4.19.296/drivers/char/agp/
Disoch.c25 struct agp_3_5_dev *cur, *n = list_entry(new, struct agp_3_5_dev, list); in agp_3_5_dev_list_insert() local
29 cur = list_entry(pos, struct agp_3_5_dev, list); in agp_3_5_dev_list_insert()
30 if (cur->maxbw > n->maxbw) in agp_3_5_dev_list_insert()
38 struct agp_3_5_dev *cur; in agp_3_5_dev_list_sort() local
46 cur = list_entry(pos, struct agp_3_5_dev, list); in agp_3_5_dev_list_sort()
47 dev = cur->dev; in agp_3_5_dev_list_sort()
49 pci_read_config_dword(dev, cur->capndx+AGPNISTAT, &nistat); in agp_3_5_dev_list_sort()
50 cur->maxbw = (nistat >> 16) & 0xff; in agp_3_5_dev_list_sort()
82 struct agp_3_5_dev *cur; in agp_3_5_isochronous_node_enable() local
138 cur = list_entry(pos, struct agp_3_5_dev, list); in agp_3_5_isochronous_node_enable()
[all …]
/linux-4.19.296/drivers/misc/ocxl/
Dpasid.c15 struct id_range *cur; in dump_list() local
18 list_for_each_entry(cur, head, list) { in dump_list()
19 pr_debug("Range %d->%d\n", cur->start, cur->end); in dump_list()
28 struct id_range *cur, *new; in range_alloc() local
37 list_for_each_entry(cur, head, list) { in range_alloc()
38 if ((cur->start - last_end) > size) in range_alloc()
40 last_end = cur->end; in range_alloc()
41 pos = &cur->list; in range_alloc()
65 struct id_range *cur, *tmp; in range_free() local
67 list_for_each_entry_safe(cur, tmp, head, list) { in range_free()
[all …]
/linux-4.19.296/drivers/clk/rockchip/
Dclk-pll.c165 struct rockchip_pll_rate_table cur; in rockchip_rk3036_pll_recalc_rate() local
168 rockchip_rk3036_pll_get_params(pll, &cur); in rockchip_rk3036_pll_recalc_rate()
170 rate64 *= cur.fbdiv; in rockchip_rk3036_pll_recalc_rate()
171 do_div(rate64, cur.refdiv); in rockchip_rk3036_pll_recalc_rate()
173 if (cur.dsmpd == 0) { in rockchip_rk3036_pll_recalc_rate()
175 u64 frac_rate64 = prate * cur.frac; in rockchip_rk3036_pll_recalc_rate()
177 do_div(frac_rate64, cur.refdiv); in rockchip_rk3036_pll_recalc_rate()
181 do_div(rate64, cur.postdiv1); in rockchip_rk3036_pll_recalc_rate()
182 do_div(rate64, cur.postdiv2); in rockchip_rk3036_pll_recalc_rate()
192 struct rockchip_pll_rate_table cur; in rockchip_rk3036_pll_set_params() local
[all …]
/linux-4.19.296/lib/
Derrseq.c60 errseq_t cur, old; in errseq_set() local
89 cur = new; in errseq_set()
94 cur = cmpxchg(eseq, old, new); in errseq_set()
100 if (likely(cur == old || cur == new)) in errseq_set()
104 old = cur; in errseq_set()
106 return cur; in errseq_set()
146 errseq_t cur = READ_ONCE(*eseq); in errseq_check() local
148 if (likely(cur == since)) in errseq_check()
150 return -(cur & MAX_ERRNO); in errseq_check()
Dtest_list_sort.c73 struct list_head *cur; in list_sort_test() local
99 for (cur = head.next; cur->next != &head; cur = cur->next) { in list_sort_test()
103 if (cur->next->prev != cur) { in list_sort_test()
108 cmp_result = cmp(NULL, cur, cur->next); in list_sort_test()
114 el = container_of(cur, struct debug_el, list); in list_sort_test()
115 el1 = container_of(cur->next, struct debug_el, list); in list_sort_test()
128 if (head.prev != cur) { in list_sort_test()
/linux-4.19.296/drivers/misc/vmw_vmci/
Dvmci_event.c57 struct vmci_subscription *cur, *p2; in vmci_event_exit() local
58 list_for_each_entry_safe(cur, p2, &subscriber_array[e], node) { in vmci_event_exit()
66 list_del(&cur->node); in vmci_event_exit()
67 kfree(cur); in vmci_event_exit()
80 struct vmci_subscription *cur; in event_find() local
81 list_for_each_entry(cur, &subscriber_array[e], node) { in event_find()
82 if (cur->id == sub_id) in event_find()
83 return cur; in event_find()
95 struct vmci_subscription *cur; in event_deliver() local
100 list_for_each_entry_rcu(cur, subscriber_list, node) { in event_deliver()
[all …]
/linux-4.19.296/fs/xfs/
Dxfs_trans_ail.c155 struct xfs_ail_cursor *cur) in xfs_trans_ail_cursor_init() argument
157 cur->item = NULL; in xfs_trans_ail_cursor_init()
158 list_add_tail(&cur->list, &ailp->ail_cursors); in xfs_trans_ail_cursor_init()
168 struct xfs_ail_cursor *cur) in xfs_trans_ail_cursor_next() argument
170 struct xfs_log_item *lip = cur->item; in xfs_trans_ail_cursor_next()
175 cur->item = xfs_ail_next(ailp, lip); in xfs_trans_ail_cursor_next()
185 struct xfs_ail_cursor *cur) in xfs_trans_ail_cursor_done() argument
187 cur->item = NULL; in xfs_trans_ail_cursor_done()
188 list_del_init(&cur->list); in xfs_trans_ail_cursor_done()
204 struct xfs_ail_cursor *cur; in xfs_trans_ail_cursor_clear() local
[all …]

123456789