Home
last modified time | relevance | path

Searched refs:sc (Results 1 – 25 of 83) sorted by relevance

1234

/linux-4.19.296/fs/xfs/scrub/
Dagheader.c32 struct xfs_scrub *sc, in xchk_superblock_xref() argument
36 struct xfs_mount *mp = sc->mp; in xchk_superblock_xref()
37 xfs_agnumber_t agno = sc->sm->sm_agno; in xchk_superblock_xref()
41 if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT) in xchk_superblock_xref()
46 error = xchk_ag_init(sc, agno, &sc->sa); in xchk_superblock_xref()
47 if (!xchk_xref_process_error(sc, agno, agbno, &error)) in xchk_superblock_xref()
50 xchk_xref_is_used_space(sc, agbno, 1); in xchk_superblock_xref()
51 xchk_xref_is_not_inode_chunk(sc, agbno, 1); in xchk_superblock_xref()
53 xchk_xref_is_owned_by(sc, agbno, 1, &oinfo); in xchk_superblock_xref()
54 xchk_xref_is_not_shared(sc, agbno, 1); in xchk_superblock_xref()
[all …]
Dcommon.c72 struct xfs_scrub *sc, in __xchk_process_error() argument
84 trace_xchk_deadlock_retry(sc->ip, sc->sm, *error); in __xchk_process_error()
89 sc->sm->sm_flags |= errflag; in __xchk_process_error()
93 trace_xchk_op_error(sc, agno, bno, *error, in __xchk_process_error()
102 struct xfs_scrub *sc, in xchk_process_error() argument
107 return __xchk_process_error(sc, agno, bno, error, in xchk_process_error()
113 struct xfs_scrub *sc, in xchk_xref_process_error() argument
118 return __xchk_process_error(sc, agno, bno, error, in xchk_xref_process_error()
125 struct xfs_scrub *sc, in __xchk_fblock_process_error() argument
137 trace_xchk_deadlock_retry(sc->ip, sc->sm, *error); in __xchk_fblock_process_error()
[all …]
Dinode.c41 struct xfs_scrub *sc, in xchk_setup_inode() argument
50 error = xchk_get_inode(sc, ip); in xchk_setup_inode()
56 return xchk_trans_alloc(sc, 0); in xchk_setup_inode()
62 sc->ilock_flags = XFS_IOLOCK_EXCL | XFS_MMAPLOCK_EXCL; in xchk_setup_inode()
63 xfs_ilock(sc->ip, sc->ilock_flags); in xchk_setup_inode()
64 error = xchk_trans_alloc(sc, 0); in xchk_setup_inode()
67 sc->ilock_flags |= XFS_ILOCK_EXCL; in xchk_setup_inode()
68 xfs_ilock(sc->ip, XFS_ILOCK_EXCL); in xchk_setup_inode()
80 struct xfs_scrub *sc, in xchk_inode_extsize() argument
88 fa = xfs_inode_validate_extsize(sc->mp, be32_to_cpu(dip->di_extsize), in xchk_inode_extsize()
[all …]
Drtbitmap.c29 struct xfs_scrub *sc, in xchk_setup_rt() argument
34 error = xchk_setup_fs(sc, ip); in xchk_setup_rt()
38 sc->ilock_flags = XFS_ILOCK_EXCL | XFS_ILOCK_RTBITMAP; in xchk_setup_rt()
39 sc->ip = sc->mp->m_rbmip; in xchk_setup_rt()
40 xfs_ilock(sc->ip, sc->ilock_flags); in xchk_setup_rt()
54 struct xfs_scrub *sc = priv; in xchk_rtbitmap_rec() local
62 !xfs_verify_rtbno(sc->mp, startblock) || in xchk_rtbitmap_rec()
63 !xfs_verify_rtbno(sc->mp, startblock + blockcount - 1)) in xchk_rtbitmap_rec()
64 xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, 0); in xchk_rtbitmap_rec()
71 struct xfs_scrub *sc) in xchk_rtbitmap() argument
[all …]
Drepair.c47 struct xfs_scrub *sc, in xrep_attempt() argument
52 trace_xrep_attempt(ip, sc->sm, error); in xrep_attempt()
54 xchk_ag_btcur_free(&sc->sa); in xrep_attempt()
57 ASSERT(sc->ops->repair); in xrep_attempt()
58 error = sc->ops->repair(sc); in xrep_attempt()
59 trace_xrep_done(ip, sc->sm, error); in xrep_attempt()
66 sc->sm->sm_flags &= ~XFS_SCRUB_FLAGS_OUT; in xrep_attempt()
72 if (!sc->try_harder) { in xrep_attempt()
73 sc->try_harder = true; in xrep_attempt()
110 struct xfs_scrub *sc) in xrep_probe() argument
[all …]
Dbmap.c37 struct xfs_scrub *sc, in xchk_setup_inode_bmap() argument
42 error = xchk_get_inode(sc, ip); in xchk_setup_inode_bmap()
46 sc->ilock_flags = XFS_IOLOCK_EXCL | XFS_MMAPLOCK_EXCL; in xchk_setup_inode_bmap()
47 xfs_ilock(sc->ip, sc->ilock_flags); in xchk_setup_inode_bmap()
54 if (S_ISREG(VFS_I(sc->ip)->i_mode) && in xchk_setup_inode_bmap()
55 sc->sm->sm_type == XFS_SCRUB_TYPE_BMBTD) { in xchk_setup_inode_bmap()
56 struct address_space *mapping = VFS_I(sc->ip)->i_mapping; in xchk_setup_inode_bmap()
58 inode_dio_wait(VFS_I(sc->ip)); in xchk_setup_inode_bmap()
81 error = xchk_trans_alloc(sc, 0); in xchk_setup_inode_bmap()
84 sc->ilock_flags |= XFS_ILOCK_EXCL; in xchk_setup_inode_bmap()
[all …]
Dcommon.h16 struct xfs_scrub *sc, in xchk_should_terminate() argument
34 int xchk_trans_alloc(struct xfs_scrub *sc, uint resblks);
35 bool xchk_process_error(struct xfs_scrub *sc, xfs_agnumber_t agno,
37 bool xchk_fblock_process_error(struct xfs_scrub *sc, int whichfork,
40 bool xchk_xref_process_error(struct xfs_scrub *sc,
42 bool xchk_fblock_xref_process_error(struct xfs_scrub *sc,
45 void xchk_block_set_preen(struct xfs_scrub *sc,
47 void xchk_ino_set_preen(struct xfs_scrub *sc, xfs_ino_t ino);
49 void xchk_block_set_corrupt(struct xfs_scrub *sc,
51 void xchk_ino_set_corrupt(struct xfs_scrub *sc, xfs_ino_t ino);
[all …]
Dquota.c34 struct xfs_scrub *sc) in xchk_quota_to_dqtype() argument
36 switch (sc->sm->sm_type) { in xchk_quota_to_dqtype()
51 struct xfs_scrub *sc, in xchk_setup_quota() argument
57 if (!XFS_IS_QUOTA_RUNNING(sc->mp) || !XFS_IS_QUOTA_ON(sc->mp)) in xchk_setup_quota()
60 dqtype = xchk_quota_to_dqtype(sc); in xchk_setup_quota()
63 sc->has_quotaofflock = true; in xchk_setup_quota()
64 mutex_lock(&sc->mp->m_quotainfo->qi_quotaofflock); in xchk_setup_quota()
65 if (!xfs_this_quota_on(sc->mp, dqtype)) in xchk_setup_quota()
67 error = xchk_setup_fs(sc, ip); in xchk_setup_quota()
70 sc->ip = xfs_quota_inode(sc->mp, dqtype); in xchk_setup_quota()
[all …]
Dscrub.h73 int xchk_tester(struct xfs_scrub *sc);
74 int xchk_superblock(struct xfs_scrub *sc);
75 int xchk_agf(struct xfs_scrub *sc);
76 int xchk_agfl(struct xfs_scrub *sc);
77 int xchk_agi(struct xfs_scrub *sc);
78 int xchk_bnobt(struct xfs_scrub *sc);
79 int xchk_cntbt(struct xfs_scrub *sc);
80 int xchk_inobt(struct xfs_scrub *sc);
81 int xchk_finobt(struct xfs_scrub *sc);
82 int xchk_rmapbt(struct xfs_scrub *sc);
[all …]
Drmap.c33 struct xfs_scrub *sc, in xchk_setup_ag_rmapbt() argument
36 return xchk_setup_ag_btree(sc, ip, false); in xchk_setup_ag_rmapbt()
44 struct xfs_scrub *sc, in xchk_rmapbt_xref_refc() argument
55 if (!sc->sa.refc_cur || xchk_skip_xref(sc->sm)) in xchk_rmapbt_xref_refc()
64 error = xfs_refcount_find_shared(sc->sa.refc_cur, irec->rm_startblock, in xchk_rmapbt_xref_refc()
66 if (!xchk_should_check_xref(sc, &error, &sc->sa.refc_cur)) in xchk_rmapbt_xref_refc()
69 xchk_btree_xref_set_corrupt(sc, sc->sa.refc_cur, 0); in xchk_rmapbt_xref_refc()
75 struct xfs_scrub *sc, in xchk_rmapbt_xref() argument
81 if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT) in xchk_rmapbt_xref()
84 xchk_xref_is_used_space(sc, agbno, len); in xchk_rmapbt_xref()
[all …]
Dparent.c31 struct xfs_scrub *sc, in xchk_setup_parent() argument
34 return xchk_setup_inode_contents(sc, ip, 0); in xchk_setup_parent()
68 struct xfs_scrub *sc, in xchk_parent_count_parent_dentries() argument
75 .ino = sc->ip->i_ino, in xchk_parent_count_parent_dentries()
105 error = xfs_readdir(sc->tp, parent, &spc.dc, bufsize); in xchk_parent_count_parent_dentries()
124 struct xfs_scrub *sc, in xchk_parent_validate() argument
128 struct xfs_mount *mp = sc->mp; in xchk_parent_validate()
136 if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT) in xchk_parent_validate()
140 if (sc->ip->i_ino == dnum) { in xchk_parent_validate()
141 xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, 0); in xchk_parent_validate()
[all …]
Dalloc.c31 struct xfs_scrub *sc, in xchk_setup_ag_allocbt() argument
34 return xchk_setup_ag_btree(sc, ip, false); in xchk_setup_ag_allocbt()
44 struct xfs_scrub *sc, in xchk_allocbt_xref_other() argument
54 if (sc->sm->sm_type == XFS_SCRUB_TYPE_BNOBT) in xchk_allocbt_xref_other()
55 pcur = &sc->sa.cnt_cur; in xchk_allocbt_xref_other()
57 pcur = &sc->sa.bno_cur; in xchk_allocbt_xref_other()
58 if (!*pcur || xchk_skip_xref(sc->sm)) in xchk_allocbt_xref_other()
62 if (!xchk_should_check_xref(sc, &error, pcur)) in xchk_allocbt_xref_other()
65 xchk_btree_xref_set_corrupt(sc, *pcur, 0); in xchk_allocbt_xref_other()
70 if (!xchk_should_check_xref(sc, &error, pcur)) in xchk_allocbt_xref_other()
[all …]
Dagheader_repair.c39 struct xfs_scrub *sc) in xrep_superblock() argument
41 struct xfs_mount *mp = sc->mp; in xrep_superblock()
47 agno = sc->sm->sm_agno; in xrep_superblock()
51 error = xfs_sb_get_secondary(mp, sc->tp, agno, &bp); in xrep_superblock()
60 xfs_trans_buf_set_type(sc->tp, bp, XFS_BLFT_SB_BUF); in xrep_superblock()
61 xfs_trans_log_buf(sc->tp, bp, 0, BBTOB(bp->b_length) - 1); in xrep_superblock()
68 struct xfs_scrub *sc; member
83 if (xchk_should_terminate(raa->sc, &error)) in xrep_agf_walk_allocbt()
99 struct xfs_scrub *sc = priv; in xrep_agf_check_agfl_block() local
101 if (!xfs_verify_agbno(mp, sc->sa.agno, agbno)) in xrep_agf_check_agfl_block()
[all …]
Drefcount.c32 struct xfs_scrub *sc, in xchk_setup_ag_refcountbt() argument
35 return xchk_setup_ag_btree(sc, ip, false); in xchk_setup_ag_refcountbt()
82 struct xfs_scrub *sc; member
113 if (xchk_should_terminate(refchk->sc, &error)) in xchk_refcountbt_rmap_check()
121 xchk_btree_xref_set_corrupt(refchk->sc, cur, 0); in xchk_refcountbt_rmap_check()
279 struct xfs_scrub *sc, in xchk_refcountbt_xref_rmap() argument
285 .sc = sc, in xchk_refcountbt_xref_rmap()
297 if (!sc->sa.rmap_cur || xchk_skip_xref(sc->sm)) in xchk_refcountbt_xref_rmap()
307 error = xfs_rmap_query_range(sc->sa.rmap_cur, &low, &high, in xchk_refcountbt_xref_rmap()
309 if (!xchk_should_check_xref(sc, &error, &sc->sa.rmap_cur)) in xchk_refcountbt_xref_rmap()
[all …]
Ddir.c35 struct xfs_scrub *sc, in xchk_setup_directory() argument
38 return xchk_setup_inode_contents(sc, ip, 0); in xchk_setup_directory()
49 struct xfs_scrub *sc; member
60 struct xfs_mount *mp = sdc->sc->mp; in xchk_dir_check_ftype()
67 xchk_fblock_set_corrupt(sdc->sc, XFS_DATA_FORK, in xchk_dir_check_ftype()
80 error = xfs_iget(mp, sdc->sc->tp, inum, 0, 0, &ip); in xchk_dir_check_ftype()
81 if (!xchk_fblock_xref_process_error(sdc->sc, XFS_DATA_FORK, offset, in xchk_dir_check_ftype()
89 xchk_fblock_set_corrupt(sdc->sc, XFS_DATA_FORK, offset); in xchk_dir_check_ftype()
121 ip = sdc->sc->ip; in xchk_dir_actor()
128 xchk_fblock_set_corrupt(sdc->sc, XFS_DATA_FORK, offset); in xchk_dir_actor()
[all …]
Dialloc.c39 struct xfs_scrub *sc, in xchk_setup_ag_iallocbt() argument
42 return xchk_setup_ag_btree(sc, ip, sc->try_harder); in xchk_setup_ag_iallocbt()
54 struct xfs_scrub *sc, in xchk_iallocbt_chunk_xref_other() argument
62 if (sc->sm->sm_type == XFS_SCRUB_TYPE_FINOBT) in xchk_iallocbt_chunk_xref_other()
63 pcur = &sc->sa.ino_cur; in xchk_iallocbt_chunk_xref_other()
65 pcur = &sc->sa.fino_cur; in xchk_iallocbt_chunk_xref_other()
69 if (!xchk_should_check_xref(sc, &error, pcur)) in xchk_iallocbt_chunk_xref_other()
73 xchk_btree_xref_set_corrupt(sc, *pcur, 0); in xchk_iallocbt_chunk_xref_other()
79 struct xfs_scrub *sc, in xchk_iallocbt_chunk_xref() argument
87 if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT) in xchk_iallocbt_chunk_xref()
[all …]
Dscrub.c154 struct xfs_scrub *sc) in xchk_probe() argument
158 if (xchk_should_terminate(sc, &error)) in xchk_probe()
169 struct xfs_scrub *sc, in xchk_teardown() argument
173 xchk_ag_free(sc, &sc->sa); in xchk_teardown()
174 if (sc->tp) { in xchk_teardown()
175 if (error == 0 && (sc->sm->sm_flags & XFS_SCRUB_IFLAG_REPAIR)) in xchk_teardown()
176 error = xfs_trans_commit(sc->tp); in xchk_teardown()
178 xfs_trans_cancel(sc->tp); in xchk_teardown()
179 sc->tp = NULL; in xchk_teardown()
181 if (sc->ip) { in xchk_teardown()
[all …]
Drepair.h9 static inline int xrep_notsupported(struct xfs_scrub *sc) in xrep_notsupported() argument
18 int xrep_attempt(struct xfs_inode *ip, struct xfs_scrub *sc, bool *fixed);
20 int xrep_roll_ag_trans(struct xfs_scrub *sc);
23 xfs_extlen_t xrep_calc_ag_resblks(struct xfs_scrub *sc);
24 int xrep_alloc_ag_block(struct xfs_scrub *sc, struct xfs_owner_info *oinfo,
26 int xrep_init_btblock(struct xfs_scrub *sc, xfs_fsblock_t fsb,
32 int xrep_fix_freelist(struct xfs_scrub *sc, bool can_shrink);
33 int xrep_invalidate_blocks(struct xfs_scrub *sc, struct xfs_bitmap *btlist);
34 int xrep_reap_extents(struct xfs_scrub *sc, struct xfs_bitmap *exlist,
52 int xrep_find_ag_btree_roots(struct xfs_scrub *sc, struct xfs_buf *agf_bp,
[all …]
Dsymlink.c29 struct xfs_scrub *sc, in xchk_setup_symlink() argument
33 sc->buf = kmem_zalloc_large(XFS_SYMLINK_MAXLEN + 1, KM_SLEEP); in xchk_setup_symlink()
34 if (!sc->buf) in xchk_setup_symlink()
37 return xchk_setup_inode_contents(sc, ip, 0); in xchk_setup_symlink()
44 struct xfs_scrub *sc) in xchk_symlink() argument
46 struct xfs_inode *ip = sc->ip; in xchk_symlink()
58 xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, 0); in xchk_symlink()
66 xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, 0); in xchk_symlink()
71 error = xfs_readlink_bmap_ilocked(sc->ip, sc->buf); in xchk_symlink()
72 if (!xchk_fblock_process_error(sc, XFS_DATA_FORK, 0, &error)) in xchk_symlink()
[all …]
Dbtree.c33 struct xfs_scrub *sc, in __xchk_btree_process_error() argument
46 trace_xchk_deadlock_retry(sc->ip, sc->sm, *error); in __xchk_btree_process_error()
51 sc->sm->sm_flags |= errflag; in __xchk_btree_process_error()
56 trace_xchk_ifork_btree_op_error(sc, cur, level, in __xchk_btree_process_error()
59 trace_xchk_btree_op_error(sc, cur, level, in __xchk_btree_process_error()
68 struct xfs_scrub *sc, in xchk_btree_process_error() argument
73 return __xchk_btree_process_error(sc, cur, level, error, in xchk_btree_process_error()
79 struct xfs_scrub *sc, in xchk_btree_xref_process_error() argument
84 return __xchk_btree_process_error(sc, cur, level, error, in xchk_btree_xref_process_error()
91 struct xfs_scrub *sc, in __xchk_btree_set_corrupt() argument
[all …]
Dattr.c36 struct xfs_scrub *sc, in xchk_setup_xattr() argument
48 BITS_TO_LONGS(sc->mp->m_attr_geo->blksize)); in xchk_setup_xattr()
49 sc->buf = kmem_zalloc_large(sz, KM_SLEEP); in xchk_setup_xattr()
50 if (!sc->buf) in xchk_setup_xattr()
53 return xchk_setup_inode_contents(sc, ip, 0); in xchk_setup_xattr()
60 struct xfs_scrub *sc; member
87 xchk_ino_set_preen(sx->sc, context->dp->i_ino); in xchk_xattr_listent()
103 args.value = sx->sc->buf; in xchk_xattr_listent()
109 if (!xchk_fblock_process_error(sx->sc, XFS_ATTR_FORK, args.blkno, in xchk_xattr_listent()
113 xchk_fblock_set_corrupt(sx->sc, XFS_ATTR_FORK, in xchk_xattr_listent()
[all …]
/linux-4.19.296/drivers/clk/qcom/
Dgdsc.c59 static int gdsc_check_status(struct gdsc *sc, enum gdsc_status status) in gdsc_check_status() argument
65 if (sc->flags & POLL_CFG_GDSCR) in gdsc_check_status()
66 reg = sc->gdscr + CFG_GDSCR_OFFSET; in gdsc_check_status()
67 else if (sc->gds_hw_ctrl) in gdsc_check_status()
68 reg = sc->gds_hw_ctrl; in gdsc_check_status()
70 reg = sc->gdscr; in gdsc_check_status()
72 ret = regmap_read(sc->regmap, reg, &val); in gdsc_check_status()
76 if (sc->flags & POLL_CFG_GDSCR) { in gdsc_check_status()
95 static int gdsc_hwctrl(struct gdsc *sc, bool en) in gdsc_hwctrl() argument
99 return regmap_update_bits(sc->regmap, sc->gdscr, HW_CONTROL_MASK, val); in gdsc_hwctrl()
[all …]
/linux-4.19.296/drivers/hid/
Dhid-sony.c585 static void sony_set_leds(struct sony_sc *sc);
587 static inline void sony_schedule_work(struct sony_sc *sc, in sony_schedule_work() argument
594 spin_lock_irqsave(&sc->lock, flags); in sony_schedule_work()
595 if (!sc->defer_initialization && sc->state_worker_initialized) in sony_schedule_work()
596 schedule_work(&sc->state_worker); in sony_schedule_work()
597 spin_unlock_irqrestore(&sc->lock, flags); in sony_schedule_work()
600 if (sc->hotplug_worker_initialized) in sony_schedule_work()
601 schedule_work(&sc->hotplug_worker); in sony_schedule_work()
611 struct sony_sc *sc = hid_get_drvdata(hdev); in ds4_show_poll_interval() local
613 return snprintf(buf, PAGE_SIZE, "%i\n", sc->ds4_bt_poll_interval); in ds4_show_poll_interval()
[all …]
/linux-4.19.296/fs/ocfs2/cluster/
Dtcp.c78 #define SC_NODEF_ARGS(sc) sc->sc_node->nd_name, sc->sc_node->nd_num, \ argument
79 &sc->sc_node->nd_ipv4_address, \
80 ntohs(sc->sc_node->nd_ipv4_port)
97 #define sclog(sc, fmt, args...) do { \ argument
98 typeof(sc) __sc = (sc); \
144 static void o2net_sc_postpone_idle(struct o2net_sock_container *sc);
145 static void o2net_sc_reset_idle_timer(struct o2net_sock_container *sc);
174 struct o2net_sock_container *sc) in o2net_set_nst_sock_container() argument
176 nst->st_sc = sc; in o2net_set_nst_sock_container()
185 static inline void o2net_set_sock_timer(struct o2net_sock_container *sc) in o2net_set_sock_timer() argument
[all …]
Dnetdebug.c213 void o2net_debug_add_sc(struct o2net_sock_container *sc) in o2net_debug_add_sc() argument
216 list_add(&sc->sc_net_debug_item, &sock_containers); in o2net_debug_add_sc()
220 void o2net_debug_del_sc(struct o2net_sock_container *sc) in o2net_debug_del_sc() argument
223 list_del_init(&sc->sc_net_debug_item); in o2net_debug_del_sc()
235 struct o2net_sock_container *sc, *ret = NULL; in next_sc() local
239 list_for_each_entry(sc, &sc_start->sc_net_debug_item, in next_sc()
242 if (&sc->sc_net_debug_item == &sock_containers) in next_sc()
246 if (sc->sc_page != NULL) { in next_sc()
247 ret = sc; in next_sc()
258 struct o2net_sock_container *sc, *dummy_sc = sd->dbg_sock; in sc_seq_start() local
[all …]

1234