/linux-4.19.296/fs/xfs/scrub/ |
D | agheader.c | 32 struct xfs_scrub *sc, in xchk_superblock_xref() argument 36 struct xfs_mount *mp = sc->mp; in xchk_superblock_xref() 37 xfs_agnumber_t agno = sc->sm->sm_agno; in xchk_superblock_xref() 41 if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT) in xchk_superblock_xref() 46 error = xchk_ag_init(sc, agno, &sc->sa); in xchk_superblock_xref() 47 if (!xchk_xref_process_error(sc, agno, agbno, &error)) in xchk_superblock_xref() 50 xchk_xref_is_used_space(sc, agbno, 1); in xchk_superblock_xref() 51 xchk_xref_is_not_inode_chunk(sc, agbno, 1); in xchk_superblock_xref() 53 xchk_xref_is_owned_by(sc, agbno, 1, &oinfo); in xchk_superblock_xref() 54 xchk_xref_is_not_shared(sc, agbno, 1); in xchk_superblock_xref() [all …]
|
D | common.c | 72 struct xfs_scrub *sc, in __xchk_process_error() argument 84 trace_xchk_deadlock_retry(sc->ip, sc->sm, *error); in __xchk_process_error() 89 sc->sm->sm_flags |= errflag; in __xchk_process_error() 93 trace_xchk_op_error(sc, agno, bno, *error, in __xchk_process_error() 102 struct xfs_scrub *sc, in xchk_process_error() argument 107 return __xchk_process_error(sc, agno, bno, error, in xchk_process_error() 113 struct xfs_scrub *sc, in xchk_xref_process_error() argument 118 return __xchk_process_error(sc, agno, bno, error, in xchk_xref_process_error() 125 struct xfs_scrub *sc, in __xchk_fblock_process_error() argument 137 trace_xchk_deadlock_retry(sc->ip, sc->sm, *error); in __xchk_fblock_process_error() [all …]
|
D | inode.c | 41 struct xfs_scrub *sc, in xchk_setup_inode() argument 50 error = xchk_get_inode(sc, ip); in xchk_setup_inode() 56 return xchk_trans_alloc(sc, 0); in xchk_setup_inode() 62 sc->ilock_flags = XFS_IOLOCK_EXCL | XFS_MMAPLOCK_EXCL; in xchk_setup_inode() 63 xfs_ilock(sc->ip, sc->ilock_flags); in xchk_setup_inode() 64 error = xchk_trans_alloc(sc, 0); in xchk_setup_inode() 67 sc->ilock_flags |= XFS_ILOCK_EXCL; in xchk_setup_inode() 68 xfs_ilock(sc->ip, XFS_ILOCK_EXCL); in xchk_setup_inode() 80 struct xfs_scrub *sc, in xchk_inode_extsize() argument 88 fa = xfs_inode_validate_extsize(sc->mp, be32_to_cpu(dip->di_extsize), in xchk_inode_extsize() [all …]
|
D | rtbitmap.c | 29 struct xfs_scrub *sc, in xchk_setup_rt() argument 34 error = xchk_setup_fs(sc, ip); in xchk_setup_rt() 38 sc->ilock_flags = XFS_ILOCK_EXCL | XFS_ILOCK_RTBITMAP; in xchk_setup_rt() 39 sc->ip = sc->mp->m_rbmip; in xchk_setup_rt() 40 xfs_ilock(sc->ip, sc->ilock_flags); in xchk_setup_rt() 54 struct xfs_scrub *sc = priv; in xchk_rtbitmap_rec() local 62 !xfs_verify_rtbno(sc->mp, startblock) || in xchk_rtbitmap_rec() 63 !xfs_verify_rtbno(sc->mp, startblock + blockcount - 1)) in xchk_rtbitmap_rec() 64 xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, 0); in xchk_rtbitmap_rec() 71 struct xfs_scrub *sc) in xchk_rtbitmap() argument [all …]
|
D | repair.c | 47 struct xfs_scrub *sc, in xrep_attempt() argument 52 trace_xrep_attempt(ip, sc->sm, error); in xrep_attempt() 54 xchk_ag_btcur_free(&sc->sa); in xrep_attempt() 57 ASSERT(sc->ops->repair); in xrep_attempt() 58 error = sc->ops->repair(sc); in xrep_attempt() 59 trace_xrep_done(ip, sc->sm, error); in xrep_attempt() 66 sc->sm->sm_flags &= ~XFS_SCRUB_FLAGS_OUT; in xrep_attempt() 72 if (!sc->try_harder) { in xrep_attempt() 73 sc->try_harder = true; in xrep_attempt() 110 struct xfs_scrub *sc) in xrep_probe() argument [all …]
|
D | bmap.c | 37 struct xfs_scrub *sc, in xchk_setup_inode_bmap() argument 42 error = xchk_get_inode(sc, ip); in xchk_setup_inode_bmap() 46 sc->ilock_flags = XFS_IOLOCK_EXCL | XFS_MMAPLOCK_EXCL; in xchk_setup_inode_bmap() 47 xfs_ilock(sc->ip, sc->ilock_flags); in xchk_setup_inode_bmap() 54 if (S_ISREG(VFS_I(sc->ip)->i_mode) && in xchk_setup_inode_bmap() 55 sc->sm->sm_type == XFS_SCRUB_TYPE_BMBTD) { in xchk_setup_inode_bmap() 56 struct address_space *mapping = VFS_I(sc->ip)->i_mapping; in xchk_setup_inode_bmap() 58 inode_dio_wait(VFS_I(sc->ip)); in xchk_setup_inode_bmap() 81 error = xchk_trans_alloc(sc, 0); in xchk_setup_inode_bmap() 84 sc->ilock_flags |= XFS_ILOCK_EXCL; in xchk_setup_inode_bmap() [all …]
|
D | common.h | 16 struct xfs_scrub *sc, in xchk_should_terminate() argument 34 int xchk_trans_alloc(struct xfs_scrub *sc, uint resblks); 35 bool xchk_process_error(struct xfs_scrub *sc, xfs_agnumber_t agno, 37 bool xchk_fblock_process_error(struct xfs_scrub *sc, int whichfork, 40 bool xchk_xref_process_error(struct xfs_scrub *sc, 42 bool xchk_fblock_xref_process_error(struct xfs_scrub *sc, 45 void xchk_block_set_preen(struct xfs_scrub *sc, 47 void xchk_ino_set_preen(struct xfs_scrub *sc, xfs_ino_t ino); 49 void xchk_block_set_corrupt(struct xfs_scrub *sc, 51 void xchk_ino_set_corrupt(struct xfs_scrub *sc, xfs_ino_t ino); [all …]
|
D | quota.c | 34 struct xfs_scrub *sc) in xchk_quota_to_dqtype() argument 36 switch (sc->sm->sm_type) { in xchk_quota_to_dqtype() 51 struct xfs_scrub *sc, in xchk_setup_quota() argument 57 if (!XFS_IS_QUOTA_RUNNING(sc->mp) || !XFS_IS_QUOTA_ON(sc->mp)) in xchk_setup_quota() 60 dqtype = xchk_quota_to_dqtype(sc); in xchk_setup_quota() 63 sc->has_quotaofflock = true; in xchk_setup_quota() 64 mutex_lock(&sc->mp->m_quotainfo->qi_quotaofflock); in xchk_setup_quota() 65 if (!xfs_this_quota_on(sc->mp, dqtype)) in xchk_setup_quota() 67 error = xchk_setup_fs(sc, ip); in xchk_setup_quota() 70 sc->ip = xfs_quota_inode(sc->mp, dqtype); in xchk_setup_quota() [all …]
|
D | scrub.h | 73 int xchk_tester(struct xfs_scrub *sc); 74 int xchk_superblock(struct xfs_scrub *sc); 75 int xchk_agf(struct xfs_scrub *sc); 76 int xchk_agfl(struct xfs_scrub *sc); 77 int xchk_agi(struct xfs_scrub *sc); 78 int xchk_bnobt(struct xfs_scrub *sc); 79 int xchk_cntbt(struct xfs_scrub *sc); 80 int xchk_inobt(struct xfs_scrub *sc); 81 int xchk_finobt(struct xfs_scrub *sc); 82 int xchk_rmapbt(struct xfs_scrub *sc); [all …]
|
D | rmap.c | 33 struct xfs_scrub *sc, in xchk_setup_ag_rmapbt() argument 36 return xchk_setup_ag_btree(sc, ip, false); in xchk_setup_ag_rmapbt() 44 struct xfs_scrub *sc, in xchk_rmapbt_xref_refc() argument 55 if (!sc->sa.refc_cur || xchk_skip_xref(sc->sm)) in xchk_rmapbt_xref_refc() 64 error = xfs_refcount_find_shared(sc->sa.refc_cur, irec->rm_startblock, in xchk_rmapbt_xref_refc() 66 if (!xchk_should_check_xref(sc, &error, &sc->sa.refc_cur)) in xchk_rmapbt_xref_refc() 69 xchk_btree_xref_set_corrupt(sc, sc->sa.refc_cur, 0); in xchk_rmapbt_xref_refc() 75 struct xfs_scrub *sc, in xchk_rmapbt_xref() argument 81 if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT) in xchk_rmapbt_xref() 84 xchk_xref_is_used_space(sc, agbno, len); in xchk_rmapbt_xref() [all …]
|
D | parent.c | 31 struct xfs_scrub *sc, in xchk_setup_parent() argument 34 return xchk_setup_inode_contents(sc, ip, 0); in xchk_setup_parent() 68 struct xfs_scrub *sc, in xchk_parent_count_parent_dentries() argument 75 .ino = sc->ip->i_ino, in xchk_parent_count_parent_dentries() 105 error = xfs_readdir(sc->tp, parent, &spc.dc, bufsize); in xchk_parent_count_parent_dentries() 124 struct xfs_scrub *sc, in xchk_parent_validate() argument 128 struct xfs_mount *mp = sc->mp; in xchk_parent_validate() 136 if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT) in xchk_parent_validate() 140 if (sc->ip->i_ino == dnum) { in xchk_parent_validate() 141 xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, 0); in xchk_parent_validate() [all …]
|
D | alloc.c | 31 struct xfs_scrub *sc, in xchk_setup_ag_allocbt() argument 34 return xchk_setup_ag_btree(sc, ip, false); in xchk_setup_ag_allocbt() 44 struct xfs_scrub *sc, in xchk_allocbt_xref_other() argument 54 if (sc->sm->sm_type == XFS_SCRUB_TYPE_BNOBT) in xchk_allocbt_xref_other() 55 pcur = &sc->sa.cnt_cur; in xchk_allocbt_xref_other() 57 pcur = &sc->sa.bno_cur; in xchk_allocbt_xref_other() 58 if (!*pcur || xchk_skip_xref(sc->sm)) in xchk_allocbt_xref_other() 62 if (!xchk_should_check_xref(sc, &error, pcur)) in xchk_allocbt_xref_other() 65 xchk_btree_xref_set_corrupt(sc, *pcur, 0); in xchk_allocbt_xref_other() 70 if (!xchk_should_check_xref(sc, &error, pcur)) in xchk_allocbt_xref_other() [all …]
|
D | agheader_repair.c | 39 struct xfs_scrub *sc) in xrep_superblock() argument 41 struct xfs_mount *mp = sc->mp; in xrep_superblock() 47 agno = sc->sm->sm_agno; in xrep_superblock() 51 error = xfs_sb_get_secondary(mp, sc->tp, agno, &bp); in xrep_superblock() 60 xfs_trans_buf_set_type(sc->tp, bp, XFS_BLFT_SB_BUF); in xrep_superblock() 61 xfs_trans_log_buf(sc->tp, bp, 0, BBTOB(bp->b_length) - 1); in xrep_superblock() 68 struct xfs_scrub *sc; member 83 if (xchk_should_terminate(raa->sc, &error)) in xrep_agf_walk_allocbt() 99 struct xfs_scrub *sc = priv; in xrep_agf_check_agfl_block() local 101 if (!xfs_verify_agbno(mp, sc->sa.agno, agbno)) in xrep_agf_check_agfl_block() [all …]
|
D | refcount.c | 32 struct xfs_scrub *sc, in xchk_setup_ag_refcountbt() argument 35 return xchk_setup_ag_btree(sc, ip, false); in xchk_setup_ag_refcountbt() 82 struct xfs_scrub *sc; member 113 if (xchk_should_terminate(refchk->sc, &error)) in xchk_refcountbt_rmap_check() 121 xchk_btree_xref_set_corrupt(refchk->sc, cur, 0); in xchk_refcountbt_rmap_check() 279 struct xfs_scrub *sc, in xchk_refcountbt_xref_rmap() argument 285 .sc = sc, in xchk_refcountbt_xref_rmap() 297 if (!sc->sa.rmap_cur || xchk_skip_xref(sc->sm)) in xchk_refcountbt_xref_rmap() 307 error = xfs_rmap_query_range(sc->sa.rmap_cur, &low, &high, in xchk_refcountbt_xref_rmap() 309 if (!xchk_should_check_xref(sc, &error, &sc->sa.rmap_cur)) in xchk_refcountbt_xref_rmap() [all …]
|
D | dir.c | 35 struct xfs_scrub *sc, in xchk_setup_directory() argument 38 return xchk_setup_inode_contents(sc, ip, 0); in xchk_setup_directory() 49 struct xfs_scrub *sc; member 60 struct xfs_mount *mp = sdc->sc->mp; in xchk_dir_check_ftype() 67 xchk_fblock_set_corrupt(sdc->sc, XFS_DATA_FORK, in xchk_dir_check_ftype() 80 error = xfs_iget(mp, sdc->sc->tp, inum, 0, 0, &ip); in xchk_dir_check_ftype() 81 if (!xchk_fblock_xref_process_error(sdc->sc, XFS_DATA_FORK, offset, in xchk_dir_check_ftype() 89 xchk_fblock_set_corrupt(sdc->sc, XFS_DATA_FORK, offset); in xchk_dir_check_ftype() 121 ip = sdc->sc->ip; in xchk_dir_actor() 128 xchk_fblock_set_corrupt(sdc->sc, XFS_DATA_FORK, offset); in xchk_dir_actor() [all …]
|
D | ialloc.c | 39 struct xfs_scrub *sc, in xchk_setup_ag_iallocbt() argument 42 return xchk_setup_ag_btree(sc, ip, sc->try_harder); in xchk_setup_ag_iallocbt() 54 struct xfs_scrub *sc, in xchk_iallocbt_chunk_xref_other() argument 62 if (sc->sm->sm_type == XFS_SCRUB_TYPE_FINOBT) in xchk_iallocbt_chunk_xref_other() 63 pcur = &sc->sa.ino_cur; in xchk_iallocbt_chunk_xref_other() 65 pcur = &sc->sa.fino_cur; in xchk_iallocbt_chunk_xref_other() 69 if (!xchk_should_check_xref(sc, &error, pcur)) in xchk_iallocbt_chunk_xref_other() 73 xchk_btree_xref_set_corrupt(sc, *pcur, 0); in xchk_iallocbt_chunk_xref_other() 79 struct xfs_scrub *sc, in xchk_iallocbt_chunk_xref() argument 87 if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT) in xchk_iallocbt_chunk_xref() [all …]
|
D | scrub.c | 154 struct xfs_scrub *sc) in xchk_probe() argument 158 if (xchk_should_terminate(sc, &error)) in xchk_probe() 169 struct xfs_scrub *sc, in xchk_teardown() argument 173 xchk_ag_free(sc, &sc->sa); in xchk_teardown() 174 if (sc->tp) { in xchk_teardown() 175 if (error == 0 && (sc->sm->sm_flags & XFS_SCRUB_IFLAG_REPAIR)) in xchk_teardown() 176 error = xfs_trans_commit(sc->tp); in xchk_teardown() 178 xfs_trans_cancel(sc->tp); in xchk_teardown() 179 sc->tp = NULL; in xchk_teardown() 181 if (sc->ip) { in xchk_teardown() [all …]
|
D | repair.h | 9 static inline int xrep_notsupported(struct xfs_scrub *sc) in xrep_notsupported() argument 18 int xrep_attempt(struct xfs_inode *ip, struct xfs_scrub *sc, bool *fixed); 20 int xrep_roll_ag_trans(struct xfs_scrub *sc); 23 xfs_extlen_t xrep_calc_ag_resblks(struct xfs_scrub *sc); 24 int xrep_alloc_ag_block(struct xfs_scrub *sc, struct xfs_owner_info *oinfo, 26 int xrep_init_btblock(struct xfs_scrub *sc, xfs_fsblock_t fsb, 32 int xrep_fix_freelist(struct xfs_scrub *sc, bool can_shrink); 33 int xrep_invalidate_blocks(struct xfs_scrub *sc, struct xfs_bitmap *btlist); 34 int xrep_reap_extents(struct xfs_scrub *sc, struct xfs_bitmap *exlist, 52 int xrep_find_ag_btree_roots(struct xfs_scrub *sc, struct xfs_buf *agf_bp, [all …]
|
D | symlink.c | 29 struct xfs_scrub *sc, in xchk_setup_symlink() argument 33 sc->buf = kmem_zalloc_large(XFS_SYMLINK_MAXLEN + 1, KM_SLEEP); in xchk_setup_symlink() 34 if (!sc->buf) in xchk_setup_symlink() 37 return xchk_setup_inode_contents(sc, ip, 0); in xchk_setup_symlink() 44 struct xfs_scrub *sc) in xchk_symlink() argument 46 struct xfs_inode *ip = sc->ip; in xchk_symlink() 58 xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, 0); in xchk_symlink() 66 xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, 0); in xchk_symlink() 71 error = xfs_readlink_bmap_ilocked(sc->ip, sc->buf); in xchk_symlink() 72 if (!xchk_fblock_process_error(sc, XFS_DATA_FORK, 0, &error)) in xchk_symlink() [all …]
|
D | btree.c | 33 struct xfs_scrub *sc, in __xchk_btree_process_error() argument 46 trace_xchk_deadlock_retry(sc->ip, sc->sm, *error); in __xchk_btree_process_error() 51 sc->sm->sm_flags |= errflag; in __xchk_btree_process_error() 56 trace_xchk_ifork_btree_op_error(sc, cur, level, in __xchk_btree_process_error() 59 trace_xchk_btree_op_error(sc, cur, level, in __xchk_btree_process_error() 68 struct xfs_scrub *sc, in xchk_btree_process_error() argument 73 return __xchk_btree_process_error(sc, cur, level, error, in xchk_btree_process_error() 79 struct xfs_scrub *sc, in xchk_btree_xref_process_error() argument 84 return __xchk_btree_process_error(sc, cur, level, error, in xchk_btree_xref_process_error() 91 struct xfs_scrub *sc, in __xchk_btree_set_corrupt() argument [all …]
|
D | attr.c | 36 struct xfs_scrub *sc, in xchk_setup_xattr() argument 48 BITS_TO_LONGS(sc->mp->m_attr_geo->blksize)); in xchk_setup_xattr() 49 sc->buf = kmem_zalloc_large(sz, KM_SLEEP); in xchk_setup_xattr() 50 if (!sc->buf) in xchk_setup_xattr() 53 return xchk_setup_inode_contents(sc, ip, 0); in xchk_setup_xattr() 60 struct xfs_scrub *sc; member 87 xchk_ino_set_preen(sx->sc, context->dp->i_ino); in xchk_xattr_listent() 103 args.value = sx->sc->buf; in xchk_xattr_listent() 109 if (!xchk_fblock_process_error(sx->sc, XFS_ATTR_FORK, args.blkno, in xchk_xattr_listent() 113 xchk_fblock_set_corrupt(sx->sc, XFS_ATTR_FORK, in xchk_xattr_listent() [all …]
|
/linux-4.19.296/drivers/clk/qcom/ |
D | gdsc.c | 59 static int gdsc_check_status(struct gdsc *sc, enum gdsc_status status) in gdsc_check_status() argument 65 if (sc->flags & POLL_CFG_GDSCR) in gdsc_check_status() 66 reg = sc->gdscr + CFG_GDSCR_OFFSET; in gdsc_check_status() 67 else if (sc->gds_hw_ctrl) in gdsc_check_status() 68 reg = sc->gds_hw_ctrl; in gdsc_check_status() 70 reg = sc->gdscr; in gdsc_check_status() 72 ret = regmap_read(sc->regmap, reg, &val); in gdsc_check_status() 76 if (sc->flags & POLL_CFG_GDSCR) { in gdsc_check_status() 95 static int gdsc_hwctrl(struct gdsc *sc, bool en) in gdsc_hwctrl() argument 99 return regmap_update_bits(sc->regmap, sc->gdscr, HW_CONTROL_MASK, val); in gdsc_hwctrl() [all …]
|
/linux-4.19.296/drivers/hid/ |
D | hid-sony.c | 585 static void sony_set_leds(struct sony_sc *sc); 587 static inline void sony_schedule_work(struct sony_sc *sc, in sony_schedule_work() argument 594 spin_lock_irqsave(&sc->lock, flags); in sony_schedule_work() 595 if (!sc->defer_initialization && sc->state_worker_initialized) in sony_schedule_work() 596 schedule_work(&sc->state_worker); in sony_schedule_work() 597 spin_unlock_irqrestore(&sc->lock, flags); in sony_schedule_work() 600 if (sc->hotplug_worker_initialized) in sony_schedule_work() 601 schedule_work(&sc->hotplug_worker); in sony_schedule_work() 611 struct sony_sc *sc = hid_get_drvdata(hdev); in ds4_show_poll_interval() local 613 return snprintf(buf, PAGE_SIZE, "%i\n", sc->ds4_bt_poll_interval); in ds4_show_poll_interval() [all …]
|
/linux-4.19.296/fs/ocfs2/cluster/ |
D | tcp.c | 78 #define SC_NODEF_ARGS(sc) sc->sc_node->nd_name, sc->sc_node->nd_num, \ argument 79 &sc->sc_node->nd_ipv4_address, \ 80 ntohs(sc->sc_node->nd_ipv4_port) 97 #define sclog(sc, fmt, args...) do { \ argument 98 typeof(sc) __sc = (sc); \ 144 static void o2net_sc_postpone_idle(struct o2net_sock_container *sc); 145 static void o2net_sc_reset_idle_timer(struct o2net_sock_container *sc); 174 struct o2net_sock_container *sc) in o2net_set_nst_sock_container() argument 176 nst->st_sc = sc; in o2net_set_nst_sock_container() 185 static inline void o2net_set_sock_timer(struct o2net_sock_container *sc) in o2net_set_sock_timer() argument [all …]
|
D | netdebug.c | 213 void o2net_debug_add_sc(struct o2net_sock_container *sc) in o2net_debug_add_sc() argument 216 list_add(&sc->sc_net_debug_item, &sock_containers); in o2net_debug_add_sc() 220 void o2net_debug_del_sc(struct o2net_sock_container *sc) in o2net_debug_del_sc() argument 223 list_del_init(&sc->sc_net_debug_item); in o2net_debug_del_sc() 235 struct o2net_sock_container *sc, *ret = NULL; in next_sc() local 239 list_for_each_entry(sc, &sc_start->sc_net_debug_item, in next_sc() 242 if (&sc->sc_net_debug_item == &sock_containers) in next_sc() 246 if (sc->sc_page != NULL) { in next_sc() 247 ret = sc; in next_sc() 258 struct o2net_sock_container *sc, *dummy_sc = sd->dbg_sock; in sc_seq_start() local [all …]
|