Home
last modified time | relevance | path

Searched full:mp (Results 1 – 25 of 504) sorted by relevance

12345678910>>...21

/linux/fs/xfs/libxfs/
H A Dxfs_trans_space.h14 #define XFS_MAX_CONTIG_BMAPS_PER_BLOCK(mp) \ argument
15 (((mp)->m_bmap_dmxr[0]) - ((mp)->m_bmap_dmnr[0]))
18 #define XFS_MAX_CONTIG_RTRMAPS_PER_BLOCK(mp) \ argument
19 (((mp)->m_rtrmap_mxr[0]) - ((mp)->m_rtrmap_mnr[0]))
22 #define XFS_RTRMAPADD_SPACE_RES(mp) ((mp)->m_rtrmap_maxlevels) argument
25 #define XFS_NRTRMAPADD_SPACE_RES(mp, b) \ argument
26 ((((b) + XFS_MAX_CONTIG_RTRMAPS_PER_BLOCK(mp) - 1) / \
27 XFS_MAX_CONTIG_RTRMAPS_PER_BLOCK(mp)) * \
28 XFS_RTRMAPADD_SPACE_RES(mp))
31 #define XFS_MAX_CONTIG_RMAPS_PER_BLOCK(mp) \ argument
[all …]
H A Dxfs_rtbitmap.h13 struct xfs_mount *mp; member
28 struct xfs_mount *mp = rtg_mount(rtg); in xfs_rtx_to_rtb() local
31 if (mp->m_rtxblklog >= 0) in xfs_rtx_to_rtb()
32 return start + (rtx << mp->m_rtxblklog); in xfs_rtx_to_rtb()
33 return start + (rtx * mp->m_sb.sb_rextsize); in xfs_rtx_to_rtb()
39 struct xfs_mount *mp, in xfs_rgbno_to_rtx() argument
42 if (likely(mp->m_rtxblklog >= 0)) in xfs_rgbno_to_rtx()
43 return rgbno >> mp->m_rtxblklog; in xfs_rgbno_to_rtx()
44 return rgbno / mp->m_sb.sb_rextsize; in xfs_rgbno_to_rtx()
49 struct xfs_mount *mp, in xfs_rtbxlen_to_blen() argument
[all …]
H A Dxfs_ag.c45 struct xfs_mount *mp, in xfs_initialize_perag_data() argument
50 struct xfs_sb *sbp = &mp->m_sb; in xfs_initialize_perag_data()
64 pag = xfs_perag_get(mp, index); in xfs_initialize_perag_data()
89 xfs_alert(mp, "AGF corruption. Please run xfs_repair."); in xfs_initialize_perag_data()
90 xfs_fs_mark_sick(mp, XFS_SICK_FS_COUNTERS); in xfs_initialize_perag_data()
96 spin_lock(&mp->m_sb_lock); in xfs_initialize_perag_data()
100 spin_unlock(&mp->m_sb_lock); in xfs_initialize_perag_data()
102 xfs_reinit_percpu_counters(mp); in xfs_initialize_perag_data()
104 xfs_fs_mark_healthy(mp, XFS_SICK_FS_COUNTERS); in xfs_initialize_perag_data()
125 struct xfs_mount *mp, in xfs_free_perag_range() argument
[all …]
H A Dxfs_metafile.c89 struct xfs_mount *mp, in xfs_metafile_resv_can_cover() argument
98 if (mp->m_metafile_resv_avail >= rhs) in xfs_metafile_resv_can_cover()
105 return xfs_compare_freecounter(mp, XC_FREE_BLOCKS, in xfs_metafile_resv_can_cover()
106 rhs - mp->m_metafile_resv_avail, 2048) >= 0; in xfs_metafile_resv_can_cover()
116 struct xfs_mount *mp) in xfs_metafile_resv_critical() argument
118 ASSERT(xfs_has_metadir(mp)); in xfs_metafile_resv_critical()
120 trace_xfs_metafile_resv_critical(mp, 0); in xfs_metafile_resv_critical()
122 if (!xfs_metafile_resv_can_cover(mp, mp->m_rtbtree_maxlevels)) in xfs_metafile_resv_critical()
125 if (!xfs_metafile_resv_can_cover(mp, in xfs_metafile_resv_critical()
126 div_u64(mp->m_metafile_resv_target, 10))) in xfs_metafile_resv_critical()
[all …]
H A Dxfs_sb.c195 struct xfs_mount *mp, in xfs_validate_sb_read() argument
206 xfs_warn(mp, in xfs_validate_sb_read()
209 xfs_warn(mp, in xfs_validate_sb_read()
214 xfs_alert(mp, in xfs_validate_sb_read()
218 if (!xfs_is_readonly(mp)) { in xfs_validate_sb_read()
219 xfs_warn(mp, in xfs_validate_sb_read()
221 xfs_warn(mp, in xfs_validate_sb_read()
228 xfs_warn(mp, in xfs_validate_sb_read()
232 xfs_warn(mp, in xfs_validate_sb_read()
325 struct xfs_mount *mp, in xfs_validate_sb_write() argument
[all …]
H A Dxfs_rtgroup.h115 struct xfs_mount *mp, in xfs_rtgroup_get() argument
118 return to_rtg(xfs_group_get(mp, rgno, XG_TYPE_RTG)); in xfs_rtgroup_get()
138 struct xfs_mount *mp, in xfs_rtgroup_grab() argument
141 return to_rtg(xfs_group_grab(mp, rgno, XG_TYPE_RTG)); in xfs_rtgroup_grab()
153 struct xfs_mount *mp, in xfs_rtgroup_next_range() argument
158 return to_rtg(xfs_group_next_range(mp, rtg ? rtg_group(rtg) : NULL, in xfs_rtgroup_next_range()
164 struct xfs_mount *mp, in xfs_rtgroup_next() argument
167 return xfs_rtgroup_next_range(mp, rtg, 0, mp->m_sb.sb_rgcount - 1); in xfs_rtgroup_next()
206 struct xfs_mount *mp, in xfs_rtb_to_rgno() argument
209 return xfs_fsb_to_gno(mp, rtbno, XG_TYPE_RTG); in xfs_rtb_to_rgno()
[all …]
/linux/fs/xfs/
H A Dxfs_super.c80 struct xfs_mount *mp, in xfs_mount_set_dax_mode() argument
85 mp->m_features &= ~(XFS_FEAT_DAX_ALWAYS | XFS_FEAT_DAX_NEVER); in xfs_mount_set_dax_mode()
88 mp->m_features |= XFS_FEAT_DAX_ALWAYS; in xfs_mount_set_dax_mode()
89 mp->m_features &= ~XFS_FEAT_DAX_NEVER; in xfs_mount_set_dax_mode()
92 mp->m_features |= XFS_FEAT_DAX_NEVER; in xfs_mount_set_dax_mode()
93 mp->m_features &= ~XFS_FEAT_DAX_ALWAYS; in xfs_mount_set_dax_mode()
207 struct xfs_mount *mp = XFS_M(root->d_sb); in xfs_fs_show_options() local
211 if (mp->m_features & xfs_infop->flag) in xfs_fs_show_options()
215 seq_printf(m, ",inode%d", xfs_has_small_inums(mp) ? 32 : 64); in xfs_fs_show_options()
217 if (xfs_has_allocsize(mp)) in xfs_fs_show_options()
[all …]
H A Dxfs_fsops.c54 struct xfs_mount *mp = tp->t_mountp; in xfs_resizefs_init_new_ags() local
55 xfs_rfsblock_t nb = mp->m_sb.sb_dblocks + delta; in xfs_resizefs_init_new_ags()
67 (xfs_rfsblock_t)mp->m_sb.sb_agblocks); in xfs_resizefs_init_new_ags()
69 id->agsize = mp->m_sb.sb_agblocks; in xfs_resizefs_init_new_ags()
71 error = xfs_ag_init_headers(mp, id); in xfs_resizefs_init_new_ags()
94 struct xfs_mount *mp, /* mount point for filesystem */ in xfs_growfs_data_private() argument
97 xfs_agnumber_t oagcount = mp->m_sb.sb_agcount; in xfs_growfs_data_private()
109 error = xfs_sb_validate_fsb_count(&mp->m_sb, nb); in xfs_growfs_data_private()
113 if (nb > mp->m_sb.sb_dblocks) { in xfs_growfs_data_private()
114 error = xfs_buf_read_uncached(mp->m_ddev_targp, in xfs_growfs_data_private()
[all …]
H A Dxfs_qm.c40 STATIC int xfs_qm_init_quotainos(struct xfs_mount *mp);
41 STATIC int xfs_qm_init_quotainfo(struct xfs_mount *mp);
55 struct xfs_mount *mp, in xfs_qm_dquot_walk() argument
60 struct xfs_quotainfo *qi = mp->m_quotainfo; in xfs_qm_dquot_walk()
205 struct xfs_mount *mp) in xfs_qm_dqpurge_all() argument
207 xfs_qm_dquot_walk(mp, XFS_DQTYPE_USER, xfs_qm_dqpurge, NULL); in xfs_qm_dqpurge_all()
208 xfs_qm_dquot_walk(mp, XFS_DQTYPE_GROUP, xfs_qm_dqpurge, NULL); in xfs_qm_dqpurge_all()
209 xfs_qm_dquot_walk(mp, XFS_DQTYPE_PROJ, xfs_qm_dqpurge, NULL); in xfs_qm_dqpurge_all()
217 struct xfs_mount *mp) in xfs_qm_unmount() argument
219 if (mp->m_quotainfo) { in xfs_qm_unmount()
[all …]
H A Dxfs_rtalloc.c108 for (log = oargs->mp->m_rsumlevels - 1; log >= 0; log--) { in xfs_rtcopy_summary()
109 for (bbno = oargs->mp->m_sb.sb_rbmblocks - 1; in xfs_rtcopy_summary()
115 if (XFS_IS_CORRUPT(oargs->mp, sum < 0)) { in xfs_rtcopy_summary()
144 struct xfs_mount *mp = args->mp; in xfs_rtallocate_range() local
174 xfs_rtx_to_rbmblock(mp, preblock), -1); in xfs_rtallocate_range()
185 xfs_rtx_to_rbmblock(mp, preblock), 1); in xfs_rtallocate_range()
197 xfs_rtx_to_rbmblock(mp, end + 1), 1); in xfs_rtallocate_range()
253 struct xfs_mount *mp = args->mp; in xfs_rtallocate_extent_block() local
267 end = min(args->rtg->rtg_extents, xfs_rbmblock_to_rtx(mp, bbno + 1)) - in xfs_rtallocate_extent_block()
269 for (i = xfs_rbmblock_to_rtx(mp, bbno); i <= end; i++) { in xfs_rtallocate_extent_block()
[all …]
H A Dxfs_notify_failure.c41 struct xfs_mount *mp, in xfs_failure_pgoff() argument
45 loff_t pos = XFS_FSB_TO_B(mp, rec->rm_offset); in xfs_failure_pgoff()
48 pos += XFS_FSB_TO_B(mp, in xfs_failure_pgoff()
55 struct xfs_mount *mp, in xfs_failure_pgcnt() argument
70 return XFS_FSB_TO_B(mp, end_cross - start_cross) >> PAGE_SHIFT; in xfs_failure_pgcnt()
79 struct xfs_mount *mp = cur->bc_mp; in xfs_dax_failure_fn() local
97 error = xfs_iget(mp, cur->bc_tp, rec->rm_owner, XFS_IGET_INCORE, in xfs_dax_failure_fn()
108 pgoff = xfs_failure_pgoff(mp, rec, notify); in xfs_dax_failure_fn()
109 pgcnt = xfs_failure_pgcnt(mp, rec, notify); in xfs_dax_failure_fn()
130 struct xfs_mount *mp) in xfs_dax_notify_failure_freeze() argument
[all …]
H A Dxfs_zone_alloc.c49 struct xfs_mount *mp, in xfs_zone_bucket() argument
53 mp->m_groups[XG_TYPE_RTG].blocks; in xfs_zone_bucket()
82 struct xfs_mount *mp = rtg_mount(rtg); in xfs_zone_account_reclaimable() local
83 struct xfs_zone_info *zi = mp->m_zone_info; in xfs_zone_account_reclaimable()
86 uint32_t from_bucket = xfs_zone_bucket(mp, used + freed); in xfs_zone_account_reclaimable()
87 uint32_t to_bucket = xfs_zone_bucket(mp, used); in xfs_zone_account_reclaimable()
127 if (zi->zi_gc_thread && xfs_zoned_need_gc(mp)) in xfs_zone_account_reclaimable()
168 struct xfs_mount *mp = rtg_mount(rtg); in xfs_open_zone_mark_full() local
169 struct xfs_zone_info *zi = mp->m_zone_info; in xfs_open_zone_mark_full()
199 struct xfs_mount *mp = tp->t_mountp; in xfs_zone_record_blocks() local
[all …]
H A Dxfs_icache.c51 static int xfs_icwalk(struct xfs_mount *mp,
88 struct xfs_mount *mp, in xfs_inode_alloc() argument
97 ip = alloc_inode_sb(mp->m_super, xfs_inode_cache, GFP_KERNEL | __GFP_NOFAIL); in xfs_inode_alloc()
99 if (inode_init_always(mp->m_super, VFS_I(ip))) { in xfs_inode_alloc()
107 M_IGEO(mp)->min_folio_order); in xfs_inode_alloc()
109 XFS_STATS_INC(mp, xs_inodes_active); in xfs_inode_alloc()
115 ip->i_mount = mp; in xfs_inode_alloc()
123 ip->i_diflags2 = mp->m_ino_geo.new_diflags2; in xfs_inode_alloc()
208 struct xfs_mount *mp) in xfs_reclaim_work_queue()
212 if (xfs_group_marked(mp, XG_TYPE_A in xfs_reclaim_work_queue()
209 xfs_reclaim_work_queue(struct xfs_mount * mp) xfs_reclaim_work_queue() argument
228 struct xfs_mount *mp = pag_mount(pag); xfs_blockgc_queue() local
310 xfs_perag_grab_next_tag(struct xfs_mount * mp,struct xfs_perag * pag,int tag) xfs_perag_grab_next_tag() argument
329 xfs_reinit_inode(struct xfs_mount * mp,struct inode * inode) xfs_reinit_inode() argument
366 struct xfs_mount *mp = ip->i_mount; xfs_iget_recycle() local
459 xfs_inodegc_queue_all(struct xfs_mount * mp) xfs_inodegc_queue_all() argument
479 xfs_inodegc_wait_all(struct xfs_mount * mp) xfs_inodegc_wait_all() argument
509 struct xfs_mount *mp = ip->i_mount; xfs_iget_cache_hit() local
626 xfs_iget_cache_miss(struct xfs_mount * mp,struct xfs_perag * pag,xfs_trans_t * tp,xfs_ino_t ino,struct xfs_inode ** ipp,int flags,int lock_flags) xfs_iget_cache_miss() argument
768 xfs_iget(struct xfs_mount * mp,struct xfs_trans * tp,xfs_ino_t ino,uint flags,uint lock_flags,struct xfs_inode ** ipp) xfs_iget() argument
850 struct xfs_mount *mp = tp->t_mountp; xfs_trans_metafile_iget() local
891 xfs_metafile_iget(struct xfs_mount * mp,xfs_ino_t ino,enum xfs_metafile_type metafile_type,struct xfs_inode ** ipp) xfs_metafile_iget() argument
1064 xfs_want_reclaim_sick(struct xfs_mount * mp) xfs_want_reclaim_sick() argument
1072 xfs_reclaim_inodes(struct xfs_mount * mp) xfs_reclaim_inodes() argument
1096 xfs_reclaim_inodes_nr(struct xfs_mount * mp,unsigned long nr_to_scan) xfs_reclaim_inodes_nr() argument
1121 xfs_reclaim_inodes_count(struct xfs_mount * mp) xfs_reclaim_inodes_count() argument
1221 struct xfs_mount *mp = container_of(to_delayed_work(work), xfs_reclaim_worker() local
1276 struct xfs_mount *mp = ip->i_mount; xfs_blockgc_set_iflag() local
1314 struct xfs_mount *mp = ip->i_mount; xfs_blockgc_clear_iflag() local
1466 xfs_blockgc_stop(struct xfs_mount * mp) xfs_blockgc_stop() argument
1481 xfs_blockgc_start(struct xfs_mount * mp) xfs_blockgc_start() argument
1565 struct xfs_mount *mp = pag_mount(pag); xfs_blockgc_worker() local
1583 xfs_blockgc_free_space(struct xfs_mount * mp,struct xfs_icwalk * icw) xfs_blockgc_free_space() argument
1603 xfs_blockgc_flush_all(struct xfs_mount * mp) xfs_blockgc_flush_all() argument
1634 xfs_blockgc_free_dquots(struct xfs_mount * mp,struct xfs_dquot * udqp,struct xfs_dquot * gdqp,struct xfs_dquot * pdqp,unsigned int iwalk_flags) xfs_blockgc_free_dquots() argument
1753 struct xfs_mount *mp = pag_mount(pag); xfs_icwalk_ag() local
1858 xfs_icwalk(struct xfs_mount * mp,enum xfs_icwalk_goal goal,struct xfs_icwalk * icw) xfs_icwalk() argument
1911 struct xfs_mount *mp = ip->i_mount; xfs_inodegc_set_reclaimable() local
1961 struct xfs_mount *mp = gc->mp; xfs_inodegc_worker() local
2007 xfs_inodegc_push(struct xfs_mount * mp) xfs_inodegc_push() argument
2021 xfs_inodegc_flush(struct xfs_mount * mp) xfs_inodegc_flush() argument
2035 xfs_inodegc_stop(struct xfs_mount * mp) xfs_inodegc_stop() argument
2068 xfs_inodegc_start(struct xfs_mount * mp) xfs_inodegc_start() argument
2082 struct xfs_mount *mp = ip->i_mount; xfs_inodegc_want_queue_rt_file() local
2110 struct xfs_mount *mp = ip->i_mount; xfs_inodegc_want_queue_work() local
2179 struct xfs_mount *mp = ip->i_mount; xfs_inodegc_queue() local
2244 struct xfs_mount *mp = ip->i_mount; xfs_inode_mark_reclaimable() local
2283 struct xfs_mount *mp = shrink->private_data; xfs_inodegc_shrinker_count() local
2304 struct xfs_mount *mp = shrink->private_data; xfs_inodegc_shrinker_scan() local
2338 xfs_inodegc_register_shrinker(struct xfs_mount * mp) xfs_inodegc_register_shrinker() argument
[all...]
H A Dxfs_ioctl.c58 xfs_bulkstat_to_bstat(breq->mp, &bs1, bstat); in xfs_fsbulkstat_one_fmt()
83 struct xfs_mount *mp = XFS_I(file_inode(file))->i_mount; in xfs_ioc_fsbulkstat() local
86 .mp = mp, in xfs_ioc_fsbulkstat()
99 if (xfs_is_shutdown(mp)) in xfs_ioc_fsbulkstat()
175 struct xfs_mount *mp, in xfs_bulk_ireq_setup() argument
202 breq->startino = mp->m_sb.sb_rootino; in xfs_bulk_ireq_setup()
216 if (hdr->agno >= mp->m_sb.sb_agcount) in xfs_bulk_ireq_setup()
220 breq->startino = XFS_AGINO_TO_INO(mp, hdr->agno, 0); in xfs_bulk_ireq_setup()
221 else if (XFS_INO_TO_AGNO(mp, breq->startino) < hdr->agno) in xfs_bulk_ireq_setup()
227 if (XFS_INO_TO_AGNO(mp, breq->startino) > hdr->agno) in xfs_bulk_ireq_setup()
[all …]
H A Dxfs_zone_gc.c127 struct xfs_mount *mp;
168 struct xfs_mount *mp) in xfs_zoned_need_gc()
173 if (!xfs_zoned_have_reclaimable(mp->m_zone_info)) in xfs_zoned_need_gc()
176 available = xfs_estimate_freecounter(mp, XC_FREE_RTAVAILABLE); in xfs_zoned_need_gc()
179 xfs_rtgs_to_rfsbs(mp, mp->m_max_open_zones - XFS_OPEN_GC_ZONES)) in xfs_zoned_need_gc()
182 free = xfs_estimate_freecounter(mp, XC_FREE_RTEXTENTS); in xfs_zoned_need_gc()
185 threshold = threshold * mp->m_zonegc_low_space + in xfs_zoned_need_gc()
186 remainder * div_s64(mp->m_zonegc_low_space, 100); in xfs_zoned_need_gc()
196 struct xfs_mount *mp) in xfs_zone_gc_data_alloc()
128 struct xfs_mount *mp; global() member
169 xfs_zoned_need_gc(struct xfs_mount * mp) xfs_zoned_need_gc() argument
197 xfs_zone_gc_data_alloc(struct xfs_mount * mp) xfs_zone_gc_data_alloc() argument
304 xfs_zone_gc_query(struct xfs_mount * mp,struct xfs_zone_gc_iter * iter) xfs_zone_gc_query() argument
367 xfs_zone_gc_iter_next(struct xfs_mount * mp,struct xfs_zone_gc_iter * iter,struct xfs_rmap_irec * chunk_rec,struct xfs_inode ** ipp) xfs_zone_gc_iter_next() argument
431 xfs_zone_gc_pick_victim_from(struct xfs_mount * mp,uint32_t bucket) xfs_zone_gc_pick_victim_from() argument
498 struct xfs_mount *mp = data->mp; xfs_zone_gc_select_victim() local
543 xfs_zone_gc_select_target(struct xfs_mount * mp) xfs_zone_gc_select_target() argument
573 xfs_zone_gc_ensure_target(struct xfs_mount * mp) xfs_zone_gc_ensure_target() argument
601 struct xfs_mount *mp = data->mp; xfs_zone_gc_alloc_blocks() local
666 struct xfs_mount *mp = data->mp; xfs_zone_gc_start_chunk() local
809 struct xfs_mount *mp = chunk->ip->i_mount; xfs_zone_gc_write_chunk() local
838 struct xfs_mount *mp = ip->i_mount; xfs_zone_gc_finish_chunk() local
883 struct xfs_mount *mp = rtg_mount(rtg); xfs_zone_gc_finish_reset() local
907 struct xfs_mount *mp = rtg_mount(rtg); xfs_submit_zone_reset_bio() local
1094 struct xfs_mount *mp = data->mp; xfs_zoned_gcd() local
1150 xfs_zone_gc_start(struct xfs_mount * mp) xfs_zone_gc_start() argument
1158 xfs_zone_gc_stop(struct xfs_mount * mp) xfs_zone_gc_stop() argument
1166 xfs_zone_gc_mount(struct xfs_mount * mp) xfs_zone_gc_mount() argument
1223 xfs_zone_gc_unmount(struct xfs_mount * mp) xfs_zone_gc_unmount() argument
[all...]
/linux/drivers/media/usb/pvrusb2/
H A Dpvrusb2-context.c30 static void pvr2_context_set_notify(struct pvr2_context *mp, int fl) in pvr2_context_set_notify() argument
35 if (!mp->notify_flag) { in pvr2_context_set_notify()
37 mp->notify_prev = pvr2_context_notify_last; in pvr2_context_set_notify()
38 mp->notify_next = NULL; in pvr2_context_set_notify()
39 pvr2_context_notify_last = mp; in pvr2_context_set_notify()
40 if (mp->notify_prev) { in pvr2_context_set_notify()
41 mp->notify_prev->notify_next = mp; in pvr2_context_set_notify()
43 pvr2_context_notify_first = mp; in pvr2_context_set_notify()
45 mp->notify_flag = !0; in pvr2_context_set_notify()
48 if (mp->notify_flag) { in pvr2_context_set_notify()
[all …]
/linux/fs/jfs/
H A Djfs_metapage.c34 #define metapage_locked(mp) test_bit(META_locked, &(mp)->flag) argument
35 #define trylock_metapage(mp) test_and_set_bit_lock(META_locked, &(mp)->flag) argument
37 static inline void unlock_metapage(struct metapage *mp) in unlock_metapage() argument
39 clear_bit_unlock(META_locked, &mp->flag); in unlock_metapage()
40 wake_up(&mp->wait); in unlock_metapage()
43 static inline void __lock_metapage(struct metapage *mp) in __lock_metapage() argument
47 add_wait_queue_exclusive(&mp->wait, &wait); in __lock_metapage()
50 if (metapage_locked(mp)) { in __lock_metapage()
51 folio_unlock(mp->folio); in __lock_metapage()
53 folio_lock(mp->folio); in __lock_metapage()
[all …]
H A Djfs_metapage.h45 #define mark_metapage_dirty(mp) set_bit(META_dirty, &(mp)->flag) argument
72 static inline void write_metapage(struct metapage *mp) in write_metapage() argument
74 set_bit(META_dirty, &mp->flag); in write_metapage()
75 release_metapage(mp); in write_metapage()
78 static inline void flush_metapage(struct metapage *mp) in flush_metapage() argument
80 set_bit(META_sync, &mp->flag); in flush_metapage()
81 write_metapage(mp); in flush_metapage()
84 static inline void discard_metapage(struct metapage *mp) in discard_metapage() argument
86 clear_bit(META_dirty, &mp->flag); in discard_metapage()
87 set_bit(META_discard, &mp->flag); in discard_metapage()
[all …]
/linux/drivers/net/ethernet/apple/
H A Dmace.c91 static inline void mace_clean_rings(struct mace_data *mp);
112 struct mace_data *mp; in mace_probe() local
155 mp = netdev_priv(dev); in mace_probe()
156 mp->mdev = mdev; in mace_probe()
160 mp->mace = ioremap(dev->base_addr, 0x1000); in mace_probe()
161 if (mp->mace == NULL) { in mace_probe()
173 mp->chipid = (in_8(&mp->mace->chipid_hi) << 8) | in mace_probe()
174 in_8(&mp->mace->chipid_lo); in mace_probe()
177 mp = netdev_priv(dev); in mace_probe()
178 mp->maccc = ENXMT | ENRCV; in mace_probe()
[all …]
/linux/drivers/net/ethernet/marvell/
H A Dmv643xx_eth.c419 static inline u32 rdl(struct mv643xx_eth_private *mp, int offset) in rdl() argument
421 return readl(mp->shared->base + offset); in rdl()
424 static inline u32 rdlp(struct mv643xx_eth_private *mp, int offset) in rdlp() argument
426 return readl(mp->base + offset); in rdlp()
429 static inline void wrl(struct mv643xx_eth_private *mp, int offset, u32 data) in wrl() argument
431 writel(data, mp->shared->base + offset); in wrl()
434 static inline void wrlp(struct mv643xx_eth_private *mp, int offset, u32 data) in wrlp() argument
436 writel(data, mp->base + offset); in wrlp()
453 struct mv643xx_eth_private *mp = rxq_to_mp(rxq); in rxq_enable() local
454 wrlp(mp, RXQ_COMMAND, 1 << rxq->index); in rxq_enable()
[all …]
/linux/drivers/pci/controller/dwc/
H A Dpci-meson.c74 static struct reset_control *meson_pcie_get_reset(struct meson_pcie *mp, in meson_pcie_get_reset() argument
78 struct device *dev = mp->pci.dev; in meson_pcie_get_reset()
89 static int meson_pcie_get_resets(struct meson_pcie *mp) in meson_pcie_get_resets() argument
91 struct meson_pcie_rc_reset *mrst = &mp->mrst; in meson_pcie_get_resets()
93 mrst->port = meson_pcie_get_reset(mp, "port", PCIE_NORMAL_RESET); in meson_pcie_get_resets()
98 mrst->apb = meson_pcie_get_reset(mp, "apb", PCIE_SHARED_RESET); in meson_pcie_get_resets()
107 struct meson_pcie *mp) in meson_pcie_get_mems() argument
109 struct dw_pcie *pci = &mp->pci; in meson_pcie_get_mems()
127 mp->cfg_base = devm_platform_ioremap_resource_byname(pdev, "cfg"); in meson_pcie_get_mems()
128 if (IS_ERR(mp->cfg_base)) in meson_pcie_get_mems()
[all …]
/linux/drivers/isdn/capi/
H A Dcapi.c147 static int capiminor_add_ack(struct capiminor *mp, u16 datahandle) in capiminor_add_ack() argument
158 spin_lock_bh(&mp->ackqlock); in capiminor_add_ack()
159 list_add_tail(&n->list, &mp->ackqueue); in capiminor_add_ack()
160 mp->nack++; in capiminor_add_ack()
161 spin_unlock_bh(&mp->ackqlock); in capiminor_add_ack()
165 static int capiminor_del_ack(struct capiminor *mp, u16 datahandle) in capiminor_del_ack() argument
169 spin_lock_bh(&mp->ackqlock); in capiminor_del_ack()
170 list_for_each_entry_safe(p, tmp, &mp->ackqueue, list) { in capiminor_del_ack()
173 mp->nack--; in capiminor_del_ack()
174 spin_unlock_bh(&mp->ackqlock); in capiminor_del_ack()
[all …]
/linux/drivers/scsi/sym53c8xx_2/
H A Dsym_malloc.c47 static void *___sym_malloc(m_pool_p mp, int size) in ___sym_malloc() argument
53 m_link_p h = mp->h; in ___sym_malloc()
93 static void ___sym_mfree(m_pool_p mp, void *ptr, int size) in ___sym_mfree() argument
99 m_link_p h = mp->h; in ___sym_mfree()
145 static void *__sym_calloc2(m_pool_p mp, int size, char *name, int uflags) in __sym_calloc2() argument
149 p = ___sym_malloc(mp, size); in __sym_calloc2()
161 #define __sym_calloc(mp, s, n) __sym_calloc2(mp, s, n, SYM_MEM_WARN) argument
166 static void __sym_mfree(m_pool_p mp, void *ptr, int size, char *name) in __sym_mfree() argument
171 ___sym_mfree(mp, ptr, size); in __sym_mfree()
180 static void *___mp0_get_mem_cluster(m_pool_p mp) in ___mp0_get_mem_cluster() argument
[all …]
/linux/fs/xfs/scrub/
H A Dagheader.c39 struct xfs_mount *mp = sc->mp; in xchk_superblock_xref() local
47 agbno = XFS_SB_BLOCK(mp); in xchk_superblock_xref()
70 struct xfs_mount *mp) in xchk_superblock_ondisk_size() argument
72 if (xfs_has_zoned(mp)) in xchk_superblock_ondisk_size()
74 if (xfs_has_metadir(mp)) in xchk_superblock_ondisk_size()
76 if (xfs_has_metauuid(mp)) in xchk_superblock_ondisk_size()
78 if (xfs_has_crc(mp)) in xchk_superblock_ondisk_size()
80 if (xfs_sb_version_hasmorebits(&mp->m_sb)) in xchk_superblock_ondisk_size()
82 if (xfs_has_logv2(mp)) in xchk_superblock_ondisk_size()
84 if (xfs_has_sector(mp)) in xchk_superblock_ondisk_size()
[all …]
/linux/drivers/net/ethernet/mellanox/mlx5/core/lag/
H A Dmp.c7 #include "lag/mp.h"
103 struct lag_mp *mp = container_of(nb, struct lag_mp, fib_nb); in mlx5_lag_fib_event_flush() local
105 flush_workqueue(mp->wq); in mlx5_lag_fib_event_flush()
108 static void mlx5_lag_fib_set(struct lag_mp *mp, struct fib_info *fi, u32 dst, int dst_len) in mlx5_lag_fib_set() argument
110 mp->fib.mfi = fi; in mlx5_lag_fib_set()
111 mp->fib.priority = fi->fib_priority; in mlx5_lag_fib_set()
112 mp->fib.dst = dst; in mlx5_lag_fib_set()
113 mp->fib.dst_len = dst_len; in mlx5_lag_fib_set()
161 struct lag_mp *mp = &ldev->lag_mp; in mlx5_lag_fib_route_event() local
170 if (mp->fib.mfi == fi) in mlx5_lag_fib_route_event()
[all …]

12345678910>>...21