Home
last modified time | relevance | path

Searched refs:i_inode (Results 1 – 22 of 22) sorted by relevance

/Linux-v6.6/fs/gfs2/
Dxattr.c91 struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); in ea_foreach_i()
95 if (gfs2_metatype_check(GFS2_SB(&ip->i_inode), bh, GFS2_METATYPE_EA)) in ea_foreach_i()
140 if (gfs2_metatype_check(GFS2_SB(&ip->i_inode), bh, GFS2_METATYPE_IN)) { in ea_foreach()
146 end = eablk + GFS2_SB(&ip->i_inode)->sd_inptrs; in ea_foreach()
234 struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); in ea_dealloc_unstuffed()
296 gfs2_add_inode_blocks(&ip->i_inode, -1); in ea_dealloc_unstuffed()
314 inode_set_ctime_current(&ip->i_inode); in ea_dealloc_unstuffed()
315 __mark_inode_dirty(&ip->i_inode, I_DIRTY_DATASYNC); in ea_dealloc_unstuffed()
330 error = gfs2_rindex_update(GFS2_SB(&ip->i_inode)); in ea_remove_unstuffed()
354 struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); in ea_list_i()
[all …]
Dinode.c363 error = gfs2_permission(&nop_mnt_idmap, &dip->i_inode, in create_ok()
369 if (!dip->i_inode.i_nlink) in create_ok()
374 if (S_ISDIR(mode) && dip->i_inode.i_nlink == (u32)-1) in create_ok()
383 if (GFS2_SB(&dip->i_inode)->sd_args.ar_suiddir && in munge_mode_uid_gid()
384 (dip->i_inode.i_mode & S_ISUID) && in munge_mode_uid_gid()
385 !uid_eq(dip->i_inode.i_uid, GLOBAL_ROOT_UID)) { in munge_mode_uid_gid()
388 else if (!uid_eq(dip->i_inode.i_uid, current_fsuid())) in munge_mode_uid_gid()
390 inode->i_uid = dip->i_inode.i_uid; in munge_mode_uid_gid()
394 if (dip->i_inode.i_mode & S_ISGID) { in munge_mode_uid_gid()
397 inode->i_gid = dip->i_inode.i_gid; in munge_mode_uid_gid()
[all …]
Ddir.c111 if (gfs2_metatype_check(GFS2_SB(&ip->i_inode), bh, GFS2_METATYPE_JD)) { in gfs2_dir_get_existing_buffer()
131 if (ip->i_inode.i_size < offset + size) in gfs2_dir_write_stuffed()
132 i_size_write(&ip->i_inode, offset + size); in gfs2_dir_write_stuffed()
133 ip->i_inode.i_mtime = inode_set_ctime_current(&ip->i_inode); in gfs2_dir_write_stuffed()
155 struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); in gfs2_dir_write_data()
193 error = gfs2_alloc_extent(&ip->i_inode, lblock, &dblock, in gfs2_dir_write_data()
228 if (ip->i_inode.i_size < offset + copied) in gfs2_dir_write_data()
229 i_size_write(&ip->i_inode, offset + copied); in gfs2_dir_write_data()
230 ip->i_inode.i_mtime = inode_set_ctime_current(&ip->i_inode); in gfs2_dir_write_data()
270 struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); in gfs2_dir_read_data()
[all …]
Dglops.c306 int isreg = ip && S_ISREG(ip->i_inode.i_mode); in inode_go_sync()
312 unmap_shared_mapping_range(ip->i_inode.i_mapping, 0, 0); in inode_go_sync()
313 inode_dio_wait(&ip->i_inode); in inode_go_sync()
324 struct address_space *mapping = ip->i_inode.i_mapping; in inode_go_sync()
367 forget_all_cached_acls(&ip->i_inode); in inode_go_inval()
368 security_inode_invalidate_secctx(&ip->i_inode); in inode_go_inval()
379 if (ip && S_ISREG(ip->i_inode.i_mode)) in inode_go_inval()
380 truncate_inode_pages(ip->i_inode.i_mapping, 0); in inode_go_inval()
404 struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); in gfs2_dinode_in()
409 struct inode *inode = &ip->i_inode; in gfs2_dinode_in()
[all …]
Dquota.h43 struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); in gfs2_quota_lock_check()
56 ret = gfs2_quota_check(ip, ip->i_inode.i_uid, ip->i_inode.i_gid, ap); in gfs2_quota_lock_check()
Dbmap.c58 struct inode *inode = &ip->i_inode; in gfs2_unstuffer_page()
104 if (i_size_read(&ip->i_inode)) { in __gfs2_unstuff_inode()
113 gfs2_trans_remove_revoke(GFS2_SB(&ip->i_inode), block, 1); in __gfs2_unstuff_inode()
133 if (i_size_read(&ip->i_inode)) { in __gfs2_unstuff_inode()
135 gfs2_add_inode_blocks(&ip->i_inode, 1); in __gfs2_unstuff_inode()
136 di->di_blocks = cpu_to_be64(gfs2_get_inode_blocks(&ip->i_inode)); in __gfs2_unstuff_inode()
159 struct inode *inode = &ip->i_inode; in gfs2_unstuff_dinode()
773 gfs2_add_inode_blocks(&ip->i_inode, alloced); in __gfs2_iomap_alloc()
1389 ip->i_inode.i_mtime = inode_set_ctime_current(&ip->i_inode); in trunc_start()
1454 struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); in sweep_bh_for_rgrps()
[all …]
Dquota.c562 struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); in gfs2_qa_get()
563 struct inode *inode = &ip->i_inode; in gfs2_qa_get()
590 struct inode *inode = &ip->i_inode; in gfs2_qa_put()
602 struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); in gfs2_quota_hold()
622 error = qdsb_get(sdp, make_kqid_uid(ip->i_inode.i_uid), qd); in gfs2_quota_hold()
628 error = qdsb_get(sdp, make_kqid_gid(ip->i_inode.i_gid), qd); in gfs2_quota_hold()
635 !uid_eq(uid, ip->i_inode.i_uid)) { in gfs2_quota_hold()
644 !gid_eq(gid, ip->i_inode.i_gid)) { in gfs2_quota_hold()
661 struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); in gfs2_quota_unhold()
737 struct inode *inode = &ip->i_inode; in gfs2_write_buf_to_page()
[all …]
Drgrp.c603 struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); in check_and_update_goal()
687 struct inode *inode = &ip->i_inode; in gfs2_rs_delete()
897 struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); in read_rindex_entry()
903 if (pos >= i_size_read(&ip->i_inode)) in read_rindex_entry()
997 struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); in gfs2_ri_update()
1571 struct inode *inode = &ip->i_inode; in rg_mblk_search()
1989 const struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); in gfs2_orlov_skip()
2047 struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); in gfs2_inplace_reserve()
2072 if (S_ISDIR(ip->i_inode.i_mode) && (ap->aflags & GFS2_AF_ORLOV)) in gfs2_inplace_reserve()
2422 struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); in gfs2_alloc_blocks()
[all …]
Dmeta_io.c406 struct address_space *mapping = ip->i_inode.i_mapping; in gfs2_getjdatabuf()
407 struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); in gfs2_getjdatabuf()
441 struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); in gfs2_journal_wipe()
488 struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); in gfs2_meta_buffer()
Dbmap.h33 const struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); in gfs2_write_calc_reserv()
Dlog.h60 struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); in gfs2_ordered_add_inode()
Dincore.h385 struct inode i_inode; member
414 return container_of(inode, struct gfs2_inode, i_inode); in GFS2_I()
870 return GFS2_SB(&ip->i_inode)->sd_sb.sb_bsize - sizeof(struct gfs2_dinode); in gfs2_max_stuffed_size()
Dfile.c594 !IS_NOATIME(&ip->i_inode)) { in gfs2_mmap()
680 if (S_ISREG(ip->i_inode.i_mode)) { in gfs2_open()
907 if (iocb->ki_pos + iov_iter_count(from) > i_size_read(&ip->i_inode)) in gfs2_file_direct_write()
1226 const struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); in calc_max_reserv()
1498 error = gfs2_glock_get(GFS2_SB(&ip->i_inode), ip->i_no_addr, in do_flock()
1521 gfs2_assert_warn(GFS2_SB(&ip->i_inode), !error); in do_flock()
Dlog.c734 if (ip->i_inode.i_mapping->nrpages == 0) { in gfs2_ordered_write()
740 filemap_fdatawrite(ip->i_inode.i_mapping); in gfs2_ordered_write()
755 if (ip->i_inode.i_mapping->nrpages == 0) in gfs2_ordered_wait()
758 filemap_fdatawait(ip->i_inode.i_mapping); in gfs2_ordered_wait()
766 struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); in gfs2_ordered_del_inode()
Dinode.h42 return S_ISDIR(ip->i_inode.i_mode); in gfs2_is_dir()
Dmain.c38 inode_init_once(&ip->i_inode); in gfs2_init_inode_once()
Dsuper.c399 const struct inode *inode = &ip->i_inode; in gfs2_dinode_out()
1191 struct inode *inode = &ip->i_inode; in gfs2_final_release_pages()
1211 struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); in gfs2_dinode_dealloc()
1216 if (gfs2_get_inode_blocks(&ip->i_inode) != 1) { in gfs2_dinode_dealloc()
1571 return &ip->i_inode; in gfs2_alloc_inode()
Daops.c415 u64 dsize = i_size_read(&ip->i_inode); in stuffed_readpage()
485 struct address_space *mapping = ip->i_inode.i_mapping; in gfs2_internal_read()
Dglock.c951 if (ip && !igrab(&ip->i_inode)) in gfs2_try_evict()
957 d_prune_aliases(&ip->i_inode); in gfs2_try_evict()
958 iput(&ip->i_inode); in gfs2_try_evict()
965 if (!igrab(&ip->i_inode)) in gfs2_try_evict()
971 iput(&ip->i_inode); in gfs2_try_evict()
Dutil.c447 struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); in gfs2_consist_inode_i()
Drecovery.c42 error = gfs2_get_extent(&ip->i_inode, blk, &dblock, &extlen); in gfs2_replay_read_block()
Dtrace_gfs2.h528 (iomap->addr >> ip->i_inode.i_blkbits);