Lines Matching +full:9 +full:v
147 bch2_fs_inconsistent_on((s64) inode->v.i_blocks + sectors < 0, c, in __bch2_i_sectors_acct()
149 inode->v.i_ino, (u64) inode->v.i_blocks, sectors, in __bch2_i_sectors_acct()
151 inode->v.i_blocks += sectors; in __bch2_i_sectors_acct()
223 struct bch_fs *c = inode->v.i_sb->s_fs_info; in bch2_fsync()
231 ret = sync_inode_metadata(&inode->v, 1); in bch2_fsync()
263 struct bch_fs *c = inode->v.i_sb->s_fs_info; in __bch2_truncate_folio()
264 struct address_space *mapping = inode->v.i_mapping; in __bch2_truncate_folio()
281 POS(inode->v.i_ino, (index << PAGE_SECTORS_SHIFT)), in __bch2_truncate_folio()
282 POS(inode->v.i_ino, (index << PAGE_SECTORS_SHIFT) + PAGE_SECTORS)); in __bch2_truncate_folio()
323 for (i = round_up(start_offset, block_bytes(c)) >> 9; in __bch2_truncate_folio()
324 i < round_down(end_offset, block_bytes(c)) >> 9; in __bch2_truncate_folio()
344 WARN_ON_ONCE(folio_pos(folio) >= inode->v.i_size); in __bch2_truncate_folio()
346 if (inode->v.i_size > folio_pos(folio)) in __bch2_truncate_folio()
347 end_pos = min_t(u64, inode->v.i_size, end_pos); in __bch2_truncate_folio()
399 struct address_space *mapping = inode->v.i_mapping; in bch2_extend()
411 truncate_setsize(&inode->v, iattr->ia_size); in bch2_extend()
419 struct bch_fs *c = inode->v.i_sb->s_fs_info; in bchfs_truncate()
420 struct address_space *mapping = inode->v.i_mapping; in bchfs_truncate()
430 if (iattr->ia_size != inode->v.i_size) { in bchfs_truncate()
438 inode_dio_wait(&inode->v); in bchfs_truncate()
455 inode->v.i_size < inode_u.bi_size, in bchfs_truncate()
457 (u64) inode->v.i_size, inode_u.bi_size); in bchfs_truncate()
459 if (iattr->ia_size > inode->v.i_size) { in bchfs_truncate()
471 truncate_setsize(&inode->v, iattr->ia_size); in bchfs_truncate()
505 bch2_fs_inconsistent_on(!inode->v.i_size && inode->v.i_blocks && in bchfs_truncate()
508 inode->v.i_ino, (u64) inode->v.i_blocks, in bchfs_truncate()
523 struct bch_fs *c = inode->v.i_sb->s_fs_info; in inode_update_times_fn()
531 struct bch_fs *c = inode->v.i_sb->s_fs_info; in bchfs_fpunch()
544 truncate_pagecache_range(&inode->v, offset, end - 1); in bchfs_fpunch()
550 block_start >> 9, block_end >> 9, in bchfs_fpunch()
556 if (end >= inode->v.i_size && !truncated_last_page) { in bchfs_fpunch()
557 ret = bch2_write_inode_size(c, inode, inode->v.i_size, in bchfs_fpunch()
572 struct bch_fs *c = inode->v.i_sb->s_fs_info; in bchfs_fcollapse_finsert()
573 struct address_space *mapping = inode->v.i_mapping; in bchfs_fcollapse_finsert()
581 if (offset >= inode->v.i_size) in bchfs_fcollapse_finsert()
584 if (offset + len >= inode->v.i_size) in bchfs_fcollapse_finsert()
593 i_size_write(&inode->v, inode->v.i_size + len); in bchfs_fcollapse_finsert()
595 ret = bch2_fcollapse_finsert(c, inode_inum(inode), offset >> 9, len >> 9, in bchfs_fcollapse_finsert()
598 i_size_write(&inode->v, inode->v.i_size - len); in bchfs_fcollapse_finsert()
607 struct bch_fs *c = inode->v.i_sb->s_fs_info; in __bchfs_fallocate()
610 struct bpos end_pos = POS(inode->v.i_ino, end_sector); in __bchfs_fallocate()
617 POS(inode->v.i_ino, start_sector), in __bchfs_fallocate()
667 if (bch2_clamp_data_hole(&inode->v, in __bchfs_fallocate()
672 (bch2_clamp_data_hole(&inode->v, in __bchfs_fallocate()
737 struct bch_fs *c = inode->v.i_sb->s_fs_info; in bchfs_fallocate()
744 if (!(mode & FALLOC_FL_KEEP_SIZE) && end > inode->v.i_size) { in bchfs_fallocate()
745 ret = inode_newsize_ok(&inode->v, end); in bchfs_fallocate()
757 truncate_pagecache_range(&inode->v, offset, end - 1); in bchfs_fallocate()
763 ret = __bchfs_fallocate(inode, mode, block_start >> 9, block_end >> 9); in bchfs_fallocate()
773 if (mode & FALLOC_FL_KEEP_SIZE && end > inode->v.i_size) in bchfs_fallocate()
774 end = inode->v.i_size; in bchfs_fallocate()
776 if (end >= inode->v.i_size && in bchfs_fallocate()
779 spin_lock(&inode->v.i_lock); in bchfs_fallocate()
780 i_size_write(&inode->v, end); in bchfs_fallocate()
781 spin_unlock(&inode->v.i_lock); in bchfs_fallocate()
795 struct bch_fs *c = inode->v.i_sb->s_fs_info; in bch2_fallocate_dispatch()
801 inode_lock(&inode->v); in bch2_fallocate_dispatch()
802 inode_dio_wait(&inode->v); in bch2_fallocate_dispatch()
821 inode_unlock(&inode->v); in bch2_fallocate_dispatch()
835 struct bch_fs *c = inode->v.i_sb->s_fs_info; in quota_reserve_range()
841 POS(inode->v.i_ino, start), in quota_reserve_range()
842 POS(inode->v.i_ino, end - 1), in quota_reserve_range()
863 struct bch_fs *c = src->v.i_sb->s_fs_info; in bch2_remap_file_range()
880 lock_two_nondirectories(&src->v, &dst->v); in bch2_remap_file_range()
883 inode_dio_wait(&src->v); in bch2_remap_file_range()
884 inode_dio_wait(&dst->v); in bch2_remap_file_range()
894 ret = bch2_write_invalidate_inode_pages_range(dst->v.i_mapping, in bch2_remap_file_range()
899 ret = quota_reserve_range(dst, "a_res, pos_dst >> 9, in bch2_remap_file_range()
900 (pos_dst + aligned_len) >> 9); in bch2_remap_file_range()
907 bch2_mark_pagecache_unallocated(src, pos_src >> 9, in bch2_remap_file_range()
908 (pos_src + aligned_len) >> 9); in bch2_remap_file_range()
917 inode_inum(dst), pos_dst >> 9, in bch2_remap_file_range()
918 inode_inum(src), pos_src >> 9, in bch2_remap_file_range()
919 aligned_len >> 9, in bch2_remap_file_range()
928 ret = min((u64) ret << 9, (u64) len); in bch2_remap_file_range()
932 spin_lock(&dst->v.i_lock); in bch2_remap_file_range()
933 if (pos_dst + ret > dst->v.i_size) in bch2_remap_file_range()
934 i_size_write(&dst->v, pos_dst + ret); in bch2_remap_file_range()
935 spin_unlock(&dst->v.i_lock); in bch2_remap_file_range()
943 unlock_two_nondirectories(&src->v, &dst->v); in bch2_remap_file_range()
953 struct bch_fs *c = inode->v.i_sb->s_fs_info; in bch2_seek_data()
957 isize = i_size_read(&inode->v); in bch2_seek_data()
963 POS(inode->v.i_ino, offset >> 9), in bch2_seek_data()
964 POS(inode->v.i_ino, U64_MAX), in bch2_seek_data()
967 next_data = max(offset, bkey_start_offset(k.k) << 9); in bch2_seek_data()
969 } else if (k.k->p.offset >> 9 > isize) in bch2_seek_data()
977 next_data = bch2_seek_pagecache_data(&inode->v, in bch2_seek_data()
989 struct bch_fs *c = inode->v.i_sb->s_fs_info; in bch2_seek_hole()
993 isize = i_size_read(&inode->v); in bch2_seek_hole()
999 POS(inode->v.i_ino, offset >> 9), in bch2_seek_hole()
1000 POS(inode->v.i_ino, U64_MAX), in bch2_seek_hole()
1002 if (k.k->p.inode != inode->v.i_ino) { in bch2_seek_hole()
1003 next_hole = bch2_seek_pagecache_hole(&inode->v, in bch2_seek_hole()
1007 next_hole = bch2_seek_pagecache_hole(&inode->v, in bch2_seek_hole()
1008 max(offset, bkey_start_offset(k.k) << 9), in bch2_seek_hole()
1009 k.k->p.offset << 9, 0, false); in bch2_seek_hole()
1011 if (next_hole < k.k->p.offset << 9) in bch2_seek_hole()
1014 offset = max(offset, bkey_start_offset(k.k) << 9); in bch2_seek_hole()