4 * Copyright (C) 1992 Rick Sladkey
7 #include <linux/file.h>
8 #include <linux/falloc.h>
9 #include <linux/nfs_fs.h>
10 #include <uapi/linux/btrfs.h> /* BTRFS_IOC_CLONE/BTRFS_IOC_CLONE_RANGE */
11 #include "delegation.h"
19 #ifdef CONFIG_NFS_V4_2
23 #define NFSDBG_FACILITY NFSDBG_FILE
26 nfs4_file_open(struct inode *inode, struct file *filp)
28 struct nfs_open_context *ctx;
29 struct dentry *dentry = filp->f_path.dentry;
30 struct dentry *parent = NULL;
32 unsigned openflags = filp->f_flags;
37 * If no cached dentry exists or if it's negative, NFSv4 handled the
38 * opens in ->lookup() or ->create().
40 * We only get this far for a cached positive dentry. We skipped
41 * revalidation, so handle it here by dropping the dentry and returning
42 * -EOPENSTALE. The VFS will retry the lookup/create/open.
45 dprintk("NFS: open file(%pd2)\n", dentry);
47 err = nfs_check_flags(openflags);
51 if ((openflags & O_ACCMODE) == 3)
54 /* We can't create new files here */
55 openflags &= ~(O_CREAT|O_EXCL);
57 parent = dget_parent(dentry);
58 dir = d_inode(parent);
60 ctx = alloc_nfs_open_context(filp->f_path.dentry, filp->f_mode);
65 attr.ia_valid = ATTR_OPEN;
66 if (openflags & O_TRUNC) {
67 attr.ia_valid |= ATTR_SIZE;
69 nfs_sync_inode(inode);
72 inode = NFS_PROTO(dir)->open_context(dir, ctx, openflags, &attr, NULL);
86 if (inode != d_inode(dentry))
89 nfs_set_verifier(dentry, nfs_save_change_attribute(dir));
90 nfs_file_set_open_context(filp, ctx);
91 nfs_fscache_open_file(inode, filp);
95 put_nfs_open_context(ctx);
107 * Flush all dirty pages, and check for write errors.
110 nfs4_file_flush(struct file *file, fl_owner_t id)
112 struct inode *inode = file_inode(file);
114 dprintk("NFS: flush(%pD2)\n", file);
116 nfs_inc_stats(inode, NFSIOS_VFSFLUSH);
117 if ((file->f_mode & FMODE_WRITE) == 0)
121 * If we're holding a write delegation, then check if we're required
122 * to flush the i/o on close. If not, then just start the i/o now.
124 if (!nfs4_delegation_flush_on_close(inode))
125 return filemap_fdatawrite(file->f_mapping);
127 /* Flush writes to the server and return any errors */
128 return vfs_fsync(file, 0);
132 nfs4_file_fsync(struct file *file, loff_t start, loff_t end, int datasync)
135 struct inode *inode = file_inode(file);
137 trace_nfs_fsync_enter(inode);
139 nfs_inode_dio_wait(inode);
141 ret = filemap_write_and_wait_range(inode->i_mapping, start, end);
144 mutex_lock(&inode->i_mutex);
145 ret = nfs_file_fsync_commit(file, start, end, datasync);
147 ret = pnfs_sync_inode(inode, !!datasync);
148 mutex_unlock(&inode->i_mutex);
150 * If nfs_file_fsync_commit detected a server reboot, then
151 * resend all dirty pages that might have been covered by
152 * the NFS_CONTEXT_RESEND_WRITES flag
156 } while (ret == -EAGAIN);
158 trace_nfs_fsync_exit(inode, ret);
162 #ifdef CONFIG_NFS_V4_2
163 static loff_t nfs4_file_llseek(struct file *filep, loff_t offset, int whence)
170 ret = nfs42_proc_llseek(filep, offset, whence);
171 if (ret != -ENOTSUPP)
174 return nfs_file_llseek(filep, offset, whence);
178 static long nfs42_fallocate(struct file *filep, int mode, loff_t offset, loff_t len)
180 struct inode *inode = file_inode(filep);
183 if (!S_ISREG(inode->i_mode))
186 if ((mode != 0) && (mode != (FALLOC_FL_PUNCH_HOLE | FALLOC_FL_KEEP_SIZE)))
189 ret = inode_newsize_ok(inode, offset + len);
193 if (mode & FALLOC_FL_PUNCH_HOLE)
194 return nfs42_proc_deallocate(filep, offset, len);
195 return nfs42_proc_allocate(filep, offset, len);
198 static int nfs42_clone_file_range(struct file *src_file, loff_t src_off,
199 struct file *dst_file, loff_t dst_off, u64 count)
201 struct inode *dst_inode = file_inode(dst_file);
202 struct nfs_server *server = NFS_SERVER(dst_inode);
203 struct inode *src_inode = file_inode(src_file);
204 unsigned int bs = server->clone_blksize;
205 bool same_inode = false;
208 /* check alignment w.r.t. clone_blksize */
211 if (!IS_ALIGNED(src_off, bs) || !IS_ALIGNED(dst_off, bs))
213 if (!IS_ALIGNED(count, bs) && i_size_read(src_inode) != (src_off + count))
217 if (src_inode == dst_inode)
220 /* XXX: do we lock at all? what if server needs CB_RECALL_LAYOUT? */
222 mutex_lock(&src_inode->i_mutex);
223 } else if (dst_inode < src_inode) {
224 mutex_lock_nested(&dst_inode->i_mutex, I_MUTEX_PARENT);
225 mutex_lock_nested(&src_inode->i_mutex, I_MUTEX_CHILD);
227 mutex_lock_nested(&src_inode->i_mutex, I_MUTEX_PARENT);
228 mutex_lock_nested(&dst_inode->i_mutex, I_MUTEX_CHILD);
231 /* flush all pending writes on both src and dst so that server
232 * has the latest data */
233 ret = nfs_sync_inode(src_inode);
236 ret = nfs_sync_inode(dst_inode);
240 ret = nfs42_proc_clone(src_file, dst_file, src_off, dst_off, count);
242 /* truncate inode page cache of the dst range so that future reads can fetch
243 * new data from server */
245 truncate_inode_pages_range(&dst_inode->i_data, dst_off, dst_off + count - 1);
249 mutex_unlock(&src_inode->i_mutex);
250 } else if (dst_inode < src_inode) {
251 mutex_unlock(&src_inode->i_mutex);
252 mutex_unlock(&dst_inode->i_mutex);
254 mutex_unlock(&dst_inode->i_mutex);
255 mutex_unlock(&src_inode->i_mutex);
260 #endif /* CONFIG_NFS_V4_2 */
262 const struct file_operations nfs4_file_operations = {
263 .read_iter = nfs_file_read,
264 .write_iter = nfs_file_write,
265 .mmap = nfs_file_mmap,
266 .open = nfs4_file_open,
267 .flush = nfs4_file_flush,
268 .release = nfs_file_release,
269 .fsync = nfs4_file_fsync,
272 .splice_read = nfs_file_splice_read,
273 .splice_write = iter_file_splice_write,
274 .check_flags = nfs_check_flags,
275 .setlease = simple_nosetlease,
276 #ifdef CONFIG_NFS_V4_2
277 .llseek = nfs4_file_llseek,
278 .fallocate = nfs42_fallocate,
279 .clone_file_range = nfs42_clone_file_range,
281 .llseek = nfs_file_llseek,