• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * (C) 2001 Clemson University and The University of Chicago
4  * Copyright 2018 Omnibond Systems, L.L.C.
5  *
6  * See COPYING in top-level directory.
7  */
8 
9 /*
10  *  Linux VFS file operations.
11  */
12 
13 #include "protocol.h"
14 #include "orangefs-kernel.h"
15 #include "orangefs-bufmap.h"
16 #include <linux/fs.h>
17 #include <linux/pagemap.h>
18 
flush_racache(struct inode * inode)19 static int flush_racache(struct inode *inode)
20 {
21 	struct orangefs_inode_s *orangefs_inode = ORANGEFS_I(inode);
22 	struct orangefs_kernel_op_s *new_op;
23 	int ret;
24 
25 	gossip_debug(GOSSIP_UTILS_DEBUG,
26 	    "%s: %pU: Handle is %pU | fs_id %d\n", __func__,
27 	    get_khandle_from_ino(inode), &orangefs_inode->refn.khandle,
28 	    orangefs_inode->refn.fs_id);
29 
30 	new_op = op_alloc(ORANGEFS_VFS_OP_RA_FLUSH);
31 	if (!new_op)
32 		return -ENOMEM;
33 	new_op->upcall.req.ra_cache_flush.refn = orangefs_inode->refn;
34 
35 	ret = service_operation(new_op, "orangefs_flush_racache",
36 	    get_interruptible_flag(inode));
37 
38 	gossip_debug(GOSSIP_UTILS_DEBUG, "%s: got return value of %d\n",
39 	    __func__, ret);
40 
41 	op_release(new_op);
42 	return ret;
43 }
44 
45 /*
46  * Post and wait for the I/O upcall to finish
47  */
wait_for_direct_io(enum ORANGEFS_io_type type,struct inode * inode,loff_t * offset,struct iov_iter * iter,size_t total_size,loff_t readahead_size,struct orangefs_write_range * wr,int * index_return)48 ssize_t wait_for_direct_io(enum ORANGEFS_io_type type, struct inode *inode,
49     loff_t *offset, struct iov_iter *iter, size_t total_size,
50     loff_t readahead_size, struct orangefs_write_range *wr, int *index_return)
51 {
52 	struct orangefs_inode_s *orangefs_inode = ORANGEFS_I(inode);
53 	struct orangefs_khandle *handle = &orangefs_inode->refn.khandle;
54 	struct orangefs_kernel_op_s *new_op = NULL;
55 	int buffer_index;
56 	ssize_t ret;
57 	size_t copy_amount;
58 
59 	new_op = op_alloc(ORANGEFS_VFS_OP_FILE_IO);
60 	if (!new_op)
61 		return -ENOMEM;
62 
63 	/* synchronous I/O */
64 	new_op->upcall.req.io.readahead_size = readahead_size;
65 	new_op->upcall.req.io.io_type = type;
66 	new_op->upcall.req.io.refn = orangefs_inode->refn;
67 
68 populate_shared_memory:
69 	/* get a shared buffer index */
70 	buffer_index = orangefs_bufmap_get();
71 	if (buffer_index < 0) {
72 		ret = buffer_index;
73 		gossip_debug(GOSSIP_FILE_DEBUG,
74 			     "%s: orangefs_bufmap_get failure (%zd)\n",
75 			     __func__, ret);
76 		goto out;
77 	}
78 	gossip_debug(GOSSIP_FILE_DEBUG,
79 		     "%s(%pU): GET op %p -> buffer_index %d\n",
80 		     __func__,
81 		     handle,
82 		     new_op,
83 		     buffer_index);
84 
85 	new_op->uses_shared_memory = 1;
86 	new_op->upcall.req.io.buf_index = buffer_index;
87 	new_op->upcall.req.io.count = total_size;
88 	new_op->upcall.req.io.offset = *offset;
89 	if (type == ORANGEFS_IO_WRITE && wr) {
90 		new_op->upcall.uid = from_kuid(&init_user_ns, wr->uid);
91 		new_op->upcall.gid = from_kgid(&init_user_ns, wr->gid);
92 	}
93 
94 	gossip_debug(GOSSIP_FILE_DEBUG,
95 		     "%s(%pU): offset: %llu total_size: %zd\n",
96 		     __func__,
97 		     handle,
98 		     llu(*offset),
99 		     total_size);
100 	/*
101 	 * Stage 1: copy the buffers into client-core's address space
102 	 */
103 	if (type == ORANGEFS_IO_WRITE && total_size) {
104 		ret = orangefs_bufmap_copy_from_iovec(iter, buffer_index,
105 		    total_size);
106 		if (ret < 0) {
107 			gossip_err("%s: Failed to copy-in buffers. Please make sure that the pvfs2-client is running. %ld\n",
108 			    __func__, (long)ret);
109 			goto out;
110 		}
111 	}
112 
113 	gossip_debug(GOSSIP_FILE_DEBUG,
114 		     "%s(%pU): Calling post_io_request with tag (%llu)\n",
115 		     __func__,
116 		     handle,
117 		     llu(new_op->tag));
118 
119 	/* Stage 2: Service the I/O operation */
120 	ret = service_operation(new_op,
121 				type == ORANGEFS_IO_WRITE ?
122 					"file_write" :
123 					"file_read",
124 				get_interruptible_flag(inode));
125 
126 	/*
127 	 * If service_operation() returns -EAGAIN #and# the operation was
128 	 * purged from orangefs_request_list or htable_ops_in_progress, then
129 	 * we know that the client was restarted, causing the shared memory
130 	 * area to be wiped clean.  To restart a  write operation in this
131 	 * case, we must re-copy the data from the user's iovec to a NEW
132 	 * shared memory location. To restart a read operation, we must get
133 	 * a new shared memory location.
134 	 */
135 	if (ret == -EAGAIN && op_state_purged(new_op)) {
136 		orangefs_bufmap_put(buffer_index);
137 		if (type == ORANGEFS_IO_WRITE)
138 			iov_iter_revert(iter, total_size);
139 		gossip_debug(GOSSIP_FILE_DEBUG,
140 			     "%s:going to repopulate_shared_memory.\n",
141 			     __func__);
142 		goto populate_shared_memory;
143 	}
144 
145 	if (ret < 0) {
146 		if (ret == -EINTR) {
147 			/*
148 			 * We can't return EINTR if any data was written,
149 			 * it's not POSIX. It is minimally acceptable
150 			 * to give a partial write, the way NFS does.
151 			 *
152 			 * It would be optimal to return all or nothing,
153 			 * but if a userspace write is bigger than
154 			 * an IO buffer, and the interrupt occurs
155 			 * between buffer writes, that would not be
156 			 * possible.
157 			 */
158 			switch (new_op->op_state - OP_VFS_STATE_GIVEN_UP) {
159 			/*
160 			 * If the op was waiting when the interrupt
161 			 * occurred, then the client-core did not
162 			 * trigger the write.
163 			 */
164 			case OP_VFS_STATE_WAITING:
165 				if (*offset == 0)
166 					ret = -EINTR;
167 				else
168 					ret = 0;
169 				break;
170 			/*
171 			 * If the op was in progress when the interrupt
172 			 * occurred, then the client-core was able to
173 			 * trigger the write.
174 			 */
175 			case OP_VFS_STATE_INPROGR:
176 				if (type == ORANGEFS_IO_READ)
177 					ret = -EINTR;
178 				else
179 					ret = total_size;
180 				break;
181 			default:
182 				gossip_err("%s: unexpected op state :%d:.\n",
183 					   __func__,
184 					   new_op->op_state);
185 				ret = 0;
186 				break;
187 			}
188 			gossip_debug(GOSSIP_FILE_DEBUG,
189 				     "%s: got EINTR, state:%d: %p\n",
190 				     __func__,
191 				     new_op->op_state,
192 				     new_op);
193 		} else {
194 			gossip_err("%s: error in %s handle %pU, returning %zd\n",
195 				__func__,
196 				type == ORANGEFS_IO_READ ?
197 					"read from" : "write to",
198 				handle, ret);
199 		}
200 		if (orangefs_cancel_op_in_progress(new_op))
201 			return ret;
202 
203 		goto out;
204 	}
205 
206 	/*
207 	 * Stage 3: Post copy buffers from client-core's address space
208 	 */
209 	if (type == ORANGEFS_IO_READ && new_op->downcall.resp.io.amt_complete) {
210 		/*
211 		 * NOTE: the iovector can either contain addresses which
212 		 *       can futher be kernel-space or user-space addresses.
213 		 *       or it can pointers to struct page's
214 		 */
215 
216 		/*
217 		 * When reading, readahead_size will only be zero when
218 		 * we're doing O_DIRECT, otherwise we got here from
219 		 * orangefs_readpage.
220 		 *
221 		 * If we got here from orangefs_readpage we want to
222 		 * copy either a page or the whole file into the io
223 		 * vector, whichever is smaller.
224 		 */
225 		if (readahead_size)
226 			copy_amount =
227 				min(new_op->downcall.resp.io.amt_complete,
228 					(__s64)PAGE_SIZE);
229 		else
230 			copy_amount = new_op->downcall.resp.io.amt_complete;
231 
232 		ret = orangefs_bufmap_copy_to_iovec(iter, buffer_index,
233 			copy_amount);
234 		if (ret < 0) {
235 			gossip_err("%s: Failed to copy-out buffers. Please make sure that the pvfs2-client is running (%ld)\n",
236 			    __func__, (long)ret);
237 			goto out;
238 		}
239 	}
240 	gossip_debug(GOSSIP_FILE_DEBUG,
241 	    "%s(%pU): Amount %s, returned by the sys-io call:%d\n",
242 	    __func__,
243 	    handle,
244 	    type == ORANGEFS_IO_READ ?  "read" : "written",
245 	    (int)new_op->downcall.resp.io.amt_complete);
246 
247 	ret = new_op->downcall.resp.io.amt_complete;
248 
249 out:
250 	if (buffer_index >= 0) {
251 		if ((readahead_size) && (type == ORANGEFS_IO_READ)) {
252 			/* readpage */
253 			*index_return = buffer_index;
254 			gossip_debug(GOSSIP_FILE_DEBUG,
255 				"%s: hold on to buffer_index :%d:\n",
256 				__func__, buffer_index);
257 		} else {
258 			/* O_DIRECT */
259 			orangefs_bufmap_put(buffer_index);
260 			gossip_debug(GOSSIP_FILE_DEBUG,
261 				"%s(%pU): PUT buffer_index %d\n",
262 				__func__, handle, buffer_index);
263 		}
264 	}
265 	op_release(new_op);
266 	return ret;
267 }
268 
orangefs_revalidate_mapping(struct inode * inode)269 int orangefs_revalidate_mapping(struct inode *inode)
270 {
271 	struct orangefs_inode_s *orangefs_inode = ORANGEFS_I(inode);
272 	struct address_space *mapping = inode->i_mapping;
273 	unsigned long *bitlock = &orangefs_inode->bitlock;
274 	int ret;
275 
276 	while (1) {
277 		ret = wait_on_bit(bitlock, 1, TASK_KILLABLE);
278 		if (ret)
279 			return ret;
280 		spin_lock(&inode->i_lock);
281 		if (test_bit(1, bitlock)) {
282 			spin_unlock(&inode->i_lock);
283 			continue;
284 		}
285 		if (!time_before(jiffies, orangefs_inode->mapping_time))
286 			break;
287 		spin_unlock(&inode->i_lock);
288 		return 0;
289 	}
290 
291 	set_bit(1, bitlock);
292 	smp_wmb();
293 	spin_unlock(&inode->i_lock);
294 
295 	unmap_mapping_range(mapping, 0, 0, 0);
296 	ret = filemap_write_and_wait(mapping);
297 	if (!ret)
298 		ret = invalidate_inode_pages2(mapping);
299 
300 	orangefs_inode->mapping_time = jiffies +
301 	    orangefs_cache_timeout_msecs*HZ/1000;
302 
303 	clear_bit(1, bitlock);
304 	smp_mb__after_atomic();
305 	wake_up_bit(bitlock, 1);
306 
307 	return ret;
308 }
309 
orangefs_file_read_iter(struct kiocb * iocb,struct iov_iter * iter)310 static ssize_t orangefs_file_read_iter(struct kiocb *iocb,
311     struct iov_iter *iter)
312 {
313 	int ret;
314 	orangefs_stats.reads++;
315 
316 	down_read(&file_inode(iocb->ki_filp)->i_rwsem);
317 	ret = orangefs_revalidate_mapping(file_inode(iocb->ki_filp));
318 	if (ret)
319 		goto out;
320 
321 	ret = generic_file_read_iter(iocb, iter);
322 out:
323 	up_read(&file_inode(iocb->ki_filp)->i_rwsem);
324 	return ret;
325 }
326 
orangefs_file_write_iter(struct kiocb * iocb,struct iov_iter * iter)327 static ssize_t orangefs_file_write_iter(struct kiocb *iocb,
328     struct iov_iter *iter)
329 {
330 	int ret;
331 	orangefs_stats.writes++;
332 
333 	if (iocb->ki_pos > i_size_read(file_inode(iocb->ki_filp))) {
334 		ret = orangefs_revalidate_mapping(file_inode(iocb->ki_filp));
335 		if (ret)
336 			return ret;
337 	}
338 
339 	ret = generic_file_write_iter(iocb, iter);
340 	return ret;
341 }
342 
orangefs_getflags(struct inode * inode,unsigned long * uval)343 static int orangefs_getflags(struct inode *inode, unsigned long *uval)
344 {
345 	__u64 val = 0;
346 	int ret;
347 
348 	ret = orangefs_inode_getxattr(inode,
349 				      "user.pvfs2.meta_hint",
350 				      &val, sizeof(val));
351 	if (ret < 0 && ret != -ENODATA)
352 		return ret;
353 	else if (ret == -ENODATA)
354 		val = 0;
355 	*uval = val;
356 	return 0;
357 }
358 
359 /*
360  * Perform a miscellaneous operation on a file.
361  */
orangefs_ioctl(struct file * file,unsigned int cmd,unsigned long arg)362 static long orangefs_ioctl(struct file *file, unsigned int cmd, unsigned long arg)
363 {
364 	struct inode *inode = file_inode(file);
365 	int ret = -ENOTTY;
366 	__u64 val = 0;
367 	unsigned long uval;
368 
369 	gossip_debug(GOSSIP_FILE_DEBUG,
370 		     "orangefs_ioctl: called with cmd %d\n",
371 		     cmd);
372 
373 	/*
374 	 * we understand some general ioctls on files, such as the immutable
375 	 * and append flags
376 	 */
377 	if (cmd == FS_IOC_GETFLAGS) {
378 		ret = orangefs_getflags(inode, &uval);
379 		if (ret)
380 			return ret;
381 		gossip_debug(GOSSIP_FILE_DEBUG,
382 			     "orangefs_ioctl: FS_IOC_GETFLAGS: %llu\n",
383 			     (unsigned long long)uval);
384 		return put_user(uval, (int __user *)arg);
385 	} else if (cmd == FS_IOC_SETFLAGS) {
386 		unsigned long old_uval;
387 
388 		ret = 0;
389 		if (get_user(uval, (int __user *)arg))
390 			return -EFAULT;
391 		/*
392 		 * ORANGEFS_MIRROR_FL is set internally when the mirroring mode
393 		 * is turned on for a file. The user is not allowed to turn
394 		 * on this bit, but the bit is present if the user first gets
395 		 * the flags and then updates the flags with some new
396 		 * settings. So, we ignore it in the following edit. bligon.
397 		 */
398 		if ((uval & ~ORANGEFS_MIRROR_FL) &
399 		    (~(FS_IMMUTABLE_FL | FS_APPEND_FL | FS_NOATIME_FL))) {
400 			gossip_err("orangefs_ioctl: the FS_IOC_SETFLAGS only supports setting one of FS_IMMUTABLE_FL|FS_APPEND_FL|FS_NOATIME_FL\n");
401 			return -EINVAL;
402 		}
403 		ret = orangefs_getflags(inode, &old_uval);
404 		if (ret)
405 			return ret;
406 		ret = vfs_ioc_setflags_prepare(inode, old_uval, uval);
407 		if (ret)
408 			return ret;
409 		val = uval;
410 		gossip_debug(GOSSIP_FILE_DEBUG,
411 			     "orangefs_ioctl: FS_IOC_SETFLAGS: %llu\n",
412 			     (unsigned long long)val);
413 		ret = orangefs_inode_setxattr(inode,
414 					      "user.pvfs2.meta_hint",
415 					      &val, sizeof(val), 0);
416 	}
417 
418 	return ret;
419 }
420 
orangefs_fault(struct vm_fault * vmf)421 static vm_fault_t orangefs_fault(struct vm_fault *vmf)
422 {
423 	struct file *file = vmf->vma->vm_file;
424 	int ret;
425 	ret = orangefs_inode_getattr(file->f_mapping->host,
426 	    ORANGEFS_GETATTR_SIZE);
427 	if (ret == -ESTALE)
428 		ret = -EIO;
429 	if (ret) {
430 		gossip_err("%s: orangefs_inode_getattr failed, "
431 		    "ret:%d:.\n", __func__, ret);
432 		return VM_FAULT_SIGBUS;
433 	}
434 	return filemap_fault(vmf);
435 }
436 
437 static const struct vm_operations_struct orangefs_file_vm_ops = {
438 	.fault = orangefs_fault,
439 	.map_pages = filemap_map_pages,
440 	.page_mkwrite = orangefs_page_mkwrite,
441 };
442 
443 /*
444  * Memory map a region of a file.
445  */
orangefs_file_mmap(struct file * file,struct vm_area_struct * vma)446 static int orangefs_file_mmap(struct file *file, struct vm_area_struct *vma)
447 {
448 	int ret;
449 
450 	ret = orangefs_revalidate_mapping(file_inode(file));
451 	if (ret)
452 		return ret;
453 
454 	gossip_debug(GOSSIP_FILE_DEBUG,
455 		     "orangefs_file_mmap: called on %s\n",
456 		     (file ?
457 			(char *)file->f_path.dentry->d_name.name :
458 			(char *)"Unknown"));
459 
460 	/* set the sequential readahead hint */
461 	vma->vm_flags |= VM_SEQ_READ;
462 	vma->vm_flags &= ~VM_RAND_READ;
463 
464 	file_accessed(file);
465 	vma->vm_ops = &orangefs_file_vm_ops;
466 	return 0;
467 }
468 
469 #define mapping_nrpages(idata) ((idata)->nrpages)
470 
471 /*
472  * Called to notify the module that there are no more references to
473  * this file (i.e. no processes have it open).
474  *
475  * \note Not called when each file is closed.
476  */
orangefs_file_release(struct inode * inode,struct file * file)477 static int orangefs_file_release(struct inode *inode, struct file *file)
478 {
479 	gossip_debug(GOSSIP_FILE_DEBUG,
480 		     "orangefs_file_release: called on %pD\n",
481 		     file);
482 
483 	/*
484 	 * remove all associated inode pages from the page cache and
485 	 * readahead cache (if any); this forces an expensive refresh of
486 	 * data for the next caller of mmap (or 'get_block' accesses)
487 	 */
488 	if (file_inode(file) &&
489 	    file_inode(file)->i_mapping &&
490 	    mapping_nrpages(&file_inode(file)->i_data)) {
491 		if (orangefs_features & ORANGEFS_FEATURE_READAHEAD) {
492 			gossip_debug(GOSSIP_INODE_DEBUG,
493 			    "calling flush_racache on %pU\n",
494 			    get_khandle_from_ino(inode));
495 			flush_racache(inode);
496 			gossip_debug(GOSSIP_INODE_DEBUG,
497 			    "flush_racache finished\n");
498 		}
499 
500 	}
501 	return 0;
502 }
503 
504 /*
505  * Push all data for a specific file onto permanent storage.
506  */
orangefs_fsync(struct file * file,loff_t start,loff_t end,int datasync)507 static int orangefs_fsync(struct file *file,
508 		       loff_t start,
509 		       loff_t end,
510 		       int datasync)
511 {
512 	int ret;
513 	struct orangefs_inode_s *orangefs_inode =
514 		ORANGEFS_I(file_inode(file));
515 	struct orangefs_kernel_op_s *new_op = NULL;
516 
517 	ret = filemap_write_and_wait_range(file_inode(file)->i_mapping,
518 	    start, end);
519 	if (ret < 0)
520 		return ret;
521 
522 	new_op = op_alloc(ORANGEFS_VFS_OP_FSYNC);
523 	if (!new_op)
524 		return -ENOMEM;
525 	new_op->upcall.req.fsync.refn = orangefs_inode->refn;
526 
527 	ret = service_operation(new_op,
528 			"orangefs_fsync",
529 			get_interruptible_flag(file_inode(file)));
530 
531 	gossip_debug(GOSSIP_FILE_DEBUG,
532 		     "orangefs_fsync got return value of %d\n",
533 		     ret);
534 
535 	op_release(new_op);
536 	return ret;
537 }
538 
539 /*
540  * Change the file pointer position for an instance of an open file.
541  *
542  * \note If .llseek is overriden, we must acquire lock as described in
543  *       Documentation/filesystems/locking.rst.
544  *
545  * Future upgrade could support SEEK_DATA and SEEK_HOLE but would
546  * require much changes to the FS
547  */
orangefs_file_llseek(struct file * file,loff_t offset,int origin)548 static loff_t orangefs_file_llseek(struct file *file, loff_t offset, int origin)
549 {
550 	int ret = -EINVAL;
551 	struct inode *inode = file_inode(file);
552 
553 	if (origin == SEEK_END) {
554 		/*
555 		 * revalidate the inode's file size.
556 		 * NOTE: We are only interested in file size here,
557 		 * so we set mask accordingly.
558 		 */
559 		ret = orangefs_inode_getattr(file->f_mapping->host,
560 		    ORANGEFS_GETATTR_SIZE);
561 		if (ret == -ESTALE)
562 			ret = -EIO;
563 		if (ret) {
564 			gossip_debug(GOSSIP_FILE_DEBUG,
565 				     "%s:%s:%d calling make bad inode\n",
566 				     __FILE__,
567 				     __func__,
568 				     __LINE__);
569 			return ret;
570 		}
571 	}
572 
573 	gossip_debug(GOSSIP_FILE_DEBUG,
574 		     "orangefs_file_llseek: offset is %ld | origin is %d"
575 		     " | inode size is %lu\n",
576 		     (long)offset,
577 		     origin,
578 		     (unsigned long)i_size_read(inode));
579 
580 	return generic_file_llseek(file, offset, origin);
581 }
582 
583 /*
584  * Support local locks (locks that only this kernel knows about)
585  * if Orangefs was mounted -o local_lock.
586  */
orangefs_lock(struct file * filp,int cmd,struct file_lock * fl)587 static int orangefs_lock(struct file *filp, int cmd, struct file_lock *fl)
588 {
589 	int rc = -EINVAL;
590 
591 	if (ORANGEFS_SB(file_inode(filp)->i_sb)->flags & ORANGEFS_OPT_LOCAL_LOCK) {
592 		if (cmd == F_GETLK) {
593 			rc = 0;
594 			posix_test_lock(filp, fl);
595 		} else {
596 			rc = posix_lock_file(filp, fl, NULL);
597 		}
598 	}
599 
600 	return rc;
601 }
602 
orangefs_flush(struct file * file,fl_owner_t id)603 static int orangefs_flush(struct file *file, fl_owner_t id)
604 {
605 	/*
606 	 * This is vfs_fsync_range(file, 0, LLONG_MAX, 0) without the
607 	 * service_operation in orangefs_fsync.
608 	 *
609 	 * Do not send fsync to OrangeFS server on a close.  Do send fsync
610 	 * on an explicit fsync call.  This duplicates historical OrangeFS
611 	 * behavior.
612 	 */
613 	struct inode *inode = file->f_mapping->host;
614 	int r;
615 
616 	if (inode->i_state & I_DIRTY_TIME) {
617 		spin_lock(&inode->i_lock);
618 		inode->i_state &= ~I_DIRTY_TIME;
619 		spin_unlock(&inode->i_lock);
620 		mark_inode_dirty_sync(inode);
621 	}
622 
623 	r = filemap_write_and_wait_range(file->f_mapping, 0, LLONG_MAX);
624 	if (r > 0)
625 		return 0;
626 	else
627 		return r;
628 }
629 
630 /** ORANGEFS implementation of VFS file operations */
631 const struct file_operations orangefs_file_operations = {
632 	.llseek		= orangefs_file_llseek,
633 	.read_iter	= orangefs_file_read_iter,
634 	.write_iter	= orangefs_file_write_iter,
635 	.lock		= orangefs_lock,
636 	.unlocked_ioctl	= orangefs_ioctl,
637 	.mmap		= orangefs_file_mmap,
638 	.open		= generic_file_open,
639 	.flush		= orangefs_flush,
640 	.release	= orangefs_file_release,
641 	.fsync		= orangefs_fsync,
642 };
643