/fs/notify/fanotify/ |
D | fanotify.h | 147 static inline void fanotify_init_event(struct fanotify_event *event, in fanotify_init_event() argument 150 fsnotify_init_event(&event->fse, id); in fanotify_init_event() 151 event->mask = mask; in fanotify_init_event() 152 event->pid = NULL; in fanotify_init_event() 164 FANOTIFY_FE(struct fanotify_event *event) in FANOTIFY_FE() argument 166 return container_of(event, struct fanotify_fid_event, fae); in FANOTIFY_FE() 176 FANOTIFY_NE(struct fanotify_event *event) in FANOTIFY_NE() argument 178 return container_of(event, struct fanotify_name_event, fae); in FANOTIFY_NE() 181 static inline __kernel_fsid_t *fanotify_event_fsid(struct fanotify_event *event) in fanotify_event_fsid() argument 183 if (event->type == FANOTIFY_EVENT_TYPE_FID) in fanotify_event_fsid() [all …]
|
D | fanotify.c | 136 static int fanotify_merge(struct list_head *list, struct fsnotify_event *event) in fanotify_merge() argument 142 pr_debug("%s: list=%p event=%p\n", __func__, list, event); in fanotify_merge() 143 new = FANOTIFY_E(event); in fanotify_merge() 156 if (fanotify_should_merge(test_event, event)) { in fanotify_merge() 173 struct fanotify_perm_event *event, in fanotify_get_response() argument 178 pr_debug("%s: group=%p event=%p\n", __func__, group, event); in fanotify_get_response() 181 event->state == FAN_EVENT_ANSWERED); in fanotify_get_response() 186 if (event->state == FAN_EVENT_REPORTED) { in fanotify_get_response() 188 event->state = FAN_EVENT_CANCELED; in fanotify_get_response() 193 if (event->state == FAN_EVENT_INIT) in fanotify_get_response() [all …]
|
D | fanotify_user.c | 68 struct fanotify_event *event) in fanotify_event_info_len() argument 70 struct fanotify_info *info = fanotify_event_info(event); in fanotify_event_info_len() 71 int dir_fh_len = fanotify_event_dir_fh_len(event); in fanotify_event_info_len() 72 int fh_len = fanotify_event_object_fh_len(event); in fanotify_event_info_len() 78 } else if ((fid_mode & FAN_REPORT_NAME) && (event->mask & FAN_ONDIR)) { in fanotify_event_info_len() 102 struct fanotify_event *event = NULL; in get_one_event() local 117 event = ERR_PTR(-EINVAL); in get_one_event() 120 event = FANOTIFY_E(fsnotify_remove_first_event(group)); in get_one_event() 121 if (fanotify_is_perm_event(event->mask)) in get_one_event() 122 FANOTIFY_PERM(event)->state = FAN_EVENT_REPORTED; in get_one_event() [all …]
|
D | Kconfig | 11 the event.
|
/fs/notify/ |
D | notification.c | 58 struct fsnotify_event *event) in fsnotify_destroy_event() argument 61 if (!event || event == group->overflow_event) in fsnotify_destroy_event() 69 if (!list_empty(&event->list)) { in fsnotify_destroy_event() 71 WARN_ON(!list_empty(&event->list)); in fsnotify_destroy_event() 74 group->ops->free_event(event); in fsnotify_destroy_event() 85 struct fsnotify_event *event, in fsnotify_add_event() argument 92 pr_debug("%s: group=%p event=%p\n", __func__, group, event); in fsnotify_add_event() 101 if (event == group->overflow_event || in fsnotify_add_event() 109 event = group->overflow_event; in fsnotify_add_event() 114 ret = merge(list, event); in fsnotify_add_event() [all …]
|
/fs/fscache/ |
D | object.c | 173 int event = -1; in fscache_object_sm_dispatcher() local 196 event = fls(events & t->events) - 1; in fscache_object_sm_dispatcher() 197 __clear_bit(event, &object->oob_event_mask); in fscache_object_sm_dispatcher() 198 clear_bit(event, &object->events); in fscache_object_sm_dispatcher() 211 event = fls(events & t->events) - 1; in fscache_object_sm_dispatcher() 213 true, false, event); in fscache_object_sm_dispatcher() 214 clear_bit(event, &object->events); in fscache_object_sm_dispatcher() 216 object->debug_id, event, in fscache_object_sm_dispatcher() 233 trace_fscache_osm(object, state, false, oob, event); in fscache_object_sm_dispatcher() 234 new_state = state->work(object, event); in fscache_object_sm_dispatcher() [all …]
|
D | internal.h | 309 unsigned event) in fscache_raise_event() argument 311 BUG_ON(event >= NR_FSCACHE_OBJECT_EVENTS); in fscache_raise_event() 314 object->debug_id, object->event_mask, (1 << event)); in fscache_raise_event() 316 if (!test_and_set_bit(event, &object->events) && in fscache_raise_event() 317 test_bit(event, &object->event_mask)) in fscache_raise_event()
|
/fs/notify/inotify/ |
D | inotify_fsnotify.c | 50 struct fsnotify_event *event) in inotify_merge() argument 55 return event_compare(last_event, event); in inotify_merge() 63 struct inotify_event_info *event; in inotify_handle_inode_event() local 95 event = kmalloc(alloc_len, GFP_KERNEL_ACCOUNT | __GFP_RETRY_MAYFAIL); in inotify_handle_inode_event() 98 if (unlikely(!event)) { in inotify_handle_inode_event() 116 fsn_event = &event->fse; in inotify_handle_inode_event() 118 event->mask = mask; in inotify_handle_inode_event() 119 event->wd = wd; in inotify_handle_inode_event() 120 event->sync_cookie = cookie; in inotify_handle_inode_event() 121 event->name_len = len; in inotify_handle_inode_event() [all …]
|
D | inotify_user.c | 119 struct inotify_event_info *event; in round_event_name_len() local 121 event = INOTIFY_E(fsn_event); in round_event_name_len() 122 if (!event->name_len) in round_event_name_len() 124 return roundup(event->name_len + 1, sizeof(struct inotify_event)); in round_event_name_len() 138 struct fsnotify_event *event; in get_one_event() local 143 event = fsnotify_peek_first_event(group); in get_one_event() 145 pr_debug("%s: group=%p event=%p\n", __func__, group, event); in get_one_event() 147 event_size += round_event_name_len(event); in get_one_event() 155 return event; in get_one_event() 169 struct inotify_event_info *event; in copy_event_to_user() local [all …]
|
/fs/ |
D | eventpoll.c | 176 struct epoll_event event; member 894 pt->_key = epi->event.events; in ep_item_poll() 896 return vfs_poll(epi->ffd.file, pt) & epi->event.events; in ep_item_poll() 904 locked) & epi->event.events; in ep_item_poll() 963 epi->ffd.fd, epi->event.events, in ep_show_fdinfo() 964 (long long)epi->event.data, in ep_show_fdinfo() 1234 if (!(epi->event.events & ~EP_PRIVATE_BITS)) in ep_poll_callback() 1243 if (pollflags && !(pollflags & epi->event.events)) in ep_poll_callback() 1266 if ((epi->event.events & EPOLLEXCLUSIVE) && in ep_poll_callback() 1270 if (epi->event.events & EPOLLIN) in ep_poll_callback() [all …]
|
D | proc_namespace.c | 27 int event; in mounts_poll() local 31 event = READ_ONCE(ns->event); in mounts_poll() 32 if (m->poll_event != event) { in mounts_poll() 33 m->poll_event = event; in mounts_poll() 277 m->poll_event = ns->event; in mounts_open_common()
|
D | userfaultfd.c | 208 msg.event = UFFD_EVENT_PAGEFAULT; in userfault_msg() 602 if (ewq->msg.event == 0) in userfaultfd_event_wait_completion() 613 if (ewq->msg.event == UFFD_EVENT_FORK) { in userfaultfd_event_wait_completion() 664 ewq->msg.event = 0; in userfaultfd_event_complete() 730 ewq.msg.event = UFFD_EVENT_FORK; in dup_fctx() 786 ewq.msg.event = UFFD_EVENT_REMAP; in mremap_userfaultfd_complete() 812 ewq.msg.event = UFFD_EVENT_REMOVE; in userfaultfd_remove() 871 ewq.msg.event = UFFD_EVENT_UNMAP; in userfaultfd_unmap_complete() 1110 if (uwq->msg.event == UFFD_EVENT_FORK) { in userfaultfd_ctx_read() 1149 if (!ret && msg->event == UFFD_EVENT_FORK) { in userfaultfd_ctx_read()
|
D | aio.c | 1104 struct io_event *ev_page, *event; in aio_complete() local 1122 event = ev_page + pos % AIO_EVENTS_PER_PAGE; in aio_complete() 1124 *event = iocb->ki_res; in aio_complete() 1186 struct io_event __user *event, long nr) in aio_read_events_ring() argument 1239 copy_ret = copy_to_user(event + ret, ev + pos, in aio_read_events_ring() 1266 struct io_event __user *event, long *i) in aio_read_events() argument 1268 long ret = aio_read_events_ring(ctx, event + *i, nr - *i); in aio_read_events() 1283 struct io_event __user *event, in read_events() argument 1303 aio_read_events(ctx, min_nr, nr, event, &ret); in read_events() 1306 aio_read_events(ctx, min_nr, nr, event, &ret), in read_events()
|
D | mount.h | 23 u64 event; member
|
D | namespace.c | 65 static u64 event; variable 801 ns->event = ++event; in touch_mnt_namespace() 811 if (ns && ns->event != event) { in __touch_mnt_namespace() 812 ns->event = event; in __touch_mnt_namespace() 1635 event++; in do_umount() 1676 event++; in __detach_mounts()
|
/fs/ocfs2/cluster/ |
D | heartbeat.c | 754 struct o2hb_node_event *event; in o2hb_run_event_list() local 764 event = list_entry(o2hb_node_events.next, in o2hb_run_event_list() 767 list_del_init(&event->hn_item); in o2hb_run_event_list() 771 event->hn_event_type == O2HB_NODE_UP_CB ? "UP" : "DOWN", in o2hb_run_event_list() 772 event->hn_node_num); in o2hb_run_event_list() 774 hbcall = hbcall_from_type(event->hn_event_type); in o2hb_run_event_list() 781 o2hb_fire_callbacks(hbcall, event->hn_node, event->hn_node_num); in o2hb_run_event_list() 790 static void o2hb_queue_node_event(struct o2hb_node_event *event, in o2hb_queue_node_event() argument 799 event->hn_event_type = type; in o2hb_queue_node_event() 800 event->hn_node = node; in o2hb_queue_node_event() [all …]
|
/fs/kernfs/ |
D | file.c | 37 atomic_t event; member 166 of->event = atomic_read(&of->kn->attr.open->event); in kernfs_seq_show() 210 of->event = atomic_read(&of->kn->attr.open->event); in kernfs_file_read_iter() 557 atomic_set(&new_on->event, 1); in kernfs_get_open_node() 822 if (of->event != atomic_read(&on->event)) in kernfs_generic_poll() 928 atomic_inc(&on->event); in kernfs_notify()
|
/fs/cifs/ |
D | smbdirect.c | 176 struct rdma_cm_id *id, struct rdma_cm_event *event) in smbd_conn_upcall() argument 181 event->event, event->status); in smbd_conn_upcall() 183 switch (event->event) { in smbd_conn_upcall() 201 log_rdma_event(INFO, "connected event=%d\n", event->event); in smbd_conn_upcall() 209 log_rdma_event(INFO, "connecting failed event=%d\n", event->event); in smbd_conn_upcall() 238 smbd_qp_async_error_upcall(struct ib_event *event, void *context) in smbd_qp_async_error_upcall() argument 243 ib_event_msg(event->event), event->device->name, info); in smbd_qp_async_error_upcall() 245 switch (event->event) { in smbd_qp_async_error_upcall()
|
/fs/lockd/ |
D | svc.c | 300 unsigned long event, void *ptr) in lockd_inetaddr_event() argument 305 if ((event != NETDEV_DOWN) || in lockd_inetaddr_event() 330 unsigned long event, void *ptr) in lockd_inet6addr_event() argument 335 if ((event != NETDEV_DOWN) || in lockd_inet6addr_event()
|
/fs/ocfs2/dlmfs/ |
D | dlmfs.c | 210 __poll_t event = 0; in dlmfs_file_poll() local 218 event = EPOLLIN | EPOLLRDNORM; in dlmfs_file_poll() 221 return event; in dlmfs_file_poll()
|
/fs/nfs/blocklayout/ |
D | rpc_pipefs.c | 165 static int rpc_pipefs_event(struct notifier_block *nb, unsigned long event, in rpc_pipefs_event() argument 182 switch (event) { in rpc_pipefs_event()
|
/fs/nfsd/ |
D | nfssvc.c | 439 static int nfsd_inetaddr_event(struct notifier_block *this, unsigned long event, in nfsd_inetaddr_event() argument 448 if ((event != NETDEV_DOWN) || in nfsd_inetaddr_event() 471 unsigned long event, void *ptr) in nfsd_inet6addr_event() argument 479 if ((event != NETDEV_DOWN) || in nfsd_inet6addr_event()
|
/fs/nfs/ |
D | dns_resolve.c | 425 static int rpc_pipefs_event(struct notifier_block *nb, unsigned long event, in rpc_pipefs_event() argument 440 switch (event) { in rpc_pipefs_event()
|
/fs/jfs/ |
D | jfs_txnmgr.c | 118 static inline void TXN_SLEEP_DROP_LOCK(wait_queue_head_t * event) in TXN_SLEEP_DROP_LOCK() argument 122 add_wait_queue(event, &wait); in TXN_SLEEP_DROP_LOCK() 126 remove_wait_queue(event, &wait); in TXN_SLEEP_DROP_LOCK() 129 #define TXN_SLEEP(event)\ argument 131 TXN_SLEEP_DROP_LOCK(event);\ 135 #define TXN_WAKEUP(event) wake_up_all(event) argument
|
/fs/proc/ |
D | proc_sysctl.c | 59 atomic_inc(&poll->event); in proc_sys_poll_notify() 644 unsigned long event; in proc_sys_poll() local 656 event = (unsigned long)filp->private_data; in proc_sys_poll() 659 if (event != atomic_read(&table->poll->event)) { in proc_sys_poll()
|