/fs/nfsd/ |
D | nfs4recover.c | 174 struct nfsd_net *nn = net_generic(clp->net, nfsd_net_id); in nfsd4_create_clid_dir() local 180 if (!nn->rec_file) in nfsd4_create_clid_dir() 191 status = mnt_want_write_file(nn->rec_file); in nfsd4_create_clid_dir() 195 dir = nn->rec_file->f_path.dentry; in nfsd4_create_clid_dir() 220 if (nn->in_grace) { in nfsd4_create_clid_dir() 221 crp = nfs4_client_to_reclaim(dname, nn); in nfsd4_create_clid_dir() 225 vfs_fsync(nn->rec_file, 0); in nfsd4_create_clid_dir() 232 mnt_drop_write_file(nn->rec_file); in nfsd4_create_clid_dir() 267 nfsd4_list_rec_dir(recdir_func *f, struct nfsd_net *nn) in nfsd4_list_rec_dir() argument 270 struct dentry *dir = nn->rec_file->f_path.dentry; in nfsd4_list_rec_dir() [all …]
|
D | nfssvc.c | 177 struct nfsd_net *nn = net_generic(net, nfsd_net_id); in nfsd_nrthreads() local 180 if (nn->nfsd_serv) in nfsd_nrthreads() 181 rv = nn->nfsd_serv->sv_nrthreads; in nfsd_nrthreads() 189 struct nfsd_net *nn = net_generic(net, nfsd_net_id); in nfsd_init_socks() local 191 if (!list_empty(&nn->nfsd_serv->sv_permsocks)) in nfsd_init_socks() 194 error = svc_create_xprt(nn->nfsd_serv, "udp", net, PF_INET, NFS_PORT, in nfsd_init_socks() 199 error = svc_create_xprt(nn->nfsd_serv, "tcp", net, PF_INET, NFS_PORT, in nfsd_init_socks() 245 struct nfsd_net *nn = net_generic(net, nfsd_net_id); in nfsd_startup_net() local 248 if (nn->nfsd_net_up) in nfsd_startup_net() 264 nn->nfsd_net_up = true; in nfsd_startup_net() [all …]
|
D | nfs4state.c | 147 struct nfsd_net *nn = net_generic(clp->net, nfsd_net_id); in mark_client_expired() local 150 spin_lock(&nn->client_lock); in mark_client_expired() 152 spin_unlock(&nn->client_lock); in mark_client_expired() 168 struct nfsd_net *nn = net_generic(clp->net, nfsd_net_id); in renew_client_locked() local 182 list_move_tail(&clp->cl_lru, &nn->client_lru); in renew_client_locked() 189 struct nfsd_net *nn = net_generic(clp->net, nfsd_net_id); in renew_client() local 191 spin_lock(&nn->client_lock); in renew_client() 193 spin_unlock(&nn->client_lock); in renew_client() 206 struct nfsd_net *nn = net_generic(clp->net, nfsd_net_id); in put_client_renew() local 208 if (!atomic_dec_and_lock(&clp->cl_refcount, &nn->client_lock)) in put_client_renew() [all …]
|
D | nfsctl.c | 133 struct nfsd_net *nn = net_generic(net, nfsd_net_id); in exports_net_open() local 140 seq->private = nn->svc_export_cache; in exports_net_open() 542 struct nfsd_net *nn = net_generic(net, nfsd_net_id); in __write_versions() local 545 if (nn->nfsd_serv) in __write_versions() 680 struct nfsd_net *nn = net_generic(net, nfsd_net_id); in __write_ports_names() local 682 if (nn->nfsd_serv == NULL) in __write_ports_names() 684 return svc_xprt_names(nn->nfsd_serv, buf, SIMPLE_TRANSACTION_LIMIT); in __write_ports_names() 696 struct nfsd_net *nn = net_generic(net, nfsd_net_id); in __write_ports_addfd() local 706 err = svc_addsock(nn->nfsd_serv, fd, buf, SIMPLE_TRANSACTION_LIMIT); in __write_ports_addfd() 713 nn->nfsd_serv->sv_nrthreads--; in __write_ports_addfd() [all …]
|
D | nfs4idmap.c | 459 struct nfsd_net *nn = net_generic(net, nfsd_net_id); in nfsd_idmap_init() local 461 nn->idtoname_cache = cache_create_net(&idtoname_cache_template, net); in nfsd_idmap_init() 462 if (IS_ERR(nn->idtoname_cache)) in nfsd_idmap_init() 463 return PTR_ERR(nn->idtoname_cache); in nfsd_idmap_init() 464 rv = cache_register_net(nn->idtoname_cache, net); in nfsd_idmap_init() 467 nn->nametoid_cache = cache_create_net(&nametoid_cache_template, net); in nfsd_idmap_init() 468 if (IS_ERR(nn->nametoid_cache)) { in nfsd_idmap_init() 469 rv = PTR_ERR(nn->nametoid_cache); in nfsd_idmap_init() 472 rv = cache_register_net(nn->nametoid_cache, net); in nfsd_idmap_init() 478 cache_destroy_net(nn->nametoid_cache, net); in nfsd_idmap_init() [all …]
|
D | export.c | 839 struct nfsd_net *nn = net_generic(net, nfsd_net_id); in exp_rootfh() local 840 struct cache_detail *cd = nn->svc_export_cache; in exp_rootfh() 880 struct nfsd_net *nn = net_generic(cd->net, nfsd_net_id); in exp_find() local 881 struct svc_expkey *ek = exp_find_key(nn->svc_expkey_cache, clp, fsid_type, fsidv, reqp); in exp_find() 886 cache_put(&ek->h, nn->svc_expkey_cache); in exp_find() 928 struct nfsd_net *nn = net_generic(SVC_NET(rqstp), nfsd_net_id); in rqst_exp_get_by_name() local 929 struct cache_detail *cd = nn->svc_export_cache; in rqst_exp_get_by_name() 959 struct nfsd_net *nn = net_generic(SVC_NET(rqstp), nfsd_net_id); in rqst_exp_find() local 960 struct cache_detail *cd = nn->svc_export_cache; in rqst_exp_find() 1239 struct nfsd_net *nn = net_generic(net, nfsd_net_id); in nfsd_export_init() local [all …]
|
D | state.h | 464 struct nfsd_net *nn); 465 extern __be32 nfs4_check_open_reclaim(clientid_t *clid, bool sessions, struct nfsd_net *nn); 479 struct nfsd_net *nn); 480 extern bool nfs4_has_reclaimed_state(const char *name, struct nfsd_net *nn); 489 extern void nfsd4_record_grace_done(struct nfsd_net *nn, time_t boot_time);
|
D | netns.h | 108 #define nfsd_netns_ready(nn) ((nn)->sessionid_hashtbl) argument
|
D | nfs3xdr.c | 725 struct nfsd_net *nn = net_generic(SVC_NET(rqstp), nfsd_net_id); in nfs3svc_encode_writeres() local 731 *p++ = htonl(nn->nfssvc_boot.tv_sec); in nfs3svc_encode_writeres() 732 *p++ = htonl(nn->nfssvc_boot.tv_usec); in nfs3svc_encode_writeres() 1090 struct nfsd_net *nn = net_generic(SVC_NET(rqstp), nfsd_net_id); in nfs3svc_encode_commitres() local 1095 *p++ = htonl(nn->nfssvc_boot.tv_sec); in nfs3svc_encode_commitres() 1096 *p++ = htonl(nn->nfssvc_boot.tv_usec); in nfs3svc_encode_commitres()
|
D | nfs4proc.c | 328 struct nfsd_net *nn = net_generic(net, nfsd_net_id); in nfsd4_open() local 356 status = nfsd4_process_open1(&resp->cstate, open, nn); in nfsd4_open() 400 nn); in nfsd4_open() 541 struct nfsd_net *nn = net_generic(net, nfsd_net_id); in gen_boot_verifier() local 543 verf[0] = (__be32)nn->nfssvc_boot.tv_sec; in gen_boot_verifier() 544 verf[1] = (__be32)nn->nfssvc_boot.tv_usec; in gen_boot_verifier()
|
D | nfs4callback.c | 608 struct nfsd_net *nn = net_generic(net, nfsd_net_id); in max_cb_time() local 609 return max(nn->nfsd4_lease/10, (time_t)1) * HZ; in max_cb_time()
|
/fs/ocfs2/cluster/ |
D | tcp.c | 297 static u8 o2net_num_from_nn(struct o2net_node *nn) in o2net_num_from_nn() argument 299 BUG_ON(nn == NULL); in o2net_num_from_nn() 300 return nn - o2net_nodes; in o2net_num_from_nn() 305 static int o2net_prep_nsw(struct o2net_node *nn, struct o2net_status_wait *nsw) in o2net_prep_nsw() argument 309 spin_lock(&nn->nn_lock); in o2net_prep_nsw() 310 ret = idr_alloc(&nn->nn_status_idr, nsw, 0, 0, GFP_ATOMIC); in o2net_prep_nsw() 313 list_add_tail(&nsw->ns_node_item, &nn->nn_status_list); in o2net_prep_nsw() 315 spin_unlock(&nn->nn_lock); in o2net_prep_nsw() 325 static void o2net_complete_nsw_locked(struct o2net_node *nn, in o2net_complete_nsw_locked() argument 330 assert_spin_locked(&nn->nn_lock); in o2net_complete_nsw_locked() [all …]
|
/fs/nfs/ |
D | nfs4client.c | 29 struct nfs_net *nn = net_generic(clp->cl_net, nfs_net_id); in nfs_get_cb_ident_idr() local 34 spin_lock(&nn->nfs_client_lock); in nfs_get_cb_ident_idr() 35 ret = idr_alloc(&nn->cb_ident_idr, clp, 0, 0, GFP_NOWAIT); in nfs_get_cb_ident_idr() 38 spin_unlock(&nn->nfs_client_lock); in nfs_get_cb_ident_idr() 269 struct nfs_net *nn = net_generic(keep->cl_net, nfs_net_id); in nfs4_swap_callback_idents() local 278 spin_lock(&nn->nfs_client_lock); in nfs4_swap_callback_idents() 280 idr_replace(&nn->cb_ident_idr, keep, drop->cl_cb_ident); in nfs4_swap_callback_idents() 283 idr_replace(&nn->cb_ident_idr, drop, save); in nfs4_swap_callback_idents() 286 spin_unlock(&nn->nfs_client_lock); in nfs4_swap_callback_idents() 306 struct nfs_net *nn = net_generic(new->cl_net, nfs_net_id); in nfs40_walk_client_list() local [all …]
|
D | client.c | 202 struct nfs_net *nn = net_generic(net, nfs_net_id); in nfs_cleanup_cb_ident_idr() local 204 idr_destroy(&nn->cb_ident_idr); in nfs_cleanup_cb_ident_idr() 210 struct nfs_net *nn = net_generic(clp->cl_net, nfs_net_id); in nfs_cb_idr_remove_locked() local 213 idr_remove(&nn->cb_ident_idr, clp->cl_cb_ident); in nfs_cb_idr_remove_locked() 266 struct nfs_net *nn; in nfs_put_client() local 272 nn = net_generic(clp->cl_net, nfs_net_id); in nfs_put_client() 274 if (atomic_dec_and_lock(&clp->cl_count, &nn->nfs_client_lock)) { in nfs_put_client() 277 spin_unlock(&nn->nfs_client_lock); in nfs_put_client() 404 struct nfs_net *nn = net_generic(data->net, nfs_net_id); in nfs_match_client() local 406 list_for_each_entry(clp, &nn->nfs_client_list, cl_share_link) { in nfs_match_client() [all …]
|
D | dns_resolve.c | 340 struct nfs_net *nn = net_generic(net, nfs_net_id); in nfs_dns_resolve_name() local 342 ret = do_cache_lookup_wait(nn->nfs_dns_resolve, &key, &item); in nfs_dns_resolve_name() 349 cache_put(&item->h, nn->nfs_dns_resolve); in nfs_dns_resolve_name() 375 struct nfs_net *nn = net_generic(net, nfs_net_id); in nfs_dns_resolver_cache_init() local 377 nn->nfs_dns_resolve = cache_create_net(&nfs_dns_resolve_template, net); in nfs_dns_resolver_cache_init() 378 if (IS_ERR(nn->nfs_dns_resolve)) in nfs_dns_resolver_cache_init() 379 return PTR_ERR(nn->nfs_dns_resolve); in nfs_dns_resolver_cache_init() 381 err = nfs_cache_register_net(net, nn->nfs_dns_resolve); in nfs_dns_resolver_cache_init() 387 cache_destroy_net(nn->nfs_dns_resolve, net); in nfs_dns_resolver_cache_init() 393 struct nfs_net *nn = net_generic(net, nfs_net_id); in nfs_dns_resolver_cache_destroy() local [all …]
|
D | callback.c | 45 struct nfs_net *nn = net_generic(net, nfs_net_id); in nfs4_callback_up_net() local 51 nn->nfs_callback_tcpport = ret; in nfs4_callback_up_net() 53 nn->nfs_callback_tcpport, PF_INET, net); in nfs4_callback_up_net() 58 nn->nfs_callback_tcpport6 = ret; in nfs4_callback_up_net() 60 nn->nfs_callback_tcpport6, PF_INET6, net); in nfs4_callback_up_net() 255 struct nfs_net *nn = net_generic(net, nfs_net_id); in nfs_callback_down_net() local 257 if (--nn->cb_users[minorversion]) in nfs_callback_down_net() 266 struct nfs_net *nn = net_generic(net, nfs_net_id); in nfs_callback_up_net() local 269 if (nn->cb_users[minorversion]++) in nfs_callback_up_net()
|
D | idmap.c | 540 struct nfs_net *nn = net_generic(net, nfs_net_id); in nfs_get_client_for_event() local 546 spin_lock(&nn->nfs_client_lock); in nfs_get_client_for_event() 547 list_for_each_entry(clp, &nn->nfs_client_list, cl_share_link) { in nfs_get_client_for_event() 551 spin_unlock(&nn->nfs_client_lock); in nfs_get_client_for_event() 569 spin_unlock(&nn->nfs_client_lock); in nfs_get_client_for_event() 572 spin_unlock(&nn->nfs_client_lock); in nfs_get_client_for_event()
|
D | nfs4state.c | 79 struct nfs_net *nn = net_generic(clp->cl_net, nfs_net_id); in nfs4_init_clientid() local 83 port = nn->nfs_callback_tcpport; in nfs4_init_clientid() 85 port = nn->nfs_callback_tcpport6; in nfs4_init_clientid() 124 struct nfs_net *nn = net_generic(clp->cl_net, nfs_net_id); in nfs40_discover_server_trunking() local 128 port = nn->nfs_callback_tcpport; in nfs40_discover_server_trunking() 130 port = nn->nfs_callback_tcpport6; in nfs40_discover_server_trunking()
|
/fs/nfs/blocklayout/ |
D | blocklayoutdm.c | 52 struct nfs_net *nn = net_generic(net, nfs_net_id); in dev_remove() local 56 bl_pipe_msg.bl_wq = &nn->bl_wq; in dev_remove() 71 add_wait_queue(&nn->bl_wq, &wq); in dev_remove() 72 if (rpc_queue_upcall(nn->bl_device_pipe, msg) < 0) { in dev_remove() 73 remove_wait_queue(&nn->bl_wq, &wq); in dev_remove() 80 remove_wait_queue(&nn->bl_wq, &wq); in dev_remove()
|
D | blocklayoutdev.c | 69 struct nfs_net *nn = net_generic(filp->f_dentry->d_sb->s_fs_info, in bl_pipe_downcall() local 75 if (copy_from_user(&nn->bl_mount_reply, src, mlen) != 0) in bl_pipe_downcall() 78 wake_up(&nn->bl_wq); in bl_pipe_downcall() 111 struct nfs_net *nn = net_generic(net, nfs_net_id); in nfs4_blk_decode_device() local 112 struct bl_dev_msg *reply = &nn->bl_mount_reply; in nfs4_blk_decode_device() 118 bl_pipe_msg.bl_wq = &nn->bl_wq; in nfs4_blk_decode_device() 139 add_wait_queue(&nn->bl_wq, &wq); in nfs4_blk_decode_device() 140 rc = rpc_queue_upcall(nn->bl_device_pipe, msg); in nfs4_blk_decode_device() 142 remove_wait_queue(&nn->bl_wq, &wq); in nfs4_blk_decode_device() 150 remove_wait_queue(&nn->bl_wq, &wq); in nfs4_blk_decode_device()
|
D | blocklayout.c | 1322 struct nfs_net *nn = net_generic(net, nfs_net_id); in rpc_pipefs_event() local 1329 if (nn->bl_device_pipe == NULL) { in rpc_pipefs_event() 1336 dentry = nfs4blocklayout_register_sb(sb, nn->bl_device_pipe); in rpc_pipefs_event() 1341 nn->bl_device_pipe->dentry = dentry; in rpc_pipefs_event() 1344 if (nn->bl_device_pipe->dentry) in rpc_pipefs_event() 1345 nfs4blocklayout_unregister_sb(sb, nn->bl_device_pipe); in rpc_pipefs_event() 1387 struct nfs_net *nn = net_generic(net, nfs_net_id); in nfs4blocklayout_net_init() local 1390 init_waitqueue_head(&nn->bl_wq); in nfs4blocklayout_net_init() 1391 nn->bl_device_pipe = rpc_mkpipe_data(&bl_upcall_ops, 0); in nfs4blocklayout_net_init() 1392 if (IS_ERR(nn->bl_device_pipe)) in nfs4blocklayout_net_init() [all …]
|
/fs/ubifs/ |
D | tnc.c | 613 int nn = *n; in tnc_next() local 615 nn += 1; in tnc_next() 616 if (nn < znode->child_cnt) { in tnc_next() 617 *n = nn; in tnc_next() 626 nn = znode->iip + 1; in tnc_next() 628 if (nn < znode->child_cnt) { in tnc_next() 629 znode = get_znode(c, znode, nn); in tnc_next() 637 nn = 0; in tnc_next() 642 *n = nn; in tnc_next() 658 int nn = *n; in tnc_prev() local [all …]
|
D | file.c | 612 int i = 0, nn = *n, offs = bu->zbranch[0].offs, hole = 0, read = 0; in populate_page() local 635 if (nn >= bu->cnt) { in populate_page() 638 } else if (key_block(c, &bu->zbranch[nn].key) == page_block) { in populate_page() 641 dn = bu->buf + (bu->zbranch[nn].offs - offs); in populate_page() 660 nn += 1; in populate_page() 662 } else if (key_block(c, &bu->zbranch[nn].key) < page_block) { in populate_page() 663 nn += 1; in populate_page() 692 *n = nn; in populate_page()
|
/fs/gfs2/ |
D | trace_gfs2.h | 19 #define dlm_state_name(nn) { DLM_LOCK_##nn, #nn } argument
|
/fs/ocfs2/dlm/ |
D | dlmdomain.c | 70 unsigned int nn; in byte_copymap() local 76 for (nn = 0 ; nn < sz; nn++) in byte_copymap() 77 if (test_bit(nn, smap)) in byte_copymap() 78 byte_set_bit(nn, dmap); in byte_copymap()
|