/net/ceph/crush/ |
D | mapper.c | 415 int item, int x) in is_out() argument 417 if (item >= weight_max) in is_out() 419 if (weight[item] >= 0x10000) in is_out() 421 if (weight[item] == 0) in is_out() 423 if ((crush_hash32_2(CRUSH_HASH_RJENKINS1, x, item) & 0xffff) in is_out() 424 < weight[item]) in is_out() 473 int item = 0; in crush_choose_firstn() local 509 item = bucket_perm_choose( in crush_choose_firstn() 513 item = crush_bucket_choose( in crush_choose_firstn() 519 if (item >= map->max_devices) { in crush_choose_firstn() [all …]
|
/net/tipc/ |
D | name_distr.c | 91 struct distr_item *item; in tipc_named_publish() local 105 item = (struct distr_item *)msg_data(buf_msg(buf)); in tipc_named_publish() 106 publ_to_item(item, publ); in tipc_named_publish() 116 struct distr_item *item; in tipc_named_withdraw() local 129 item = (struct distr_item *)msg_data(buf_msg(buf)); in tipc_named_withdraw() 130 publ_to_item(item, publ); in tipc_named_withdraw() 145 struct distr_item *item = NULL; in named_distribute() local 160 item = (struct distr_item *)msg_data(buf_msg(skb)); in named_distribute() 164 publ_to_item(item, publ); in named_distribute() 165 item++; in named_distribute() [all …]
|
D | name_table.c | 1062 struct u32_item *item; in u32_find() local 1064 list_for_each_entry(item, l, list) { in u32_find() 1065 if (item->value == value) in u32_find() 1073 struct u32_item *item; in u32_push() local 1075 list_for_each_entry(item, l, list) { in u32_push() 1076 if (item->value == value) in u32_push() 1079 item = kmalloc(sizeof(*item), GFP_ATOMIC); in u32_push() 1080 if (unlikely(!item)) in u32_push() 1083 item->value = value; in u32_push() 1084 list_add(&item->list, l); in u32_push() [all …]
|
/net/netfilter/ |
D | nf_conntrack_ecache.c | 138 struct nf_ct_event item = { in nf_conntrack_eventmask_report() local 149 ret = notify->fcn(eventmask | missed, &item); in nf_conntrack_eventmask_report() 184 struct nf_ct_event item; in nf_ct_deliver_cached_events() local 209 item.ct = ct; in nf_ct_deliver_cached_events() 210 item.portid = 0; in nf_ct_deliver_cached_events() 211 item.report = 0; in nf_ct_deliver_cached_events() 213 ret = notify->fcn(events | missed, &item); in nf_ct_deliver_cached_events() 249 struct nf_exp_event item = { in nf_ct_expect_event_report() local 254 notify->fcn(1 << event, &item); in nf_ct_expect_event_report()
|
D | nf_conntrack_h323_main.c | 919 &setup->fastStart.item[i]); in process_setup() 952 &callproc->fastStart.item[i]); in process_callproceeding() 984 &connect->fastStart.item[i]); in process_connect() 1016 &alert->fastStart.item[i]); in process_alerting() 1057 &facility->fastStart.item[i]); in process_facility() 1089 &progress->fastStart.item[i]); in process_progress() 1148 &pdu->h245Control.item[i]); in process_q931() 1407 rrq->callSignalAddress.item, in process_rrq() 1416 rrq->rasAddress.item, in process_rrq() 1449 rcf->callSignalAddress.item, in process_rcf() [all …]
|
D | nf_conntrack_netlink.c | 625 ctnetlink_conntrack_event(unsigned int events, struct nf_ct_event *item) in ctnetlink_conntrack_event() argument 632 struct nf_conn *ct = item->ct; in ctnetlink_conntrack_event() 652 if (!item->report && !nfnetlink_has_listeners(net, group)) in ctnetlink_conntrack_event() 660 nlh = nlmsg_put(skb, item->portid, 0, type, sizeof(*nfmsg), flags); in ctnetlink_conntrack_event() 741 err = nfnetlink_send(skb, net, item->portid, group, item->report, in ctnetlink_conntrack_event() 2650 ctnetlink_expect_event(unsigned int events, struct nf_exp_event *item) in ctnetlink_expect_event() argument 2652 struct nf_conntrack_expect *exp = item->exp; in ctnetlink_expect_event() 2670 if (!item->report && !nfnetlink_has_listeners(net, group)) in ctnetlink_expect_event() 2678 nlh = nlmsg_put(skb, item->portid, 0, type, sizeof(*nfmsg), flags); in ctnetlink_expect_event() 2691 nfnetlink_send(skb, net, item->portid, group, item->report, GFP_ATOMIC); in ctnetlink_expect_event()
|
/net/sunrpc/auth_gss/ |
D | svcauth_gss.c | 82 static struct rsi *rsi_lookup(struct cache_detail *cd, struct rsi *item); 99 static inline int rsi_hash(struct rsi *item) in rsi_hash() argument 101 return hash_mem(item->in_handle.data, item->in_handle.len, RSI_HASHBITS) in rsi_hash() 102 ^ hash_mem(item->in_token.data, item->in_token.len, RSI_HASHBITS); in rsi_hash() 107 struct rsi *item = container_of(a, struct rsi, h); in rsi_match() local 109 return netobj_equal(&item->in_handle, &tmp->in_handle) && in rsi_match() 110 netobj_equal(&item->in_token, &tmp->in_token); in rsi_match() 130 struct rsi *item = container_of(citem, struct rsi, h); in rsi_init() local 136 new->in_handle.len = item->in_handle.len; in rsi_init() 137 item->in_handle.len = 0; in rsi_init() [all …]
|
/net/sunrpc/ |
D | svcauth_unix.c | 98 struct cache_head *item = container_of(kref, struct cache_head, ref); in ip_map_put() local 99 struct ip_map *im = container_of(item, struct ip_map,h); in ip_map_put() 101 if (test_bit(CACHE_VALID, &item->flags) && in ip_map_put() 102 !test_bit(CACHE_NEGATIVE, &item->flags)) in ip_map_put() 121 struct ip_map *item = container_of(citem, struct ip_map, h); in ip_map_init() local 123 strcpy(new->m_class, item->m_class); in ip_map_init() 124 new->m_addr = item->m_addr; in ip_map_init() 129 struct ip_map *item = container_of(citem, struct ip_map, h); in update() local 131 kref_get(&item->m_client->h.ref); in update() 132 new->m_client = item->m_client; in update() [all …]
|
D | cache.c | 41 static bool cache_defer_req(struct cache_req *req, struct cache_head *item); 42 static void cache_revisit_request(struct cache_head *item); 543 #define DFR_HASH(item) ((((long)item)>>4 ^ (((long)item)>>13)) % DFR_HASHSIZE) argument 561 static void __hash_deferred_req(struct cache_deferred_req *dreq, struct cache_head *item) in __hash_deferred_req() argument 563 int hash = DFR_HASH(item); in __hash_deferred_req() 570 struct cache_head *item, in setup_deferral() argument 574 dreq->item = item; in setup_deferral() 578 __hash_deferred_req(dreq, item); in setup_deferral() 601 static void cache_wait_req(struct cache_req *req, struct cache_head *item) in cache_wait_req() argument 609 setup_deferral(dreq, item, 0); in cache_wait_req() [all …]
|
/net/ceph/ |
D | mon_client.c | 284 le64_to_cpu(monc->subs[i].item.start), in __send_subscribe() 285 monc->subs[i].item.flags); in __send_subscribe() 287 memcpy(p, &monc->subs[i].item, sizeof(monc->subs[i].item)); in __send_subscribe() 288 p += sizeof(monc->subs[i].item); in __send_subscribe() 346 monc->subs[sub].item.start == start && in __ceph_monc_want_map() 347 monc->subs[sub].item.flags == flags) in __ceph_monc_want_map() 350 monc->subs[sub].item.start = start; in __ceph_monc_want_map() 351 monc->subs[sub].item.flags = flags; in __ceph_monc_want_map() 381 if (monc->subs[sub].item.flags & CEPH_SUBSCRIBE_ONETIME) in __ceph_monc_got_map() 384 monc->subs[sub].item.start = cpu_to_le64(epoch + 1); in __ceph_monc_got_map()
|
D | debugfs.c | 151 le64_to_cpu(monc->subs[i].item.start), in monc_show() 152 (monc->subs[i].item.flags & in monc_show()
|
D | osd_client.c | 4711 static int decode_watcher(void **p, void *end, struct ceph_watch_item *item) in decode_watcher() argument 4722 ceph_decode_copy(p, &item->name, sizeof(item->name)); in decode_watcher() 4723 item->cookie = ceph_decode_64(p); in decode_watcher() 4726 ceph_decode_copy(p, &item->addr, sizeof(item->addr)); in decode_watcher() 4727 ceph_decode_addr(&item->addr); in decode_watcher() 4731 ENTITY_NAME(item->name), item->cookie, in decode_watcher() 4732 ceph_pr_addr(&item->addr.in_addr)); in decode_watcher()
|
/net/qrtr/ |
D | qrtr.c | 111 struct list_head item; member 129 list_del(&node->item); in __qrtr_node_release() 410 list_add(&node->item, &qrtr_all_nodes); in qrtr_endpoint_register() 654 list_for_each_entry(node, &qrtr_all_nodes, item) { in qrtr_bcast_enqueue()
|