/kernel/linux/linux-5.10/arch/mips/include/asm/netlogic/xlp-hal/ |
D | iomap.h | 50 #define XLP_IO_DEV(node, dev) ((dev) + (node) * 8) argument 53 #define XLP_HDR_OFFSET(node, bus, dev, fn) \ argument 54 XLP_IO_PCI_OFFSET(bus, XLP_IO_DEV(node, dev), fn) 56 #define XLP_IO_BRIDGE_OFFSET(node) XLP_HDR_OFFSET(node, 0, 0, 0) argument 58 #define XLP_IO_CIC0_OFFSET(node) XLP_HDR_OFFSET(node, 0, 0, 1) argument 59 #define XLP_IO_CIC1_OFFSET(node) XLP_HDR_OFFSET(node, 0, 0, 2) argument 60 #define XLP_IO_CIC2_OFFSET(node) XLP_HDR_OFFSET(node, 0, 0, 3) argument 61 #define XLP_IO_PIC_OFFSET(node) XLP_HDR_OFFSET(node, 0, 0, 4) argument 63 #define XLP_IO_PCIE_OFFSET(node, i) XLP_HDR_OFFSET(node, 0, 1, i) argument 64 #define XLP_IO_PCIE0_OFFSET(node) XLP_HDR_OFFSET(node, 0, 1, 0) argument [all …]
|
/kernel/linux/linux-5.10/fs/hfs/ |
D | bnode.c | 9 * Handle basic btree node operations 18 void hfs_bnode_read(struct hfs_bnode *node, void *buf, int off, int len) in hfs_bnode_read() argument 26 off += node->page_offset; in hfs_bnode_read() 31 if (pagenum >= node->tree->pages_per_bnode) in hfs_bnode_read() 33 page = node->page[pagenum]; in hfs_bnode_read() 45 u16 hfs_bnode_read_u16(struct hfs_bnode *node, int off) in hfs_bnode_read_u16() argument 49 hfs_bnode_read(node, &data, off, 2); in hfs_bnode_read_u16() 53 u8 hfs_bnode_read_u8(struct hfs_bnode *node, int off) in hfs_bnode_read_u8() argument 57 hfs_bnode_read(node, &data, off, 1); in hfs_bnode_read_u8() 61 void hfs_bnode_read_key(struct hfs_bnode *node, void *key, int off) in hfs_bnode_read_key() argument [all …]
|
D | brec.c | 18 /* Get the length and offset of the given record in the given node */ 19 u16 hfs_brec_lenoff(struct hfs_bnode *node, u16 rec, u16 *off) in hfs_brec_lenoff() argument 24 dataoff = node->tree->node_size - (rec + 2) * 2; in hfs_brec_lenoff() 25 hfs_bnode_read(node, retval, dataoff, 4); in hfs_brec_lenoff() 31 u16 hfs_brec_keylen(struct hfs_bnode *node, u16 rec) in hfs_brec_keylen() argument 35 if (node->type != HFS_NODE_INDEX && node->type != HFS_NODE_LEAF) in hfs_brec_keylen() 38 if ((node->type == HFS_NODE_INDEX) && in hfs_brec_keylen() 39 !(node->tree->attributes & HFS_TREE_VARIDXKEYS)) { in hfs_brec_keylen() 40 if (node->tree->attributes & HFS_TREE_BIGKEYS) in hfs_brec_keylen() 41 retval = node->tree->max_key_len + 2; in hfs_brec_keylen() [all …]
|
/kernel/linux/linux-5.10/scripts/gdb/linux/ |
D | rbtree.py | 15 node = root.address.cast(rb_root_type.get_type().pointer()) 19 node = root['rb_node'] 20 if node == 0: 23 while node['rb_left']: 24 node = node['rb_left'] 26 return node 31 node = root.address.cast(rb_root_type.get_type().pointer()) 35 node = root['rb_node'] 36 if node == 0: 39 while node['rb_right']: [all …]
|
/kernel/linux/linux-5.10/scripts/dtc/ |
D | livetree.c | 88 struct node *build_node(struct property *proplist, struct node *children, in build_node() 91 struct node *new = xmalloc(sizeof(*new)); in build_node() 92 struct node *child; in build_node() 107 struct node *build_node_delete(struct srcpos *srcpos) in build_node_delete() 109 struct node *new = xmalloc(sizeof(*new)); in build_node_delete() 119 struct node *name_node(struct node *node, char *name) in name_node() argument 121 assert(node->name == NULL); in name_node() 123 node->name = name; in name_node() 125 return node; in name_node() 128 struct node *omit_node_if_unused(struct node *node) in omit_node_if_unused() argument [all …]
|
D | checks.c | 29 typedef void (*check_fn)(struct check *c, struct dt_info *dti, struct node *node); 62 struct node *node, in check_msg() argument 76 else if (node && node->srcpos) in check_msg() 77 pos = node->srcpos; in check_msg() 92 if (node) { in check_msg() 94 xasprintf_append(&str, "%s:%s: ", node->fullpath, prop->name); in check_msg() 96 xasprintf_append(&str, "%s: ", node->fullpath); in check_msg() 106 pos = node->srcpos; in check_msg() 119 #define FAIL(c, dti, node, ...) \ argument 123 check_msg((c), dti, node, NULL, __VA_ARGS__); \ [all …]
|
/kernel/linux/linux-5.10/include/linux/ |
D | bootconfig.h | 19 /* XBC tree node */ 36 /* Node tree access raw APIs */ 38 int __init xbc_node_index(struct xbc_node *node); 39 struct xbc_node * __init xbc_node_get_parent(struct xbc_node *node); 40 struct xbc_node * __init xbc_node_get_child(struct xbc_node *node); 41 struct xbc_node * __init xbc_node_get_next(struct xbc_node *node); 42 const char * __init xbc_node_get_data(struct xbc_node *node); 45 * xbc_node_is_value() - Test the node is a value node 46 * @node: An XBC node. 48 * Test the @node is a value node and return true if a value node, false if not. [all …]
|
/kernel/linux/linux-5.10/lib/ |
D | bootconfig.c | 21 * xbc_parse() parses the text to build a simple tree. Each tree node is 22 * simply a key word or a value. A key node may have a next key node or/and 23 * a child node (both key and value). A value node may have a next value 24 * node (for array). 46 * xbc_root_node() - Get the root node of extended boot config 48 * Return the address of root node of extended boot config. If the 60 * xbc_node_index() - Get the index of XBC node 61 * @node: A target node of getting index. 63 * Return the index number of @node in XBC node list. 65 int __init xbc_node_index(struct xbc_node *node) in xbc_node_index() argument [all …]
|
D | radix-tree.c | 31 * Radix tree node cache. 43 * of RADIX_TREE_MAX_PATH size to be created, with only the root node shared. 98 static inline void tag_set(struct radix_tree_node *node, unsigned int tag, in tag_set() argument 101 __set_bit(offset, node->tags[tag]); in tag_set() 104 static inline void tag_clear(struct radix_tree_node *node, unsigned int tag, in tag_clear() argument 107 __clear_bit(offset, node->tags[tag]); in tag_clear() 110 static inline int tag_get(const struct radix_tree_node *node, unsigned int tag, in tag_get() argument 113 return test_bit(offset, node->tags[tag]); in tag_get() 147 * Returns 1 if any slot in the node has this tag set. 150 static inline int any_tag_set(const struct radix_tree_node *node, in any_tag_set() argument [all …]
|
D | btree.c | 16 * well is that access to a random tree node is much faster than a large number 17 * of operations within each node. 35 * values are to the right, not to the left. All used slots within a node 95 unsigned long *node; in btree_node_alloc() local 97 node = mempool_alloc(head->mempool, gfp); in btree_node_alloc() 98 if (likely(node)) in btree_node_alloc() 99 memset(node, 0, NODESIZE); in btree_node_alloc() 100 return node; in btree_node_alloc() 148 static unsigned long *bkey(struct btree_geo *geo, unsigned long *node, int n) in bkey() argument 150 return &node[n * geo->keylen]; in bkey() [all …]
|
/kernel/linux/linux-5.10/fs/hfsplus/ |
D | bnode.c | 9 * Handle basic btree node operations 21 /* Copy a specified range of bytes from the raw data of a node */ 22 void hfs_bnode_read(struct hfs_bnode *node, void *buf, int off, int len) in hfs_bnode_read() argument 27 off += node->page_offset; in hfs_bnode_read() 28 pagep = node->page + (off >> PAGE_SHIFT); in hfs_bnode_read() 43 u16 hfs_bnode_read_u16(struct hfs_bnode *node, int off) in hfs_bnode_read_u16() argument 47 hfs_bnode_read(node, &data, off, 2); in hfs_bnode_read_u16() 51 u8 hfs_bnode_read_u8(struct hfs_bnode *node, int off) in hfs_bnode_read_u8() argument 55 hfs_bnode_read(node, &data, off, 1); in hfs_bnode_read_u8() 59 void hfs_bnode_read_key(struct hfs_bnode *node, void *key, int off) in hfs_bnode_read_key() argument [all …]
|
D | brec.c | 19 /* Get the length and offset of the given record in the given node */ 20 u16 hfs_brec_lenoff(struct hfs_bnode *node, u16 rec, u16 *off) in hfs_brec_lenoff() argument 25 dataoff = node->tree->node_size - (rec + 2) * 2; in hfs_brec_lenoff() 26 hfs_bnode_read(node, retval, dataoff, 4); in hfs_brec_lenoff() 32 u16 hfs_brec_keylen(struct hfs_bnode *node, u16 rec) in hfs_brec_keylen() argument 36 if (node->type != HFS_NODE_INDEX && node->type != HFS_NODE_LEAF) in hfs_brec_keylen() 39 if ((node->type == HFS_NODE_INDEX) && in hfs_brec_keylen() 40 !(node->tree->attributes & HFS_TREE_VARIDXKEYS) && in hfs_brec_keylen() 41 (node->tree->cnid != HFSPLUS_ATTR_CNID)) { in hfs_brec_keylen() 42 retval = node->tree->max_key_len + 2; in hfs_brec_keylen() [all …]
|
/kernel/linux/linux-5.10/kernel/gcov/ |
D | fs.c | 33 * @list: list head for child node list 36 * @parent: parent node 141 * Return a profiling data set associated with the given node. This is 145 static struct gcov_info *get_node_info(struct gcov_node *node) in get_node_info() argument 147 if (node->num_loaded > 0) in get_node_info() 148 return node->loaded_info[0]; in get_node_info() 150 return node->unloaded_info; in get_node_info() 155 * all profiling data associated with the given node. 157 static struct gcov_info *get_accumulated_info(struct gcov_node *node) in get_accumulated_info() argument 162 if (node->unloaded_info) in get_accumulated_info() [all …]
|
/kernel/linux/linux-5.10/drivers/net/wireguard/ |
D | allowedips.c | 23 static void copy_and_assign_cidr(struct allowedips_node *node, const u8 *src, in copy_and_assign_cidr() argument 26 node->cidr = cidr; in copy_and_assign_cidr() 27 node->bit_at_a = cidr / 8U; in copy_and_assign_cidr() 29 node->bit_at_a ^= (bits / 8U - 1U) % 8U; in copy_and_assign_cidr() 31 node->bit_at_b = 7U - (cidr % 8U); in copy_and_assign_cidr() 32 node->bitlen = bits; in copy_and_assign_cidr() 33 memcpy(node->bits, src, bits / 8U); in copy_and_assign_cidr() 36 static inline u8 choose(struct allowedips_node *node, const u8 *key) in choose() argument 38 return (key[node->bit_at_a] >> node->bit_at_b) & 1; in choose() 58 struct allowedips_node *node, *stack[MAX_ALLOWEDIPS_BITS] = { in root_free_rcu() local [all …]
|
/kernel/linux/linux-5.10/drivers/block/zram/zram_group/ |
D | zlist.c | 25 static inline void zlist_node_lock(struct zlist_node *node) in zlist_node_lock() argument 27 bit_spin_lock(ZLIST_LOCK_BIT, (unsigned long *)node); in zlist_node_lock() 30 static inline void zlist_node_unlock(struct zlist_node *node) in zlist_node_unlock() argument 32 bit_spin_unlock(ZLIST_LOCK_BIT, (unsigned long *)node); in zlist_node_unlock() 37 struct zlist_node *prev, struct zlist_node *node, in zlist_before_add_check() argument 42 assert(idx2node(node->prev, tab) == node); in zlist_before_add_check() 43 assert(idx2node(node->next, tab) == node); in zlist_before_add_check() 47 struct zlist_node *prev, struct zlist_node *node, in zlist_after_add_check() argument 50 assert(idx2node(prev->next, tab) == node); in zlist_after_add_check() 51 assert(idx2node(next->prev, tab) == node); in zlist_after_add_check() [all …]
|
/kernel/linux/linux-5.10/arch/ia64/mm/ |
D | discontig.c | 33 * Track per-node information needed to setup the boot memory allocator, the 34 * per-node areas, and the real VM. 50 * To prevent cache aliasing effects, align per-node structures so that they 51 * start at addresses that are strided by node number. 54 #define NODEDATA_ALIGN(addr, node) \ argument 56 (((node)*PERCPU_PAGE_SIZE) & (MAX_NODE_ALIGN_OFFSET - 1))) 59 * build_node_maps - callback to setup mem_data structs for each node 62 * @node: node where this range resides 65 * treat as a virtually contiguous block (i.e. each node). Each such block 71 int node) in build_node_maps() argument [all …]
|
/kernel/linux/linux-5.10/kernel/bpf/ |
D | bpf_lru_list.c | 42 static bool bpf_lru_node_is_ref(const struct bpf_lru_node *node) in bpf_lru_node_is_ref() argument 44 return node->ref; in bpf_lru_node_is_ref() 62 struct bpf_lru_node *node, in __bpf_lru_node_move_to_free() argument 66 if (WARN_ON_ONCE(IS_LOCAL_LIST_TYPE(node->type))) in __bpf_lru_node_move_to_free() 69 /* If the removing node is the next_inactive_rotation candidate, in __bpf_lru_node_move_to_free() 72 if (&node->list == l->next_inactive_rotation) in __bpf_lru_node_move_to_free() 75 bpf_lru_list_count_dec(l, node->type); in __bpf_lru_node_move_to_free() 77 node->type = tgt_free_type; in __bpf_lru_node_move_to_free() 78 list_move(&node->list, free_list); in __bpf_lru_node_move_to_free() 83 struct bpf_lru_node *node, in __bpf_lru_node_move_in() argument [all …]
|
D | lpm_trie.c | 18 /* Intermediate node */ 48 * lead to more nodes containing more specific matches. Each node also stores 58 * As the trie is empty initially, the new node (1) will be places as root 59 * node, denoted as (R) in the example below. As there are no other node, both 69 * Next, let's add a new node (2) matching 192.168.0.0/24. As there is already 70 * a node with the same data and a smaller prefix (ie, a less specific one), 71 * node (2) will become a child of (1). In child index depends on the next bit 89 * The child[1] slot of (1) could be filled with another node which has bit #17 107 * Let's add another node (4) to the game for 192.168.1.0/24. In order to place 108 * it, node (1) is looked at first, and because (4) of the semantics laid out [all …]
|
/kernel/linux/linux-5.10/fs/hmdfs/comm/ |
D | connection.c | 81 body->case_sensitive = conn_impl->node->sbi->s_case_sensitive; in hs_fill_case_sense_data() 88 __u8 sensitive = conn_impl->node->sbi->s_case_sensitive ? 1 : 0; in hs_parse_case_sense_data() 111 body->features = cpu_to_le64(conn_impl->node->sbi->s_features); in hs_fill_feature_data() 125 conn_impl->node->features = le64_to_cpu(body->features); in hs_parse_feature_data() 286 data_len, conn_impl->node->device_id); in hs_proc_msg_data() 295 req_len, data_len, conn_impl->node->device_id); in hs_proc_msg_data() 306 ops, data_len, conn_impl->node->device_id, hs_req->len); in hs_proc_msg_data() 409 static int hmdfs_node_waiting_evt_sum(const struct hmdfs_peer *node) in hmdfs_node_waiting_evt_sum() argument 415 sum += node->waiting_evt[i]; in hmdfs_node_waiting_evt_sum() 420 static int hmdfs_update_node_waiting_evt(struct hmdfs_peer *node, int evt, in hmdfs_update_node_waiting_evt() argument [all …]
|
/kernel/linux/linux-5.10/drivers/base/ |
D | node.c | 3 * Basic Node interface support 12 #include <linux/node.h> 25 .name = "node", 26 .dev_name = "node", 34 struct node *node_dev = to_node(dev); in node_read_cpumap() 71 * @list_node: List element in the node's access list 109 static void node_remove_accesses(struct node *node) in node_remove_accesses() argument 113 list_for_each_entry_safe(c, cnext, &node->access_list, list_node) { in node_remove_accesses() 124 static struct node_access_nodes *node_init_node_access(struct node *node, in node_init_node_access() argument 130 list_for_each_entry(access_node, &node->access_list, list_node) in node_init_node_access() [all …]
|
/kernel/linux/linux-5.10/net/hsr/ |
D | hsr_framereg.c | 43 struct hsr_node *node; in hsr_addr_is_self() local 45 node = list_first_or_null_rcu(&hsr->self_node_db, struct hsr_node, in hsr_addr_is_self() 47 if (!node) { in hsr_addr_is_self() 48 WARN_ONCE(1, "HSR: No self node\n"); in hsr_addr_is_self() 52 if (ether_addr_equal(addr, node->macaddress_A)) in hsr_addr_is_self() 54 if (ether_addr_equal(addr, node->macaddress_B)) in hsr_addr_is_self() 65 struct hsr_node *node; in find_node_by_addr_A() local 67 list_for_each_entry_rcu(node, node_db, mac_list) { in find_node_by_addr_A() 68 if (ether_addr_equal(node->macaddress_A, addr)) in find_node_by_addr_A() 69 return node; in find_node_by_addr_A() [all …]
|
/kernel/linux/linux-5.10/fs/ocfs2/cluster/ |
D | quorum.c | 14 * that a node is broken and should be recovered. They can't both recover each 19 * So we declare that a node which has given up on connecting to a majority 22 * There are huge opportunities for races here. After we give up on a node's 24 * to declare the node as truly dead. We also need to be careful with the 25 * race between when we see a node start heartbeating and when we connect 85 * go away as our node would be fenced externally before other nodes 117 mlog(ML_ERROR, "fencing this node because it is " in o2quo_make_decision() 128 * the lowest numbered node */ in o2quo_make_decision() 131 mlog(ML_ERROR, "fencing this node because it is " in o2quo_make_decision() 140 mlog(ML_ERROR, "fencing this node because it is " in o2quo_make_decision() [all …]
|
/kernel/linux/linux-5.10/include/drm/ |
D | drm_vma_manager.h | 72 struct drm_vma_offset_node *node, unsigned long pages); 74 struct drm_vma_offset_node *node); 76 int drm_vma_node_allow(struct drm_vma_offset_node *node, struct drm_file *tag); 77 void drm_vma_node_revoke(struct drm_vma_offset_node *node, 79 bool drm_vma_node_is_allowed(struct drm_vma_offset_node *node, 83 * drm_vma_offset_exact_lookup_locked() - Look up node by exact address 88 * Same as drm_vma_offset_lookup_locked() but does not allow any offset into the node. 92 * Node at exact start address @start. 99 struct drm_vma_offset_node *node; in drm_vma_offset_exact_lookup_locked() local 101 node = drm_vma_offset_lookup_locked(mgr, start, pages); in drm_vma_offset_exact_lookup_locked() [all …]
|
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/gt/ |
D | intel_gt_buffer_pool.c | 34 static void node_free(struct intel_gt_buffer_pool_node *node) in node_free() argument 36 i915_gem_object_put(node->obj); in node_free() 37 i915_active_fini(&node->active); in node_free() 38 kfree_rcu(node, rcu); in node_free() 43 struct intel_gt_buffer_pool_node *node, *stale = NULL; in pool_free_older_than() local 61 node = list_entry(pos, typeof(*node), link); in pool_free_older_than() 63 age = READ_ONCE(node->age); in pool_free_older_than() 67 /* Check we are the first to claim this node */ in pool_free_older_than() 68 if (!xchg(&node->age, 0)) in pool_free_older_than() 71 node->free = stale; in pool_free_older_than() [all …]
|
/kernel/linux/linux-5.10/drivers/gpu/drm/ |
D | drm_mm.c | 79 * steep cliff not a real concern. Removing a node again is O(1). 107 static noinline void save_stack(struct drm_mm_node *node) in save_stack() argument 115 node->stack = stack_depot_save(entries, n, GFP_NOWAIT); in save_stack() 120 struct drm_mm_node *node; in show_leaks() local 129 list_for_each_entry(node, drm_mm_nodes(mm), node_list) { in show_leaks() 130 if (!node->stack) { in show_leaks() 131 DRM_ERROR("node [%08llx + %08llx]: unknown owner\n", in show_leaks() 132 node->start, node->size); in show_leaks() 136 nr_entries = stack_depot_fetch(node->stack, &entries); in show_leaks() 138 DRM_ERROR("node [%08llx + %08llx]: inserted at\n%s", in show_leaks() [all …]
|