/kernel/liteos_m/arch/risc-v/nuclei/gcc/nmsis/Core/Include/ |
D | core_feature_pmp.h | 61 __STATIC_INLINE uint8_t __get_PMPxCFG(uint32_t idx) in __get_PMPxCFG() argument 65 if (idx >= __PMP_ENTRY_NUM) return 0; in __get_PMPxCFG() 67 if (idx < 4) { in __get_PMPxCFG() 69 } else if ((idx >=4) && (idx < 8)) { in __get_PMPxCFG() 70 idx -= 4; in __get_PMPxCFG() 72 } else if ((idx >=8) && (idx < 12)) { in __get_PMPxCFG() 73 idx -= 8; in __get_PMPxCFG() 76 idx -= 12; in __get_PMPxCFG() 80 idx = idx << 3; in __get_PMPxCFG() 81 return (uint8_t)((pmpcfg>>idx) & 0xFF); in __get_PMPxCFG() [all …]
|
/kernel/linux/linux-5.10/drivers/net/ethernet/huawei/hinic/ |
D | hinic_hw_csr.h | 21 #define HINIC_CSR_DMA_ATTR_ADDR(idx) \ argument 22 (HINIC_DMA_ATTR_BASE + (idx) * HINIC_DMA_ATTR_STRIDE) 27 #define HINIC_CSR_PPF_ELECTION_ADDR(idx) \ argument 28 (HINIC_ELECTION_BASE + (idx) * HINIC_PPF_ELECTION_STRIDE) 35 #define HINIC_CSR_API_CMD_CHAIN_HEAD_HI_ADDR(idx) \ argument 36 (HINIC_CSR_API_CMD_BASE + 0x0 + (idx) * HINIC_CSR_API_CMD_STRIDE) 38 #define HINIC_CSR_API_CMD_CHAIN_HEAD_LO_ADDR(idx) \ argument 39 (HINIC_CSR_API_CMD_BASE + 0x4 + (idx) * HINIC_CSR_API_CMD_STRIDE) 41 #define HINIC_CSR_API_CMD_STATUS_HI_ADDR(idx) \ argument 42 (HINIC_CSR_API_CMD_BASE + 0x8 + (idx) * HINIC_CSR_API_CMD_STRIDE) [all …]
|
/kernel/linux/linux-5.10/drivers/net/can/sja1000/ |
D | sja1000_isa.c | 122 int idx = pdev->id; in sja1000_isa_probe() local 126 idx, port[idx], mem[idx], irq[idx]); in sja1000_isa_probe() 128 if (mem[idx]) { in sja1000_isa_probe() 129 if (!request_mem_region(mem[idx], iosize, DRV_NAME)) { in sja1000_isa_probe() 133 base = ioremap(mem[idx], iosize); in sja1000_isa_probe() 139 if (indirect[idx] > 0 || in sja1000_isa_probe() 140 (indirect[idx] == -1 && indirect[0] > 0)) in sja1000_isa_probe() 142 if (!request_region(port[idx], iosize, DRV_NAME)) { in sja1000_isa_probe() 155 dev->irq = irq[idx]; in sja1000_isa_probe() 157 if (mem[idx]) { in sja1000_isa_probe() [all …]
|
/kernel/linux/linux-5.10/drivers/crypto/ccree/ |
D | cc_aead.c | 303 unsigned int idx = 0; in hmac_setkey() local 309 hw_desc_init(&desc[idx]); in hmac_setkey() 310 set_cipher_mode(&desc[idx], hash_mode); in hmac_setkey() 311 set_din_sram(&desc[idx], in hmac_setkey() 315 set_flow_mode(&desc[idx], S_DIN_to_HASH); in hmac_setkey() 316 set_setup_mode(&desc[idx], SETUP_LOAD_STATE0); in hmac_setkey() 317 idx++; in hmac_setkey() 320 hw_desc_init(&desc[idx]); in hmac_setkey() 321 set_cipher_mode(&desc[idx], hash_mode); in hmac_setkey() 322 set_din_const(&desc[idx], 0, ctx->hash_len); in hmac_setkey() [all …]
|
D | cc_hash.c | 342 int idx) in cc_fin_result() argument 350 hw_desc_init(&desc[idx]); in cc_fin_result() 351 set_hash_cipher_mode(&desc[idx], ctx->hw_mode, ctx->hash_mode); in cc_fin_result() 352 set_dout_dlli(&desc[idx], state->digest_result_dma_addr, digestsize, in cc_fin_result() 354 set_queue_last_ind(ctx->drvdata, &desc[idx]); in cc_fin_result() 355 set_flow_mode(&desc[idx], S_HASH_to_DOUT); in cc_fin_result() 356 set_setup_mode(&desc[idx], SETUP_WRITE_STATE0); in cc_fin_result() 357 set_cipher_config1(&desc[idx], HASH_PADDING_DISABLED); in cc_fin_result() 358 cc_set_endianity(ctx->hash_mode, &desc[idx]); in cc_fin_result() 359 idx++; in cc_fin_result() [all …]
|
/kernel/linux/linux-5.10/drivers/net/can/cc770/ |
D | cc770_isa.c | 167 int idx = pdev->id; in cc770_isa_probe() local 172 idx, port[idx], mem[idx], irq[idx]); in cc770_isa_probe() 173 if (mem[idx]) { in cc770_isa_probe() 174 if (!request_mem_region(mem[idx], iosize, KBUILD_MODNAME)) { in cc770_isa_probe() 178 base = ioremap(mem[idx], iosize); in cc770_isa_probe() 184 if (indirect[idx] > 0 || in cc770_isa_probe() 185 (indirect[idx] == -1 && indirect[0] > 0)) in cc770_isa_probe() 187 if (!request_region(port[idx], iosize, KBUILD_MODNAME)) { in cc770_isa_probe() 200 dev->irq = irq[idx]; in cc770_isa_probe() 202 if (mem[idx]) { in cc770_isa_probe() [all …]
|
/kernel/linux/linux-5.10/drivers/block/zram/zram_group/ |
D | zlist.h | 25 struct zlist_node *(*idx2node)(u32 idx, void *priv); 29 static inline struct zlist_node *idx2node(u32 idx, struct zlist_table *tab) in idx2node() argument 31 return tab->idx2node(idx, tab->private); in idx2node() 34 static inline u32 next_idx(u32 idx, struct zlist_table *tab) in next_idx() argument 36 return idx2node(idx, tab)->next; in next_idx() 39 static inline u32 prev_idx(u32 idx, struct zlist_table *tab) in prev_idx() argument 41 return idx2node(idx, tab)->prev; in prev_idx() 52 void zlist_lock(u32 idx, struct zlist_table *tab); 53 void zlist_unlock(u32 idx, struct zlist_table *tab); 55 void zlist_add_nolock(u32 hid, u32 idx, struct zlist_table *tab); [all …]
|
D | zlist.c | 103 void zlist_lock(u32 idx, struct zlist_table *tab) in zlist_lock() argument 105 zlist_node_lock(idx2node(idx, tab)); in zlist_lock() 108 void zlist_unlock(u32 idx, struct zlist_table *tab) in zlist_unlock() argument 110 zlist_node_unlock(idx2node(idx, tab)); in zlist_unlock() 113 void zlist_add_nolock(u32 hid, u32 idx, struct zlist_table *tab) in zlist_add_nolock() argument 115 struct zlist_node *node = idx2node(idx, tab); in zlist_add_nolock() 121 if (idx != hid) in zlist_add_nolock() 125 if (idx != hid) in zlist_add_nolock() 127 head->next = idx; in zlist_add_nolock() 130 next->prev = idx; in zlist_add_nolock() [all …]
|
/kernel/linux/linux-5.10/tools/testing/selftests/kvm/lib/ |
D | sparsebit.c | 172 sparsebit_idx_t idx; /* index of least-significant bit in mask */ member 287 root->idx = subtree->idx; in node_copy_subtree() 310 static struct node *node_find(struct sparsebit *s, sparsebit_idx_t idx) in node_find() argument 316 nodep = nodep->idx > idx ? nodep->left : nodep->right) { in node_find() 317 if (idx >= nodep->idx && in node_find() 318 idx <= nodep->idx + MASK_BITS + nodep->num_after - 1) in node_find() 333 static struct node *node_add(struct sparsebit *s, sparsebit_idx_t idx) in node_add() argument 344 nodep->idx = idx & -MASK_BITS; in node_add() 358 if (idx < parentp->idx) { in node_add() 366 assert(idx > parentp->idx + MASK_BITS + parentp->num_after - 1); in node_add() [all …]
|
/kernel/linux/linux-5.10/kernel/sched/ |
D | cpudeadline.c | 26 static void cpudl_heapify_down(struct cpudl *cp, int idx) in cpudl_heapify_down() argument 30 int orig_cpu = cp->elements[idx].cpu; in cpudl_heapify_down() 31 u64 orig_dl = cp->elements[idx].dl; in cpudl_heapify_down() 33 if (left_child(idx) >= cp->size) in cpudl_heapify_down() 40 l = left_child(idx); in cpudl_heapify_down() 41 r = right_child(idx); in cpudl_heapify_down() 42 largest = idx; in cpudl_heapify_down() 54 if (largest == idx) in cpudl_heapify_down() 58 cp->elements[idx].cpu = cp->elements[largest].cpu; in cpudl_heapify_down() 59 cp->elements[idx].dl = cp->elements[largest].dl; in cpudl_heapify_down() [all …]
|
/kernel/linux/linux-5.10/arch/x86/um/ |
D | tls_32.c | 66 int idx; in get_free_idx() local 68 for (idx = 0; idx < GDT_ENTRY_TLS_ENTRIES; idx++) in get_free_idx() 69 if (!t->arch.tls_array[idx].present) in get_free_idx() 70 return idx + GDT_ENTRY_TLS_MIN; in get_free_idx() 92 int idx; in load_TLS() local 94 for (idx = GDT_ENTRY_TLS_MIN; idx < GDT_ENTRY_TLS_MAX; idx++) { in load_TLS() 96 &to->thread.arch.tls_array[idx - GDT_ENTRY_TLS_MIN]; in load_TLS() 105 curr->tls.entry_number = idx; in load_TLS() 201 int idx, int flushed) in set_tls_entry() argument 205 if (idx < GDT_ENTRY_TLS_MIN || idx > GDT_ENTRY_TLS_MAX) in set_tls_entry() [all …]
|
/kernel/linux/linux-5.10/drivers/clk/uniphier/ |
D | clk-uniphier-sys.c | 23 #define UNIPHIER_LD4_SYS_CLK_NAND(idx) \ argument 25 UNIPHIER_CLK_GATE("nand", (idx), "nand-50m", 0x2104, 2) 27 #define UNIPHIER_PRO5_SYS_CLK_NAND(idx) \ argument 29 UNIPHIER_CLK_GATE("nand", (idx), "nand-50m", 0x2104, 2) 31 #define UNIPHIER_LD11_SYS_CLK_NAND(idx) \ argument 33 UNIPHIER_CLK_GATE("nand", (idx), "nand-50m", 0x210c, 0) 35 #define UNIPHIER_SYS_CLK_NAND_4X(idx) \ argument 36 UNIPHIER_CLK_FACTOR("nand-4x", (idx), "nand", 4, 1) 38 #define UNIPHIER_LD11_SYS_CLK_EMMC(idx) \ argument 39 UNIPHIER_CLK_GATE("emmc", (idx), NULL, 0x210c, 2) [all …]
|
/kernel/linux/linux-5.10/mm/ |
D | hugetlb_cgroup.c | 30 #define hugetlb_cgroup_from_counter(counter, idx) \ argument 31 container_of(counter, struct hugetlb_cgroup, hugepage[idx]) 36 __hugetlb_cgroup_counter_from_cgroup(struct hugetlb_cgroup *h_cg, int idx, in __hugetlb_cgroup_counter_from_cgroup() argument 40 return &h_cg->rsvd_hugepage[idx]; in __hugetlb_cgroup_counter_from_cgroup() 41 return &h_cg->hugepage[idx]; in __hugetlb_cgroup_counter_from_cgroup() 45 hugetlb_cgroup_counter_from_cgroup(struct hugetlb_cgroup *h_cg, int idx) in hugetlb_cgroup_counter_from_cgroup() argument 47 return __hugetlb_cgroup_counter_from_cgroup(h_cg, idx, false); in hugetlb_cgroup_counter_from_cgroup() 51 hugetlb_cgroup_counter_from_cgroup_rsvd(struct hugetlb_cgroup *h_cg, int idx) in hugetlb_cgroup_counter_from_cgroup_rsvd() argument 53 return __hugetlb_cgroup_counter_from_cgroup(h_cg, idx, true); in hugetlb_cgroup_counter_from_cgroup_rsvd() 81 int idx; in hugetlb_cgroup_have_usage() local [all …]
|
/kernel/linux/linux-5.10/tools/perf/scripts/perl/Perf-Trace-Util/lib/Perf/Trace/ |
D | Core.pm | 39 foreach my $idx (sort {$a <=> $b} keys %trace_flags) { 40 if (!$value && !$idx) { 45 if ($idx && ($value & $idx) == $idx) { 49 $string .= "$trace_flags{$idx}"; 51 $value &= ~$idx; 69 foreach my $idx (sort {$a <=> $b} keys %{$flag_fields{$event_name}{$field_name}{"values"}}) { 70 if (!$value && !$idx) { 71 $string .= "$flag_fields{$event_name}{$field_name}{'values'}{$idx}"; 74 if ($idx && ($value & $idx) == $idx) { 78 $string .= "$flag_fields{$event_name}{$field_name}{'values'}{$idx}"; [all …]
|
/kernel/linux/linux-5.10/include/asm-generic/ |
D | fixmap.h | 30 static __always_inline unsigned long fix_to_virt(const unsigned int idx) in fix_to_virt() argument 32 BUILD_BUG_ON(idx >= __end_of_fixed_addresses); in fix_to_virt() 33 return __fix_to_virt(idx); in fix_to_virt() 64 #define set_fixmap(idx, phys) \ argument 65 __set_fixmap(idx, phys, FIXMAP_PAGE_NORMAL) 69 #define clear_fixmap(idx) \ argument 70 __set_fixmap(idx, 0, FIXMAP_PAGE_CLEAR) 74 #define __set_fixmap_offset(idx, phys, flags) \ argument 77 __set_fixmap(idx, phys, flags); \ 78 ________addr = fix_to_virt(idx) + ((phys) & (PAGE_SIZE - 1)); \ [all …]
|
/kernel/linux/linux-5.10/arch/x86/events/intel/ |
D | uncore_nhmex.c | 249 if (hwc->idx == UNCORE_PMC_IDX_FIXED) in nhmex_uncore_msr_enable_event() 371 reg1->idx = 0; in nhmex_bbox_hw_config() 383 if (reg1->idx != EXTRA_REG_NONE) { in nhmex_bbox_msr_enable_event() 456 reg1->idx = 0; in nhmex_sbox_hw_config() 468 if (reg1->idx != EXTRA_REG_NONE) { in nhmex_sbox_msr_enable_event() 553 static bool nhmex_mbox_get_shared_reg(struct intel_uncore_box *box, int idx, u64 config) in nhmex_mbox_get_shared_reg() argument 560 if (idx < EXTRA_REG_NHMEX_M_ZDP_CTL_FVC) { in nhmex_mbox_get_shared_reg() 561 er = &box->shared_regs[idx]; in nhmex_mbox_get_shared_reg() 577 idx -= EXTRA_REG_NHMEX_M_ZDP_CTL_FVC; in nhmex_mbox_get_shared_reg() 578 if (WARN_ON_ONCE(idx >= 4)) in nhmex_mbox_get_shared_reg() [all …]
|
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/include/nvkm/core/ |
D | device.h | 205 int (*acr )(struct nvkm_device *, int idx, struct nvkm_acr **); 206 int (*bar )(struct nvkm_device *, int idx, struct nvkm_bar **); 207 int (*bios )(struct nvkm_device *, int idx, struct nvkm_bios **); 208 int (*bus )(struct nvkm_device *, int idx, struct nvkm_bus **); 209 int (*clk )(struct nvkm_device *, int idx, struct nvkm_clk **); 210 int (*devinit )(struct nvkm_device *, int idx, struct nvkm_devinit **); 211 int (*fault )(struct nvkm_device *, int idx, struct nvkm_fault **); 212 int (*fb )(struct nvkm_device *, int idx, struct nvkm_fb **); 213 int (*fuse )(struct nvkm_device *, int idx, struct nvkm_fuse **); 214 int (*gpio )(struct nvkm_device *, int idx, struct nvkm_gpio **); [all …]
|
/kernel/linux/linux-5.10/include/linux/ |
D | memcontrol.h | 360 static __always_inline bool memcg_stat_item_in_bytes(int idx) in memcg_stat_item_in_bytes() argument 362 if (idx == MEMCG_PERCPU_B) in memcg_stat_item_in_bytes() 364 return vmstat_item_in_bytes(idx); in memcg_stat_item_in_bytes() 707 static inline unsigned long memcg_page_state(struct mem_cgroup *memcg, int idx) in memcg_page_state() argument 709 long x = atomic_long_read(&memcg->vmstats[idx]); in memcg_page_state() 722 int idx) in memcg_page_state_local() argument 728 x += per_cpu(memcg->vmstats_local->stat[idx], cpu); in memcg_page_state_local() 736 void __mod_memcg_state(struct mem_cgroup *memcg, int idx, int val); 740 int idx, int val) in mod_memcg_state() argument 745 __mod_memcg_state(memcg, idx, val); in mod_memcg_state() [all …]
|
/kernel/linux/linux-5.10/drivers/gpu/drm/radeon/ |
D | evergreen_cs.c | 756 unsigned idx) in evergreen_cs_track_validate_texture() argument 764 texdw[0] = radeon_get_ib_value(p, idx + 0); in evergreen_cs_track_validate_texture() 765 texdw[1] = radeon_get_ib_value(p, idx + 1); in evergreen_cs_track_validate_texture() 766 texdw[2] = radeon_get_ib_value(p, idx + 2); in evergreen_cs_track_validate_texture() 767 texdw[3] = radeon_get_ib_value(p, idx + 3); in evergreen_cs_track_validate_texture() 768 texdw[4] = radeon_get_ib_value(p, idx + 4); in evergreen_cs_track_validate_texture() 769 texdw[5] = radeon_get_ib_value(p, idx + 5); in evergreen_cs_track_validate_texture() 770 texdw[6] = radeon_get_ib_value(p, idx + 6); in evergreen_cs_track_validate_texture() 771 texdw[7] = radeon_get_ib_value(p, idx + 7); in evergreen_cs_track_validate_texture() 1050 unsigned idx, unsigned reg) in evergreen_packet0_check() argument [all …]
|
D | r600_cs.c | 838 r = radeon_cs_packet_parse(p, &wait_reg_mem, p->idx); in r600_cs_common_vline_parse() 849 wait_reg_mem_info = radeon_get_ib_value(p, wait_reg_mem.idx + 1); in r600_cs_common_vline_parse() 865 if ((radeon_get_ib_value(p, wait_reg_mem.idx + 2) << 2) != vline_status[0]) { in r600_cs_common_vline_parse() 870 if (radeon_get_ib_value(p, wait_reg_mem.idx + 5) != RADEON_VLINE_STAT) { in r600_cs_common_vline_parse() 876 r = radeon_cs_packet_parse(p, &p3reloc, p->idx + wait_reg_mem.count + 2); in r600_cs_common_vline_parse() 880 h_idx = p->idx - 2; in r600_cs_common_vline_parse() 881 p->idx += wait_reg_mem.count + 2; in r600_cs_common_vline_parse() 882 p->idx += p3reloc.count + 2; in r600_cs_common_vline_parse() 919 unsigned idx, unsigned reg) in r600_packet0_check() argument 928 idx, reg); in r600_packet0_check() [all …]
|
/kernel/linux/linux-5.10/drivers/gpu/drm/msm/disp/dpu1/ |
D | dpu_hw_sspp.c | 138 u32 *idx) in _sspp_subblk_offset() argument 150 *idx = sblk->src_blk.base; in _sspp_subblk_offset() 155 *idx = sblk->scaler_blk.base; in _sspp_subblk_offset() 159 *idx = sblk->csc_blk.base; in _sspp_subblk_offset() 173 u32 idx; in dpu_hw_sspp_setup_multirect() local 175 if (_sspp_subblk_offset(ctx, DPU_SSPP_SRC, &idx)) in dpu_hw_sspp_setup_multirect() 186 mode_mask = DPU_REG_READ(&ctx->hw, SSPP_MULTIRECT_OPMODE + idx); in dpu_hw_sspp_setup_multirect() 194 DPU_REG_WRITE(&ctx->hw, SSPP_MULTIRECT_OPMODE + idx, mode_mask); in dpu_hw_sspp_setup_multirect() 200 u32 idx; in _sspp_setup_opmode() local 204 _sspp_subblk_offset(ctx, DPU_SSPP_SCALER_QSEED2, &idx) || in _sspp_setup_opmode() [all …]
|
/kernel/linux/linux-5.10/drivers/net/ethernet/chelsio/cxgb/ |
D | fpga_defs.h | 215 #define MAC_REG_ADDR(idx, reg) (MAC_REG_BASE + (idx) * 128 + (reg)) argument 217 #define MAC_REG_IDLO(idx) MAC_REG_ADDR(idx, A_GMAC_MACID_LO) argument 218 #define MAC_REG_IDHI(idx) MAC_REG_ADDR(idx, A_GMAC_MACID_HI) argument 219 #define MAC_REG_CSR(idx) MAC_REG_ADDR(idx, A_GMAC_CSR) argument 220 #define MAC_REG_IFS(idx) MAC_REG_ADDR(idx, A_GMAC_IFS) argument 221 #define MAC_REG_LARGEFRAMELENGTH(idx) MAC_REG_ADDR(idx, A_GMAC_JUMBO_FRAME_LEN) argument 222 #define MAC_REG_LINKDLY(idx) MAC_REG_ADDR(idx, A_GMAC_LNK_DLY) argument 223 #define MAC_REG_PAUSETIME(idx) MAC_REG_ADDR(idx, A_GMAC_PAUSETIME) argument 224 #define MAC_REG_CASTLO(idx) MAC_REG_ADDR(idx, A_GMAC_MCAST_LO) argument 225 #define MAC_REG_MCASTHI(idx) MAC_REG_ADDR(idx, A_GMAC_MCAST_HI) argument [all …]
|
/kernel/linux/linux-5.10/tools/perf/scripts/python/Perf-Trace-Util/lib/Perf/Trace/ |
D | Core.py | 35 for idx in sorted(flag_fields[event_name][field_name]['values']): 36 if not value and not idx: 37 string += flag_fields[event_name][field_name]['values'][idx] 39 if idx and (value & idx) == idx: 42 string += flag_fields[event_name][field_name]['values'][idx] 44 value &= ~idx 52 for idx in sorted(symbolic_fields[event_name][field_name]['values']): 53 if not value and not idx: 54 string = symbolic_fields[event_name][field_name]['values'][idx] 56 if (value == idx): [all …]
|
/kernel/linux/linux-5.10/include/trace/events/ |
D | cpuhp.h | 14 int idx, 17 TP_ARGS(cpu, target, idx, fun), 22 __field( int, idx ) 29 __entry->idx = idx; 34 __entry->cpu, __entry->target, __entry->idx, __entry->fun) 41 int idx, 45 TP_ARGS(cpu, target, idx, fun, node), 50 __field( int, idx ) 57 __entry->idx = idx; 62 __entry->cpu, __entry->target, __entry->idx, __entry->fun) [all …]
|
/kernel/linux/linux-5.10/drivers/net/ethernet/ti/ |
D | cpsw_ale.c | 107 int idx; in cpsw_ale_get_field() local 109 idx = start / 32; in cpsw_ale_get_field() 110 start -= idx * 32; in cpsw_ale_get_field() 111 idx = 2 - idx; /* flip */ in cpsw_ale_get_field() 112 return (ale_entry[idx] >> start) & BITMASK(bits); in cpsw_ale_get_field() 118 int idx; in cpsw_ale_set_field() local 121 idx = start / 32; in cpsw_ale_set_field() 122 start -= idx * 32; in cpsw_ale_set_field() 123 idx = 2 - idx; /* flip */ in cpsw_ale_set_field() 124 ale_entry[idx] &= ~(BITMASK(bits) << start); in cpsw_ale_set_field() [all …]
|