/arch/arm/mach-cns3xxx/ |
D | pm.c | 15 void cns3xxx_pwr_clk_en(unsigned int block) in cns3xxx_pwr_clk_en() argument 19 reg |= (block & PM_CLK_GATE_REG_MASK); in cns3xxx_pwr_clk_en() 24 void cns3xxx_pwr_clk_dis(unsigned int block) in cns3xxx_pwr_clk_dis() argument 28 reg &= ~(block & PM_CLK_GATE_REG_MASK); in cns3xxx_pwr_clk_dis() 33 void cns3xxx_pwr_power_up(unsigned int block) in cns3xxx_pwr_power_up() argument 37 reg &= ~(block & CNS3XXX_PWR_PLL_ALL); in cns3xxx_pwr_power_up() 45 void cns3xxx_pwr_power_down(unsigned int block) in cns3xxx_pwr_power_down() argument 50 reg |= (block & CNS3XXX_PWR_PLL_ALL); in cns3xxx_pwr_power_down() 55 static void cns3xxx_pwr_soft_rst_force(unsigned int block) in cns3xxx_pwr_soft_rst_force() argument 63 if (block & 0x30000001) { in cns3xxx_pwr_soft_rst_force() [all …]
|
D | pm.h | 13 void cns3xxx_pwr_clk_en(unsigned int block); 14 void cns3xxx_pwr_clk_dis(unsigned int block); 15 void cns3xxx_pwr_power_up(unsigned int block); 16 void cns3xxx_pwr_power_down(unsigned int block);
|
/arch/mips/cavium-octeon/crypto/ |
D | octeon-md5.c | 55 const u64 *block = _block; in octeon_md5_transform() local 57 write_octeon_64bit_block_dword(block[0], 0); in octeon_md5_transform() 58 write_octeon_64bit_block_dword(block[1], 1); in octeon_md5_transform() 59 write_octeon_64bit_block_dword(block[2], 2); in octeon_md5_transform() 60 write_octeon_64bit_block_dword(block[3], 3); in octeon_md5_transform() 61 write_octeon_64bit_block_dword(block[4], 4); in octeon_md5_transform() 62 write_octeon_64bit_block_dword(block[5], 5); in octeon_md5_transform() 63 write_octeon_64bit_block_dword(block[6], 6); in octeon_md5_transform() 64 octeon_md5_start(block[7]); in octeon_md5_transform() 85 const u32 avail = sizeof(mctx->block) - (mctx->byte_count & 0x3f); in octeon_md5_update() [all …]
|
D | octeon-sha512.c | 57 const u64 *block = _block; in octeon_sha512_transform() local 59 write_octeon_64bit_block_sha512(block[0], 0); in octeon_sha512_transform() 60 write_octeon_64bit_block_sha512(block[1], 1); in octeon_sha512_transform() 61 write_octeon_64bit_block_sha512(block[2], 2); in octeon_sha512_transform() 62 write_octeon_64bit_block_sha512(block[3], 3); in octeon_sha512_transform() 63 write_octeon_64bit_block_sha512(block[4], 4); in octeon_sha512_transform() 64 write_octeon_64bit_block_sha512(block[5], 5); in octeon_sha512_transform() 65 write_octeon_64bit_block_sha512(block[6], 6); in octeon_sha512_transform() 66 write_octeon_64bit_block_sha512(block[7], 7); in octeon_sha512_transform() 67 write_octeon_64bit_block_sha512(block[8], 8); in octeon_sha512_transform() [all …]
|
D | octeon-sha1.c | 62 const u64 *block = _block; in octeon_sha1_transform() local 64 write_octeon_64bit_block_dword(block[0], 0); in octeon_sha1_transform() 65 write_octeon_64bit_block_dword(block[1], 1); in octeon_sha1_transform() 66 write_octeon_64bit_block_dword(block[2], 2); in octeon_sha1_transform() 67 write_octeon_64bit_block_dword(block[3], 3); in octeon_sha1_transform() 68 write_octeon_64bit_block_dword(block[4], 4); in octeon_sha1_transform() 69 write_octeon_64bit_block_dword(block[5], 5); in octeon_sha1_transform() 70 write_octeon_64bit_block_dword(block[6], 6); in octeon_sha1_transform() 71 octeon_sha1_start(block[7]); in octeon_sha1_transform()
|
D | octeon-sha256.c | 54 const u64 *block = _block; in octeon_sha256_transform() local 56 write_octeon_64bit_block_dword(block[0], 0); in octeon_sha256_transform() 57 write_octeon_64bit_block_dword(block[1], 1); in octeon_sha256_transform() 58 write_octeon_64bit_block_dword(block[2], 2); in octeon_sha256_transform() 59 write_octeon_64bit_block_dword(block[3], 3); in octeon_sha256_transform() 60 write_octeon_64bit_block_dword(block[4], 4); in octeon_sha256_transform() 61 write_octeon_64bit_block_dword(block[5], 5); in octeon_sha256_transform() 62 write_octeon_64bit_block_dword(block[6], 6); in octeon_sha256_transform() 63 octeon_sha256_start(block[7]); in octeon_sha256_transform()
|
/arch/mips/alchemy/common/ |
D | vss.c | 23 static inline void __enable_block(int block) in __enable_block() argument 25 void __iomem *base = (void __iomem *)VSS_ADDR(block); in __enable_block() 54 static inline void __disable_block(int block) in __disable_block() argument 56 void __iomem *base = (void __iomem *)VSS_ADDR(block); in __disable_block() 70 void au1300_vss_block_control(int block, int enable) in au1300_vss_block_control() argument 80 __enable_block(block); in au1300_vss_block_control() 82 __disable_block(block); in au1300_vss_block_control()
|
D | usb.c | 265 static inline int au1300_usb_control(int block, int enable) in au1300_usb_control() argument 271 switch (block) { in au1300_usb_control() 360 static inline int au1200_usb_control(int block, int enable) in au1200_usb_control() argument 365 switch (block) { in au1200_usb_control() 458 static inline int au1000_usb_control(int block, int enable, unsigned long rb, in au1000_usb_control() argument 463 switch (block) { in au1000_usb_control() 478 int alchemy_usb_control(int block, int enable) in alchemy_usb_control() argument 488 ret = au1000_usb_control(block, enable, in alchemy_usb_control() 492 ret = au1000_usb_control(block, enable, in alchemy_usb_control() 496 ret = au1200_usb_control(block, enable); in alchemy_usb_control() [all …]
|
/arch/x86/crypto/ |
D | blake2s-glue.c | 20 const u8 *block, const size_t nblocks, 23 const u8 *block, const size_t nblocks, 29 void blake2s_compress(struct blake2s_state *state, const u8 *block, in blake2s_compress() argument 36 blake2s_compress_generic(state, block, nblocks, inc); in blake2s_compress() 47 blake2s_compress_avx512(state, block, blocks, inc); in blake2s_compress() 49 blake2s_compress_ssse3(state, block, blocks, inc); in blake2s_compress() 53 block += blocks * BLAKE2S_BLOCK_SIZE; in blake2s_compress()
|
/arch/powerpc/mm/kasan/ |
D | 8xx.c | 10 kasan_init_shadow_8M(unsigned long k_start, unsigned long k_end, void *block) in kasan_init_shadow_8M() argument 15 for (k_cur = k_start; k_cur != k_end; k_cur = k_next, pmd += 2, block += SZ_8M) { in kasan_init_shadow_8M() 27 *new = pte_val(pte_mkhuge(pfn_pte(PHYS_PFN(__pa(block)), PAGE_KERNEL))); in kasan_init_shadow_8M() 41 void *block; in kasan_init_region() local 43 block = memblock_alloc(k_end - k_start, SZ_8M); in kasan_init_region() 44 if (!block) in kasan_init_region() 48 kasan_init_shadow_8M(k_start, ALIGN_DOWN(k_end, SZ_8M), block); in kasan_init_region() 62 void *va = block + k_cur - k_start; in kasan_init_region()
|
/arch/s390/boot/ |
D | mem_detect.c | 51 struct mem_detect_block *block; in add_mem_detect_block() local 54 block = __get_mem_detect_block_ptr(mem_detect.count - 1); in add_mem_detect_block() 55 if (block->end == start) { in add_mem_detect_block() 56 block->end = end; in add_mem_detect_block() 61 block = __get_mem_detect_block_ptr(mem_detect.count); in add_mem_detect_block() 62 block->start = start; in add_mem_detect_block() 63 block->end = end; in add_mem_detect_block()
|
/arch/s390/kernel/ |
D | sthyi.c | 229 struct diag204_x_phys_block *block, in fill_diag_mac() argument 234 for (i = 0; i < block->hdr.cpus; i++) { in fill_diag_mac() 235 switch (cpu_id(block->cpus[i].ctidx, diag224_buf)) { in fill_diag_mac() 237 if (block->cpus[i].weight == DED_WEIGHT) in fill_diag_mac() 243 if (block->cpus[i].weight == DED_WEIGHT) in fill_diag_mac() 257 struct diag204_x_part_block *block) in lpar_cpu_inf() argument 262 for (i = 0; i < block->hdr.rcpus; i++) { in lpar_cpu_inf() 263 if (!(block->cpus[i].cflag & DIAG204_CPU_ONLINE)) in lpar_cpu_inf() 266 switch (cpu_id(block->cpus[i].ctidx, diag224_buf)) { in lpar_cpu_inf() 269 if (block->cpus[i].cur_weight < DED_WEIGHT) in lpar_cpu_inf() [all …]
|
/arch/powerpc/crypto/ |
D | md5-glue.c | 54 memcpy((char *)sctx->block + offset, src, len); in ppc_md5_update() 59 memcpy((char *)sctx->block + offset, src, avail); in ppc_md5_update() 60 ppc_md5_transform(sctx->hash, (const u8 *)sctx->block, 1); in ppc_md5_update() 71 memcpy((char *)sctx->block, src, len); in ppc_md5_update() 79 const u8 *src = (const u8 *)sctx->block; in ppc_md5_final() 82 __le64 *pbits = (__le64 *)((char *)sctx->block + 56); in ppc_md5_final() 90 p = (char *)sctx->block; in ppc_md5_final()
|
/arch/arm/crypto/ |
D | blake2b-neon-glue.c | 19 const u8 *block, size_t nblocks, u32 inc); 22 const u8 *block, size_t nblocks, u32 inc) in blake2b_compress_arch() argument 25 blake2b_compress_generic(state, block, nblocks, inc); in blake2b_compress_arch() 34 blake2b_compress_neon(state, block, blocks, inc); in blake2b_compress_arch() 38 block += blocks * BLAKE2B_BLOCK_SIZE; in blake2b_compress_arch()
|
/arch/x86/kernel/cpu/mce/ |
D | amd.c | 374 b->bank, b->block, b->address, hi, lo); in lvt_off_valid() 389 b->cpu, apic, b->bank, b->block, b->address, hi, lo); in lvt_off_valid() 502 static u32 smca_get_block_address(unsigned int bank, unsigned int block, in smca_get_block_address() argument 505 if (!block) in smca_get_block_address() 511 return MSR_AMD64_SMCA_MCx_MISCy(bank, block - 1); in smca_get_block_address() 515 unsigned int bank, unsigned int block, in get_block_address() argument 520 if ((bank >= per_cpu(mce_num_banks, cpu)) || (block >= NR_BLOCKS)) in get_block_address() 524 return smca_get_block_address(bank, block, cpu); in get_block_address() 527 switch (block) { in get_block_address() 543 prepare_threshold_block(unsigned int bank, unsigned int block, u32 addr, in prepare_threshold_block() argument [all …]
|
/arch/powerpc/lib/ |
D | rheap.c | 47 rh_block_t *block, *blk; in grow() local 57 block = kmalloc_array(max_blocks, sizeof(rh_block_t), GFP_ATOMIC); in grow() 58 if (block == NULL) in grow() 64 memcpy(block, info->block, in grow() 67 delta = (char *)block - (char *)info->block; in grow() 70 blks = (unsigned long)info->block; in grow() 71 blke = (unsigned long)(info->block + info->max_blocks); in grow() 73 for (i = 0, blk = block; i < info->max_blocks; i++, blk++) in grow() 82 kfree(info->block); in grow() 85 info->block = block; in grow() [all …]
|
/arch/mips/cavium-octeon/executive/ |
D | cvmx-interrupt-rsl.c | 42 void __cvmx_interrupt_gmxx_rxx_int_en_enable(int index, int block); 50 void __cvmx_interrupt_asxx_enable(int block) in __cvmx_interrupt_asxx_enable() argument 65 csr.u64 = cvmx_read_csr(CVMX_ASXX_INT_EN(block)); in __cvmx_interrupt_asxx_enable() 69 cvmx_write_csr(CVMX_ASXX_INT_EN(block), csr.u64); in __cvmx_interrupt_asxx_enable()
|
D | cvmx-interrupt-decodes.c | 53 void __cvmx_interrupt_gmxx_rxx_int_en_enable(int index, int block) in __cvmx_interrupt_gmxx_rxx_int_en_enable() argument 56 cvmx_write_csr(CVMX_GMXX_RXX_INT_REG(index, block), in __cvmx_interrupt_gmxx_rxx_int_en_enable() 57 cvmx_read_csr(CVMX_GMXX_RXX_INT_REG(index, block))); in __cvmx_interrupt_gmxx_rxx_int_en_enable() 229 cvmx_write_csr(CVMX_GMXX_RXX_INT_EN(index, block), gmx_rx_int_en.u64); in __cvmx_interrupt_gmxx_rxx_int_en_enable() 236 void __cvmx_interrupt_pcsx_intx_en_reg_enable(int index, int block) in __cvmx_interrupt_pcsx_intx_en_reg_enable() argument 239 cvmx_write_csr(CVMX_PCSX_INTX_REG(index, block), in __cvmx_interrupt_pcsx_intx_en_reg_enable() 240 cvmx_read_csr(CVMX_PCSX_INTX_REG(index, block))); in __cvmx_interrupt_pcsx_intx_en_reg_enable() 272 cvmx_write_csr(CVMX_PCSX_INTX_EN_REG(index, block), pcs_int_en_reg.u64); in __cvmx_interrupt_pcsx_intx_en_reg_enable()
|
/arch/sparc/crypto/ |
D | md5_glue.c | 54 memcpy((u8 *)sctx->block + partial, data, done); in __md5_sparc64_update() 55 md5_sparc64_transform(sctx->hash, (u8 *)sctx->block, 1); in __md5_sparc64_update() 64 memcpy(sctx->block, data + done, len - done); in __md5_sparc64_update() 76 memcpy((u8 *)sctx->block + partial, data, len); in md5_sparc64_update() 101 memcpy((u8 *)sctx->block + index, padding, padlen); in md5_sparc64_final()
|
/arch/riscv/boot/dts/microchip/ |
D | microchip-mpfs.dtsi | 20 i-cache-block-size = <64>; 37 d-cache-block-size = <64>; 43 i-cache-block-size = <64>; 64 d-cache-block-size = <64>; 70 i-cache-block-size = <64>; 91 d-cache-block-size = <64>; 97 i-cache-block-size = <64>; 118 d-cache-block-size = <64>; 124 i-cache-block-size = <64>; 150 cache-block-size = <64>;
|
/arch/riscv/boot/dts/sifive/ |
D | fu540-c000.dtsi | 28 i-cache-block-size = <64>; 42 d-cache-block-size = <64>; 48 i-cache-block-size = <64>; 66 d-cache-block-size = <64>; 72 i-cache-block-size = <64>; 90 d-cache-block-size = <64>; 96 i-cache-block-size = <64>; 114 d-cache-block-size = <64>; 120 i-cache-block-size = <64>; 262 cache-block-size = <64>;
|
D | fu740-c000.dtsi | 28 i-cache-block-size = <64>; 43 d-cache-block-size = <64>; 49 i-cache-block-size = <64>; 67 d-cache-block-size = <64>; 73 i-cache-block-size = <64>; 91 d-cache-block-size = <64>; 97 i-cache-block-size = <64>; 115 d-cache-block-size = <64>; 121 i-cache-block-size = <64>; 294 cache-block-size = <64>;
|
/arch/arm/lib/ |
D | memmove.S | 39 UNWIND( .save {r0, r4, lr} ) @ in first stmfd block 56 UNWIND( .save {r5 - r8} ) @ in second stmfd block 109 UNWIND( .fnend ) @ end of second stmfd block 112 UNWIND( .save {r0, r4, lr} ) @ still in first stmfd block 146 UNWIND( .save {r0, r4, lr} ) @ still in first stmfd block 160 UNWIND( .save {r5 - r9} ) @ in new second stmfd block 195 UNWIND( .fnend ) @ end of the second stmfd block 198 UNWIND( .save {r0, r4, lr} ) @ still in first stmfd block
|
/arch/arm/kernel/ |
D | setup.c | 451 int block; in cpuid_init_hwcaps() local 457 block = cpuid_feature_extract(CPUID_EXT_ISAR0, 24); in cpuid_init_hwcaps() 458 if (block >= 2) in cpuid_init_hwcaps() 460 if (block >= 1) in cpuid_init_hwcaps() 464 block = cpuid_feature_extract(CPUID_EXT_MMFR0, 0); in cpuid_init_hwcaps() 465 if (block >= 5) in cpuid_init_hwcaps() 471 block = cpuid_feature_extract_field(isar5, 4); in cpuid_init_hwcaps() 472 if (block >= 2) in cpuid_init_hwcaps() 474 if (block >= 1) in cpuid_init_hwcaps() 477 block = cpuid_feature_extract_field(isar5, 8); in cpuid_init_hwcaps() [all …]
|
/arch/x86/boot/ |
D | header.S | 476 # 1 bit (last block flag) 477 # 2 bits (block type) 479 # 1 block occurs every 32K -1 bytes or when there 50% compression 480 # has been achieved. The smallest block type encoding is always used. 506 # The worst case at the block level is a growth of the compressed data 509 # The worst case internal to a compressed block is very hard to figure. 515 # to be safe. To avoid problems at the block level allocating 5 extra bytes 517 # block adding an extra 32767 bytes (the worst case uncompressed block size) 519 # block will stop the byte before the compressed data for a block begins. 544 # byte fixed overhead but has a maximum block size of 128K, so it needs a
|