/lib/ |
D | cpumask.c | 34 int cpumask_any_but(const struct cpumask *mask, unsigned int cpu) in cpumask_any_but() argument 39 for_each_cpu(i, mask) in cpumask_any_but() 57 int cpumask_next_wrap(int n, const struct cpumask *mask, int start, bool wrap) in cpumask_next_wrap() argument 62 next = cpumask_next(n, mask); in cpumask_next_wrap() 93 bool alloc_cpumask_var_node(cpumask_var_t *mask, gfp_t flags, int node) in alloc_cpumask_var_node() argument 95 *mask = kmalloc_node(cpumask_size(), flags, node); in alloc_cpumask_var_node() 98 if (!*mask) { in alloc_cpumask_var_node() 104 return *mask != NULL; in alloc_cpumask_var_node() 108 bool zalloc_cpumask_var_node(cpumask_var_t *mask, gfp_t flags, int node) in zalloc_cpumask_var_node() argument 110 return alloc_cpumask_var_node(mask, flags | __GFP_ZERO, node); in zalloc_cpumask_var_node() [all …]
|
D | kfifo.c | 35 return (fifo->mask + 1) - (fifo->in - fifo->out); in kfifo_unused() 53 fifo->mask = 0; in __kfifo_alloc() 60 fifo->mask = 0; in __kfifo_alloc() 63 fifo->mask = size - 1; in __kfifo_alloc() 76 fifo->mask = 0; in __kfifo_free() 94 fifo->mask = 0; in __kfifo_init() 97 fifo->mask = size - 1; in __kfifo_init() 106 unsigned int size = fifo->mask + 1; in kfifo_copy_in() 110 off &= fifo->mask; in kfifo_copy_in() 145 unsigned int size = fifo->mask + 1; in kfifo_copy_out() [all …]
|
D | stmp_device.c | 30 static int stmp_clear_poll_bit(void __iomem *addr, u32 mask) in stmp_clear_poll_bit() argument 34 writel(mask, addr + STMP_OFFSET_REG_CLR); in stmp_clear_poll_bit() 36 while ((readl(addr) & mask) && --timeout) in stmp_clear_poll_bit()
|
D | devres.c | 336 int pcim_iomap_regions(struct pci_dev *pdev, int mask, const char *name) in pcim_iomap_regions() argument 348 if (!(mask & (1 << i))) in pcim_iomap_regions() 371 if (!(mask & (1 << i))) in pcim_iomap_regions() 389 int pcim_iomap_regions_request_all(struct pci_dev *pdev, int mask, in pcim_iomap_regions_request_all() argument 392 int request_mask = ((1 << 6) - 1) & ~mask; in pcim_iomap_regions_request_all() 399 rc = pcim_iomap_regions(pdev, mask, name); in pcim_iomap_regions_request_all() 413 void pcim_iounmap_regions(struct pci_dev *pdev, int mask) in pcim_iounmap_regions() argument 423 if (!(mask & (1 << i))) in pcim_iounmap_regions()
|
D | bitmap.c | 89 unsigned long mask = BITMAP_LAST_WORD_MASK(nbits); in __bitmap_shift_right() local 102 upper &= mask; in __bitmap_shift_right() 107 lower &= mask; in __bitmap_shift_right() 954 unsigned long mask; /* bitmask for one long of region */ in __reg_op() local 972 mask = (1UL << (nbitsinlong - 1)); in __reg_op() 973 mask += mask - 1; in __reg_op() 974 mask <<= offset; in __reg_op() 979 if (bitmap[index + i] & mask) in __reg_op() 987 bitmap[index + i] |= mask; in __reg_op() 992 bitmap[index + i] &= ~mask; in __reg_op()
|
D | iommu-common.c | 101 unsigned long mask, in iommu_tbl_range_alloc() argument 153 if (limit + shift > mask) { in iommu_tbl_range_alloc() 154 limit = mask - shift + 1; in iommu_tbl_range_alloc() 159 if ((start & mask) >= limit || pass > 0) { in iommu_tbl_range_alloc() 165 start &= mask; in iommu_tbl_range_alloc()
|
D | cpu_rmap.c | 98 const struct cpumask *mask, u16 dist) in cpu_rmap_copy_neigh() argument 102 for_each_cpu(neigh, mask) { in cpu_rmap_copy_neigh() 253 irq_cpu_rmap_notify(struct irq_affinity_notify *notify, const cpumask_t *mask) in irq_cpu_rmap_notify() argument 259 rc = cpu_rmap_update(glue->rmap, glue->index, mask); in irq_cpu_rmap_notify()
|
D | swiotlb.c | 436 unsigned long mask; in swiotlb_tbl_map_single() local 443 mask = dma_get_seg_boundary(hwdev); in swiotlb_tbl_map_single() 445 tbl_dma_addr &= mask; in swiotlb_tbl_map_single() 452 max_slots = mask + 1 in swiotlb_tbl_map_single() 453 ? ALIGN(mask + 1, 1 << IO_TLB_SHIFT) >> IO_TLB_SHIFT in swiotlb_tbl_map_single() 1000 swiotlb_dma_supported(struct device *hwdev, u64 mask) in swiotlb_dma_supported() argument 1002 return phys_to_dma(hwdev, io_tlb_end - 1) <= mask; in swiotlb_dma_supported()
|
D | bch.c | 436 unsigned int tmp, mask; in solve_linear_system() local 440 mask = 1 << m; in solve_linear_system() 448 if (rows[r] & mask) { in solve_linear_system() 462 if (rows[r] & mask) in solve_linear_system() 469 mask >>= 1; in solve_linear_system() 496 mask = rows[r] & (tmp|1); in solve_linear_system() 497 tmp |= parity(mask) << (m-r); in solve_linear_system() 514 unsigned int mask = 0xff, t, rows[16] = {0,}; in find_affine4_roots() local 532 for (j = 8; j != 0; j >>= 1, mask ^= (mask << j)) { in find_affine4_roots() 534 t = ((rows[k] >> j)^rows[k+j]) & mask; in find_affine4_roots()
|
D | dynamic_debug.c | 139 unsigned int flags, unsigned int mask) in ddebug_change() argument 188 newflags = (dp->flags & mask) | flags; in ddebug_change() 438 unsigned int flags = 0, mask = 0; in ddebug_exec_query() local 450 if (ddebug_parse_flags(words[nwords-1], &flags, &mask)) { in ddebug_exec_query() 459 nfound = ddebug_change(&query, flags, mask); in ddebug_exec_query()
|
D | nmi_backtrace.c | 53 void (*raise)(cpumask_t *mask)) in nmi_trigger_all_cpu_backtrace() argument
|
D | vsprintf.c | 434 int mask = spec.base - 1; in number() local 440 tmp[i++] = (hex_asc_upper[((unsigned char)num) & mask] | locase); in number()
|
/lib/zlib_inflate/ |
D | inftrees.c | 38 unsigned mask; /* mask for low root bits */ in zlib_inflate_table() local 195 mask = used - 1; /* mask for comparing low */ in zlib_inflate_table() 246 if (len > root && (huff & mask) != low) { in zlib_inflate_table() 270 low = huff & mask; in zlib_inflate_table() 289 if (drop != 0 && (huff & mask) != low) { in zlib_inflate_table()
|
/lib/xz/ |
D | xz_dec_bcj.c | 214 uint32_t mask; in bcj_ia64() local 223 mask = branch_table[buf[i] & 0x1F]; in bcj_ia64() 225 if (((mask >> slot) & 1) == 0) in bcj_ia64()
|
D | xz_dec_lzma2.c | 572 uint32_t mask; in rc_direct() local 578 mask = (uint32_t)0 - (rc->code >> 31); in rc_direct() 579 rc->code += rc->range & mask; in rc_direct() 580 *dest = (*dest << 1) + (mask + 1); in rc_direct()
|