Searched refs:mask (Results 1 – 4 of 4) sorted by relevance
708 void blk_queue_dma_pad(struct request_queue *q, unsigned int mask) in blk_queue_dma_pad() argument710 q->dma_pad_mask = mask; in blk_queue_dma_pad()724 void blk_queue_update_dma_pad(struct request_queue *q, unsigned int mask) in blk_queue_update_dma_pad() argument726 if (mask > q->dma_pad_mask) in blk_queue_update_dma_pad()727 q->dma_pad_mask = mask; in blk_queue_update_dma_pad()773 void blk_queue_segment_boundary(struct request_queue *q, unsigned long mask) in blk_queue_segment_boundary() argument775 if (mask < PAGE_CACHE_SIZE - 1) { in blk_queue_segment_boundary()776 mask = PAGE_CACHE_SIZE - 1; in blk_queue_segment_boundary()778 __func__, mask); in blk_queue_segment_boundary()781 q->limits.seg_boundary_mask = mask; in blk_queue_segment_boundary()[all …]
1562 void disk_flush_events(struct gendisk *disk, unsigned int mask) in disk_flush_events() argument1570 ev->clearing |= mask; in disk_flush_events()1588 unsigned int disk_clear_events(struct gendisk *disk, unsigned int mask) in disk_clear_events() argument1593 unsigned int clearing = mask; in disk_clear_events()1597 if ((mask & DISK_EVENT_MEDIA_CHANGE) && in disk_clear_events()1624 pending = ev->pending & mask; in disk_clear_events()1625 ev->pending &= ~mask; in disk_clear_events()1627 WARN_ON_ONCE(clearing & mask); in disk_clear_events()
853 unsigned int mask = 0; in bsg_poll() local860 mask |= POLLIN | POLLRDNORM; in bsg_poll()862 mask |= POLLOUT; in bsg_poll()865 return mask; in bsg_poll()
75 unsigned mask = queue_logical_block_size(q) - 1; in get_max_io_size() local78 sectors &= ~(mask >> 9); in get_max_io_size()