/drivers/iommu/ |
D | msm_iommu_hw-8xxx.h | 33 #define SET_GLOBAL_REG_N(b, n, r, v) SET_GLOBAL_REG(b, ((r) + (n << 2)), (v)) argument 34 #define GET_GLOBAL_REG_N(b, n, r) GET_GLOBAL_REG(b, ((r) + (n << 2))) argument 37 #define GET_GLOBAL_FIELD(b, r, F) GET_FIELD(((b) + (r)), F##_MASK, F##_SHIFT) argument 38 #define GET_CONTEXT_FIELD(b, c, r, F) \ argument 39 GET_FIELD(((b) + (r) + ((c) << CTX_SHIFT)), F##_MASK, F##_SHIFT) 41 #define SET_GLOBAL_FIELD(b, r, F, v) \ argument 42 SET_FIELD(((b) + (r)), F##_MASK, F##_SHIFT, (v)) 43 #define SET_CONTEXT_FIELD(b, c, r, F, v) \ argument 44 SET_FIELD(((b) + (r) + ((c) << CTX_SHIFT)), F##_MASK, F##_SHIFT, (v)) 97 #define SET_M2VCBR_N(b, N, v) SET_GLOBAL_REG_N(M2VCBR_N, N, (b), (v)) argument [all …]
|
/drivers/md/bcache/ |
D | btree.c | 95 #define insert_lock(s, b) ((b)->level <= (s)->lock) argument 115 #define btree(fn, key, b, op, ...) \ argument 117 int _r, l = (b)->level - 1; \ 119 struct btree *_child = bch_btree_node_get((b)->c, op, key, l, \ 120 _w, b); \ 156 static inline struct bset *write_block(struct btree *b) in write_block() argument 158 return ((void *) btree_bset_first(b)) + b->written * block_bytes(b->c); in write_block() 161 static void bch_btree_init_next(struct btree *b) in bch_btree_init_next() argument 164 if (b->level && b->keys.nsets) in bch_btree_init_next() 165 bch_btree_sort(&b->keys, &b->c->sort); in bch_btree_init_next() [all …]
|
D | alloc.c | 73 uint8_t bch_inc_gen(struct cache *ca, struct bucket *b) in bch_inc_gen() argument 75 uint8_t ret = ++b->gen; in bch_inc_gen() 77 ca->set->need_gc = max(ca->set->need_gc, bucket_gc_gen(b)); in bch_inc_gen() 86 struct bucket *b; in bch_rescale_priorities() local 105 for_each_bucket(b, ca) in bch_rescale_priorities() 106 if (b->prio && in bch_rescale_priorities() 107 b->prio != BTREE_PRIO && in bch_rescale_priorities() 108 !atomic_read(&b->pin)) { in bch_rescale_priorities() 109 b->prio--; in bch_rescale_priorities() 110 c->min_prio = min(c->min_prio, b->prio); in bch_rescale_priorities() [all …]
|
D | bset.c | 19 void bch_dump_bset(struct btree_keys *b, struct bset *i, unsigned set) in bch_dump_bset() argument 29 if (b->ops->key_dump) in bch_dump_bset() 30 b->ops->key_dump(b, k); in bch_dump_bset() 35 bkey_cmp(k, b->ops->is_extents ? in bch_dump_bset() 41 void bch_dump_bucket(struct btree_keys *b) in bch_dump_bucket() argument 46 for (i = 0; i <= b->nsets; i++) in bch_dump_bucket() 47 bch_dump_bset(b, b->set[i].data, in bch_dump_bucket() 48 bset_sector_offset(b, b->set[i].data)); in bch_dump_bucket() 52 int __bch_count_data(struct btree_keys *b) in __bch_count_data() argument 58 if (b->ops->is_extents) in __bch_count_data() [all …]
|
D | bset.h | 226 static inline struct bset_tree *bset_tree_last(struct btree_keys *b) in bset_tree_last() argument 228 return b->set + b->nsets; in bset_tree_last() 231 static inline bool bset_written(struct btree_keys *b, struct bset_tree *t) in bset_written() argument 233 return t <= b->set + b->nsets - b->last_set_unwritten; in bset_written() 236 static inline bool bkey_written(struct btree_keys *b, struct bkey *k) in bkey_written() argument 238 return !b->last_set_unwritten || k < b->set[b->nsets].data->start; in bkey_written() 241 static inline unsigned bset_byte_offset(struct btree_keys *b, struct bset *i) in bset_byte_offset() argument 243 return ((size_t) i) - ((size_t) b->set->data); in bset_byte_offset() 246 static inline unsigned bset_sector_offset(struct btree_keys *b, struct bset *i) in bset_sector_offset() argument 248 return bset_byte_offset(b, i) >> 9; in bset_sector_offset() [all …]
|
D | extents.c | 127 struct btree *b = container_of(keys, struct btree, keys); in bch_bkey_dump() local 135 size_t n = PTR_BUCKET_NR(b->c, k, j); in bch_bkey_dump() 138 if (n >= b->c->sb.first_bucket && n < b->c->sb.nbuckets) in bch_bkey_dump() 140 PTR_BUCKET(b->c, k, j)->prio); in bch_bkey_dump() 143 printk(" %s\n", bch_ptr_status(b->c, k)); in bch_bkey_dump() 167 struct btree *b = container_of(bk, struct btree, keys); in bch_btree_ptr_invalid() local 168 return __bch_btree_ptr_invalid(b->c, k); in bch_btree_ptr_invalid() 171 static bool btree_ptr_bad_expensive(struct btree *b, const struct bkey *k) in btree_ptr_bad_expensive() argument 177 if (mutex_trylock(&b->c->bucket_lock)) { in btree_ptr_bad_expensive() 179 if (ptr_available(b->c, k, i)) { in btree_ptr_bad_expensive() [all …]
|
D | btree.h | 150 static inline bool btree_node_ ## flag(struct btree *b) \ 151 { return test_bit(BTREE_NODE_ ## flag, &b->flags); } \ 153 static inline void set_btree_node_ ## flag(struct btree *b) \ 154 { set_bit(BTREE_NODE_ ## flag, &b->flags); } \ 166 static inline struct btree_write *btree_current_write(struct btree *b) in btree_current_write() argument 168 return b->writes + btree_node_write_idx(b); in btree_current_write() 171 static inline struct btree_write *btree_prev_write(struct btree *b) in btree_prev_write() argument 173 return b->writes + (btree_node_write_idx(b) ^ 1); in btree_prev_write() 176 static inline struct bset *btree_bset_first(struct btree *b) in btree_bset_first() argument 178 return b->keys.set->data; in btree_bset_first() [all …]
|
D | debug.c | 23 #define for_each_written_bset(b, start, i) \ argument 25 (void *) i < (void *) (start) + (KEY_SIZE(&b->key) << 9) &&\ 27 i = (void *) i + set_blocks(i, block_bytes(b->c)) * \ 28 block_bytes(b->c)) 30 void bch_btree_verify(struct btree *b) in bch_btree_verify() argument 32 struct btree *v = b->c->verify_data; in bch_btree_verify() 36 if (!b->c->verify || !b->c->verify_ondisk) in bch_btree_verify() 39 down(&b->io_mutex); in bch_btree_verify() 40 mutex_lock(&b->c->verify_lock); in bch_btree_verify() 42 ondisk = b->c->verify_ondisk; in bch_btree_verify() [all …]
|
/drivers/misc/ |
D | vmw_balloon.c | 258 void (*add_page)(struct vmballoon *b, int idx, struct page *p); 259 int (*lock)(struct vmballoon *b, unsigned int num_pages, 261 int (*unlock)(struct vmballoon *b, unsigned int num_pages, 322 static bool vmballoon_send_start(struct vmballoon *b, unsigned long req_caps) in vmballoon_send_start() argument 327 STATS_INC(b->stats.start); in vmballoon_send_start() 333 b->capabilities = capabilities; in vmballoon_send_start() 337 b->capabilities = VMW_BALLOON_BASIC_CMDS; in vmballoon_send_start() 344 if (b->capabilities & VMW_BALLOON_BATCHED_2M_CMDS) in vmballoon_send_start() 345 b->supported_page_sizes = 2; in vmballoon_send_start() 347 b->supported_page_sizes = 1; in vmballoon_send_start() [all …]
|
/drivers/md/ |
D | dm-bufio.c | 250 static void buffer_record_stack(struct dm_buffer *b) in buffer_record_stack() argument 252 b->stack_trace.nr_entries = 0; in buffer_record_stack() 253 b->stack_trace.max_entries = MAX_STACK; in buffer_record_stack() 254 b->stack_trace.entries = b->stack_entries; in buffer_record_stack() 255 b->stack_trace.skip = 2; in buffer_record_stack() 256 save_stack_trace(&b->stack_trace); in buffer_record_stack() 266 struct dm_buffer *b; in __find() local 269 b = container_of(n, struct dm_buffer, node); in __find() 271 if (b->block == block) in __find() 272 return b; in __find() [all …]
|
D | dm-cache-block-types.h | 24 static inline dm_oblock_t to_oblock(dm_block_t b) in to_oblock() argument 26 return (__force dm_oblock_t) b; in to_oblock() 29 static inline dm_block_t from_oblock(dm_oblock_t b) in from_oblock() argument 31 return (__force dm_block_t) b; in from_oblock() 34 static inline dm_cblock_t to_cblock(uint32_t b) in to_cblock() argument 36 return (__force dm_cblock_t) b; in to_cblock() 39 static inline uint32_t from_cblock(dm_cblock_t b) in from_cblock() argument 41 return (__force uint32_t) b; in from_cblock() 44 static inline dm_dblock_t to_dblock(dm_block_t b) in to_dblock() argument 46 return (__force dm_dblock_t) b; in to_dblock() [all …]
|
/drivers/block/drbd/ |
D | drbd_bitmap.c | 127 struct drbd_bitmap *b = device->bitmap; in __bm_print_lock_info() local 132 func, b->bm_why ?: "?", in __bm_print_lock_info() 133 b->bm_task->comm, task_pid_nr(b->bm_task)); in __bm_print_lock_info() 138 struct drbd_bitmap *b = device->bitmap; in drbd_bm_lock() local 141 if (!b) { in drbd_bm_lock() 146 trylock_failed = !mutex_trylock(&b->bm_change); in drbd_bm_lock() 151 why, b->bm_why ?: "?", in drbd_bm_lock() 152 b->bm_task->comm, task_pid_nr(b->bm_task)); in drbd_bm_lock() 153 mutex_lock(&b->bm_change); in drbd_bm_lock() 155 if (BM_LOCKED_MASK & b->bm_flags) in drbd_bm_lock() [all …]
|
D | drbd_vli.h | 150 #define LEVEL(t,b,v) \ in vli_decode_bits() argument 152 if ((in & ((1 << b) -1)) == v) { \ in vli_decode_bits() 153 *out = ((in & ((~0ULL) >> (64-t))) >> b) + adj; \ in vli_decode_bits() 156 adj += 1ULL << (t - b); \ in vli_decode_bits() 176 #define LEVEL(t,b,v) do { \ in __vli_encode_bits() argument 177 max += 1ULL << (t - b); \ in __vli_encode_bits() 180 *out = ((in - adj) << b) | v; \ in __vli_encode_bits() 205 u8 *b; member 213 cur->b = s; in bitstream_cursor_reset() 222 cur->b = cur->b + (bits >> 3); in bitstream_cursor_advance() [all …]
|
/drivers/gpu/drm/i915/ |
D | intel_breadcrumbs.c | 32 struct intel_breadcrumbs *b = &engine->breadcrumbs; in intel_breadcrumbs_hangcheck() local 34 if (!b->irq_enabled) in intel_breadcrumbs_hangcheck() 37 if (time_before(jiffies, b->timeout)) { in intel_breadcrumbs_hangcheck() 38 mod_timer(&b->hangcheck, b->timeout); in intel_breadcrumbs_hangcheck() 100 static void __intel_breadcrumbs_enable_irq(struct intel_breadcrumbs *b) in __intel_breadcrumbs_enable_irq() argument 103 container_of(b, struct intel_engine_cs, breadcrumbs); in __intel_breadcrumbs_enable_irq() 106 assert_spin_locked(&b->lock); in __intel_breadcrumbs_enable_irq() 107 if (b->rpm_wakelock) in __intel_breadcrumbs_enable_irq() 116 b->rpm_wakelock = true; in __intel_breadcrumbs_enable_irq() 122 b->irq_enabled = true; in __intel_breadcrumbs_enable_irq() [all …]
|
/drivers/media/v4l2-core/ |
D | videobuf2-v4l2.c | 56 static int __verify_planes_array(struct vb2_buffer *vb, const struct v4l2_buffer *b) in __verify_planes_array() argument 58 if (!V4L2_TYPE_IS_MULTIPLANAR(b->type)) in __verify_planes_array() 62 if (b->m.planes == NULL) { in __verify_planes_array() 68 if (b->length < vb->num_planes || b->length > VB2_MAX_PLANES) { in __verify_planes_array() 70 "expected %d, got %d\n", vb->num_planes, b->length); in __verify_planes_array() 86 static int __verify_length(struct vb2_buffer *vb, const struct v4l2_buffer *b) in __verify_length() argument 92 if (!V4L2_TYPE_IS_OUTPUT(b->type)) in __verify_length() 95 if (V4L2_TYPE_IS_MULTIPLANAR(b->type)) { in __verify_length() 97 length = (b->memory == VB2_MEMORY_USERPTR || in __verify_length() 98 b->memory == VB2_MEMORY_DMABUF) in __verify_length() [all …]
|
/drivers/media/platform/s5p-mfc/ |
D | s5p_mfc_opr.c | 41 struct s5p_mfc_priv_buf *b) in s5p_mfc_alloc_priv_buf() argument 43 mfc_debug(3, "Allocating priv: %zu\n", b->size); in s5p_mfc_alloc_priv_buf() 45 b->virt = dma_alloc_coherent(dev, b->size, &b->dma, GFP_KERNEL); in s5p_mfc_alloc_priv_buf() 47 if (!b->virt) { in s5p_mfc_alloc_priv_buf() 52 if (b->dma < base) { in s5p_mfc_alloc_priv_buf() 55 &b->dma, &base); in s5p_mfc_alloc_priv_buf() 56 dma_free_coherent(dev, b->size, b->virt, b->dma); in s5p_mfc_alloc_priv_buf() 60 mfc_debug(3, "Allocated addr %p %pad\n", b->virt, &b->dma); in s5p_mfc_alloc_priv_buf() 65 struct s5p_mfc_priv_buf *b) in s5p_mfc_release_priv_buf() argument 67 if (b->virt) { in s5p_mfc_release_priv_buf() [all …]
|
/drivers/video/fbdev/matrox/ |
D | i2c-matroxfb.c | 72 struct i2c_bit_adapter* b = data; in matroxfb_gpio_setsda() local 73 matroxfb_i2c_set(b->minfo, b->mask.data, state); in matroxfb_gpio_setsda() 77 struct i2c_bit_adapter* b = data; in matroxfb_gpio_setscl() local 78 matroxfb_i2c_set(b->minfo, b->mask.clock, state); in matroxfb_gpio_setscl() 82 struct i2c_bit_adapter* b = data; in matroxfb_gpio_getsda() local 83 return (matroxfb_read_gpio(b->minfo) & b->mask.data) ? 1 : 0; in matroxfb_gpio_getsda() 87 struct i2c_bit_adapter* b = data; in matroxfb_gpio_getscl() local 88 return (matroxfb_read_gpio(b->minfo) & b->mask.clock) ? 1 : 0; in matroxfb_gpio_getscl() 101 static int i2c_bus_reg(struct i2c_bit_adapter* b, struct matrox_fb_info* minfo, in i2c_bus_reg() argument 107 b->minfo = minfo; in i2c_bus_reg() [all …]
|
/drivers/md/persistent-data/ |
D | dm-space-map-common.c | 23 struct dm_block *b, in index_prepare_for_write() argument 26 struct disk_metadata_index *mi_le = dm_block_data(b); in index_prepare_for_write() 28 mi_le->blocknr = cpu_to_le64(dm_block_location(b)); in index_prepare_for_write() 35 struct dm_block *b, in index_check() argument 38 struct disk_metadata_index *mi_le = dm_block_data(b); in index_check() 41 if (dm_block_location(b) != le64_to_cpu(mi_le->blocknr)) { in index_check() 43 le64_to_cpu(mi_le->blocknr), dm_block_location(b)); in index_check() 73 struct dm_block *b, in bitmap_prepare_for_write() argument 76 struct disk_bitmap_header *disk_header = dm_block_data(b); in bitmap_prepare_for_write() 78 disk_header->blocknr = cpu_to_le64(dm_block_location(b)); in bitmap_prepare_for_write() [all …]
|
D | dm-space-map.h | 42 int (*get_count)(struct dm_space_map *sm, dm_block_t b, uint32_t *result); 43 int (*count_is_more_than_one)(struct dm_space_map *sm, dm_block_t b, 45 int (*set_count)(struct dm_space_map *sm, dm_block_t b, uint32_t count); 49 int (*inc_block)(struct dm_space_map *sm, dm_block_t b); 50 int (*dec_block)(struct dm_space_map *sm, dm_block_t b); 55 int (*new_block)(struct dm_space_map *sm, dm_block_t *b); 97 static inline int dm_sm_get_count(struct dm_space_map *sm, dm_block_t b, in dm_sm_get_count() argument 100 return sm->get_count(sm, b, result); in dm_sm_get_count() 104 dm_block_t b, int *result) in dm_sm_count_is_more_than_one() argument 106 return sm->count_is_more_than_one(sm, b, result); in dm_sm_count_is_more_than_one() [all …]
|
/drivers/gpu/drm/nouveau/include/nvif/ |
D | object.h | 38 #define nvif_rd(a,f,b,c) ({ \ argument 44 _data = nvif_object_rd(_object, (b), (c)); \ 47 #define nvif_wr(a,f,b,c,d) ({ \ argument 52 nvif_object_wr(_object, (b), (c), (d)); \ 54 #define nvif_rd08(a,b) ({ ((u8)nvif_rd((a), ioread8, 1, (b))); }) argument 55 #define nvif_rd16(a,b) ({ ((u16)nvif_rd((a), ioread16_native, 2, (b))); }) argument 56 #define nvif_rd32(a,b) ({ ((u32)nvif_rd((a), ioread32_native, 4, (b))); }) argument 57 #define nvif_wr08(a,b,c) nvif_wr((a), iowrite8, 1, (b), (u8)(c)) argument 58 #define nvif_wr16(a,b,c) nvif_wr((a), iowrite16_native, 2, (b), (u16)(c)) argument 59 #define nvif_wr32(a,b,c) nvif_wr((a), iowrite32_native, 4, (b), (u32)(c)) argument [all …]
|
/drivers/net/fddi/skfp/h/ |
D | cmtdef.h | 57 #define DB_PR(flag,a,b,c) { if (flag) printf(a,b,c) ; } argument 59 #define DB_PR(flag,a,b,c) argument 63 #define DB_ECM(a,b,c) DB_PR((smc->debug.d_smt&1),a,b,c) argument 64 #define DB_ECMN(n,a,b,c) DB_PR((smc->debug.d_ecm >=(n)),a,b,c) argument 65 #define DB_RMT(a,b,c) DB_PR((smc->debug.d_smt&2),a,b,c) argument 66 #define DB_RMTN(n,a,b,c) DB_PR((smc->debug.d_rmt >=(n)),a,b,c) argument 67 #define DB_CFM(a,b,c) DB_PR((smc->debug.d_smt&4),a,b,c) argument 68 #define DB_CFMN(n,a,b,c) DB_PR((smc->debug.d_cfm >=(n)),a,b,c) argument 69 #define DB_PCM(a,b,c) DB_PR((smc->debug.d_smt&8),a,b,c) argument 70 #define DB_PCMN(n,a,b,c) DB_PR((smc->debug.d_pcm >=(n)),a,b,c) argument [all …]
|
/drivers/staging/skein/ |
D | skein_base.c | 26 u8 b[SKEIN_256_STATE_BYTES]; in skein_256_init() member 66 skein_256_process_block(ctx, cfg.b, 1, SKEIN_CFG_STR_LEN); in skein_256_init() 86 u8 b[SKEIN_256_STATE_BYTES]; in skein_256_init_ext() member 98 skein_assert(sizeof(cfg.b) >= sizeof(ctx->x)); in skein_256_init_ext() 109 skein_256_final_pad(ctx, cfg.b); in skein_256_init_ext() 111 memcpy(ctx->x, cfg.b, sizeof(cfg.b)); in skein_256_init_ext() 130 skein_256_process_block(ctx, cfg.b, 1, SKEIN_CFG_STR_LEN); in skein_256_init_ext() 158 memcpy(&ctx->b[ctx->h.b_cnt], msg, n); in skein_256_update() 164 skein_256_process_block(ctx, ctx->b, 1, in skein_256_update() 187 memcpy(&ctx->b[ctx->h.b_cnt], msg, msg_byte_cnt); in skein_256_update() [all …]
|
/drivers/net/wan/ |
D | sealevel.c | 213 struct slvl_board *b; in slvl_init() local 224 b = kzalloc(sizeof(struct slvl_board), GFP_KERNEL); in slvl_init() 225 if (!b) in slvl_init() 228 b->dev[0].chan = &b->board.chanA; in slvl_init() 229 b->dev[0].channel = 0; in slvl_init() 231 b->dev[1].chan = &b->board.chanB; in slvl_init() 232 b->dev[1].channel = 1; in slvl_init() 234 dev = &b->board; in slvl_init() 242 b->iobase = iobase; in slvl_init() 263 outb(3 | (1 << 7), b->iobase + 4); in slvl_init() [all …]
|
/drivers/media/pci/saa7164/ |
D | saa7164-bus.c | 29 struct tmComResBusInfo *b = &dev->bus; in saa7164_bus_setup() local 31 mutex_init(&b->lock); in saa7164_bus_setup() 33 b->Type = TYPE_BUS_PCIe; in saa7164_bus_setup() 34 b->m_wMaxReqSize = SAA_DEVICE_MAXREQUESTSIZE; in saa7164_bus_setup() 36 b->m_pdwSetRing = (u8 __iomem *)(dev->bmmio + in saa7164_bus_setup() 39 b->m_dwSizeSetRing = SAA_DEVICE_BUFFERBLOCKSIZE; in saa7164_bus_setup() 41 b->m_pdwGetRing = (u8 __iomem *)(dev->bmmio + in saa7164_bus_setup() 44 b->m_dwSizeGetRing = SAA_DEVICE_BUFFERBLOCKSIZE; in saa7164_bus_setup() 46 b->m_dwSetWritePos = ((u32)dev->intfdesc.BARLocation) + in saa7164_bus_setup() 48 b->m_dwSetReadPos = b->m_dwSetWritePos + (1 * sizeof(u32)); in saa7164_bus_setup() [all …]
|
/drivers/media/pci/bt8xx/ |
D | bttv.h | 269 #define _MUXSELe(a, b...) 0##a << 28 | _MUXSELf(b) argument 270 #define _MUXSELd(a, b...) 0##a << 26 | _MUXSELe(b) argument 271 #define _MUXSELc(a, b...) 0##a << 24 | _MUXSELd(b) argument 272 #define _MUXSELb(a, b...) 0##a << 22 | _MUXSELc(b) argument 273 #define _MUXSELa(a, b...) 0##a << 20 | _MUXSELb(b) argument 274 #define _MUXSEL9(a, b...) 0##a << 18 | _MUXSELa(b) argument 275 #define _MUXSEL8(a, b...) 0##a << 16 | _MUXSEL9(b) argument 276 #define _MUXSEL7(a, b...) 0##a << 14 | _MUXSEL8(b) argument 277 #define _MUXSEL6(a, b...) 0##a << 12 | _MUXSEL7(b) argument 278 #define _MUXSEL5(a, b...) 0##a << 10 | _MUXSEL6(b) argument [all …]
|