Home
last modified time | relevance | path

Searched refs:b (Results 1 – 25 of 991) sorted by relevance

12345678910>>...40

/drivers/iommu/
Dmsm_iommu_hw-8xxx.h33 #define SET_GLOBAL_REG_N(b, n, r, v) SET_GLOBAL_REG(b, ((r) + (n << 2)), (v)) argument
34 #define GET_GLOBAL_REG_N(b, n, r) GET_GLOBAL_REG(b, ((r) + (n << 2))) argument
37 #define GET_GLOBAL_FIELD(b, r, F) GET_FIELD(((b) + (r)), F##_MASK, F##_SHIFT) argument
38 #define GET_CONTEXT_FIELD(b, c, r, F) \ argument
39 GET_FIELD(((b) + (r) + ((c) << CTX_SHIFT)), F##_MASK, F##_SHIFT)
41 #define SET_GLOBAL_FIELD(b, r, F, v) \ argument
42 SET_FIELD(((b) + (r)), F##_MASK, F##_SHIFT, (v))
43 #define SET_CONTEXT_FIELD(b, c, r, F, v) \ argument
44 SET_FIELD(((b) + (r) + ((c) << CTX_SHIFT)), F##_MASK, F##_SHIFT, (v))
97 #define SET_M2VCBR_N(b, N, v) SET_GLOBAL_REG_N(M2VCBR_N, N, (b), (v)) argument
[all …]
/drivers/md/bcache/
Dbtree.c95 #define insert_lock(s, b) ((b)->level <= (s)->lock) argument
115 #define btree(fn, key, b, op, ...) \ argument
117 int _r, l = (b)->level - 1; \
119 struct btree *_child = bch_btree_node_get((b)->c, op, key, l, \
120 _w, b); \
156 static inline struct bset *write_block(struct btree *b) in write_block() argument
158 return ((void *) btree_bset_first(b)) + b->written * block_bytes(b->c); in write_block()
161 static void bch_btree_init_next(struct btree *b) in bch_btree_init_next() argument
164 if (b->level && b->keys.nsets) in bch_btree_init_next()
165 bch_btree_sort(&b->keys, &b->c->sort); in bch_btree_init_next()
[all …]
Dalloc.c73 uint8_t bch_inc_gen(struct cache *ca, struct bucket *b) in bch_inc_gen() argument
75 uint8_t ret = ++b->gen; in bch_inc_gen()
77 ca->set->need_gc = max(ca->set->need_gc, bucket_gc_gen(b)); in bch_inc_gen()
86 struct bucket *b; in bch_rescale_priorities() local
105 for_each_bucket(b, ca) in bch_rescale_priorities()
106 if (b->prio && in bch_rescale_priorities()
107 b->prio != BTREE_PRIO && in bch_rescale_priorities()
108 !atomic_read(&b->pin)) { in bch_rescale_priorities()
109 b->prio--; in bch_rescale_priorities()
110 c->min_prio = min(c->min_prio, b->prio); in bch_rescale_priorities()
[all …]
Dbset.c19 void bch_dump_bset(struct btree_keys *b, struct bset *i, unsigned set) in bch_dump_bset() argument
29 if (b->ops->key_dump) in bch_dump_bset()
30 b->ops->key_dump(b, k); in bch_dump_bset()
35 bkey_cmp(k, b->ops->is_extents ? in bch_dump_bset()
41 void bch_dump_bucket(struct btree_keys *b) in bch_dump_bucket() argument
46 for (i = 0; i <= b->nsets; i++) in bch_dump_bucket()
47 bch_dump_bset(b, b->set[i].data, in bch_dump_bucket()
48 bset_sector_offset(b, b->set[i].data)); in bch_dump_bucket()
52 int __bch_count_data(struct btree_keys *b) in __bch_count_data() argument
58 if (b->ops->is_extents) in __bch_count_data()
[all …]
Dbset.h226 static inline struct bset_tree *bset_tree_last(struct btree_keys *b) in bset_tree_last() argument
228 return b->set + b->nsets; in bset_tree_last()
231 static inline bool bset_written(struct btree_keys *b, struct bset_tree *t) in bset_written() argument
233 return t <= b->set + b->nsets - b->last_set_unwritten; in bset_written()
236 static inline bool bkey_written(struct btree_keys *b, struct bkey *k) in bkey_written() argument
238 return !b->last_set_unwritten || k < b->set[b->nsets].data->start; in bkey_written()
241 static inline unsigned bset_byte_offset(struct btree_keys *b, struct bset *i) in bset_byte_offset() argument
243 return ((size_t) i) - ((size_t) b->set->data); in bset_byte_offset()
246 static inline unsigned bset_sector_offset(struct btree_keys *b, struct bset *i) in bset_sector_offset() argument
248 return bset_byte_offset(b, i) >> 9; in bset_sector_offset()
[all …]
Dextents.c127 struct btree *b = container_of(keys, struct btree, keys); in bch_bkey_dump() local
135 size_t n = PTR_BUCKET_NR(b->c, k, j); in bch_bkey_dump()
138 if (n >= b->c->sb.first_bucket && n < b->c->sb.nbuckets) in bch_bkey_dump()
140 PTR_BUCKET(b->c, k, j)->prio); in bch_bkey_dump()
143 printk(" %s\n", bch_ptr_status(b->c, k)); in bch_bkey_dump()
167 struct btree *b = container_of(bk, struct btree, keys); in bch_btree_ptr_invalid() local
168 return __bch_btree_ptr_invalid(b->c, k); in bch_btree_ptr_invalid()
171 static bool btree_ptr_bad_expensive(struct btree *b, const struct bkey *k) in btree_ptr_bad_expensive() argument
177 if (mutex_trylock(&b->c->bucket_lock)) { in btree_ptr_bad_expensive()
179 if (ptr_available(b->c, k, i)) { in btree_ptr_bad_expensive()
[all …]
Dbtree.h150 static inline bool btree_node_ ## flag(struct btree *b) \
151 { return test_bit(BTREE_NODE_ ## flag, &b->flags); } \
153 static inline void set_btree_node_ ## flag(struct btree *b) \
154 { set_bit(BTREE_NODE_ ## flag, &b->flags); } \
166 static inline struct btree_write *btree_current_write(struct btree *b) in btree_current_write() argument
168 return b->writes + btree_node_write_idx(b); in btree_current_write()
171 static inline struct btree_write *btree_prev_write(struct btree *b) in btree_prev_write() argument
173 return b->writes + (btree_node_write_idx(b) ^ 1); in btree_prev_write()
176 static inline struct bset *btree_bset_first(struct btree *b) in btree_bset_first() argument
178 return b->keys.set->data; in btree_bset_first()
[all …]
Ddebug.c23 #define for_each_written_bset(b, start, i) \ argument
25 (void *) i < (void *) (start) + (KEY_SIZE(&b->key) << 9) &&\
27 i = (void *) i + set_blocks(i, block_bytes(b->c)) * \
28 block_bytes(b->c))
30 void bch_btree_verify(struct btree *b) in bch_btree_verify() argument
32 struct btree *v = b->c->verify_data; in bch_btree_verify()
36 if (!b->c->verify || !b->c->verify_ondisk) in bch_btree_verify()
39 down(&b->io_mutex); in bch_btree_verify()
40 mutex_lock(&b->c->verify_lock); in bch_btree_verify()
42 ondisk = b->c->verify_ondisk; in bch_btree_verify()
[all …]
/drivers/misc/
Dvmw_balloon.c258 void (*add_page)(struct vmballoon *b, int idx, struct page *p);
259 int (*lock)(struct vmballoon *b, unsigned int num_pages,
261 int (*unlock)(struct vmballoon *b, unsigned int num_pages,
322 static bool vmballoon_send_start(struct vmballoon *b, unsigned long req_caps) in vmballoon_send_start() argument
327 STATS_INC(b->stats.start); in vmballoon_send_start()
333 b->capabilities = capabilities; in vmballoon_send_start()
337 b->capabilities = VMW_BALLOON_BASIC_CMDS; in vmballoon_send_start()
344 if (b->capabilities & VMW_BALLOON_BATCHED_2M_CMDS) in vmballoon_send_start()
345 b->supported_page_sizes = 2; in vmballoon_send_start()
347 b->supported_page_sizes = 1; in vmballoon_send_start()
[all …]
/drivers/md/
Ddm-bufio.c250 static void buffer_record_stack(struct dm_buffer *b) in buffer_record_stack() argument
252 b->stack_trace.nr_entries = 0; in buffer_record_stack()
253 b->stack_trace.max_entries = MAX_STACK; in buffer_record_stack()
254 b->stack_trace.entries = b->stack_entries; in buffer_record_stack()
255 b->stack_trace.skip = 2; in buffer_record_stack()
256 save_stack_trace(&b->stack_trace); in buffer_record_stack()
266 struct dm_buffer *b; in __find() local
269 b = container_of(n, struct dm_buffer, node); in __find()
271 if (b->block == block) in __find()
272 return b; in __find()
[all …]
Ddm-cache-block-types.h24 static inline dm_oblock_t to_oblock(dm_block_t b) in to_oblock() argument
26 return (__force dm_oblock_t) b; in to_oblock()
29 static inline dm_block_t from_oblock(dm_oblock_t b) in from_oblock() argument
31 return (__force dm_block_t) b; in from_oblock()
34 static inline dm_cblock_t to_cblock(uint32_t b) in to_cblock() argument
36 return (__force dm_cblock_t) b; in to_cblock()
39 static inline uint32_t from_cblock(dm_cblock_t b) in from_cblock() argument
41 return (__force uint32_t) b; in from_cblock()
44 static inline dm_dblock_t to_dblock(dm_block_t b) in to_dblock() argument
46 return (__force dm_dblock_t) b; in to_dblock()
[all …]
/drivers/block/drbd/
Ddrbd_bitmap.c127 struct drbd_bitmap *b = device->bitmap; in __bm_print_lock_info() local
132 func, b->bm_why ?: "?", in __bm_print_lock_info()
133 b->bm_task->comm, task_pid_nr(b->bm_task)); in __bm_print_lock_info()
138 struct drbd_bitmap *b = device->bitmap; in drbd_bm_lock() local
141 if (!b) { in drbd_bm_lock()
146 trylock_failed = !mutex_trylock(&b->bm_change); in drbd_bm_lock()
151 why, b->bm_why ?: "?", in drbd_bm_lock()
152 b->bm_task->comm, task_pid_nr(b->bm_task)); in drbd_bm_lock()
153 mutex_lock(&b->bm_change); in drbd_bm_lock()
155 if (BM_LOCKED_MASK & b->bm_flags) in drbd_bm_lock()
[all …]
Ddrbd_vli.h150 #define LEVEL(t,b,v) \ in vli_decode_bits() argument
152 if ((in & ((1 << b) -1)) == v) { \ in vli_decode_bits()
153 *out = ((in & ((~0ULL) >> (64-t))) >> b) + adj; \ in vli_decode_bits()
156 adj += 1ULL << (t - b); \ in vli_decode_bits()
176 #define LEVEL(t,b,v) do { \ in __vli_encode_bits() argument
177 max += 1ULL << (t - b); \ in __vli_encode_bits()
180 *out = ((in - adj) << b) | v; \ in __vli_encode_bits()
205 u8 *b; member
213 cur->b = s; in bitstream_cursor_reset()
222 cur->b = cur->b + (bits >> 3); in bitstream_cursor_advance()
[all …]
/drivers/gpu/drm/i915/
Dintel_breadcrumbs.c32 struct intel_breadcrumbs *b = &engine->breadcrumbs; in intel_breadcrumbs_hangcheck() local
34 if (!b->irq_enabled) in intel_breadcrumbs_hangcheck()
37 if (time_before(jiffies, b->timeout)) { in intel_breadcrumbs_hangcheck()
38 mod_timer(&b->hangcheck, b->timeout); in intel_breadcrumbs_hangcheck()
100 static void __intel_breadcrumbs_enable_irq(struct intel_breadcrumbs *b) in __intel_breadcrumbs_enable_irq() argument
103 container_of(b, struct intel_engine_cs, breadcrumbs); in __intel_breadcrumbs_enable_irq()
106 assert_spin_locked(&b->lock); in __intel_breadcrumbs_enable_irq()
107 if (b->rpm_wakelock) in __intel_breadcrumbs_enable_irq()
116 b->rpm_wakelock = true; in __intel_breadcrumbs_enable_irq()
122 b->irq_enabled = true; in __intel_breadcrumbs_enable_irq()
[all …]
/drivers/media/v4l2-core/
Dvideobuf2-v4l2.c56 static int __verify_planes_array(struct vb2_buffer *vb, const struct v4l2_buffer *b) in __verify_planes_array() argument
58 if (!V4L2_TYPE_IS_MULTIPLANAR(b->type)) in __verify_planes_array()
62 if (b->m.planes == NULL) { in __verify_planes_array()
68 if (b->length < vb->num_planes || b->length > VB2_MAX_PLANES) { in __verify_planes_array()
70 "expected %d, got %d\n", vb->num_planes, b->length); in __verify_planes_array()
86 static int __verify_length(struct vb2_buffer *vb, const struct v4l2_buffer *b) in __verify_length() argument
92 if (!V4L2_TYPE_IS_OUTPUT(b->type)) in __verify_length()
95 if (V4L2_TYPE_IS_MULTIPLANAR(b->type)) { in __verify_length()
97 length = (b->memory == VB2_MEMORY_USERPTR || in __verify_length()
98 b->memory == VB2_MEMORY_DMABUF) in __verify_length()
[all …]
/drivers/media/platform/s5p-mfc/
Ds5p_mfc_opr.c41 struct s5p_mfc_priv_buf *b) in s5p_mfc_alloc_priv_buf() argument
43 mfc_debug(3, "Allocating priv: %zu\n", b->size); in s5p_mfc_alloc_priv_buf()
45 b->virt = dma_alloc_coherent(dev, b->size, &b->dma, GFP_KERNEL); in s5p_mfc_alloc_priv_buf()
47 if (!b->virt) { in s5p_mfc_alloc_priv_buf()
52 if (b->dma < base) { in s5p_mfc_alloc_priv_buf()
55 &b->dma, &base); in s5p_mfc_alloc_priv_buf()
56 dma_free_coherent(dev, b->size, b->virt, b->dma); in s5p_mfc_alloc_priv_buf()
60 mfc_debug(3, "Allocated addr %p %pad\n", b->virt, &b->dma); in s5p_mfc_alloc_priv_buf()
65 struct s5p_mfc_priv_buf *b) in s5p_mfc_release_priv_buf() argument
67 if (b->virt) { in s5p_mfc_release_priv_buf()
[all …]
/drivers/video/fbdev/matrox/
Di2c-matroxfb.c72 struct i2c_bit_adapter* b = data; in matroxfb_gpio_setsda() local
73 matroxfb_i2c_set(b->minfo, b->mask.data, state); in matroxfb_gpio_setsda()
77 struct i2c_bit_adapter* b = data; in matroxfb_gpio_setscl() local
78 matroxfb_i2c_set(b->minfo, b->mask.clock, state); in matroxfb_gpio_setscl()
82 struct i2c_bit_adapter* b = data; in matroxfb_gpio_getsda() local
83 return (matroxfb_read_gpio(b->minfo) & b->mask.data) ? 1 : 0; in matroxfb_gpio_getsda()
87 struct i2c_bit_adapter* b = data; in matroxfb_gpio_getscl() local
88 return (matroxfb_read_gpio(b->minfo) & b->mask.clock) ? 1 : 0; in matroxfb_gpio_getscl()
101 static int i2c_bus_reg(struct i2c_bit_adapter* b, struct matrox_fb_info* minfo, in i2c_bus_reg() argument
107 b->minfo = minfo; in i2c_bus_reg()
[all …]
/drivers/md/persistent-data/
Ddm-space-map-common.c23 struct dm_block *b, in index_prepare_for_write() argument
26 struct disk_metadata_index *mi_le = dm_block_data(b); in index_prepare_for_write()
28 mi_le->blocknr = cpu_to_le64(dm_block_location(b)); in index_prepare_for_write()
35 struct dm_block *b, in index_check() argument
38 struct disk_metadata_index *mi_le = dm_block_data(b); in index_check()
41 if (dm_block_location(b) != le64_to_cpu(mi_le->blocknr)) { in index_check()
43 le64_to_cpu(mi_le->blocknr), dm_block_location(b)); in index_check()
73 struct dm_block *b, in bitmap_prepare_for_write() argument
76 struct disk_bitmap_header *disk_header = dm_block_data(b); in bitmap_prepare_for_write()
78 disk_header->blocknr = cpu_to_le64(dm_block_location(b)); in bitmap_prepare_for_write()
[all …]
Ddm-space-map.h42 int (*get_count)(struct dm_space_map *sm, dm_block_t b, uint32_t *result);
43 int (*count_is_more_than_one)(struct dm_space_map *sm, dm_block_t b,
45 int (*set_count)(struct dm_space_map *sm, dm_block_t b, uint32_t count);
49 int (*inc_block)(struct dm_space_map *sm, dm_block_t b);
50 int (*dec_block)(struct dm_space_map *sm, dm_block_t b);
55 int (*new_block)(struct dm_space_map *sm, dm_block_t *b);
97 static inline int dm_sm_get_count(struct dm_space_map *sm, dm_block_t b, in dm_sm_get_count() argument
100 return sm->get_count(sm, b, result); in dm_sm_get_count()
104 dm_block_t b, int *result) in dm_sm_count_is_more_than_one() argument
106 return sm->count_is_more_than_one(sm, b, result); in dm_sm_count_is_more_than_one()
[all …]
/drivers/gpu/drm/nouveau/include/nvif/
Dobject.h38 #define nvif_rd(a,f,b,c) ({ \ argument
44 _data = nvif_object_rd(_object, (b), (c)); \
47 #define nvif_wr(a,f,b,c,d) ({ \ argument
52 nvif_object_wr(_object, (b), (c), (d)); \
54 #define nvif_rd08(a,b) ({ ((u8)nvif_rd((a), ioread8, 1, (b))); }) argument
55 #define nvif_rd16(a,b) ({ ((u16)nvif_rd((a), ioread16_native, 2, (b))); }) argument
56 #define nvif_rd32(a,b) ({ ((u32)nvif_rd((a), ioread32_native, 4, (b))); }) argument
57 #define nvif_wr08(a,b,c) nvif_wr((a), iowrite8, 1, (b), (u8)(c)) argument
58 #define nvif_wr16(a,b,c) nvif_wr((a), iowrite16_native, 2, (b), (u16)(c)) argument
59 #define nvif_wr32(a,b,c) nvif_wr((a), iowrite32_native, 4, (b), (u32)(c)) argument
[all …]
/drivers/net/fddi/skfp/h/
Dcmtdef.h57 #define DB_PR(flag,a,b,c) { if (flag) printf(a,b,c) ; } argument
59 #define DB_PR(flag,a,b,c) argument
63 #define DB_ECM(a,b,c) DB_PR((smc->debug.d_smt&1),a,b,c) argument
64 #define DB_ECMN(n,a,b,c) DB_PR((smc->debug.d_ecm >=(n)),a,b,c) argument
65 #define DB_RMT(a,b,c) DB_PR((smc->debug.d_smt&2),a,b,c) argument
66 #define DB_RMTN(n,a,b,c) DB_PR((smc->debug.d_rmt >=(n)),a,b,c) argument
67 #define DB_CFM(a,b,c) DB_PR((smc->debug.d_smt&4),a,b,c) argument
68 #define DB_CFMN(n,a,b,c) DB_PR((smc->debug.d_cfm >=(n)),a,b,c) argument
69 #define DB_PCM(a,b,c) DB_PR((smc->debug.d_smt&8),a,b,c) argument
70 #define DB_PCMN(n,a,b,c) DB_PR((smc->debug.d_pcm >=(n)),a,b,c) argument
[all …]
/drivers/staging/skein/
Dskein_base.c26 u8 b[SKEIN_256_STATE_BYTES]; in skein_256_init() member
66 skein_256_process_block(ctx, cfg.b, 1, SKEIN_CFG_STR_LEN); in skein_256_init()
86 u8 b[SKEIN_256_STATE_BYTES]; in skein_256_init_ext() member
98 skein_assert(sizeof(cfg.b) >= sizeof(ctx->x)); in skein_256_init_ext()
109 skein_256_final_pad(ctx, cfg.b); in skein_256_init_ext()
111 memcpy(ctx->x, cfg.b, sizeof(cfg.b)); in skein_256_init_ext()
130 skein_256_process_block(ctx, cfg.b, 1, SKEIN_CFG_STR_LEN); in skein_256_init_ext()
158 memcpy(&ctx->b[ctx->h.b_cnt], msg, n); in skein_256_update()
164 skein_256_process_block(ctx, ctx->b, 1, in skein_256_update()
187 memcpy(&ctx->b[ctx->h.b_cnt], msg, msg_byte_cnt); in skein_256_update()
[all …]
/drivers/net/wan/
Dsealevel.c213 struct slvl_board *b; in slvl_init() local
224 b = kzalloc(sizeof(struct slvl_board), GFP_KERNEL); in slvl_init()
225 if (!b) in slvl_init()
228 b->dev[0].chan = &b->board.chanA; in slvl_init()
229 b->dev[0].channel = 0; in slvl_init()
231 b->dev[1].chan = &b->board.chanB; in slvl_init()
232 b->dev[1].channel = 1; in slvl_init()
234 dev = &b->board; in slvl_init()
242 b->iobase = iobase; in slvl_init()
263 outb(3 | (1 << 7), b->iobase + 4); in slvl_init()
[all …]
/drivers/media/pci/saa7164/
Dsaa7164-bus.c29 struct tmComResBusInfo *b = &dev->bus; in saa7164_bus_setup() local
31 mutex_init(&b->lock); in saa7164_bus_setup()
33 b->Type = TYPE_BUS_PCIe; in saa7164_bus_setup()
34 b->m_wMaxReqSize = SAA_DEVICE_MAXREQUESTSIZE; in saa7164_bus_setup()
36 b->m_pdwSetRing = (u8 __iomem *)(dev->bmmio + in saa7164_bus_setup()
39 b->m_dwSizeSetRing = SAA_DEVICE_BUFFERBLOCKSIZE; in saa7164_bus_setup()
41 b->m_pdwGetRing = (u8 __iomem *)(dev->bmmio + in saa7164_bus_setup()
44 b->m_dwSizeGetRing = SAA_DEVICE_BUFFERBLOCKSIZE; in saa7164_bus_setup()
46 b->m_dwSetWritePos = ((u32)dev->intfdesc.BARLocation) + in saa7164_bus_setup()
48 b->m_dwSetReadPos = b->m_dwSetWritePos + (1 * sizeof(u32)); in saa7164_bus_setup()
[all …]
/drivers/media/pci/bt8xx/
Dbttv.h269 #define _MUXSELe(a, b...) 0##a << 28 | _MUXSELf(b) argument
270 #define _MUXSELd(a, b...) 0##a << 26 | _MUXSELe(b) argument
271 #define _MUXSELc(a, b...) 0##a << 24 | _MUXSELd(b) argument
272 #define _MUXSELb(a, b...) 0##a << 22 | _MUXSELc(b) argument
273 #define _MUXSELa(a, b...) 0##a << 20 | _MUXSELb(b) argument
274 #define _MUXSEL9(a, b...) 0##a << 18 | _MUXSELa(b) argument
275 #define _MUXSEL8(a, b...) 0##a << 16 | _MUXSEL9(b) argument
276 #define _MUXSEL7(a, b...) 0##a << 14 | _MUXSEL8(b) argument
277 #define _MUXSEL6(a, b...) 0##a << 12 | _MUXSEL7(b) argument
278 #define _MUXSEL5(a, b...) 0##a << 10 | _MUXSEL6(b) argument
[all …]

12345678910>>...40