/drivers/iommu/ |
D | msm_iommu_hw-8xxx.h | 20 #define SET_GLOBAL_REG_N(b, n, r, v) SET_GLOBAL_REG(b, ((r) + (n << 2)), (v)) argument 21 #define GET_GLOBAL_REG_N(b, n, r) GET_GLOBAL_REG(b, ((r) + (n << 2))) argument 24 #define GET_GLOBAL_FIELD(b, r, F) GET_FIELD(((b) + (r)), F##_MASK, F##_SHIFT) argument 25 #define GET_CONTEXT_FIELD(b, c, r, F) \ argument 26 GET_FIELD(((b) + (r) + ((c) << CTX_SHIFT)), F##_MASK, F##_SHIFT) 28 #define SET_GLOBAL_FIELD(b, r, F, v) \ argument 29 SET_FIELD(((b) + (r)), F##_MASK, F##_SHIFT, (v)) 30 #define SET_CONTEXT_FIELD(b, c, r, F, v) \ argument 31 SET_FIELD(((b) + (r) + ((c) << CTX_SHIFT)), F##_MASK, F##_SHIFT, (v)) 84 #define SET_M2VCBR_N(b, N, v) SET_GLOBAL_REG_N(M2VCBR_N, N, (b), (v)) argument [all …]
|
/drivers/md/bcache/ |
D | btree.c | 104 #define insert_lock(s, b) ((b)->level <= (s)->lock) argument 107 static inline struct bset *write_block(struct btree *b) in write_block() argument 109 return ((void *) btree_bset_first(b)) + b->written * block_bytes(b->c->cache); in write_block() 112 static void bch_btree_init_next(struct btree *b) in bch_btree_init_next() argument 115 if (b->level && b->keys.nsets) in bch_btree_init_next() 116 bch_btree_sort(&b->keys, &b->c->sort); in bch_btree_init_next() 118 bch_btree_sort_lazy(&b->keys, &b->c->sort); in bch_btree_init_next() 120 if (b->written < btree_blocks(b)) in bch_btree_init_next() 121 bch_bset_init_next(&b->keys, write_block(b), in bch_btree_init_next() 122 bset_magic(&b->c->cache->sb)); in bch_btree_init_next() [all …]
|
D | bset.h | 194 bool (*insert_fixup)(struct btree_keys *b, 234 static inline struct bset_tree *bset_tree_last(struct btree_keys *b) in bset_tree_last() argument 236 return b->set + b->nsets; in bset_tree_last() 239 static inline bool bset_written(struct btree_keys *b, struct bset_tree *t) in bset_written() argument 241 return t <= b->set + b->nsets - b->last_set_unwritten; in bset_written() 244 static inline bool bkey_written(struct btree_keys *b, struct bkey *k) in bkey_written() argument 246 return !b->last_set_unwritten || k < b->set[b->nsets].data->start; in bkey_written() 249 static inline unsigned int bset_byte_offset(struct btree_keys *b, in bset_byte_offset() argument 252 return ((size_t) i) - ((size_t) b->set->data); in bset_byte_offset() 255 static inline unsigned int bset_sector_offset(struct btree_keys *b, in bset_sector_offset() argument [all …]
|
D | alloc.c | 76 uint8_t bch_inc_gen(struct cache *ca, struct bucket *b) in bch_inc_gen() argument 78 uint8_t ret = ++b->gen; in bch_inc_gen() 80 ca->set->need_gc = max(ca->set->need_gc, bucket_gc_gen(b)); in bch_inc_gen() 89 struct bucket *b; in bch_rescale_priorities() local 107 for_each_bucket(b, ca) in bch_rescale_priorities() 108 if (b->prio && in bch_rescale_priorities() 109 b->prio != BTREE_PRIO && in bch_rescale_priorities() 110 !atomic_read(&b->pin)) { in bch_rescale_priorities() 111 b->prio--; in bch_rescale_priorities() 112 c->min_prio = min(c->min_prio, b->prio); in bch_rescale_priorities() [all …]
|
D | bset.c | 21 void bch_dump_bset(struct btree_keys *b, struct bset *i, unsigned int set) in bch_dump_bset() argument 31 if (b->ops->key_dump) in bch_dump_bset() 32 b->ops->key_dump(b, k); in bch_dump_bset() 37 bkey_cmp(k, b->ops->is_extents ? in bch_dump_bset() 43 void bch_dump_bucket(struct btree_keys *b) in bch_dump_bucket() argument 48 for (i = 0; i <= b->nsets; i++) in bch_dump_bucket() 49 bch_dump_bset(b, b->set[i].data, in bch_dump_bucket() 50 bset_sector_offset(b, b->set[i].data)); in bch_dump_bucket() 54 int __bch_count_data(struct btree_keys *b) in __bch_count_data() argument 60 if (b->ops->is_extents) in __bch_count_data() [all …]
|
D | btree.h | 152 static inline bool btree_node_ ## flag(struct btree *b) \ 153 { return test_bit(BTREE_NODE_ ## flag, &b->flags); } \ 155 static inline void set_btree_node_ ## flag(struct btree *b) \ 156 { set_bit(BTREE_NODE_ ## flag, &b->flags); } 170 static inline struct btree_write *btree_current_write(struct btree *b) in btree_current_write() argument 172 return b->writes + btree_node_write_idx(b); in btree_current_write() 175 static inline struct btree_write *btree_prev_write(struct btree *b) in btree_prev_write() argument 177 return b->writes + (btree_node_write_idx(b) ^ 1); in btree_prev_write() 180 static inline struct bset *btree_bset_first(struct btree *b) in btree_bset_first() argument 182 return b->keys.set->data; in btree_bset_first() [all …]
|
D | extents.c | 128 struct btree *b = container_of(keys, struct btree, keys); in bch_bkey_dump() local 136 size_t n = PTR_BUCKET_NR(b->c, k, j); in bch_bkey_dump() 139 if (n >= b->c->cache->sb.first_bucket && n < b->c->cache->sb.nbuckets) in bch_bkey_dump() 141 PTR_BUCKET(b->c, k, j)->prio); in bch_bkey_dump() 144 pr_cont(" %s\n", bch_ptr_status(b->c, k)); in bch_bkey_dump() 168 struct btree *b = container_of(bk, struct btree, keys); in bch_btree_ptr_invalid() local 170 return __bch_btree_ptr_invalid(b->c, k); in bch_btree_ptr_invalid() 173 static bool btree_ptr_bad_expensive(struct btree *b, const struct bkey *k) in btree_ptr_bad_expensive() argument 179 if (mutex_trylock(&b->c->bucket_lock)) { in btree_ptr_bad_expensive() 181 if (ptr_available(b->c, k, i)) { in btree_ptr_bad_expensive() [all …]
|
/drivers/md/ |
D | dm-bufio.c | 242 static void buffer_record_stack(struct dm_buffer *b) in buffer_record_stack() argument 244 b->stack_len = stack_trace_save(b->stack_entries, MAX_STACK, 2); in buffer_record_stack() 254 struct dm_buffer *b; in __find() local 257 b = container_of(n, struct dm_buffer, node); in __find() 259 if (b->block == block) in __find() 260 return b; in __find() 262 n = block < b->block ? n->rb_left : n->rb_right; in __find() 271 struct dm_buffer *b; in __find_next() local 275 b = container_of(n, struct dm_buffer, node); in __find_next() 277 if (b->block == block) in __find_next() [all …]
|
D | dm-cache-background-tracker.c | 34 struct background_tracker *b = kmalloc(sizeof(*b), GFP_KERNEL); in btracker_create() local 36 if (!b) { in btracker_create() 41 b->max_work = max_work; in btracker_create() 42 atomic_set(&b->pending_promotes, 0); in btracker_create() 43 atomic_set(&b->pending_writebacks, 0); in btracker_create() 44 atomic_set(&b->pending_demotes, 0); in btracker_create() 46 INIT_LIST_HEAD(&b->issued); in btracker_create() 47 INIT_LIST_HEAD(&b->queued); in btracker_create() 49 b->pending = RB_ROOT; in btracker_create() 50 b->work_cache = KMEM_CACHE(bt_work, 0); in btracker_create() [all …]
|
/drivers/gpu/drm/i915/gt/ |
D | intel_breadcrumbs.c | 18 static bool irq_enable(struct intel_breadcrumbs *b) in irq_enable() argument 20 return intel_engine_irq_enable(b->irq_engine); in irq_enable() 23 static void irq_disable(struct intel_breadcrumbs *b) in irq_disable() argument 25 intel_engine_irq_disable(b->irq_engine); in irq_disable() 28 static void __intel_breadcrumbs_arm_irq(struct intel_breadcrumbs *b) in __intel_breadcrumbs_arm_irq() argument 34 if (GEM_WARN_ON(!intel_gt_pm_get_if_awake(b->irq_engine->gt))) in __intel_breadcrumbs_arm_irq() 43 WRITE_ONCE(b->irq_armed, true); in __intel_breadcrumbs_arm_irq() 46 if (!b->irq_enabled++ && b->irq_enable(b)) in __intel_breadcrumbs_arm_irq() 47 irq_work_queue(&b->irq_work); in __intel_breadcrumbs_arm_irq() 50 static void intel_breadcrumbs_arm_irq(struct intel_breadcrumbs *b) in intel_breadcrumbs_arm_irq() argument [all …]
|
/drivers/misc/ |
D | vmw_balloon.c | 417 static inline void vmballoon_stats_op_inc(struct vmballoon *b, unsigned int op, in vmballoon_stats_op_inc() argument 421 atomic64_inc(&b->stats->ops[op][type]); in vmballoon_stats_op_inc() 424 static inline void vmballoon_stats_gen_inc(struct vmballoon *b, in vmballoon_stats_gen_inc() argument 428 atomic64_inc(&b->stats->general_stat[stat]); in vmballoon_stats_gen_inc() 431 static inline void vmballoon_stats_gen_add(struct vmballoon *b, in vmballoon_stats_gen_add() argument 436 atomic64_add(val, &b->stats->general_stat[stat]); in vmballoon_stats_gen_add() 439 static inline void vmballoon_stats_page_inc(struct vmballoon *b, in vmballoon_stats_page_inc() argument 444 atomic64_inc(&b->stats->page_stat[stat][size]); in vmballoon_stats_page_inc() 447 static inline void vmballoon_stats_page_add(struct vmballoon *b, in vmballoon_stats_page_add() argument 453 atomic64_add(val, &b->stats->page_stat[stat][size]); in vmballoon_stats_page_add() [all …]
|
/drivers/staging/media/atomisp/pci/hive_isp_css_include/ |
D | math_support.h | 32 #define IMPLIES(a, b) (!(a) || (b)) argument 36 #define MAX(a, b) (((a) > (b)) ? (a) : (b)) argument 37 #define MIN(a, b) (((a) < (b)) ? (a) : (b)) argument 39 #define ROUND_DIV(a, b) (((b) != 0) ? ((a) + ((b) >> 1)) / (b) : 0) argument 40 #define CEIL_DIV(a, b) (((b) != 0) ? ((a) + (b) - 1) / (b) : 0) argument 41 #define CEIL_MUL(a, b) (CEIL_DIV(a, b) * (b)) argument 42 #define CEIL_MUL2(a, b) (((a) + (b) - 1) & ~((b) - 1)) argument 43 #define CEIL_SHIFT(a, b) (((a) + (1 << (b)) - 1) >> (b)) argument 44 #define CEIL_SHIFT_MUL(a, b) (CEIL_SHIFT(a, b) << (b)) argument 45 #define ROUND_HALF_DOWN_DIV(a, b) (((b) != 0) ? ((a) + (b / 2) - 1) / (b) : 0) argument [all …]
|
/drivers/block/drbd/ |
D | drbd_bitmap.c | 115 struct drbd_bitmap *b = device->bitmap; in __bm_print_lock_info() local 120 func, b->bm_why ?: "?", in __bm_print_lock_info() 121 b->bm_task->comm, task_pid_nr(b->bm_task)); in __bm_print_lock_info() 126 struct drbd_bitmap *b = device->bitmap; in drbd_bm_lock() local 129 if (!b) { in drbd_bm_lock() 134 trylock_failed = !mutex_trylock(&b->bm_change); in drbd_bm_lock() 139 why, b->bm_why ?: "?", in drbd_bm_lock() 140 b->bm_task->comm, task_pid_nr(b->bm_task)); in drbd_bm_lock() 141 mutex_lock(&b->bm_change); in drbd_bm_lock() 143 if (BM_LOCKED_MASK & b->bm_flags) in drbd_bm_lock() [all …]
|
/drivers/media/platform/s5p-mfc/ |
D | s5p_mfc_opr.c | 38 struct s5p_mfc_priv_buf *b) in s5p_mfc_alloc_priv_buf() argument 41 unsigned int count = b->size >> PAGE_SHIFT; in s5p_mfc_alloc_priv_buf() 45 mfc_debug(3, "Allocating priv: %zu\n", b->size); in s5p_mfc_alloc_priv_buf() 54 b->virt = dev->mem_virt + offset; in s5p_mfc_alloc_priv_buf() 55 b->dma = dev->mem_base + offset; in s5p_mfc_alloc_priv_buf() 60 b->ctx = mem_ctx; in s5p_mfc_alloc_priv_buf() 61 b->virt = dma_alloc_coherent(mem_dev, b->size, &b->dma, GFP_KERNEL); in s5p_mfc_alloc_priv_buf() 62 if (!b->virt) in s5p_mfc_alloc_priv_buf() 64 if (b->dma < base) { in s5p_mfc_alloc_priv_buf() 66 &b->dma, &base); in s5p_mfc_alloc_priv_buf() [all …]
|
/drivers/crypto/cavium/cpt/ |
D | cpt_common.h | 49 #define CPTX_PF_MBOX_INTX(a, b) \ argument 50 (0x400ll + ((u64)(a) << 36) + ((b) << 3)) 51 #define CPTX_PF_MBOX_INT_W1SX(a, b) \ argument 52 (0x420ll + ((u64)(a) << 36) + ((b) << 3)) 53 #define CPTX_PF_MBOX_ENA_W1CX(a, b) \ argument 54 (0x440ll + ((u64)(a) << 36) + ((b) << 3)) 55 #define CPTX_PF_MBOX_ENA_W1SX(a, b) \ argument 56 (0x460ll + ((u64)(a) << 36) + ((b) << 3)) 61 #define CPTX_PF_GX_EN(a, b) \ argument 62 (0x600ll + ((u64)(a) << 36) + ((b) << 3)) [all …]
|
/drivers/crypto/marvell/octeontx/ |
D | otx_cpt_hw_types.h | 27 #define OTX_CPT_BAR_E_CPTX_VFX_BAR0_OFFSET(a, b) \ argument 28 (0x000020000000ll + 0x1000000000ll * (a) + 0x100000ll * (b)) 63 #define OTX_CPT_PF_MBOX_INTX(b) (0x400ll | (u64)(b) << 3) argument 64 #define OTX_CPT_PF_MBOX_INT_W1SX(b) (0x420ll | (u64)(b) << 3) argument 65 #define OTX_CPT_PF_MBOX_ENA_W1CX(b) (0x440ll | (u64)(b) << 3) argument 66 #define OTX_CPT_PF_MBOX_ENA_W1SX(b) (0x460ll | (u64)(b) << 3) argument 71 #define OTX_CPT_PF_GX_EN(b) (0x600ll | (u64)(b) << 3) argument 91 #define OTX_CPT_PF_EXE_DBG_CNTX(b) (0x4001100ll | (u64)(b) << 3) argument 93 #define OTX_CPT_PF_EXE_EPCI_INBX_CNT(b) (0x4001200ll | (u64)(b) << 3) argument 94 #define OTX_CPT_PF_EXE_EPCI_OUTBX_CNT(b) (0x4001240ll | (u64)(b) << 3) argument [all …]
|
/drivers/iommu/intel/ |
D | cap_audit.c | 20 struct intel_iommu *b) in check_irq_capabilities() argument 22 CHECK_FEATURE_MISMATCH(a, b, cap, pi_support, CAP_PI_MASK); in check_irq_capabilities() 23 CHECK_FEATURE_MISMATCH(a, b, ecap, eim_support, ECAP_EIM_MASK); in check_irq_capabilities() 27 struct intel_iommu *b) in check_dmar_capabilities() argument 29 MINIMAL_FEATURE_IOMMU(b, cap, CAP_MAMV_MASK); in check_dmar_capabilities() 30 MINIMAL_FEATURE_IOMMU(b, cap, CAP_NFR_MASK); in check_dmar_capabilities() 31 MINIMAL_FEATURE_IOMMU(b, cap, CAP_SLLPS_MASK); in check_dmar_capabilities() 32 MINIMAL_FEATURE_IOMMU(b, cap, CAP_FRO_MASK); in check_dmar_capabilities() 33 MINIMAL_FEATURE_IOMMU(b, cap, CAP_MGAW_MASK); in check_dmar_capabilities() 34 MINIMAL_FEATURE_IOMMU(b, cap, CAP_SAGAW_MASK); in check_dmar_capabilities() [all …]
|
/drivers/video/fbdev/matrox/ |
D | i2c-matroxfb.c | 73 struct i2c_bit_adapter* b = data; in matroxfb_gpio_setsda() local 74 matroxfb_i2c_set(b->minfo, b->mask.data, state); in matroxfb_gpio_setsda() 78 struct i2c_bit_adapter* b = data; in matroxfb_gpio_setscl() local 79 matroxfb_i2c_set(b->minfo, b->mask.clock, state); in matroxfb_gpio_setscl() 83 struct i2c_bit_adapter* b = data; in matroxfb_gpio_getsda() local 84 return (matroxfb_read_gpio(b->minfo) & b->mask.data) ? 1 : 0; in matroxfb_gpio_getsda() 88 struct i2c_bit_adapter* b = data; in matroxfb_gpio_getscl() local 89 return (matroxfb_read_gpio(b->minfo) & b->mask.clock) ? 1 : 0; in matroxfb_gpio_getscl() 102 static int i2c_bus_reg(struct i2c_bit_adapter* b, struct matrox_fb_info* minfo, in i2c_bus_reg() argument 108 b->minfo = minfo; in i2c_bus_reg() [all …]
|
/drivers/media/common/videobuf2/ |
D | videobuf2-v4l2.c | 64 static int __verify_planes_array(struct vb2_buffer *vb, const struct v4l2_buffer *b) in __verify_planes_array() argument 66 if (!V4L2_TYPE_IS_MULTIPLANAR(b->type)) in __verify_planes_array() 70 if (b->m.planes == NULL) { in __verify_planes_array() 76 if (b->length < vb->num_planes || b->length > VB2_MAX_PLANES) { in __verify_planes_array() 79 vb->num_planes, b->length); in __verify_planes_array() 95 static int __verify_length(struct vb2_buffer *vb, const struct v4l2_buffer *b) in __verify_length() argument 101 if (V4L2_TYPE_IS_CAPTURE(b->type)) in __verify_length() 104 if (V4L2_TYPE_IS_MULTIPLANAR(b->type)) { in __verify_length() 106 length = (b->memory == VB2_MEMORY_USERPTR || in __verify_length() 107 b->memory == VB2_MEMORY_DMABUF) in __verify_length() [all …]
|
/drivers/net/ethernet/marvell/octeontx2/af/ |
D | rvu_reg.h | 26 #define RVU_AF_AFPFX_MBOXX(a, b) (0x2000 | (a) << 4 | (b) << 3) argument 82 #define RVU_PF_VFX_PFVF_MBOXX(a, b) (0x0 | (a) << 12 | (b) << 3) argument 383 #define NIX_AF_TL3_TL2X_LINKX_CFG(a, b) (0x1700 | (a) << 16 | (b) << 3) argument 384 #define NIX_AF_RX_FLOW_KEY_ALGX_FIELDX(a, b) (0x1800 | (a) << 18 | (b) << 3) argument 391 #define NIX_AF_LSO_FORMATX_FIELDX(a, b) (0x1B00 | (a) << 16 | (b) << 3) argument 415 #define NIX_AF_LFX_RX_VTAG_TYPEX(a, b) (0x4200 | (a) << 17 | (b) << 3) argument 416 #define NIX_AF_LFX_LOCKX(a, b) (0x4300 | (a) << 17 | (b) << 3) argument 417 #define NIX_AF_LFX_TX_STATX(a, b) (0x4400 | (a) << 17 | (b) << 3) argument 418 #define NIX_AF_LFX_RX_STATX(a, b) (0x4500 | (a) << 17 | (b) << 3) argument 419 #define NIX_AF_LFX_RSS_GRPX(a, b) (0x4600 | (a) << 17 | (b) << 3) argument [all …]
|
/drivers/hwmon/pmbus/ |
D | lm25066.c | 48 short m, b, R; member 58 .b = 1343, 63 .b = -1833, 68 .b = -537, 73 .b = -2908, 78 .b = -5646, 83 .b = -14500, 90 .b = -1800, 95 .b = -1800, 100 .b = -5200, [all …]
|
/drivers/md/persistent-data/ |
D | dm-space-map-common.c | 25 struct dm_block *b, in index_prepare_for_write() argument 28 struct disk_metadata_index *mi_le = dm_block_data(b); in index_prepare_for_write() 30 mi_le->blocknr = cpu_to_le64(dm_block_location(b)); in index_prepare_for_write() 37 struct dm_block *b, in index_check() argument 40 struct disk_metadata_index *mi_le = dm_block_data(b); in index_check() 43 if (dm_block_location(b) != le64_to_cpu(mi_le->blocknr)) { in index_check() 45 le64_to_cpu(mi_le->blocknr), dm_block_location(b)); in index_check() 75 struct dm_block *b, in dm_bitmap_prepare_for_write() argument 78 struct disk_bitmap_header *disk_header = dm_block_data(b); in dm_bitmap_prepare_for_write() 80 disk_header->blocknr = cpu_to_le64(dm_block_location(b)); in dm_bitmap_prepare_for_write() [all …]
|
D | dm-space-map.h | 42 int (*get_count)(struct dm_space_map *sm, dm_block_t b, uint32_t *result); 43 int (*count_is_more_than_one)(struct dm_space_map *sm, dm_block_t b, 45 int (*set_count)(struct dm_space_map *sm, dm_block_t b, uint32_t count); 49 int (*inc_blocks)(struct dm_space_map *sm, dm_block_t b, dm_block_t e); 50 int (*dec_blocks)(struct dm_space_map *sm, dm_block_t b, dm_block_t e); 55 int (*new_block)(struct dm_space_map *sm, dm_block_t *b); 97 static inline int dm_sm_get_count(struct dm_space_map *sm, dm_block_t b, in dm_sm_get_count() argument 100 return sm->get_count(sm, b, result); in dm_sm_get_count() 104 dm_block_t b, int *result) in dm_sm_count_is_more_than_one() argument 106 return sm->count_is_more_than_one(sm, b, result); in dm_sm_count_is_more_than_one() [all …]
|
/drivers/crypto/marvell/octeontx2/ |
D | otx2_cpt_hw_types.h | 46 #define OTX2_CPT_PF_MBOX_INTX(b) (0x400 | (b) << 3) argument 47 #define OTX2_CPT_PF_MBOX_INT_W1SX(b) (0x420 | (b) << 3) argument 48 #define OTX2_CPT_PF_MBOX_ENA_W1CX(b) (0x440 | (b) << 3) argument 49 #define OTX2_CPT_PF_MBOX_ENA_W1SX(b) (0x460 | (b) << 3) argument 54 #define OTX2_CPT_PF_GX_EN(b) (0x600 | (b) << 3) argument 74 #define OTX2_CPT_PF_EXE_DBG_CNTX(b) (0x4001100 | (b) << 3) argument 76 #define OTX2_CPT_PF_EXE_EPCI_INBX_CNT(b) (0x4001200 | (b) << 3) argument 77 #define OTX2_CPT_PF_EXE_EPCI_OUTBX_CNT(b) (0x4001240 | (b) << 3) argument 78 #define OTX2_CPT_PF_ENGX_UCODE_BASE(b) (0x4002000 | (b) << 3) argument 79 #define OTX2_CPT_PF_QX_CTL(b) (0x8000000 | (b) << 20) argument [all …]
|
/drivers/gpu/drm/nouveau/nvkm/subdev/bios/ |
D | vpstate.c | 29 nvbios_vpstate_offset(struct nvkm_bios *b) in nvbios_vpstate_offset() argument 33 if (!bit_entry(b, 'P', &bit_P)) { in nvbios_vpstate_offset() 35 return nvbios_rd32(b, bit_P.offset + 0x38); in nvbios_vpstate_offset() 42 nvbios_vpstate_parse(struct nvkm_bios *b, struct nvbios_vpstate_header *h) in nvbios_vpstate_parse() argument 47 h->offset = nvbios_vpstate_offset(b); in nvbios_vpstate_parse() 51 h->version = nvbios_rd08(b, h->offset); in nvbios_vpstate_parse() 54 h->hlen = nvbios_rd08(b, h->offset + 0x1); in nvbios_vpstate_parse() 55 h->elen = nvbios_rd08(b, h->offset + 0x2); in nvbios_vpstate_parse() 56 h->slen = nvbios_rd08(b, h->offset + 0x3); in nvbios_vpstate_parse() 57 h->scount = nvbios_rd08(b, h->offset + 0x4); in nvbios_vpstate_parse() [all …]
|