• Home
  • Raw
  • Download

Lines Matching refs:assm_

218 ConstantPool::ConstantPool(Assembler* assm) : assm_(assm) {}  in ConstantPool()
225 return RecordKey(std::move(key), assm_->pc_offset()); in RecordEntry()
232 return RecordKey(std::move(key), assm_->pc_offset()); in RecordEntry()
270 Assembler::BlockPoolsScope block_pools(assm_, PoolEmissionCheck::kSkip); in EmitAndClear()
272 IsAlignmentRequiredIfEmittedAt(require_jump, assm_->pc_offset()); in EmitAndClear()
275 assm_->bind(&size_check); in EmitAndClear()
276 assm_->RecordConstPool(size); in EmitAndClear()
290 if (require_jump == Jump::kRequired) assm_->b(&after_pool); in EmitAndClear()
292 assm_->RecordComment("[ Constant Pool"); in EmitAndClear()
294 if (require_alignment == Alignment::kRequired) assm_->Align(kInt64Size); in EmitAndClear()
296 assm_->RecordComment("]"); in EmitAndClear()
298 if (after_pool.is_linked()) assm_->bind(&after_pool); in EmitAndClear()
300 DCHECK_EQ(assm_->SizeOfCodeGeneratedSince(&size_check), size); in EmitAndClear()
325 DCHECK(IsInImmRangeIfEmittedAt(assm_->pc_offset())); in EndBlock()
335 assm_->pc_offset() + static_cast<int>(instructions * kInstrSize); in SetNextCheckIn()
340 DCHECK(iter->first.is_value32() || IsAligned(assm_->pc_offset(), 8)); in EmitEntries()
344 SetLoadOffsetToConstPoolEntry(it->second, assm_->pc(), it->first); in EmitEntries()
354 assm_->dd(key.value32()); in Emit()
356 assm_->dq(key.value64()); in Emit()
376 size_t pool_end_32 = assm_->pc_offset() + margin + worst_case_size; in ShouldEmitNow()
442 pool_->assm_->EmitConstPoolWithJumpIfNeeded(margin); in BlockScope()
455 if (assm_->pc_offset() >= next_check_) { in MaybeCheck()
466 ConstantPool::ConstantPool(Assembler* assm) : assm_(assm) {} in ConstantPool()
473 return RecordKey(std::move(key), assm_->pc_offset()); in RecordEntry()
480 return RecordKey(std::move(key), assm_->pc_offset()); in RecordEntry()
518 Assembler::BlockPoolsScope block_pools(assm_, PoolEmissionCheck::kSkip); in EmitAndClear()
520 IsAlignmentRequiredIfEmittedAt(require_jump, assm_->pc_offset()); in EmitAndClear()
523 assm_->bind(&size_check); in EmitAndClear()
524 assm_->RecordConstPool(size); in EmitAndClear()
538 if (require_jump == Jump::kRequired) assm_->b(&after_pool); in EmitAndClear()
540 assm_->RecordComment("[ Constant Pool"); in EmitAndClear()
543 if (require_alignment == Alignment::kRequired) assm_->DataAlign(kInt64Size); in EmitAndClear()
545 assm_->RecordComment("]"); in EmitAndClear()
546 assm_->bind(&after_pool); in EmitAndClear()
549 DCHECK_LE(assm_->SizeOfCodeGeneratedSince(&size_check) - size, 3); in EmitAndClear()
574 DCHECK(IsInImmRangeIfEmittedAt(assm_->pc_offset())); in EndBlock()
584 assm_->pc_offset() + static_cast<int>(instructions * kInstrSize); in SetNextCheckIn()
589 DCHECK(iter->first.is_value32() || IsAligned(assm_->pc_offset(), 8)); in EmitEntries()
593 SetLoadOffsetToConstPoolEntry(it->second, assm_->pc(), it->first); in EmitEntries()
603 assm_->dd(key.value32()); in Emit()
605 assm_->dq(key.value64()); in Emit()
625 size_t pool_end_32 = assm_->pc_offset() + margin + worst_case_size; in ShouldEmitNow()
691 pool_->assm_->EmitConstPoolWithJumpIfNeeded(margin); in BlockScope()
704 if (assm_->pc_offset() >= next_check_) { in MaybeCheck()