• Home
  • Raw
  • Download

Lines Matching refs:vassert

156          vassert(r >= 0 && r < 31);  in ppHRegARM64()
161 vassert(r >= 0 && r < 32); in ppHRegARM64()
166 vassert(r >= 0 && r < 32); in ppHRegARM64()
217 vassert(-256 <= simm9 && simm9 <= 255); in ARM64AMode_RI9()
227 vassert(uimm12 >= 0 && uimm12 <= 4095); in ARM64AMode_RI12()
230 default: vassert(0); in ARM64AMode_RI12()
264 vassert(0); in ppARM64AMode()
310 vassert(imm12 < 4096); in ARM64RIA_I12()
311 vassert(shift == 0 || shift == 12); in ARM64RIA_I12()
331 vassert(0); in ppARM64RIA()
368 vassert(bitN < 2); in ARM64RIL_I13()
369 vassert(immR < 64); in ARM64RIL_I13()
370 vassert(immS < 64); in ARM64RIL_I13()
392 vassert(0); in ppARM64RIL()
427 vassert(imm6 > 0 && imm6 < 64); in ARM64RI6_I6()
446 vassert(0); in ppARM64RI6()
877 vassert(hregClass(src) == HRcInt64); in ARM64Instr_MovI()
878 vassert(hregClass(dst) == HRcInt64); in ARM64Instr_MovI()
967 vassert(is_sane_RetLoc(rloc)); in ARM64Instr_Call()
974 vassert(-4096 < simm && simm < 4096); in ARM64Instr_AddToSP()
975 vassert(0 == (simm & 0xF)); in ARM64Instr_AddToSP()
998 vassert(szB == 8 || szB == 4 || szB == 2 || szB == 1); in ARM64Instr_LdrEX()
1005 vassert(szB == 8 || szB == 4 || szB == 2 || szB == 1); in ARM64Instr_StrEX()
1012 vassert(szB == 8 || szB == 4 || szB == 2 || szB == 1); in ARM64Instr_CAS()
1032 vassert(uimm12 < 8192 && 0 == (uimm12 & 1)); in ARM64Instr_VLdStH()
1042 vassert(uimm12 < 16384 && 0 == (uimm12 & 3)); in ARM64Instr_VLdStS()
1052 vassert(uimm12 < 32768 && 0 == (uimm12 & 7)); in ARM64Instr_VLdStD()
1079 vassert(armRM <= 3); in ARM64Instr_VCvtF2I()
1215 vassert(dszBlg2 == 0 || dszBlg2 == 1 || dszBlg2 == 2); in ARM64Instr_VNarrowV()
1269 vassert(0); in ARM64Instr_VShiftImmV()
1271 vassert(maxSh > 0); in ARM64Instr_VShiftImmV()
1272 vassert(amt >= minSh && amt <= maxSh); in ARM64Instr_VShiftImmV()
1282 vassert(amtB >= 1 && amtB <= 15); in ARM64Instr_VExtV()
1296 vassert(0); in ARM64Instr_VImmQ()
1328 vassert(laneNo <= 1); in ARM64Instr_VXfromQ()
1347 vassert(hregClass(src) == HRcVec128); in ARM64Instr_VMov()
1348 vassert(hregClass(dst) == HRcVec128); in ARM64Instr_VMov()
1351 vassert(hregClass(src) == HRcFlt64); in ARM64Instr_VMov()
1352 vassert(hregClass(dst) == HRcFlt64); in ARM64Instr_VMov()
1561 default: vassert(0); in ppARM64Instr()
1573 default: vassert(0); in ppARM64Instr()
1925 vassert(mode64 == True); in getRegUsage_ARM64Instr()
2262 vassert(mode64 == True); in mapRegs_ARM64Instr()
2541 vassert(offsetB >= 0); in genSpill_ARM64()
2542 vassert(!hregIsVirtual(rreg)); in genSpill_ARM64()
2543 vassert(mode64 == True); in genSpill_ARM64()
2548 vassert(0 == (offsetB & 7)); in genSpill_ARM64()
2550 vassert(offsetB < 4096); in genSpill_ARM64()
2558 vassert(0 == (offsetB & 7)); in genSpill_ARM64()
2559 vassert(offsetB >= 0 && offsetB < 32768); in genSpill_ARM64()
2566 vassert(0 == (offsetB & 15)); // check sane alignment in genSpill_ARM64()
2567 vassert(offsetB < 4096); in genSpill_ARM64()
2582 vassert(offsetB >= 0); in genReload_ARM64()
2583 vassert(!hregIsVirtual(rreg)); in genReload_ARM64()
2584 vassert(mode64 == True); in genReload_ARM64()
2589 vassert(0 == (offsetB & 7)); in genReload_ARM64()
2591 vassert(offsetB < 4096); in genReload_ARM64()
2599 vassert(0 == (offsetB & 7)); in genReload_ARM64()
2600 vassert(offsetB >= 0 && offsetB < 32768); in genReload_ARM64()
2607 vassert(0 == (offsetB & 15)); // check sane alignment in genReload_ARM64()
2608 vassert(offsetB < 4096); in genReload_ARM64()
2627 vassert(hregClass(r) == HRcInt64); in iregEnc()
2628 vassert(!hregIsVirtual(r)); in iregEnc()
2630 vassert(n <= 30); in iregEnc()
2637 vassert(hregClass(r) == HRcFlt64); in dregEnc()
2638 vassert(!hregIsVirtual(r)); in dregEnc()
2640 vassert(n <= 31); in dregEnc()
2647 vassert(hregClass(r) == HRcVec128); in qregEnc()
2648 vassert(!hregIsVirtual(r)); in qregEnc()
2650 vassert(n <= 31); in qregEnc()
2779 vassert(8+19+1+4 == 32); in X_8_19_1_4()
2780 vassert(f1 < (1<<8)); in X_8_19_1_4()
2781 vassert(f2 < (1<<19)); in X_8_19_1_4()
2782 vassert(f3 < (1<<1)); in X_8_19_1_4()
2783 vassert(f4 < (1<<4)); in X_8_19_1_4()
2796 vassert(3+6+2+16+5 == 32); in X_3_6_2_16_5()
2797 vassert(f1 < (1<<3)); in X_3_6_2_16_5()
2798 vassert(f2 < (1<<6)); in X_3_6_2_16_5()
2799 vassert(f3 < (1<<2)); in X_3_6_2_16_5()
2800 vassert(f4 < (1<<16)); in X_3_6_2_16_5()
2801 vassert(f5 < (1<<5)); in X_3_6_2_16_5()
2815 vassert(2+6+2+12+5+5 == 32); in X_2_6_2_12_5_5()
2816 vassert(f1 < (1<<2)); in X_2_6_2_12_5_5()
2817 vassert(f2 < (1<<6)); in X_2_6_2_12_5_5()
2818 vassert(f3 < (1<<2)); in X_2_6_2_12_5_5()
2819 vassert(f4 < (1<<12)); in X_2_6_2_12_5_5()
2820 vassert(f5 < (1<<5)); in X_2_6_2_12_5_5()
2821 vassert(f6 < (1<<5)); in X_2_6_2_12_5_5()
2834 vassert(3+8+5+6+5+5 == 32); in X_3_8_5_6_5_5()
2835 vassert(f1 < (1<<3)); in X_3_8_5_6_5_5()
2836 vassert(f2 < (1<<8)); in X_3_8_5_6_5_5()
2837 vassert(f3 < (1<<5)); in X_3_8_5_6_5_5()
2838 vassert(f4 < (1<<6)); in X_3_8_5_6_5_5()
2839 vassert(f5 < (1<<5)); in X_3_8_5_6_5_5()
2840 vassert(f6 < (1<<5)); in X_3_8_5_6_5_5()
2853 vassert(3+8+5+6+5+5 == 32); in X_3_5_8_6_5_5()
2854 vassert(f1 < (1<<3)); in X_3_5_8_6_5_5()
2855 vassert(f2 < (1<<5)); in X_3_5_8_6_5_5()
2856 vassert(f3 < (1<<8)); in X_3_5_8_6_5_5()
2857 vassert(f4 < (1<<6)); in X_3_5_8_6_5_5()
2858 vassert(f5 < (1<<5)); in X_3_5_8_6_5_5()
2859 vassert(f6 < (1<<5)); in X_3_5_8_6_5_5()
2872 vassert(3+6+7+6+5+5 == 32); in X_3_6_7_6_5_5()
2873 vassert(f1 < (1<<3)); in X_3_6_7_6_5_5()
2874 vassert(f2 < (1<<6)); in X_3_6_7_6_5_5()
2875 vassert(f3 < (1<<7)); in X_3_6_7_6_5_5()
2876 vassert(f4 < (1<<6)); in X_3_6_7_6_5_5()
2877 vassert(f5 < (1<<5)); in X_3_6_7_6_5_5()
2878 vassert(f6 < (1<<5)); in X_3_6_7_6_5_5()
2893 vassert(2+6+3+9+2+5+5 == 32); in X_2_6_3_9_2_5_5()
2894 vassert(f1 < (1<<2)); in X_2_6_3_9_2_5_5()
2895 vassert(f2 < (1<<6)); in X_2_6_3_9_2_5_5()
2896 vassert(f3 < (1<<3)); in X_2_6_3_9_2_5_5()
2897 vassert(f4 < (1<<9)); in X_2_6_3_9_2_5_5()
2898 vassert(f5 < (1<<2)); in X_2_6_3_9_2_5_5()
2899 vassert(f6 < (1<<5)); in X_2_6_3_9_2_5_5()
2900 vassert(f7 < (1<<5)); in X_2_6_3_9_2_5_5()
2914 vassert(3+6+1+6+6+5+5 == 32); in X_3_6_1_6_6_5_5()
2915 vassert(f1 < (1<<3)); in X_3_6_1_6_6_5_5()
2916 vassert(f2 < (1<<6)); in X_3_6_1_6_6_5_5()
2917 vassert(f3 < (1<<1)); in X_3_6_1_6_6_5_5()
2918 vassert(f4 < (1<<6)); in X_3_6_1_6_6_5_5()
2919 vassert(f5 < (1<<6)); in X_3_6_1_6_6_5_5()
2920 vassert(f6 < (1<<5)); in X_3_6_1_6_6_5_5()
2921 vassert(f7 < (1<<5)); in X_3_6_1_6_6_5_5()
3007 vassert(i < 4); in imm64_to_ireg()
3084 vassert(wD <= 30); in do_load_or_store8()
3090 vassert(-256 <= simm9 && simm9 <= 255); in do_load_or_store8()
3103 vassert(scale == 1); /* failure of this is serious. Do not ignore. */ in do_load_or_store8()
3105 vassert(xN <= 30); in do_load_or_store8()
3117 vassert(xN <= 30); in do_load_or_store8()
3124 vassert(0); in do_load_or_store8()
3133 vassert(wD <= 30); in do_load_or_store16()
3139 vassert(-256 <= simm9 && simm9 <= 255); in do_load_or_store16()
3152 vassert(scale == 2); /* failure of this is serious. Do not ignore. */ in do_load_or_store16()
3154 vassert(xN <= 30); in do_load_or_store16()
3166 vassert(xN <= 30); in do_load_or_store16()
3173 vassert(0); in do_load_or_store16()
3182 vassert(wD <= 30); in do_load_or_store32()
3188 vassert(-256 <= simm9 && simm9 <= 255); in do_load_or_store32()
3201 vassert(scale == 4); /* failure of this is serious. Do not ignore. */ in do_load_or_store32()
3203 vassert(xN <= 30); in do_load_or_store32()
3215 vassert(xN <= 30); in do_load_or_store32()
3222 vassert(0); in do_load_or_store32()
3232 vassert(xD <= 30); in do_load_or_store64()
3238 vassert(-256 <= simm9 && simm9 <= 255); in do_load_or_store64()
3240 vassert(xN <= 30); in do_load_or_store64()
3252 vassert(scale == 8); /* failure of this is serious. Do not ignore. */ in do_load_or_store64()
3254 vassert(xN <= 30); in do_load_or_store64()
3266 vassert(xN <= 30); in do_load_or_store64()
3273 vassert(0); in do_load_or_store64()
3292 vassert(nbuf >= 32); in emit_ARM64Instr()
3293 vassert(mode64 == True); in emit_ARM64Instr()
3294 vassert(0 == (((HWord)buf) & 3)); in emit_ARM64Instr()
3355 vassert(rD < 31); in emit_ARM64Instr()
3356 vassert(rN < 31); in emit_ARM64Instr()
3363 vassert(opc != 0); in emit_ARM64Instr()
3381 vassert(rM < 31); in emit_ARM64Instr()
3413 vassert(rD < 31); in emit_ARM64Instr()
3414 vassert(rN < 31); in emit_ARM64Instr()
3421 vassert(sh > 0 && sh < 64); in emit_ARM64Instr()
3434 vassert(0); in emit_ARM64Instr()
3443 vassert(rM < 31); in emit_ARM64Instr()
3449 default: vassert(0); in emit_ARM64Instr()
3455 vassert(0); in emit_ARM64Instr()
3532 vassert(disp_cp_chain_me_to_slowEP != NULL); in emit_ARM64Instr()
3533 vassert(disp_cp_chain_me_to_fastEP != NULL); in emit_ARM64Instr()
3542 vassert(i->ARM64in.XDirect.cond != ARM64cc_NV); in emit_ARM64Instr()
3574 vassert(delta > 0 && delta < 40); in emit_ARM64Instr()
3575 vassert((delta & 3) == 0); in emit_ARM64Instr()
3577 vassert(notCond <= 13); /* Neither AL nor NV */ in emit_ARM64Instr()
3578 vassert(ptmp != NULL); in emit_ARM64Instr()
3596 vassert(0); //ATC in emit_ARM64Instr()
3615 vassert(0); //ATC in emit_ARM64Instr()
3636 vassert(i->ARM64in.XDirect.cond != ARM64cc_NV); in emit_ARM64Instr()
3672 vassert(trcval != 0); in emit_ARM64Instr()
3683 vassert(delta > 0 && delta < 40); in emit_ARM64Instr()
3684 vassert((delta & 3) == 0); in emit_ARM64Instr()
3686 vassert(notCond <= 13); /* Neither AL nor NV */ in emit_ARM64Instr()
3687 vassert(ptmp != NULL); in emit_ARM64Instr()
3700 vassert(dd < 31 && nn < 31 && mm < 31 && cond < 16); in emit_ARM64Instr()
3738 vassert(dist >= 2 && dist <= 5); in emit_ARM64Instr()
3739 vassert(ptmp != NULL); in emit_ARM64Instr()
3744 vassert(ptmp == NULL); in emit_ARM64Instr()
3755 vassert(-4096 < simm12 && simm12 < 4096); in emit_ARM64Instr()
3756 vassert(0 == (simm12 & 0xF)); in emit_ARM64Instr()
3768 vassert(dd < 31); in emit_ARM64Instr()
3781 vassert(dd < 31 && nn < 31 && mm < 31); in emit_ARM64Instr()
3793 vassert(0); in emit_ARM64Instr()
3863 default: vassert(0); in emit_ARM64Instr()
3900 vassert(uimm12 < 8192 && 0 == (uimm12 & 1)); in emit_ARM64Instr()
3902 vassert(uimm12 < (1<<12)); in emit_ARM64Instr()
3903 vassert(hD < 32); in emit_ARM64Instr()
3904 vassert(rN < 31); in emit_ARM64Instr()
3917 vassert(uimm12 < 16384 && 0 == (uimm12 & 3)); in emit_ARM64Instr()
3919 vassert(uimm12 < (1<<12)); in emit_ARM64Instr()
3920 vassert(sD < 32); in emit_ARM64Instr()
3921 vassert(rN < 31); in emit_ARM64Instr()
3934 vassert(uimm12 < 32768 && 0 == (uimm12 & 7)); in emit_ARM64Instr()
3936 vassert(uimm12 < (1<<12)); in emit_ARM64Instr()
3937 vassert(dD < 32); in emit_ARM64Instr()
3938 vassert(rN < 31); in emit_ARM64Instr()
3949 vassert(rQ < 32); in emit_ARM64Instr()
3950 vassert(rN < 31); in emit_ARM64Instr()
4197 vassert(b1512 < 16); in emit_ARM64Instr()
4220 vassert(b1512 < 16); in emit_ARM64Instr()
4249 vassert(cond < 16); in emit_ARM64Instr()
5086 vassert(dszBlg2 >= 0 && dszBlg2 <= 2); in emit_ARM64Instr()
5284 vassert(imm4 >= 1 && imm4 <= 15); in emit_ARM64Instr()
5292 vassert(rQ < 32); in emit_ARM64Instr()
5336 vassert(xx < 31); in emit_ARM64Instr()
5348 vassert(xx < 31); in emit_ARM64Instr()
5362 vassert(xhi < 31 && xlo < 31); in emit_ARM64Instr()
5375 vassert(dd < 31); in emit_ARM64Instr()
5376 vassert(laneNo < 2); in emit_ARM64Instr()
5389 vassert(dd < 31); in emit_ARM64Instr()
5447 vassert(evCheckSzB_ARM64() == (UChar*)p - (UChar*)p0); in emit_ARM64Instr()
5467 vassert(!(*is_profInc)); in emit_ARM64Instr()
5483 vassert(((UChar*)p) - &buf[0] <= 36); in emit_ARM64Instr()
5504 vassert(endness_host == VexEndnessLE); in chainXDirect_ARM64()
5517 vassert(0 == (3 & (HWord)p)); in chainXDirect_ARM64()
5518 vassert(is_imm64_to_ireg_EXACTLY4( in chainXDirect_ARM64()
5520 vassert(p[4] == 0xD63F0120); in chainXDirect_ARM64()
5549 vassert(endness_host == VexEndnessLE); in unchainXDirect_ARM64()
5562 vassert(0 == (3 & (HWord)p)); in unchainXDirect_ARM64()
5563 vassert(is_imm64_to_ireg_EXACTLY4( in unchainXDirect_ARM64()
5565 vassert(p[4] == 0xD61F0120); in unchainXDirect_ARM64()
5591 vassert(sizeof(ULong*) == 8); in patchProfInc_ARM64()
5592 vassert(endness_host == VexEndnessLE); in patchProfInc_ARM64()
5594 vassert(0 == (3 & (HWord)p)); in patchProfInc_ARM64()
5595 vassert(is_imm64_to_ireg_EXACTLY4(p, /*x*/9, 0x6555755585559566ULL)); in patchProfInc_ARM64()
5596 vassert(p[4] == 0xF9400128); in patchProfInc_ARM64()
5597 vassert(p[5] == 0x91000508); in patchProfInc_ARM64()
5598 vassert(p[6] == 0xF9000128); in patchProfInc_ARM64()