/third_party/pcre2/pcre2/src/sljit/ |
D | sljitNativeX86_64.c | 288 static sljit_u8* generate_far_jump_code(struct sljit_jump *jump, sljit_u8 *code_ptr) in generate_far_jump_code() argument 299 *code_ptr++ = U8(get_jump_code(type ^ 0x1) - 0x10); in generate_far_jump_code() 300 *code_ptr++ = short_addr ? (6 + 3) : (10 + 3); in generate_far_jump_code() 303 *code_ptr++ = short_addr ? REX_B : (REX_W | REX_B); in generate_far_jump_code() 304 *code_ptr++ = MOV_r_i32 | reg_lmap[TMP_REG2]; in generate_far_jump_code() 305 jump->addr = (sljit_uw)code_ptr; in generate_far_jump_code() 310 sljit_unaligned_store_s32(code_ptr, (sljit_s32)jump->u.target); in generate_far_jump_code() 312 sljit_unaligned_store_sw(code_ptr, (sljit_sw)jump->u.target); in generate_far_jump_code() 314 code_ptr += short_addr ? sizeof(sljit_s32) : sizeof(sljit_sw); in generate_far_jump_code() 316 *code_ptr++ = REX_B; in generate_far_jump_code() [all …]
|
D | sljitNativePPC_common.c | 284 static SLJIT_INLINE sljit_s32 detect_jump_type(struct sljit_jump *jump, sljit_ins *code_ptr, sljit_… in detect_jump_type() argument 310 diff = ((sljit_sw)target_addr - (sljit_sw)(code_ptr) - executable_offset) & ~0x3l; in detect_jump_type() 408 sljit_ins *code_ptr; in sljit_generate_code() local 436 code_ptr = code; in sljit_generate_code() 450 *code_ptr = *buf_ptr++; in sljit_generate_code() 460 label->addr = (sljit_uw)SLJIT_ADD_EXEC_OFFSET(code_ptr, executable_offset); in sljit_generate_code() 461 label->size = (sljit_uw)(code_ptr - code); in sljit_generate_code() 466 jump->addr = (sljit_uw)(code_ptr - 3); in sljit_generate_code() 468 jump->addr = (sljit_uw)(code_ptr - 6); in sljit_generate_code() 470 if (detect_jump_type(jump, code_ptr, code, executable_offset)) { in sljit_generate_code() [all …]
|
D | sljitNativeARM_32.c | 291 static sljit_uw patch_pc_relative_loads(sljit_uw *last_pc_patch, sljit_uw *code_ptr, sljit_uw* cons… in patch_pc_relative_loads() argument 299 SLJIT_ASSERT(const_pool - code_ptr <= CONST_POOL_ALIGNMENT); in patch_pc_relative_loads() 306 while (last_pc_patch < code_ptr) { in patch_pc_relative_loads() 417 static SLJIT_INLINE sljit_s32 detect_jump_type(struct sljit_jump *jump, sljit_uw *code_ptr, sljit_u… in detect_jump_type() argument 426 code_ptr--; in detect_jump_type() 429 diff = ((sljit_sw)jump->u.target - (sljit_sw)(code_ptr + 2) - executable_offset); in detect_jump_type() 432 diff = ((sljit_sw)(code + jump->u.label->size) - (sljit_sw)(code_ptr + 2)); in detect_jump_type() 441 *code_ptr = (BL - CONDITIONAL) | (*(code_ptr + 1) & COND_MASK); in detect_jump_type() 448 *code_ptr = (B - CONDITIONAL) | (*code_ptr & COND_MASK); in detect_jump_type() 454 diff = ((sljit_sw)jump->u.target - (sljit_sw)code_ptr - executable_offset); in detect_jump_type() [all …]
|
D | sljitNativeARM_64.c | 161 static SLJIT_INLINE sljit_sw detect_jump_type(struct sljit_jump *jump, sljit_ins *code_ptr, sljit_i… in detect_jump_type() argument 178 diff = (sljit_sw)target_addr - (sljit_sw)(code_ptr + 4) - executable_offset; in detect_jump_type() 183 code_ptr[-5] ^= (jump->flags & IS_CBZ) ? (0x1 << 24) : 0x1; in detect_jump_type() 198 code_ptr[-5] -= (2 << 5); in detect_jump_type() 199 code_ptr[-2] = code_ptr[0]; in detect_jump_type() 205 code_ptr[-5] -= (1 << 5); in detect_jump_type() 207 code_ptr[-1] = code_ptr[0]; in detect_jump_type() 235 sljit_ins *code_ptr; in sljit_generate_code() local 257 code_ptr = code; in sljit_generate_code() 271 *code_ptr = *buf_ptr++; in sljit_generate_code() [all …]
|
D | sljitNativeX86_common.c | 526 static sljit_u8* generate_far_jump_code(struct sljit_jump *jump, sljit_u8 *code_ptr, sljit_sw execu… 528 static sljit_u8* generate_far_jump_code(struct sljit_jump *jump, sljit_u8 *code_ptr); 529 static sljit_u8* generate_put_label_code(struct sljit_put_label *put_label, sljit_u8 *code_ptr, slj… 532 static sljit_u8* generate_near_jump_code(struct sljit_jump *jump, sljit_u8 *code_ptr, sljit_u8 *cod… in generate_near_jump_code() argument 545 return generate_far_jump_code(jump, code_ptr); in generate_near_jump_code() 552 *code_ptr++ = JMP_i8; in generate_near_jump_code() 554 *code_ptr++ = JMP_i32; in generate_near_jump_code() 559 *code_ptr++ = CALL_i32; in generate_near_jump_code() 563 *code_ptr++ = U8(get_jump_code(type) - 0x10); in generate_near_jump_code() 567 *code_ptr++ = GROUP_0F; in generate_near_jump_code() [all …]
|
D | sljitNativeARM_T2_32.c | 255 static SLJIT_INLINE sljit_s32 detect_jump_type(struct sljit_jump *jump, sljit_u16 *code_ptr, sljit_… in detect_jump_type() argument 266 diff = ((sljit_sw)jump->u.target - (sljit_sw)(code_ptr + 2) - executable_offset) >> 1; in detect_jump_type() 270 diff = ((sljit_sw)(code + jump->u.label->size) - (sljit_sw)(code_ptr + 2)) >> 1; in detect_jump_type() 379 sljit_u16 *code_ptr; in sljit_generate_code() local 399 code_ptr = code; in sljit_generate_code() 413 *code_ptr = *buf_ptr++; in sljit_generate_code() 422 label->addr = ((sljit_uw)SLJIT_ADD_EXEC_OFFSET(code_ptr, executable_offset)) | 0x1; in sljit_generate_code() 423 label->size = (sljit_uw)(code_ptr - code); in sljit_generate_code() 427 jump->addr = (sljit_uw)code_ptr - ((jump->flags & IS_COND) ? 10 : 8); in sljit_generate_code() 428 code_ptr -= detect_jump_type(jump, code_ptr, code, executable_offset); in sljit_generate_code() [all …]
|
D | sljitNativeX86_32.c | 232 static sljit_u8* generate_far_jump_code(struct sljit_jump *jump, sljit_u8 *code_ptr, sljit_sw execu… in generate_far_jump_code() argument 237 *code_ptr++ = JMP_i32; in generate_far_jump_code() 241 *code_ptr++ = CALL_i32; in generate_far_jump_code() 245 *code_ptr++ = GROUP_0F; in generate_far_jump_code() 246 *code_ptr++ = get_jump_code(type); in generate_far_jump_code() 253 …sljit_unaligned_store_sw(code_ptr, (sljit_sw)(jump->u.target - (jump->addr + 4) - (sljit_uw)execut… in generate_far_jump_code() 254 code_ptr += 4; in generate_far_jump_code() 256 return code_ptr; in generate_far_jump_code()
|
D | sljitNativeRISCV_common.c | 377 sljit_ins *code_ptr; in sljit_generate_code() local 398 code_ptr = code; in sljit_generate_code() 412 *code_ptr = *buf_ptr++; in sljit_generate_code() 421 label->addr = (sljit_uw)SLJIT_ADD_EXEC_OFFSET(code_ptr, executable_offset); in sljit_generate_code() 422 label->size = (sljit_uw)(code_ptr - code); in sljit_generate_code() 431 jump->addr = (sljit_uw)code_ptr; in sljit_generate_code() 432 code_ptr = detect_jump_type(jump, code, executable_offset); in sljit_generate_code() 436 const_->addr = (sljit_uw)code_ptr; in sljit_generate_code() 441 put_label->addr = (sljit_uw)code_ptr; in sljit_generate_code() 443 code_ptr += 1; in sljit_generate_code() [all …]
|
D | sljitNativeMIPS_common.c | 494 static __attribute__ ((noinline)) void sljit_cache_flush(void* code, void* code_ptr) in sljit_cache_flush() argument 496 SLJIT_CACHE_FLUSH(code, code_ptr); in sljit_cache_flush() 574 sljit_ins *code_ptr; in sljit_generate_code() local 595 code_ptr = code; in sljit_generate_code() 609 *code_ptr = *buf_ptr++; in sljit_generate_code() 618 label->addr = (sljit_uw)SLJIT_ADD_EXEC_OFFSET(code_ptr, executable_offset); in sljit_generate_code() 619 label->size = (sljit_uw)(code_ptr - code); in sljit_generate_code() 628 jump->addr = (sljit_uw)(code_ptr - 1); in sljit_generate_code() 629 code_ptr = detect_jump_type(jump, code, executable_offset); in sljit_generate_code() 633 const_->addr = (sljit_uw)code_ptr; in sljit_generate_code() [all …]
|
D | sljitNativeS390X.c | 1436 void *code, *code_ptr; in sljit_generate_code() local 1513 code_ptr = code; in sljit_generate_code() 1529 (sljit_uw)code_ptr + label->size, executable_offset); in sljit_generate_code() 1549 source = (sljit_sw)code_ptr; in sljit_generate_code() 1573 source = (sljit_sw)SLJIT_ADD_EXEC_OFFSET(code_ptr, executable_offset); in sljit_generate_code() 1580 encode_inst(&code_ptr, lgrl(tmp1, offset & 0xffffffff)); in sljit_generate_code() 1600 jump->addr = (sljit_uw)code_ptr + 2; in sljit_generate_code() 1601 source = (sljit_sw)SLJIT_ADD_EXEC_OFFSET(code_ptr, executable_offset); in sljit_generate_code() 1615 source = (sljit_sw)SLJIT_ADD_EXEC_OFFSET(code_ptr, executable_offset); in sljit_generate_code() 1618 put_label->addr = (sljit_uw)code_ptr; in sljit_generate_code() [all …]
|
/third_party/node/deps/v8/src/base/platform/ |
D | platform-darwin.cc | 53 char* code_ptr = getsectdatafromheader(header, SEG_TEXT, SECT_TEXT, &size); in GetSharedLibraryAddresses() local 56 char* code_ptr = getsectdatafromheader_64( in GetSharedLibraryAddresses() local 60 if (code_ptr == nullptr) continue; in GetSharedLibraryAddresses() 62 const uintptr_t start = reinterpret_cast<uintptr_t>(code_ptr) + slide; in GetSharedLibraryAddresses()
|
/third_party/mesa3d/src/amd/vulkan/ |
D | radv_shader.h | 496 uint8_t *code_ptr; member
|
D | radv_shader.c | 1985 shader->code_ptr = dest_ptr; in radv_shader_binary_upload() 1997 shader->code_ptr = dest_ptr; in radv_shader_binary_upload()
|
/third_party/mesa3d/src/amd/vulkan/layers/ |
D | radv_sqtt_layer.c | 895 memcpy(code, shader->code_ptr, shader->code_size); in radv_add_code_object()
|
/third_party/vk-gl-cts/external/openglcts/modules/gles31/ |
D | es31cArrayOfArraysTests.cpp | 726 const char* code_ptr = shader_source.c_str(); in compile_shader_and_get_compilation_result() local 730 context_id.getTestContext().getLog() << tcu::TestLog::KernelSource(code_ptr); in compile_shader_and_get_compilation_result() 733 gl.shaderSource(shader_object_id, 1 /* count */, &code_ptr, NULL); in compile_shader_and_get_compilation_result() 766 context_id.getTestContext().getLog() << tcu::TestLog::KernelSource(code_ptr); in compile_shader_and_get_compilation_result()
|
/third_party/node/deps/v8/src/execution/ |
D | isolate.cc | 429 uint8_t* const code_ptr = in HashIsolateForEmbeddedBlob() local 450 hash = base::hash_combine(hash, size_t{code_ptr[j]}); in HashIsolateForEmbeddedBlob()
|