/external/llvm-project/llvm/test/Transforms/DeadStoreElimination/MemDepAnalysis/ |
D | free.ll | 50 %alloc1 = tail call noalias i8* @malloc(i64 4) nounwind 54 store i8 1, i8* %alloc1 58 tail call void @free(i8* %alloc1) nounwind
|
/external/llvm/test/Transforms/DeadStoreElimination/ |
D | free.ll | 50 %alloc1 = tail call noalias i8* @malloc(i64 4) nounwind 54 store i8 1, i8* %alloc1 58 tail call void @free(i8* %alloc1) nounwind
|
/external/llvm-project/llvm/test/Transforms/DeadStoreElimination/MSSA/ |
D | free.ll | 62 %alloc1 = tail call noalias i8* @malloc(i64 4) nounwind 66 store i8 1, i8* %alloc1 70 tail call void @free(i8* %alloc1) nounwind
|
/external/python/cffi/testing/cffi1/ |
D | test_ffi_obj.py | 339 alloc1 = ffi.new_allocator() 342 p1 = alloc1("int[10]") 366 alloc1 = ffi.new_allocator(myalloc, myfree) 369 p1 = alloc1("int[10]") 397 alloc1 = ffi.new_allocator(myalloc) # no 'free' 398 p1 = alloc1("int[10]")
|
/external/llvm-project/llvm/test/CodeGen/X86/ |
D | swap.ll | 73 %alloc1 = alloca [16 x i8], align 1 75 %part1 = getelementptr inbounds [16 x i8], [16 x i8]* %alloc1, i64 0, i64 0 101 %alloc1 = alloca [16 x i8], align 1 103 %part1 = getelementptr inbounds [16 x i8], [16 x i8]* %alloc1, i64 0, i64 0
|
/external/llvm-project/llvm/test/Transforms/LoadStoreVectorizer/AMDGPU/ |
D | vect-ptr-ptr-size-mismatch.ll | 40 define void @all_to_cast(i8* nocapture readonly align 16 dereferenceable(16) %alloc1) { 42 %alloc16 = addrspacecast i8* %alloc1 to i8 addrspace(1)*
|
/external/python/cffi/testing/cffi0/ |
D | test_ffi_backend.py | 69 alloc1 = ffi.new_allocator(myalloc, myfree) 72 p1 = alloc1("int[10]") 99 alloc1 = ffi.new_allocator(myalloc) # no 'free' 100 p1 = alloc1("int[10]")
|
/external/python/cffi/c/ |
D | ffi_obj.c | 360 cffi_allocator_t alloc1; in _ffi_new_with_allocator() local 364 alloc1.ca_alloc = (my_alloc == Py_None ? NULL : my_alloc); in _ffi_new_with_allocator() 365 alloc1.ca_free = (my_free == Py_None ? NULL : my_free); in _ffi_new_with_allocator() 366 alloc1.ca_dont_clear = (PyTuple_GET_ITEM(allocator, 3) == Py_False); in _ffi_new_with_allocator() 369 args, kwds, &alloc1); in _ffi_new_with_allocator()
|
/external/tensorflow/tensorflow/compiler/xla/service/gpu/tests/ |
D | scatter.hlo | 3 …6 dereferenceable(36) %alloc0, i8* noalias align 16 dereferenceable(24) %alloc1, i8* noalias align… 73 … 16 dereferenceable(4) %alloc0, i8* noalias align 16 dereferenceable(4) %alloc1, i8* noalias align… 130 …6 dereferenceable(36) %alloc0, i8* noalias align 16 dereferenceable(24) %alloc1, i8* noalias align… 213 …16 dereferenceable(16) %alloc0, i8* noalias align 16 dereferenceable(4) %alloc1, i8* noalias align…
|
/external/llvm-project/llvm/test/Transforms/Attributor/IPConstantProp/ |
D | pthreads.ll | 57 %alloc1 = alloca i8, align 8 62 …(i64* nonnull %thread, %union.pthread_attr_t* null, i8* (i8*)* nonnull @baz, i8* nocapture %alloc1)
|
/external/llvm-project/mlir/test/Transforms/ |
D | buffer-deallocation.mlir | 1116 %alloc1 = alloc(%size) : memref<?xf32> 1117 br ^loopHeader(%inc, %alloc1 : i32, memref<?xf32>) 1144 %alloc1 = alloc() : memref<2xf32> 1145 br ^loopHeader(%inc, %alloc1 : i32, memref<2xf32>)
|
/external/angle/third_party/abseil-cpp/absl/container/ |
D | btree_test.cc | 2856 Alloc alloc1(&bytes_used1); in TEST() local 2857 Set set1(alloc1); in TEST() 2893 Alloc alloc1(&bytes_used1); in TEST() local 2894 Set set1(alloc1); in TEST()
|
/external/openscreen/third_party/abseil/src/absl/container/ |
D | btree_test.cc | 2769 Alloc alloc1(&bytes_used1); in TEST() local 2770 Set set1(alloc1); in TEST() 2806 Alloc alloc1(&bytes_used1); in TEST() local 2807 Set set1(alloc1); in TEST()
|
/external/tensorflow/tensorflow/compiler/xrt/tests/ |
D | raw_api_test.cc | 614 xrt::XLAAllocation alloc1; in TEST() local 615 *alloc1.mutable_value() = in TEST() 623 ops::Const(root.WithDevice("/device:CPU:0"), alloc1.SerializeAsString()); in TEST()
|
/external/llvm-project/mlir/docs/ |
D | Bufferization.md | 570 to free %alloc0 in bb0 and %alloc1 in bb1. Afterwards, we can continue
|