/external/tensorflow/tensorflow/core/kernels/sparse/ |
D | kernels_test.cc | 34 auto csr_row_ptr = test::AsTensor<int32>({0, 0, 0, 0, 0}); in TEST() local 39 batch_ptr.vec<int32>(), csr_row_ptr.vec<int32>(), in TEST() 43 test::ExpectTensorEqual<int32>(csr_row_ptr, in TEST() 58 Tensor csr_row_ptr(DT_INT32, {12}); in TEST() local 59 test::FillFn<int32>(&csr_row_ptr, [](int unused) { return 0; }); in TEST() 64 batch_ptr.vec<int32>(), csr_row_ptr.vec<int32>(), in TEST() 69 test::ExpectTensorEqual<int32>(csr_row_ptr, in TEST()
|
D | kernels.cc | 32 TTypes<int32>::Vec csr_row_ptr, TTypes<int32>::Vec csr_col_ind) { in operator ()() argument 39 if (csr_row_ptr.size() != batch_size * (num_rows + 1)) { in operator ()() 42 csr_row_ptr.size(), " vs. ", batch_size * (num_rows + 1)); in operator ()() 66 csr_row_ptr(indices(i, 0) + 1) += 1; in operator ()() 73 csr_row_ptr(cur_batch * (num_rows + 1) + indices(i, 1) + 1) += 1; in operator ()() 92 auto* row_ptr_batch = csr_row_ptr.data() + batch_idx * (num_rows + 1); in operator ()()
|
D | sparse_tensor_to_csr_sparse_matrix_op.cc | 80 Tensor csr_row_ptr(cpu_allocator(), DT_INT32, in Compute() local 85 set_zero(ctx->eigen_device<CPUDevice>(), csr_row_ptr.flat<int32>()); in Compute() 91 batch_ptr.vec<int32>(), csr_row_ptr.vec<int32>(), in Compute() 99 DataTypeToEnum<T>::value, dense_shape, batch_ptr, csr_row_ptr, in Compute() 227 auto csr_row_ptr = csr_row_ptr_t.vec<int32>(); in ComputeAsync() local 255 &csr_row_ptr((rows + 1) * i), rows + 1); in ComputeAsync() 309 TTypes<int>::UnalignedVec csr_row_ptr) { in operator ()() 314 /*m == rows of A*/ rows, csr_row_ptr.data()); in operator ()()
|
D | csr_sparse_matrix_to_sparse_tensor_op.cc | 94 auto csr_row_ptr = csr_sparse_matrix->row_pointers().vec<int32>(); in Compute() local 108 const int64 col_begin = csr_row_ptr(row_offset); in Compute() 109 const int64 col_end = csr_row_ptr(row_offset + 1); in Compute() 167 auto csr_row_ptr = csr_sparse_matrix->row_pointers().vec<int32>(); in Compute() local 185 &csr_row_ptr((rows + 1) * i), rows + 1); in Compute() 238 TTypes<const int>::UnalignedVec csr_row_ptr,
|
D | dense_to_csr_sparse_matrix_op.cc | 104 Tensor csr_row_ptr(cpu_allocator(), DT_INT32, in Compute() local 109 set_zero(ctx->eigen_device<Device>(), csr_row_ptr.flat<int32>()); in Compute() 115 batch_ptr.vec<int32>(), csr_row_ptr.vec<int32>(), in Compute() 120 values.dtype(), dense_shape, batch_ptr, csr_row_ptr, in Compute() 294 auto csr_row_ptr = csr_row_ptr_t.vec<int32>(); in ComputeAsync() local 322 &csr_row_ptr((rows + 1) * i), rows + 1); in ComputeAsync() 389 TTypes<int>::UnalignedVec csr_row_ptr) { in operator ()() 394 /*m == rows of A*/ rows, csr_row_ptr.data()); in operator ()()
|
D | kernels.h | 104 TTypes<int32>::UnalignedVec csr_row_ptr); 125 TTypes<int32>::Vec csr_row_ptr, 139 TTypes<int32>::UnalignedConstVec csr_row_ptr,
|
D | csr_sparse_matrix_to_dense_op.cc | 161 auto csr_row_ptr = csr_sparse_matrix->row_pointers().vec<int32>(); in Compute() local 179 &csr_row_ptr((rows + 1) * i), rows + 1); in Compute() 254 TTypes<const int>::UnalignedVec csr_row_ptr,
|
D | kernels_gpu.cu.cc | 139 OpKernelContext* c, TTypes<const int>::UnalignedVec csr_row_ptr, in operator ()() argument 144 const int m = csr_row_ptr.size() - 1; // rows in operator ()() 145 return gpu_sparse.Csr2coo(csr_row_ptr.data(), nnz, m, coo_row_ind.data()); in operator ()()
|