/external/tensorflow/tensorflow/python/kernel_tests/ |
D | decode_raw_op_test.py | 33 in_bytes = array_ops.placeholder(dtypes.string, shape=[2]) 34 decode = parsing_ops.decode_raw(in_bytes, out_type=dtypes.uint8) 37 result = decode.eval(feed_dict={in_bytes: ["A", "a"]}) 40 result = decode.eval(feed_dict={in_bytes: ["wer", "XYZ"]}) 47 decode.eval(feed_dict={in_bytes: ["short", "longer"]}) 51 in_bytes = array_ops.placeholder(dtypes.string, shape=[None]) 52 decode = parsing_ops.decode_raw(in_bytes, out_type=dtypes.int16) 55 result = decode.eval(feed_dict={in_bytes: ["AaBC"]}) 62 decode.eval(feed_dict={in_bytes: ["123", "456"]}) 66 in_bytes = array_ops.placeholder(dtypes.string, shape=[None]) [all …]
|
D | decode_compressed_op_test.py | 48 in_bytes = array_ops.placeholder(dtypes.string, shape=[2]) 50 in_bytes, compression_type=compression_type) 54 feed_dict={in_bytes: [self._compress(b"AaAA", compression_type), 61 in_bytes = array_ops.placeholder(dtypes.string, shape=[None]) 63 in_bytes, compression_type=compression_type) 67 feed_dict={in_bytes: [self._compress(b"AaBC", compression_type)]})
|
/external/curl/src/ |
D | tool_convert.c | 53 size_t res, in_bytes, out_bytes; in convert_to_network() local 65 in_bytes = out_bytes = length; in convert_to_network() 66 res = iconv(outbound_cd, &input_ptr, &in_bytes, in convert_to_network() 68 if((res == (size_t)-1) || (in_bytes != 0)) { in convert_to_network() 83 size_t res, in_bytes, out_bytes; in convert_from_network() local 95 in_bytes = out_bytes = length; in convert_from_network() 96 res = iconv(inbound_cd, &input_ptr, &in_bytes, in convert_from_network() 98 if((res == (size_t)-1) || (in_bytes != 0)) { in convert_from_network()
|
/external/curl/lib/ |
D | non-ascii.c | 102 size_t in_bytes, out_bytes, rc; in Curl_convert_to_network() local 121 in_bytes = out_bytes = length; in Curl_convert_to_network() 122 rc = iconv(*cd, &input_ptr, &in_bytes, in Curl_convert_to_network() 126 if((rc == ICONV_ERROR) || (in_bytes != 0)) { in Curl_convert_to_network() 165 size_t in_bytes, out_bytes, rc; in Curl_convert_from_network() local 184 in_bytes = out_bytes = length; in Curl_convert_from_network() 185 rc = iconv(*cd, &input_ptr, &in_bytes, in Curl_convert_from_network() 189 if((rc == ICONV_ERROR) || (in_bytes != 0)) { in Curl_convert_from_network() 229 size_t in_bytes, out_bytes, rc; in Curl_convert_from_utf8() local 248 in_bytes = out_bytes = length; in Curl_convert_from_utf8() [all …]
|
/external/eigen/unsupported/test/ |
D | cxx11_tensor_argmax_cuda.cu | 34 std::size_t in_bytes = in.size() * sizeof(double); in test_cuda_simple_argmax() local 40 cudaMalloc((void**)(&d_in), in_bytes); in test_cuda_simple_argmax() 44 cudaMemcpy(d_in, in.data(), in_bytes, cudaMemcpyHostToDevice); in test_cuda_simple_argmax() 98 std::size_t in_bytes = tensor.size() * sizeof(float); in test_cuda_argmax_dim() local 103 cudaMalloc((void**)(&d_in), in_bytes); in test_cuda_argmax_dim() 106 cudaMemcpy(d_in, tensor.data(), in_bytes, cudaMemcpyHostToDevice); in test_cuda_argmax_dim() 140 cudaMemcpy(d_in, tensor.data(), in_bytes, cudaMemcpyHostToDevice); in test_cuda_argmax_dim() 187 std::size_t in_bytes = tensor.size() * sizeof(float); in test_cuda_argmin_dim() local 192 cudaMalloc((void**)(&d_in), in_bytes); in test_cuda_argmin_dim() 195 cudaMemcpy(d_in, tensor.data(), in_bytes, cudaMemcpyHostToDevice); in test_cuda_argmin_dim() [all …]
|
D | cxx11_tensor_complex_cuda.cu | 88 std::size_t in_bytes = in.size() * sizeof(std::complex<float>); in test_cuda_sum_reductions() local 90 …std::complex<float>* gpu_in_ptr = static_cast<std::complex<float>*>(gpu_device.allocate(in_bytes)); in test_cuda_sum_reductions() 92 gpu_device.memcpyHostToDevice(gpu_in_ptr, in.data(), in_bytes); in test_cuda_sum_reductions() 125 std::size_t in_bytes = in.size() * sizeof(std::complex<float>); in test_cuda_product_reductions() local 127 …std::complex<float>* gpu_in_ptr = static_cast<std::complex<float>*>(gpu_device.allocate(in_bytes)); in test_cuda_product_reductions() 129 gpu_device.memcpyHostToDevice(gpu_in_ptr, in.data(), in_bytes); in test_cuda_product_reductions()
|
D | cxx11_tensor_reduction_cuda.cu | 37 std::size_t in_bytes = in.size() * sizeof(Type); in test_full_reductions() local 39 Type* gpu_in_ptr = static_cast<Type*>(gpu_device.allocate(in_bytes)); in test_full_reductions() 41 gpu_device.memcpyHostToDevice(gpu_in_ptr, in.data(), in_bytes); in test_full_reductions()
|
/external/eigen/test/ |
D | cuda_common.h | 38 std::ptrdiff_t in_bytes = in.size() * sizeof(typename Input::Scalar); in run_on_cuda() local 41 cudaMalloc((void**)(&d_in), in_bytes); in run_on_cuda() 44 cudaMemcpy(d_in, in.data(), in_bytes, cudaMemcpyHostToDevice); in run_on_cuda() 57 …cudaMemcpy(const_cast<typename Input::Scalar*>(in.data()), d_in, in_bytes, cudaMemcpyDeviceToHo… in run_on_cuda()
|
/external/libpng/contrib/tools/ |
D | pngfix.c | 2284 png_uint_32 in_bytes = 0; in zlib_advance() local 2289 while (in_bytes < nbytes && endrc == ZLIB_OK) in zlib_advance() 2442 in_bytes += 1 - zlib->z.avail_in; in zlib_advance() 2445 assert(in_bytes == nbytes || endrc != ZLIB_OK); in zlib_advance() 2449 zlib->compressed_digits, in_bytes - zlib->z.avail_in); in zlib_advance() 2478 if (in_bytes < nbytes) in zlib_advance() 2481 zlib->extra_bytes = nbytes - in_bytes; in zlib_advance()
|
/external/boringssl/src/ssl/ |
D | ssl_lib.cc | 1760 const uint8_t *in_bytes = reinterpret_cast<const uint8_t *>(in); in SSL_CTX_set_tlsext_ticket_keys() local 1761 OPENSSL_memcpy(ctx->tlsext_ticket_key_current->name, in_bytes, 16); in SSL_CTX_set_tlsext_ticket_keys() 1762 OPENSSL_memcpy(ctx->tlsext_ticket_key_current->hmac_key, in_bytes + 16, 16); in SSL_CTX_set_tlsext_ticket_keys() 1763 OPENSSL_memcpy(ctx->tlsext_ticket_key_current->aes_key, in_bytes + 32, 16); in SSL_CTX_set_tlsext_ticket_keys()
|