/external/tensorflow/tensorflow/core/kernels/image/ |
D | adjust_hsv_gpu.cu.h | 75 const float new_h = h * 6.0f; in hsv2rgb_cuda() local 77 const float x = chroma * (1.0f - fabsf(fmodf(new_h, 2.0f) - 1.0f)); in hsv2rgb_cuda() 79 const bool between_0_and_1 = new_h >= 0.0f && new_h < 1.0f; in hsv2rgb_cuda() 80 const bool between_1_and_2 = new_h >= 1.0f && new_h < 2.0f; in hsv2rgb_cuda() 81 const bool between_2_and_3 = new_h >= 2.0f && new_h < 3.0f; in hsv2rgb_cuda() 82 const bool between_3_and_4 = new_h >= 3.0f && new_h < 4.0f; in hsv2rgb_cuda() 83 const bool between_4_and_5 = new_h >= 4.0f && new_h < 5.0f; in hsv2rgb_cuda() 84 const bool between_5_and_6 = new_h >= 5.0f && new_h < 6.0f; in hsv2rgb_cuda() 113 float new_h = hsv.h; in adjust_hsv_nhwc() local 119 new_h = fmodf(hsv.h + delta, 1.0f); in adjust_hsv_nhwc() [all …]
|
/external/sdv/vsomeip/third_party/boost/numeric/odeint/include/boost/numeric/odeint/stepper/ |
D | bulirsch_stoer.hpp | 216 time_type new_h = dt; in try_step() local 249 new_h = h_opt[k]; in try_step() 250 … new_h *= static_cast<value_type>( m_cost[k+1] ) / static_cast<value_type>( m_cost[k] ); in try_step() 253 new_h = h_opt[k]; in try_step() 260 new_h = h_opt[k]; in try_step() 273 new_h = h_opt[m_current_k_opt]; in try_step() 278 new_h = h_opt[k]; in try_step() 279 … new_h *= static_cast<value_type>(m_cost[m_current_k_opt])/static_cast<value_type>(m_cost[k]); in try_step() 281 new_h = h_opt[m_current_k_opt]; in try_step() 287 new_h = h_opt[m_current_k_opt]; in try_step() [all …]
|
D | bulirsch_stoer_dense_out.hpp | 181 time_type new_h = dt; in try_step() local 215 … new_h = h_opt[k] * static_cast<value_type>( m_cost[k+1] ) / static_cast<value_type>( m_cost[k] ); in try_step() 218 new_h = h_opt[k]; in try_step() 225 new_h = h_opt[k]; in try_step() 238 new_h = h_opt[m_current_k_opt]; in try_step() 243 …new_h = h_opt[k]*static_cast<value_type>( m_cost[m_current_k_opt] ) / static_cast<value_type>( m_c… in try_step() 245 new_h = h_opt[m_current_k_opt]; in try_step() 251 new_h = h_opt[m_current_k_opt]; in try_step() 264 new_h = h_opt[m_current_k_opt]; in try_step() 268 new_h = h_opt[m_current_k_opt]; in try_step() [all …]
|
/external/compiler-rt/lib/tsan/rtl/ |
D | tsan_libdispatch_mac.cc | 514 fd_handler_t new_h = Block_copy(^(dispatch_data_t data, int error) { in TSAN_INTERCEPTOR() local 522 REAL(dispatch_read)(fd, length, q, new_h); in TSAN_INTERCEPTOR() 523 Block_release(new_h); in TSAN_INTERCEPTOR() 531 fd_handler_t new_h = Block_copy(^(dispatch_data_t data, int error) { in TSAN_INTERCEPTOR() local 539 REAL(dispatch_write)(fd, data, q, new_h); in TSAN_INTERCEPTOR() 540 Block_release(new_h); in TSAN_INTERCEPTOR() 548 dispatch_io_handler_t new_h = in TSAN_INTERCEPTOR() local 557 REAL(dispatch_io_read)(channel, offset, length, q, new_h); in TSAN_INTERCEPTOR() 558 Block_release(new_h); in TSAN_INTERCEPTOR() 567 dispatch_io_handler_t new_h = in TSAN_INTERCEPTOR() local [all …]
|
/external/tensorflow/tensorflow/python/keras/layers/ |
D | recurrent_v2.py | 501 (last_output, outputs, new_h, 529 last_output, outputs, new_h, runtime = gpu_gru(**gpu_gru_kwargs) 531 last_output, outputs, new_h, runtime = standard_gru( 534 last_output, outputs, new_h, runtime = gru_with_backend_selection( 537 states = [new_h] 809 (last_output, outputs, new_h, 834 last_output, outputs, new_h, runtime = defun_standard_gru(**params) 837 return last_output, outputs, new_h, runtime 1218 (last_output, outputs, new_h, new_c, 1259 last_output, outputs, new_h, new_c, runtime = gpu_lstm( [all …]
|
/external/angle/third_party/abseil-cpp/absl/synchronization/ |
D | mutex.cc | 2031 PerThreadSynch* new_h = Enqueue(nullptr, waitp, v, flags); in LockSlowLoop() local 2035 ABSL_RAW_CHECK(new_h != nullptr, "Enqueue to empty list failed"); in LockSlowLoop() 2040 v, reinterpret_cast<intptr_t>(new_h) | nv, in LockSlowLoop() 2080 PerThreadSynch* new_h = Enqueue(h, waitp, v, flags); in LockSlowLoop() local 2082 ABSL_RAW_CHECK(new_h != nullptr, "Enqueue to list failed"); in LockSlowLoop() 2091 reinterpret_cast<intptr_t>(new_h), in LockSlowLoop() 2179 PerThreadSynch* new_h = nullptr; in UnlockSlow() local 2186 new_h = Enqueue(nullptr, waitp, new_readers, kMuIsCond); in UnlockSlow() 2193 if (new_h != nullptr) { in UnlockSlow() 2194 nv |= kMuWait | reinterpret_cast<intptr_t>(new_h); in UnlockSlow() [all …]
|
/external/openscreen/third_party/abseil/src/absl/synchronization/ |
D | mutex.cc | 1937 PerThreadSynch *new_h = Enqueue(nullptr, waitp, v, flags); in LockSlowLoop() local 1940 ABSL_RAW_CHECK(new_h != nullptr, "Enqueue to empty list failed"); in LockSlowLoop() 1945 v, reinterpret_cast<intptr_t>(new_h) | nv, in LockSlowLoop() 1983 PerThreadSynch *new_h = Enqueue(h, waitp, v, flags); in LockSlowLoop() local 1985 ABSL_RAW_CHECK(new_h != nullptr, "Enqueue to list failed"); in LockSlowLoop() 1993 reinterpret_cast<intptr_t>(new_h), in LockSlowLoop() 2083 PerThreadSynch *new_h = nullptr; in UnlockSlow() local 2090 new_h = Enqueue(nullptr, waitp, new_readers, kMuIsCond); in UnlockSlow() 2097 if (new_h != nullptr) { in UnlockSlow() 2098 nv |= kMuWait | reinterpret_cast<intptr_t>(new_h); in UnlockSlow() [all …]
|
/external/cronet/tot/third_party/abseil-cpp/absl/synchronization/ |
D | mutex.cc | 2031 PerThreadSynch* new_h = Enqueue(nullptr, waitp, v, flags); in LockSlowLoop() local 2035 ABSL_RAW_CHECK(new_h != nullptr, "Enqueue to empty list failed"); in LockSlowLoop() 2040 v, reinterpret_cast<intptr_t>(new_h) | nv, in LockSlowLoop() 2080 PerThreadSynch* new_h = Enqueue(h, waitp, v, flags); in LockSlowLoop() local 2082 ABSL_RAW_CHECK(new_h != nullptr, "Enqueue to list failed"); in LockSlowLoop() 2091 reinterpret_cast<intptr_t>(new_h), in LockSlowLoop() 2179 PerThreadSynch* new_h = nullptr; in UnlockSlow() local 2186 new_h = Enqueue(nullptr, waitp, new_readers, kMuIsCond); in UnlockSlow() 2193 if (new_h != nullptr) { in UnlockSlow() 2194 nv |= kMuWait | reinterpret_cast<intptr_t>(new_h); in UnlockSlow() [all …]
|
/external/abseil-cpp/absl/synchronization/ |
D | mutex.cc | 2005 PerThreadSynch* new_h = Enqueue(nullptr, waitp, v, flags); in LockSlowLoop() local 2009 ABSL_RAW_CHECK(new_h != nullptr, "Enqueue to empty list failed"); in LockSlowLoop() 2014 v, reinterpret_cast<intptr_t>(new_h) | nv, in LockSlowLoop() 2054 PerThreadSynch* new_h = Enqueue(h, waitp, v, flags); in LockSlowLoop() local 2056 ABSL_RAW_CHECK(new_h != nullptr, "Enqueue to list failed"); in LockSlowLoop() 2065 reinterpret_cast<intptr_t>(new_h), in LockSlowLoop() 2153 PerThreadSynch* new_h = nullptr; in UnlockSlow() local 2160 new_h = Enqueue(nullptr, waitp, new_readers, kMuIsCond); in UnlockSlow() 2167 if (new_h != nullptr) { in UnlockSlow() 2168 nv |= kMuWait | reinterpret_cast<intptr_t>(new_h); in UnlockSlow() [all …]
|
/external/cronet/stable/third_party/abseil-cpp/absl/synchronization/ |
D | mutex.cc | 2031 PerThreadSynch* new_h = Enqueue(nullptr, waitp, v, flags); in LockSlowLoop() local 2035 ABSL_RAW_CHECK(new_h != nullptr, "Enqueue to empty list failed"); in LockSlowLoop() 2040 v, reinterpret_cast<intptr_t>(new_h) | nv, in LockSlowLoop() 2080 PerThreadSynch* new_h = Enqueue(h, waitp, v, flags); in LockSlowLoop() local 2082 ABSL_RAW_CHECK(new_h != nullptr, "Enqueue to list failed"); in LockSlowLoop() 2091 reinterpret_cast<intptr_t>(new_h), in LockSlowLoop() 2179 PerThreadSynch* new_h = nullptr; in UnlockSlow() local 2186 new_h = Enqueue(nullptr, waitp, new_readers, kMuIsCond); in UnlockSlow() 2193 if (new_h != nullptr) { in UnlockSlow() 2194 nv |= kMuWait | reinterpret_cast<intptr_t>(new_h); in UnlockSlow() [all …]
|
/external/rust/android-crates-io/crates/grpcio-sys/grpc/third_party/abseil-cpp/absl/synchronization/ |
D | mutex.cc | 1998 PerThreadSynch *new_h = Enqueue(nullptr, waitp, v, flags); in LockSlowLoop() local 2002 ABSL_RAW_CHECK(new_h != nullptr, "Enqueue to empty list failed"); in LockSlowLoop() 2007 v, reinterpret_cast<intptr_t>(new_h) | nv, in LockSlowLoop() 2047 PerThreadSynch *new_h = Enqueue(h, waitp, v, flags); in LockSlowLoop() local 2049 ABSL_RAW_CHECK(new_h != nullptr, "Enqueue to list failed"); in LockSlowLoop() 2057 reinterpret_cast<intptr_t>(new_h), in LockSlowLoop() 2147 PerThreadSynch *new_h = nullptr; in UnlockSlow() local 2154 new_h = Enqueue(nullptr, waitp, new_readers, kMuIsCond); in UnlockSlow() 2161 if (new_h != nullptr) { in UnlockSlow() 2162 nv |= kMuWait | reinterpret_cast<intptr_t>(new_h); in UnlockSlow() [all …]
|
/external/tensorflow/tensorflow/python/keras/layers/legacy_rnn/ |
D | rnn_cell_impl.py | 616 new_h = u * state + (1 - u) * c 617 return new_h, new_h 806 new_h = multiply(self._activation(new_c), sigmoid(o)) 809 new_state = LSTMStateTuple(new_c, new_h) 811 new_state = array_ops.concat([new_c, new_h], 1) 812 return new_h, new_state
|
/external/tensorflow/tensorflow/core/common_runtime/ |
D | bfc_allocator.cc | 910 ChunkHandle new_h = TryToCoalesce(h, (required_bytes > 0)); in MergeTimestampedChunks() local 911 InsertFreeChunkIntoBin(new_h); in MergeTimestampedChunks() 913 c = ChunkFromHandle(new_h); in MergeTimestampedChunks() 914 if (new_h != h && c->freed_at_count > 0) { in MergeTimestampedChunks() 915 timestamped_chunks_.push_back(new_h); in MergeTimestampedChunks()
|
/external/pytorch/test/ |
D | test_jit.py | 3061 new_h = torch.tanh(x + h) 3062 return new_h, new_h
|