/external/tensorflow/tensorflow/lite/experimental/microfrontend/ |
D | audio_microfrontend.cc | 133 std::vector<std::vector<T>> frame_buffer(num_frames); in GenerateFeatures() local 144 frame_buffer[frame_index].reserve(output.size); in GenerateFeatures() 147 frame_buffer[frame_index].push_back(static_cast<T>(output.values[i]) / in GenerateFeatures() 157 for (anchor = 0; anchor < frame_buffer.size(); anchor += data->frame_stride) { in GenerateFeatures() 162 if (data->zero_padding && (frame < 0 || frame >= frame_buffer.size())) { in GenerateFeatures() 165 feature = &frame_buffer[0]; in GenerateFeatures() 166 } else if (frame >= frame_buffer.size()) { in GenerateFeatures() 167 feature = &frame_buffer[frame_buffer.size() - 1]; in GenerateFeatures() 169 feature = &frame_buffer[frame]; in GenerateFeatures()
|
/external/tensorflow/tensorflow/lite/experimental/microfrontend/ops/ |
D | audio_microfrontend_op.cc | 242 std::vector<std::vector<T>> frame_buffer(num_frames); in Compute() local 252 frame_buffer[frame_index].reserve(output.size); in Compute() 255 frame_buffer[frame_index].push_back(static_cast<T>(output.values[i]) / in Compute() 266 for (anchor = 0; anchor < frame_buffer.size(); anchor += frame_stride_) { in Compute() 271 if (zero_padding_ && (frame < 0 || frame >= frame_buffer.size())) { in Compute() 274 feature = &frame_buffer[0]; in Compute() 275 } else if (frame >= frame_buffer.size()) { in Compute() 276 feature = &frame_buffer[frame_buffer.size() - 1]; in Compute() 278 feature = &frame_buffer[frame]; in Compute()
|
/external/webrtc/talk/app/webrtc/java/jni/ |
D | androidmediadecoder_jni.cc | 625 rtc::scoped_refptr<webrtc::VideoFrameBuffer> frame_buffer; in DeliverPendingOutputs() local 652 frame_buffer = surface_texture_helper_->CreateTextureFrame( in DeliverPendingOutputs() 689 frame_buffer = decoded_frame_pool_.CreateBuffer(width, height); in DeliverPendingOutputs() 701 frame_buffer->MutableData(webrtc::kYPlane), in DeliverPendingOutputs() 702 frame_buffer->stride(webrtc::kYPlane), in DeliverPendingOutputs() 703 frame_buffer->MutableData(webrtc::kUPlane), in DeliverPendingOutputs() 704 frame_buffer->stride(webrtc::kUPlane), in DeliverPendingOutputs() 705 frame_buffer->MutableData(webrtc::kVPlane), in DeliverPendingOutputs() 706 frame_buffer->stride(webrtc::kVPlane), in DeliverPendingOutputs() 715 frame_buffer->MutableData(webrtc::kYPlane), in DeliverPendingOutputs() [all …]
|
/external/webrtc/webrtc/modules/video_coding/ |
D | session_info.h | 50 uint8_t* frame_buffer, 59 size_t BuildVP8FragmentationHeader(uint8_t* frame_buffer, 119 size_t InsertBuffer(uint8_t* frame_buffer, PacketIterator packetIterator); 123 uint8_t* frame_buffer);
|
D | session_info.cc | 157 size_t VCMSessionInfo::InsertBuffer(uint8_t* frame_buffer, in InsertBuffer() argument 170 packet.dataPtr = frame_buffer + offset; in InsertBuffer() 191 uint8_t* frame_buffer_ptr = frame_buffer + offset; in InsertBuffer() 215 uint8_t* frame_buffer) { in Insert() argument 218 memcpy(frame_buffer, startCode, kH264StartCodeLengthBytes); in Insert() 220 memcpy(frame_buffer + (insert_start_code ? kH264StartCodeLengthBytes : 0), in Insert() 334 uint8_t* frame_buffer, in BuildVP8FragmentationHeader() argument 351 (*it).dataPtr - frame_buffer; in BuildVP8FragmentationHeader() 471 uint8_t* frame_buffer, in InsertPacket() argument 546 size_t returnLength = InsertBuffer(frame_buffer, packet_list_it); in InsertPacket()
|
D | video_coding.gypi | 35 'frame_buffer.h', 62 'frame_buffer.cc',
|
D | jitter_buffer.cc | 615 VCMFrameBuffer* frame_buffer = static_cast<VCMFrameBuffer*>(frame); in ReleaseFrame() local 616 if (frame_buffer) { in ReleaseFrame() 617 free_frames_.push_back(frame_buffer); in ReleaseFrame() 658 const VCMFrameBuffer* frame_buffer = in LastPacketTime() local 660 *retransmitted = (frame_buffer->GetNackCount() > 0); in LastPacketTime() 661 return frame_buffer->LatestPacketTimeMs(); in LastPacketTime()
|
D | BUILD.gn | 24 "frame_buffer.cc", 25 "frame_buffer.h",
|
/external/webrtc/webrtc/test/testsupport/ |
D | frame_writer.cc | 53 bool FrameWriterImpl::WriteFrame(uint8_t* frame_buffer) { in WriteFrame() argument 54 assert(frame_buffer); in WriteFrame() 59 size_t bytes_written = fwrite(frame_buffer, 1, frame_length_in_bytes_, in WriteFrame()
|
D | frame_writer.h | 35 virtual bool WriteFrame(uint8_t* frame_buffer) = 0; 57 bool WriteFrame(uint8_t* frame_buffer) override;
|
/external/webrtc/talk/media/base/ |
D | yuvframegenerator.cc | 78 void YuvFrameGenerator::GenerateNextFrame(uint8_t* frame_buffer, in GenerateNextFrame() argument 98 memcpy(frame_buffer, y_data_, size); in GenerateNextFrame() 99 frame_buffer += size; in GenerateNextFrame() 100 memcpy(frame_buffer, u_data_, qsize); in GenerateNextFrame() 101 frame_buffer += qsize; in GenerateNextFrame() 102 memcpy(frame_buffer, v_data_, qsize); in GenerateNextFrame()
|
D | yuvframegenerator.h | 64 void GenerateNextFrame(uint8_t* frame_buffer, int32_t barcode_value);
|
/external/u-boot/drivers/video/ |
D | tegra.c | 33 fdt_addr_t frame_buffer; /* Address of frame buffer */ member 230 win->phys_addr = priv->frame_buffer; in setup_window() 269 priv->frame_buffer = (u32)default_lcd_base; in tegra_display_probe() 321 mmu_set_region_dcache_behaviour(priv->frame_buffer, plat->size, in tegra_lcd_probe() 330 debug("LCD frame buffer at %pa, size %x\n", &priv->frame_buffer, in tegra_lcd_probe()
|
D | omap3_dss.c | 119 writel((u32) panel_cfg->frame_buffer, &dispc->gfx_ba0); in omap3_dss_panel_config() 121 if (!panel_cfg->frame_buffer) in omap3_dss_panel_config()
|
/external/libaom/libaom/av1/common/ |
D | thread_common.c | 140 YV12_BUFFER_CONFIG *frame_buffer, in loop_filter_data_reset() argument 143 lf_data->frame_buffer = frame_buffer; in loop_filter_data_reset() 267 const YV12_BUFFER_CONFIG *const frame_buffer, AV1_COMMON *const cm, in thread_loop_filter_rows() argument 288 av1_setup_dst_planes(planes, cm->seq_params.sb_size, frame_buffer, in thread_loop_filter_rows() 307 av1_setup_dst_planes(planes, cm->seq_params.sb_size, frame_buffer, in thread_loop_filter_rows() 323 thread_loop_filter_rows(lf_data->frame_buffer, lf_data->cm, lf_data->planes, in loop_filter_row_worker() 330 const YV12_BUFFER_CONFIG *const frame_buffer, AV1_COMMON *const cm, in thread_loop_filter_bitmask_rows() argument 352 av1_setup_dst_planes(planes, BLOCK_64X64, frame_buffer, mi_row, in thread_loop_filter_bitmask_rows() 371 av1_setup_dst_planes(planes, BLOCK_64X64, frame_buffer, mi_row, in thread_loop_filter_bitmask_rows() 387 thread_loop_filter_bitmask_rows(lf_data->frame_buffer, lf_data->cm, in loop_filter_bitmask_row_worker()
|
D | av1_loopfilter.c | 2352 static void loop_filter_rows(YV12_BUFFER_CONFIG *frame_buffer, AV1_COMMON *cm, in loop_filter_rows() argument 2375 av1_setup_dst_planes(pd, cm->seq_params.sb_size, frame_buffer, 0, 0, in loop_filter_rows() 2384 av1_setup_dst_planes(pd, BLOCK_64X64, frame_buffer, mi_row, mi_col, in loop_filter_rows() 2389 av1_setup_dst_planes(pd, BLOCK_64X64, frame_buffer, mi_row, in loop_filter_rows() 2395 av1_setup_dst_planes(pd, BLOCK_64X64, frame_buffer, mi_row, in loop_filter_rows() 2418 av1_setup_dst_planes(pd, cm->seq_params.sb_size, frame_buffer, mi_row, in loop_filter_rows() 2424 av1_setup_dst_planes(pd, cm->seq_params.sb_size, frame_buffer, in loop_filter_rows() 2432 av1_setup_dst_planes(pd, cm->seq_params.sb_size, frame_buffer, mi_row, in loop_filter_rows() 2441 av1_setup_dst_planes(pd, cm->seq_params.sb_size, frame_buffer, mi_row, in loop_filter_rows() 2451 av1_setup_dst_planes(pd, cm->seq_params.sb_size, frame_buffer, mi_row, in loop_filter_rows()
|
D | av1_loopfilter.h | 151 YV12_BUFFER_CONFIG *frame_buffer; member
|
/external/webrtc/webrtc/common_video/libyuv/ |
D | scaler_unittest.cc | 360 rtc::scoped_ptr<uint8_t[]> frame_buffer(new uint8_t[src_required_size]); in ScaleSequence() local 366 if (fread(frame_buffer.get(), 1, src_required_size, source_file) != in ScaleSequence() 370 input_frame.CreateFrame(frame_buffer.get(), in ScaleSequence() 371 frame_buffer.get() + size_y, in ScaleSequence() 372 frame_buffer.get() + size_y + size_uv, in ScaleSequence()
|
/external/libvpx/libvpx/vp9/common/ |
D | vp9_loopfilter.h | 141 YV12_BUFFER_CONFIG *frame_buffer; member 151 LFWorkerData *lf_data, YV12_BUFFER_CONFIG *frame_buffer,
|
D | vp9_thread_common.c | 91 const YV12_BUFFER_CONFIG *const frame_buffer, VP9_COMMON *const cm, in thread_loop_filter_rows() argument 122 vp9_setup_dst_planes(planes, frame_buffer, mi_row, mi_col); in thread_loop_filter_rows() 151 thread_loop_filter_rows(lf_data->frame_buffer, lf_data->cm, lf_data->planes, in loop_filter_row_worker() 449 thread_loop_filter_rows(lf_data->frame_buffer, lf_data->cm, lf_data->planes, in vp9_loopfilter_rows()
|
/external/webrtc/webrtc/modules/video_coding/codecs/vp8/ |
D | vp8_sequence_coder.cc | 150 rtc::scoped_ptr<uint8_t[]> frame_buffer(new uint8_t[length]); in SequenceCoder() local 166 if (fread(frame_buffer.get(), 1, length, input_file) != length) in SequenceCoder() 169 webrtc::ConvertToI420(webrtc::kI420, frame_buffer.get(), 0, 0, width, in SequenceCoder()
|
/external/webrtc/webrtc/test/testsupport/mock/ |
D | mock_frame_writer.h | 24 MOCK_METHOD1(WriteFrame, bool(uint8_t* frame_buffer));
|
/external/u-boot/board/htkw/mcx/ |
D | mcx.c | 131 lcd_cfg.frame_buffer = fb; in board_video_init()
|
/external/webrtc/webrtc/examples/objc/AppRTCDemo/third_party/SocketRocket/ |
D | SRWebSocket.m | 1319 uint8_t *frame_buffer = (uint8_t *)[frame mutableBytes]; 1322 frame_buffer[0] = SRFinMask | opcode; 1331 frame_buffer[1] |= SRMaskMask; 1346 frame_buffer[1] |= payloadLength; 1348 frame_buffer[1] |= 126; 1349 *((uint16_t *)(frame_buffer + frame_buffer_size)) = EndianU16_BtoN((uint16_t)payloadLength); 1352 frame_buffer[1] |= 127; 1353 *((uint64_t *)(frame_buffer + frame_buffer_size)) = EndianU64_BtoN((uint64_t)payloadLength); 1359 frame_buffer[frame_buffer_size] = unmasked_payload[i]; 1363 uint8_t *mask_key = frame_buffer + frame_buffer_size; [all …]
|
/external/libxaac/decoder/ |
D | ixheaacd_process.c | 380 (VOID *)(pstr_usac_data->pstr_esbr_dec->frame_buffer[ch]); in ixheaacd_usac_process() 396 (VOID *)(pstr_usac_data->pstr_esbr_dec->frame_buffer[1]); in ixheaacd_usac_process()
|