/external/libvpx/tools/ |
D | tiny_ssim.c | 327 size_t i, n_frames = 0, allocated_frames = 0; in main() local 457 if (n_frames == allocated_frames) { in main() 470 psnr_and_ssim(ssimy[n_frames], psnry[n_frames], y[0], y[1], w, h); in main() 471 psnr_and_ssim(ssimu[n_frames], psnru[n_frames], u[0], u[1], (w + 1) / 2, in main() 473 psnr_and_ssim(ssimv[n_frames], psnrv[n_frames], v[0], v[1], (w + 1) / 2, in main() 476 n_frames++; in main() 484 for (i = 0; i < n_frames; ++i) { in main() 516 ssimavg /= n_frames; in main() 517 ssimyavg /= n_frames; in main() 518 ssimuavg /= n_frames; in main() [all …]
|
/external/v4l-utils/contrib/test/ |
D | v4l2grab.c | 539 struct v4l2_format *fmt, int n_frames, in read_capture_loop() argument 550 for (i = 0; i < n_frames; i++) { in read_capture_loop() 602 static int read_capture(int fd, int n_frames, unsigned char *out_buf, char *out_dir, in read_capture() argument 612 return read_capture_loop(fd, buffers, fmt, n_frames, out_buf, out_dir); in read_capture() 620 int n_frames, unsigned char *out_buf, char *out_dir) in userptr_capture_loop() argument 631 for (i = 0; i < n_frames; i++) { in userptr_capture_loop() 680 if (i + 1 < n_frames) in userptr_capture_loop() 686 static int userptr_capture(int fd, int n_frames, unsigned char *out_buf, char *out_dir, in userptr_capture() argument 736 n_frames, out_buf, out_dir); in userptr_capture() 754 int n_frames; member [all …]
|
D | v4l2gl.c | 80 static int capture(char *dev_name, int x_res, int y_res, int n_frames, in capture() argument 212 while (i < n_frames || n_frames <= 0) { in capture() 286 static int n_frames = -1; variable 312 n_frames = val; in parse_opt() 331 return capture(dev_name, x_res, y_res, n_frames, out_dir); in main()
|
/external/deqp/external/vulkancts/data/vulkan/video/ |
D | frame_checksums.py | 71 n_frames = file_size // frame_size variable 74 print("Computing checksums for ", n_frames, " frames") 75 for frame in range(n_frames):
|
/external/libvpx/vp9/encoder/ |
D | vp9_mbgraph.c | 286 int n_frames = cpi->mbgraph_n_frames; in separate_arf_mbs() local 295 if (n_frames > cpi->rc.frames_till_gf_update_due) in separate_arf_mbs() 296 n_frames = cpi->rc.frames_till_gf_update_due; in separate_arf_mbs() 299 for (i = n_frames - 1; i >= 0; i--) { in separate_arf_mbs() 353 int i, n_frames = vp9_lookahead_depth(cpi->lookahead); in vp9_update_mbgraph_stats() local 360 if (n_frames <= cpi->rc.frames_till_gf_update_due) return; in vp9_update_mbgraph_stats() 362 if (n_frames > MAX_LAG_BUFFERS) n_frames = MAX_LAG_BUFFERS; in vp9_update_mbgraph_stats() 364 cpi->mbgraph_n_frames = n_frames; in vp9_update_mbgraph_stats() 365 for (i = 0; i < n_frames; i++) { in vp9_update_mbgraph_stats() 375 for (i = 0; i < n_frames; i++) { in vp9_update_mbgraph_stats()
|
/external/autotest/client/cros/audio/ |
D | audio_data.py | 100 n_frames = len(np_array) // self.channel 102 np_array = np_array.reshape(n_frames, self.channel)
|
/external/libvpx/test/ |
D | cq_test.cc | 93 int n_frames() const { return n_frames_; } in n_frames() function in __anon0ac915e10111::CQTest 122 static_cast<unsigned int>(file_size()) * 8 * 30 / (n_frames() * 1000); in TEST_P()
|
/external/libxaac/encoder/ |
D | ixheaace_signal_classifier.c | 961 WORD32 n_frames, n_class, avg_cls, nf; in iusace_classification() local 966 n_frames = pstr_sig_class->n_buffer_samples / ccfl; in iusace_classification() 968 for (nf = 0; nf < n_frames; nf++) { in iusace_classification() 989 pstr_sig_class->n_buf_class += n_frames; in iusace_classification() 1027 pstr_sig_class->n_buffer_samples -= ccfl * n_frames; in iusace_classification() 1033 pstr_sig_class->input_samples[i] = pstr_sig_class->input_samples[i + ccfl * n_frames]; in iusace_classification() 1038 pstr_sig_class->input_samples[i] = pstr_sig_class->input_samples[i + ccfl * n_frames]; in iusace_classification() 1043 pstr_sig_class->input_samples[i] = pstr_sig_class->input_samples[i + ccfl * n_frames]; in iusace_classification()
|
/external/mp4parser/isoparser/src/main/java/com/googlecode/mp4parser/authoring/tracks/ |
D | H264TrackImpl.java | 283 if (seiMessage.n_frames == 0) { in readSamples() 288 offset = seiMessage.n_frames - frameNrInGop; in readSamples() 578 int n_frames; field in H264TrackImpl.SEIMessage 658 n_frames = reader.readU(8, "pic_timing SEI: n_frames"); in SEIMessage() 727 ", n_frames=" + n_frames + in toString()
|
/external/mesa3d/src/vulkan/overlay-layer/ |
D | overlay.cpp | 189 uint64_t n_frames; member 815 uint32_t f_idx = data->n_frames % ARRAY_SIZE(data->frames_stats); in snapshot_swapchain_frame() 896 data->n_frames++; in snapshot_swapchain_frame() 903 if ((ARRAY_SIZE(data->frames_stats) - _idx) > data->n_frames) in get_time_stat() 906 data->n_frames < ARRAY_SIZE(data->frames_stats) ? in get_time_stat() 907 _idx - data->n_frames : in get_time_stat() 908 _idx + data->n_frames; in get_time_stat() 917 if ((ARRAY_SIZE(data->frames_stats) - _idx) > data->n_frames) in get_stat() 920 data->n_frames < ARRAY_SIZE(data->frames_stats) ? in get_stat() 921 _idx - data->n_frames : in get_stat() [all …]
|
/external/pytorch/aten/src/ATen/native/ |
D | SpectralOps.cpp | 946 int64_t n_frames = 1 + (len - n_fft) / hop_length; in stft() local 949 {batch, n_frames, n_fft}, in stft() 1057 const auto n_frames = input.size(-2); in istft() local 1060 const auto expected_output_signal_len = n_fft + hop_length * (n_frames - 1); in istft() 1149 window_tmp = window_tmp.pow(2).expand({1, n_frames, n_fft}); // size: (1, n_frames, n_fft) in istft()
|
/external/libdav1d/src/ |
D | thread_task.c | 83 unsigned frame_idx, unsigned n_frames) in reset_task_cur_async() argument 86 if (frame_idx < first) frame_idx += n_frames; in reset_task_cur_async()
|
/external/pytorch/test/ |
D | test_spectral_ops.py | 1381 n_frames = stft.size(-1) 1383 … len_expected = stft_kwargs["n_fft"] // 2 + stft_kwargs["hop_length"] * (n_frames - 1) 1385 len_expected = stft_kwargs["n_fft"] + stft_kwargs["hop_length"] * (n_frames - 1)
|
/external/libaom/av1/encoder/ |
D | encode_strategy.c | 947 static int is_in_ref_map(RefBufMapData *map, int disp_order, int n_frames) { argument 948 for (int i = 0; i < n_frames; i++) {
|
/external/python/cpython3/Lib/test/ |
D | pickletester.py | 3311 n_frames = count_opcode(pickle.FRAME, pickled) 3314 self.assertEqual(n_frames, 1) 3398 n_frames = (len(pickled) - 1) // self.FRAME_SIZE_TARGET + 1 3400 self.assertGreaterEqual(len(writer.chunks), n_frames) 3405 self.assertLessEqual(len(writer.chunks), 2 * n_frames + 3)
|
/external/deqp/external/vulkancts/modules/vulkan/video/ |
D | vktVideoTestUtils.cpp | 2289 void generateYCbCrFile(std::string fileName, uint32_t n_frames, uint32_t width, uint32_t height, ui… in generateYCbCrFile() argument 2302 max_frames = n_frames; in generateYCbCrFile()
|
D | vktVideoTestUtils.hpp | 1273 void generateYCbCrFile(std::string fileName, uint32_t n_frames, uint32_t width, uint32_t height, ui…
|
/external/flashrom/ |
D | cros_ec_commands.h | 4342 uint32_t n_frames; member
|
/external/pytorch/torch/_refs/ |
D | __init__.py | 3448 n_frames = input.size(-1) 3451 expected_output_signal_len = n_fft + hop_length_ * (n_frames - 1) 3530 window_.pow(2).expand((1, n_frames, n_fft)),
|
/external/coreboot/src/ec/google/chromeec/ |
D | ec_commands.h | 8422 uint32_t n_frames; member
|