/hardware/interfaces/automotive/sv/1.0/default/ |
D | SurroundView3dSession.cpp | 46 framesRecord.frames.svBuffers.resize(1); in SurroundView3dSession() 47 framesRecord.frames.svBuffers[0].viewId = 0; in SurroundView3dSession() 48 framesRecord.frames.svBuffers[0].hardwareBuffer.nativeHandle = new native_handle_t(); in SurroundView3dSession() 49 framesRecord.frames.svBuffers[0].hardwareBuffer.description[0] = mConfig.width; in SurroundView3dSession() 50 framesRecord.frames.svBuffers[0].hardwareBuffer.description[1] = mConfig.height; in SurroundView3dSession() 275 framesRecord.frames.timestampNs = elapsedRealtimeNano(); in generateFrames() 276 framesRecord.frames.sequenceId = sequenceId++; in generateFrames() 278 framesRecord.frames.svBuffers.resize(mViews.size()); in generateFrames() 280 framesRecord.frames.svBuffers[i].viewId = mViews[i].viewId; in generateFrames() 281 framesRecord.frames.svBuffers[i].hardwareBuffer.nativeHandle = new native_handle_t(); in generateFrames() [all …]
|
D | SurroundView2dSession.cpp | 36 framesRecord.frames.svBuffers.resize(1); in SurroundView2dSession() 37 framesRecord.frames.svBuffers[0].viewId = 0; in SurroundView2dSession() 38 framesRecord.frames.svBuffers[0].hardwareBuffer.nativeHandle = in SurroundView2dSession() 40 framesRecord.frames.svBuffers[0].hardwareBuffer.description[0] = in SurroundView2dSession() 42 framesRecord.frames.svBuffers[0].hardwareBuffer.description[1] = in SurroundView2dSession() 204 framesRecord.frames.svBuffers[0].hardwareBuffer.description[0] = in generateFrames() 206 framesRecord.frames.svBuffers[0].hardwareBuffer.description[1] = in generateFrames() 212 framesRecord.frames.timestampNs = elapsedRealtimeNano(); in generateFrames() 213 framesRecord.frames.sequenceId = sequenceId++; in generateFrames() 223 mStream->receiveFrames(framesRecord.frames); in generateFrames()
|
/hardware/interfaces/automotive/sv/1.0/default/tests/fuzzer/ |
D | AutomotiveSvV1_0Fuzzer.cpp | 121 SvFramesDesc frames; in invoke2dSessionAPI() local 122 frames.timestampNs = mFuzzedDataProvider.ConsumeIntegral<uint64_t>(); in invoke2dSessionAPI() 123 frames.sequenceId = mFuzzedDataProvider.ConsumeIntegral<uint32_t>(); in invoke2dSessionAPI() 126 frames.svBuffers.resize(numSvBuffers); in invoke2dSessionAPI() 128 frames.svBuffers[i].viewId = in invoke2dSessionAPI() 130 frames.svBuffers[i].hardwareBuffer.nativeHandle = new native_handle_t(); in invoke2dSessionAPI() 131 frames.svBuffers[i].hardwareBuffer.description[0] = in invoke2dSessionAPI() 133 frames.svBuffers[i].hardwareBuffer.description[1] = in invoke2dSessionAPI() 136 surroundView2dSession->doneWithFrames(frames); in invoke2dSessionAPI() 138 delete frames.svBuffers[i].hardwareBuffer.nativeHandle; in invoke2dSessionAPI() [all …]
|
/hardware/libhardware/modules/audio_remote_submix/tests/ |
D | remote_submix_tests.cpp | 246 uint64_t frames; in TEST_F() local 248 EXPECT_EQ(0, streamOut->get_presentation_position(streamOut, &frames, ×tamp)); in TEST_F() 249 EXPECT_EQ(uint64_t{0}, frames); in TEST_F() 250 uint64_t prevFrames = frames; in TEST_F() 253 EXPECT_EQ(0, streamOut->get_presentation_position(streamOut, &frames, ×tamp)); in TEST_F() 254 EXPECT_LE(prevFrames, frames); in TEST_F() 255 prevFrames = frames; in TEST_F() 264 uint32_t frames; in TEST_F() local 265 EXPECT_EQ(0, streamOut->get_render_position(streamOut, &frames)); in TEST_F() 266 EXPECT_EQ(0U, frames); in TEST_F() [all …]
|
/hardware/interfaces/automotive/sv/1.0/vts/functional/ |
D | VtsHalSurroundViewV1_0TargetTest.cpp | 208 SvFramesDesc frames = handler->getLastReceivedFrames(); in TEST_P() local 209 EXPECT_EQ(frames.svBuffers.size(), 1); in TEST_P() 211 SvBuffer svBuffer2d = frames.svBuffers[0]; in TEST_P() 275 SvFramesDesc frames = handler->getLastReceivedFrames(); in TEST_P() local 276 EXPECT_EQ(frames.svBuffers.size(), 1); in TEST_P() 277 SvBuffer svBuffer2d = frames.svBuffers[0]; in TEST_P() 315 SvFramesDesc frames = handler->getLastReceivedFrames(); in TEST_P() local 316 EXPECT_EQ(frames.svBuffers.size(), 1); in TEST_P() 317 SvBuffer svBuffer2d = frames.svBuffers[0]; in TEST_P() 373 SvFramesDesc frames = handler->getLastReceivedFrames(); in TEST_P() local [all …]
|
/hardware/interfaces/automotive/sv/1.0/ |
D | ISurroundViewStream.hal | 22 * This interface is to be implemented by client to receive callback for output frames and events. 26 * Receives callback of surround view 2d/3d frames. 28 * @param svFramesDesc Frames descriptor containing the output frames.
|
D | ISurroundView3dSession.hal | 40 * may continue to receive frames of previous views after setViews() call for a while and can 42 * views frames. 44 * @param views List of desired views to generate output frames. 56 * all frames received will be of the updated config. 95 * Useful for mapping points detected on individual camera frames onto the surround view 3D
|
/hardware/interfaces/audio/2.0/ |
D | IStreamIn.hal | 78 uint64_t frames; 101 * @param framesCount the number of frames in a buffer. 123 * Return the amount of input frames lost in the audio driver since the last 131 * @return framesLost the number of input audio frames lost. 136 * Return a recent count of the number of audio frames received and the 142 * @return frames the total frame count received. This must be as early in 143 * the capture pipeline as possible. In general, frames 145 * @return time is the clock monotonic time when frames was measured. In 150 generates (Result retval, uint64_t frames, uint64_t time);
|
D | IStreamOut.hal | 71 uint64_t frames; // getPresentationPosition. 95 * @param framesCount the number of frames in a buffer. 116 * Return the number of audio frames written by the audio DSP to DAC since 120 * @return dspFrames number of audio frames written. 234 * Return a recent count of the number of audio frames presented to an 235 * external observer. This excludes frames which have been written but are 242 * Example: assume that N+M frames have been presented, where M is a 'small' 248 * @return frames count of presented audio frames. 252 generates (Result retval, uint64_t frames, TimeSpec timeStamp);
|
/hardware/interfaces/audio/aidl/default/stub/ |
D | StreamMmapStub.cpp | 41 mSharedState.mmapPos.frames = StreamDescriptor::Position::UNKNOWN; in init() 65 if (mSharedState.mmapPos.frames == StreamDescriptor::Position::UNKNOWN) { in cycle() 66 mSharedState.mmapPos.frames = mLastFrames; in cycle() 68 mSharedState.mmapPos.frames += newFrames; in cycle() 69 mLastFrames = mSharedState.mmapPos.frames; in cycle() 80 mSharedState.mmapPos.frames = StreamDescriptor::Position::UNKNOWN; in cycle() 198 if (position->frames != StreamDescriptor::Position::UNKNOWN) { in getMmapPositionAndLatency() 199 position->frames += latencyFrames; in getMmapPositionAndLatency()
|
/hardware/interfaces/automotive/evs/1.1/ |
D | IEvsCameraStream.hal | 29 * Receives calls from the HAL each time video frames is ready for inspection. 38 * A camera device will deliver the same number of frames as number of 41 * frames as many as number of backing physical camera devices. 43 * @param buffer Buffer descriptors of delivered image frames.
|
D | IEvsUltrasonicsArray.hal | 40 * Up to this many data frames may be held concurrently by the client of IEvsUltrasonicsArray. 41 * If this many frames have been delivered to the receiver without being returned 42 * by doneWithFrame, the stream must skip frames until a buffer is returned for reuse. 65 * Requests to stop the delivery of the ultrasonic array data frames. 67 * Because delivery is asynchronous, frames may continue to arrive for
|
/hardware/interfaces/audio/4.0/ |
D | IStreamIn.hal | 78 uint64_t frames; 107 * @param framesCount the number of frames in a buffer. 129 * Return the amount of input frames lost in the audio driver since the last 137 * @return framesLost the number of input audio frames lost. 142 * Return a recent count of the number of audio frames received and the 148 * @return frames the total frame count received. This must be as early in 149 * the capture pipeline as possible. In general, frames 151 * @return time is the clock monotonic time when frames was measured. In 156 generates (Result retval, uint64_t frames, uint64_t time);
|
D | IStreamOut.hal | 71 uint64_t frames; // getPresentationPosition. 101 * @param framesCount the number of frames in a buffer. 122 * Return the number of audio frames written by the audio DSP to DAC since 127 * @return dspFrames number of audio frames written. 246 * Return a recent count of the number of audio frames presented to an 247 * external observer. This excludes frames which have been written but are 254 * Example: assume that N+M frames have been presented, where M is a 'small' 262 * @return frames count of presented audio frames. 266 generates (Result retval, uint64_t frames, TimeSpec timeStamp);
|
/hardware/interfaces/automotive/evs/1.0/ |
D | IEvsCamera.hal | 38 * Up to this many frames may be held concurrently by the client of IEvsCamera. 39 * If this many frames have been delivered to the receiver without being returned 40 * by doneWithFrame, the stream must skip frames until a buffer is returned for reuse. 68 * as one), and if the supply is exhausted, no further frames may be 76 * Stop the delivery of EVS camera frames. 78 * Because delivery is asynchronous, frames may continue to arrive for
|
/hardware/interfaces/audio/6.0/ |
D | IStreamIn.hal | 78 uint64_t frames; 107 * @param framesCount the number of frames in a buffer. 129 * Return the amount of input frames lost in the audio driver since the last 137 * @return framesLost the number of input audio frames lost. 142 * Return a recent count of the number of audio frames received and the 148 * @return frames the total frame count received. This must be as early in 149 * the capture pipeline as possible. In general, frames 151 * @return time is the clock monotonic time when frames was measured. In 156 generates (Result retval, uint64_t frames, uint64_t time);
|
/hardware/interfaces/audio/5.0/ |
D | IStreamIn.hal | 78 uint64_t frames; 107 * @param framesCount the number of frames in a buffer. 129 * Return the amount of input frames lost in the audio driver since the last 137 * @return framesLost the number of input audio frames lost. 142 * Return a recent count of the number of audio frames received and the 148 * @return frames the total frame count received. This must be as early in 149 * the capture pipeline as possible. In general, frames 151 * @return time is the clock monotonic time when frames was measured. In 156 generates (Result retval, uint64_t frames, uint64_t time);
|
D | IStreamOut.hal | 71 uint64_t frames; // getPresentationPosition. 101 * @param framesCount the number of frames in a buffer. 122 * Return the number of audio frames written by the audio DSP to DAC since 127 * @return dspFrames number of audio frames written. 246 * Return a recent count of the number of audio frames presented to an 247 * external observer. This excludes frames which have been written but are 254 * Example: assume that N+M frames have been presented, where M is a 'small' 262 * @return frames count of presented audio frames. 266 generates (Result retval, uint64_t frames, TimeSpec timeStamp);
|
/hardware/interfaces/audio/7.0/ |
D | IStreamIn.hal | 93 uint64_t frames; 116 * @param framesCount the number of frames in a buffer. 140 * Return the amount of input frames lost in the audio driver since the last 148 * @return framesLost the number of input audio frames lost. 153 * Return a recent count of the number of audio frames received and the 160 * @return frames the total frame count received. This must be as early in 161 * the capture pipeline as possible. In general, frames 163 * @return time is the clock monotonic time when frames was measured. In 168 generates (Result retval, uint64_t frames, uint64_t time);
|
/hardware/interfaces/input/classifier/1.0/vts/functional/ |
D | VtsHalInputClassifierV1_0TargetTest.cpp | 56 event.frames = {}; in getSimpleMotionEvent() 132 event.frames = {frame}; in TEST_P() 161 event.frames = {frame1, frame2}; in TEST_P()
|
/hardware/interfaces/audio/core/all-versions/vts/functional/7.0/ |
D | AudioPrimaryHidlHalTest.cpp | 585 uint64_t frames = framesInitial; in waitForPresentationPositionAdvance() local 587 frames <= framesInitial && !writer.hasError() && in waitForPresentationPositionAdvance() 590 ASSERT_OK(stream->getPresentationPosition(returnIn(res, frames, ts))); in waitForPresentationPositionAdvance() 595 EXPECT_GT(frames, framesInitial); in waitForPresentationPositionAdvance() 597 if (lastPosition) *lastPosition = frames; in waitForPresentationPositionAdvance() 645 uint64_t frames = std::numeric_limits<uint64_t>::max(); in TEST_P() local 646 ASSERT_NO_FATAL_FAILURE(waitForPresentationPositionAdvance(writer, &frames)); in TEST_P() 647 EXPECT_GT(frames, framesInitial); in TEST_P() 745 uint64_t frames = framesInitial; in waitForCapturePositionAdvance() local 747 frames <= framesInitial && !reader.hasError() && in waitForCapturePositionAdvance() [all …]
|
/hardware/libhardware/modules/audio_remote_submix/ |
D | audio_hw.cpp | 744 const size_t frames = bytes / frame_size; in out_write() local 758 usleep(frames * 1000000 / out_get_sample_rate(&stream->common)); in out_write() 761 out->frames_written += frames; in out_write() 762 out->frames_written_since_standby += frames; in out_write() 788 if (dont_block && availableToWrite < frames) { in out_write() 791 size_t frames_to_flush_from_source = frames - availableToWrite; in out_write() 805 written_frames = sink->write(buffer, frames); in out_write() 824 written_frames = sink->write(buffer, frames); in out_write() 846 uint64_t *frames, struct timespec *timestamp) in out_get_presentation_position() argument 848 if (stream == NULL || frames == NULL || timestamp == NULL) { in out_get_presentation_position() [all …]
|
/hardware/google/interfaces/media/c2/1.0/ |
D | IComponentListener.hal | 64 * Notify the listener that frames are rendered. 66 * @param renderedFrames List of information about renderings of frames.
|
/hardware/interfaces/wifi/1.4/ |
D | types.hal | 161 … * Specifies whether SDF (service discovery frames) are transmitted in the specified band. Indexed 314 * Num of frames per burst. 316 * For 2-sided this equals the number of FTM frames 318 * equals the number of FTM frames that the 363 * RTT preamble to be used in the RTT frames. 368 * RTT BW to be used in the RTT frames. 476 * Total RTT measurement frames attempted. 481 * Total successful RTT measurement frames. 486 * Maximum number of "FTM frames per burst" supported by 491 * many frames as originally requested.
|
/hardware/interfaces/camera/device/1.0/ |
D | ICameraDevice.hal | 187 /** Set the callback interface through which preview frames are sent */ 203 * release video recording frames sent out by the cameral HAL before and 280 * responsibility to manage the life-cycle of the video recording frames, 281 * and the client must not modify/access any video recording frames. 304 * frames sent out by the camera HAL before the camera HAL receives a call 307 * responsibility to manage the life-cycle of the video recording frames. 319 * frames sent out by the camera HAL before the camera HAL receives a call 322 * responsibility to manage the life-cycle of the video recording frames. 331 * Release a batch of record frames previously returned by CAMERA_MSG_VIDEO_FRAME 335 * frames sent out by the camera HAL before the camera HAL receives a call [all …]
|