/hardware/interfaces/automotive/sv/1.0/default/ |
D | SurroundView3dSession.cpp | 46 framesRecord.frames.svBuffers.resize(1); in SurroundView3dSession() 47 framesRecord.frames.svBuffers[0].viewId = 0; in SurroundView3dSession() 48 framesRecord.frames.svBuffers[0].hardwareBuffer.nativeHandle = new native_handle_t(); in SurroundView3dSession() 49 framesRecord.frames.svBuffers[0].hardwareBuffer.description[0] = mConfig.width; in SurroundView3dSession() 50 framesRecord.frames.svBuffers[0].hardwareBuffer.description[1] = mConfig.height; in SurroundView3dSession() 275 framesRecord.frames.timestampNs = elapsedRealtimeNano(); in generateFrames() 276 framesRecord.frames.sequenceId = sequenceId++; in generateFrames() 278 framesRecord.frames.svBuffers.resize(mViews.size()); in generateFrames() 280 framesRecord.frames.svBuffers[i].viewId = mViews[i].viewId; in generateFrames() 281 framesRecord.frames.svBuffers[i].hardwareBuffer.nativeHandle = new native_handle_t(); in generateFrames() [all …]
|
D | SurroundView2dSession.cpp | 36 framesRecord.frames.svBuffers.resize(1); in SurroundView2dSession() 37 framesRecord.frames.svBuffers[0].viewId = 0; in SurroundView2dSession() 38 framesRecord.frames.svBuffers[0].hardwareBuffer.nativeHandle = in SurroundView2dSession() 40 framesRecord.frames.svBuffers[0].hardwareBuffer.description[0] = in SurroundView2dSession() 42 framesRecord.frames.svBuffers[0].hardwareBuffer.description[1] = in SurroundView2dSession() 204 framesRecord.frames.svBuffers[0].hardwareBuffer.description[0] = in generateFrames() 206 framesRecord.frames.svBuffers[0].hardwareBuffer.description[1] = in generateFrames() 212 framesRecord.frames.timestampNs = elapsedRealtimeNano(); in generateFrames() 213 framesRecord.frames.sequenceId = sequenceId++; in generateFrames() 223 mStream->receiveFrames(framesRecord.frames); in generateFrames()
|
/hardware/interfaces/automotive/sv/1.0/default/tests/fuzzer/ |
D | AutomotiveSvV1_0Fuzzer.cpp | 121 SvFramesDesc frames; in invoke2dSessionAPI() local 122 frames.timestampNs = mFuzzedDataProvider.ConsumeIntegral<uint64_t>(); in invoke2dSessionAPI() 123 frames.sequenceId = mFuzzedDataProvider.ConsumeIntegral<uint32_t>(); in invoke2dSessionAPI() 126 frames.svBuffers.resize(numSvBuffers); in invoke2dSessionAPI() 128 frames.svBuffers[i].viewId = in invoke2dSessionAPI() 130 frames.svBuffers[i].hardwareBuffer.nativeHandle = new native_handle_t(); in invoke2dSessionAPI() 131 frames.svBuffers[i].hardwareBuffer.description[0] = in invoke2dSessionAPI() 133 frames.svBuffers[i].hardwareBuffer.description[1] = in invoke2dSessionAPI() 136 surroundView2dSession->doneWithFrames(frames); in invoke2dSessionAPI() 138 delete frames.svBuffers[i].hardwareBuffer.nativeHandle; in invoke2dSessionAPI() [all …]
|
/hardware/libhardware/modules/audio_remote_submix/tests/ |
D | remote_submix_tests.cpp | 246 uint64_t frames; in TEST_F() local 248 EXPECT_EQ(0, streamOut->get_presentation_position(streamOut, &frames, ×tamp)); in TEST_F() 249 EXPECT_EQ(uint64_t{0}, frames); in TEST_F() 250 uint64_t prevFrames = frames; in TEST_F() 253 EXPECT_EQ(0, streamOut->get_presentation_position(streamOut, &frames, ×tamp)); in TEST_F() 254 EXPECT_LE(prevFrames, frames); in TEST_F() 255 prevFrames = frames; in TEST_F() 264 uint32_t frames; in TEST_F() local 265 EXPECT_EQ(0, streamOut->get_render_position(streamOut, &frames)); in TEST_F() 266 EXPECT_EQ(0U, frames); in TEST_F() [all …]
|
/hardware/interfaces/automotive/sv/1.0/vts/functional/ |
D | VtsHalSurroundViewV1_0TargetTest.cpp | 208 SvFramesDesc frames = handler->getLastReceivedFrames(); in TEST_P() local 209 EXPECT_EQ(frames.svBuffers.size(), 1); in TEST_P() 211 SvBuffer svBuffer2d = frames.svBuffers[0]; in TEST_P() 275 SvFramesDesc frames = handler->getLastReceivedFrames(); in TEST_P() local 276 EXPECT_EQ(frames.svBuffers.size(), 1); in TEST_P() 277 SvBuffer svBuffer2d = frames.svBuffers[0]; in TEST_P() 315 SvFramesDesc frames = handler->getLastReceivedFrames(); in TEST_P() local 316 EXPECT_EQ(frames.svBuffers.size(), 1); in TEST_P() 317 SvBuffer svBuffer2d = frames.svBuffers[0]; in TEST_P() 373 SvFramesDesc frames = handler->getLastReceivedFrames(); in TEST_P() local [all …]
|
/hardware/qcom/audio/legacy/libalsa-intf/ |
D | arec.c | 185 long frames; in record_file() local 269 frames = bufsize / 2; in record_file() 271 frames = bufsize / 8; in record_file() 273 frames = bufsize / 12; in record_file() 275 frames = bufsize / 4; in record_file() 277 x.frames = frames; in record_file() 300 fprintf(stderr, "Arec:avail 1 = %d frames = %ld\n",avail, frames); in record_file() 307 if (x.frames > avail) in record_file() 308 frames = avail; in record_file() 324 x.frames -= frames; in record_file() [all …]
|
D | alsa_pcm.c | 468 long frames) in mmap_transfer() argument 478 frames = frames * channels *2 ; in mmap_transfer() 480 while (frames-- > 0) { in mmap_transfer() 489 long frames) in mmap_transfer_capture() argument 502 frames = frames * channels *2 ; in mmap_transfer_capture() 504 while (frames-- > 0) { in mmap_transfer_capture() 524 long frames; in pcm_write_mmap() local 528 frames = (pcm->flags & PCM_MONO) ? (count / 2) : (count / 4); in pcm_write_mmap() 539 pcm->sync_ptr->c.control.appl_ptr += frames; in pcm_write_mmap() 580 x.frames = (count / (channels * 2)) ; in pcm_write_nmmap() [all …]
|
/hardware/interfaces/audio/2.0/ |
D | IStreamIn.hal | 78 uint64_t frames; 101 * @param framesCount the number of frames in a buffer. 123 * Return the amount of input frames lost in the audio driver since the last 131 * @return framesLost the number of input audio frames lost. 136 * Return a recent count of the number of audio frames received and the 142 * @return frames the total frame count received. This must be as early in 143 * the capture pipeline as possible. In general, frames 145 * @return time is the clock monotonic time when frames was measured. In 150 generates (Result retval, uint64_t frames, uint64_t time);
|
D | IStreamOut.hal | 71 uint64_t frames; // getPresentationPosition. 95 * @param framesCount the number of frames in a buffer. 116 * Return the number of audio frames written by the audio DSP to DAC since 120 * @return dspFrames number of audio frames written. 234 * Return a recent count of the number of audio frames presented to an 235 * external observer. This excludes frames which have been written but are 242 * Example: assume that N+M frames have been presented, where M is a 'small' 248 * @return frames count of presented audio frames. 252 generates (Result retval, uint64_t frames, TimeSpec timeStamp);
|
/hardware/interfaces/automotive/sv/1.0/ |
D | ISurroundViewStream.hal | 22 * This interface is to be implemented by client to receive callback for output frames and events. 26 * Receives callback of surround view 2d/3d frames. 28 * @param svFramesDesc Frames descriptor containing the output frames.
|
/hardware/interfaces/audio/4.0/ |
D | IStreamIn.hal | 78 uint64_t frames; 107 * @param framesCount the number of frames in a buffer. 129 * Return the amount of input frames lost in the audio driver since the last 137 * @return framesLost the number of input audio frames lost. 142 * Return a recent count of the number of audio frames received and the 148 * @return frames the total frame count received. This must be as early in 149 * the capture pipeline as possible. In general, frames 151 * @return time is the clock monotonic time when frames was measured. In 156 generates (Result retval, uint64_t frames, uint64_t time);
|
D | IStreamOut.hal | 71 uint64_t frames; // getPresentationPosition. 101 * @param framesCount the number of frames in a buffer. 122 * Return the number of audio frames written by the audio DSP to DAC since 127 * @return dspFrames number of audio frames written. 246 * Return a recent count of the number of audio frames presented to an 247 * external observer. This excludes frames which have been written but are 254 * Example: assume that N+M frames have been presented, where M is a 'small' 262 * @return frames count of presented audio frames. 266 generates (Result retval, uint64_t frames, TimeSpec timeStamp);
|
/hardware/interfaces/automotive/evs/1.1/ |
D | IEvsCameraStream.hal | 29 * Receives calls from the HAL each time video frames is ready for inspection. 38 * A camera device will deliver the same number of frames as number of 41 * frames as many as number of backing physical camera devices. 43 * @param buffer Buffer descriptors of delivered image frames.
|
D | IEvsUltrasonicsArray.hal | 40 * Up to this many data frames may be held concurrently by the client of IEvsUltrasonicsArray. 41 * If this many frames have been delivered to the receiver without being returned 42 * by doneWithFrame, the stream must skip frames until a buffer is returned for reuse. 65 * Requests to stop the delivery of the ultrasonic array data frames. 67 * Because delivery is asynchronous, frames may continue to arrive for
|
/hardware/qcom/audio/legacy/alsa_sound/ |
D | AudioUsbALSA.cpp | 383 long frames; in RecordingThreadEntry() local 432 frames = (musbRecordingHandle->flags & PCM_MONO) ? (bufsize / 2) : (bufsize / 4); in RecordingThreadEntry() 433 x.frames = (musbRecordingHandle->flags & PCM_MONO) ? (bufsize / 2) : (bufsize / 4); in RecordingThreadEntry() 492 if (x.frames > avail) in RecordingThreadEntry() 493 frames = avail; in RecordingThreadEntry() 537 musbRecordingHandle->sync_ptr->c.control.appl_ptr += frames; in RecordingThreadEntry() 549 mproxyRecordingHandle->sync_ptr->c.control.appl_ptr += frames; in RecordingThreadEntry() 772 long frames; in PlaybackThreadEntry() local 839 frames = (mproxyPlaybackHandle->flags & PCM_MONO) ? (proxyPeriod / 2) : (proxyPeriod / 4); in PlaybackThreadEntry() 840 x.frames = (mproxyPlaybackHandle->flags & PCM_MONO) ? (proxyPeriod / 2) : (proxyPeriod / 4); in PlaybackThreadEntry() [all …]
|
/hardware/google/gfxstream/codegen/vulkan/vulkan-docs-next/proposals/ |
D | VK_EXT_frame_boundary.adoc | 9 debuggers) to group queue submissions per frames in non-trivial scenarios, 22 tools where the frames start and end. 44 ### Overlapping frames 47 different frames is being submitted in an interleaved way, a scenario we call 48 here "`overlapping frames`". 64 the erroneous grouping of queue submissions of different frames as the work of a 72 command buffers may be used concurrently by several frames, so that is not a 74 satisfactory since they cannot handle overlapping frames. 84 We want applications to be able to group queue submissions by frames. To this 144 // frameID is necessary to disambiguate overlapping frames [all …]
|
/hardware/interfaces/audio/5.0/ |
D | IStreamIn.hal | 78 uint64_t frames; 107 * @param framesCount the number of frames in a buffer. 129 * Return the amount of input frames lost in the audio driver since the last 137 * @return framesLost the number of input audio frames lost. 142 * Return a recent count of the number of audio frames received and the 148 * @return frames the total frame count received. This must be as early in 149 * the capture pipeline as possible. In general, frames 151 * @return time is the clock monotonic time when frames was measured. In 156 generates (Result retval, uint64_t frames, uint64_t time);
|
/hardware/interfaces/audio/6.0/ |
D | IStreamIn.hal | 78 uint64_t frames; 107 * @param framesCount the number of frames in a buffer. 129 * Return the amount of input frames lost in the audio driver since the last 137 * @return framesLost the number of input audio frames lost. 142 * Return a recent count of the number of audio frames received and the 148 * @return frames the total frame count received. This must be as early in 149 * the capture pipeline as possible. In general, frames 151 * @return time is the clock monotonic time when frames was measured. In 156 generates (Result retval, uint64_t frames, uint64_t time);
|
/hardware/interfaces/automotive/evs/1.0/ |
D | IEvsCamera.hal | 38 * Up to this many frames may be held concurrently by the client of IEvsCamera. 39 * If this many frames have been delivered to the receiver without being returned 40 * by doneWithFrame, the stream must skip frames until a buffer is returned for reuse. 68 * as one), and if the supply is exhausted, no further frames may be 76 * Stop the delivery of EVS camera frames. 78 * Because delivery is asynchronous, frames may continue to arrive for
|
/hardware/google/gfxstream/guest/mesa/src/util/tests/ |
D | u_debug_stack_test.cpp | 53 func_c(struct debug_stack_frame *frames) in func_c() argument 55 debug_backtrace_capture(frames, 0, 16); in func_c()
|
/hardware/google/gchips/include/libaudio/audiohal/ |
D | audio_proxy_interface.h | 96 int proxy_get_render_position(void *proxy_stream, uint32_t *frames); 97 int proxy_get_presen_position(void *proxy_stream, uint64_t *frames, struct timespec *timestamp); 112 int __proxy_get_capture_position(void *proxy_stream, int64_t *frames, int64_t *time);
|
/hardware/interfaces/audio/7.0/ |
D | IStreamIn.hal | 93 uint64_t frames; 116 * @param framesCount the number of frames in a buffer. 140 * Return the amount of input frames lost in the audio driver since the last 148 * @return framesLost the number of input audio frames lost. 153 * Return a recent count of the number of audio frames received and the 160 * @return frames the total frame count received. This must be as early in 161 * the capture pipeline as possible. In general, frames 163 * @return time is the clock monotonic time when frames was measured. In 168 generates (Result retval, uint64_t frames, uint64_t time);
|
/hardware/interfaces/audio/aidl/default/alsa/ |
D | StreamAlsa.cpp | 127 alsa::resetTransferredFrames(mAlsaDeviceProxies[0], position->frames); in refinePosition() 129 if (int ret = proxy_get_capture_position(mAlsaDeviceProxies[0].get(), &position->frames, in refinePosition() 144 position->frames = static_cast<int64_t>(hwFrames); in refinePosition()
|
/hardware/interfaces/input/classifier/1.0/vts/functional/ |
D | VtsHalInputClassifierV1_0TargetTest.cpp | 56 event.frames = {}; in getSimpleMotionEvent() 132 event.frames = {frame}; in TEST_P() 161 event.frames = {frame1, frame2}; in TEST_P()
|
/hardware/interfaces/audio/core/all-versions/vts/functional/7.0/ |
D | AudioPrimaryHidlHalTest.cpp | 585 uint64_t frames = framesInitial; in waitForPresentationPositionAdvance() local 587 frames <= framesInitial && !writer.hasError() && in waitForPresentationPositionAdvance() 590 ASSERT_OK(stream->getPresentationPosition(returnIn(res, frames, ts))); in waitForPresentationPositionAdvance() 595 EXPECT_GT(frames, framesInitial); in waitForPresentationPositionAdvance() 597 if (lastPosition) *lastPosition = frames; in waitForPresentationPositionAdvance() 645 uint64_t frames = std::numeric_limits<uint64_t>::max(); in TEST_P() local 646 ASSERT_NO_FATAL_FAILURE(waitForPresentationPositionAdvance(writer, &frames)); in TEST_P() 647 EXPECT_GT(frames, framesInitial); in TEST_P() 745 uint64_t frames = framesInitial; in waitForCapturePositionAdvance() local 747 frames <= framesInitial && !reader.hasError() && in waitForCapturePositionAdvance() [all …]
|