Home
last modified time | relevance | path

Searched refs:frames (Results 1 – 25 of 127) sorted by relevance

123456

/hardware/interfaces/automotive/sv/1.0/default/
DSurroundView3dSession.cpp46 framesRecord.frames.svBuffers.resize(1); in SurroundView3dSession()
47 framesRecord.frames.svBuffers[0].viewId = 0; in SurroundView3dSession()
48 framesRecord.frames.svBuffers[0].hardwareBuffer.nativeHandle = new native_handle_t(); in SurroundView3dSession()
49 framesRecord.frames.svBuffers[0].hardwareBuffer.description[0] = mConfig.width; in SurroundView3dSession()
50 framesRecord.frames.svBuffers[0].hardwareBuffer.description[1] = mConfig.height; in SurroundView3dSession()
275 framesRecord.frames.timestampNs = elapsedRealtimeNano(); in generateFrames()
276 framesRecord.frames.sequenceId = sequenceId++; in generateFrames()
278 framesRecord.frames.svBuffers.resize(mViews.size()); in generateFrames()
280 framesRecord.frames.svBuffers[i].viewId = mViews[i].viewId; in generateFrames()
281 framesRecord.frames.svBuffers[i].hardwareBuffer.nativeHandle = new native_handle_t(); in generateFrames()
[all …]
DSurroundView2dSession.cpp36 framesRecord.frames.svBuffers.resize(1); in SurroundView2dSession()
37 framesRecord.frames.svBuffers[0].viewId = 0; in SurroundView2dSession()
38 framesRecord.frames.svBuffers[0].hardwareBuffer.nativeHandle = in SurroundView2dSession()
40 framesRecord.frames.svBuffers[0].hardwareBuffer.description[0] = in SurroundView2dSession()
42 framesRecord.frames.svBuffers[0].hardwareBuffer.description[1] = in SurroundView2dSession()
204 framesRecord.frames.svBuffers[0].hardwareBuffer.description[0] = in generateFrames()
206 framesRecord.frames.svBuffers[0].hardwareBuffer.description[1] = in generateFrames()
212 framesRecord.frames.timestampNs = elapsedRealtimeNano(); in generateFrames()
213 framesRecord.frames.sequenceId = sequenceId++; in generateFrames()
223 mStream->receiveFrames(framesRecord.frames); in generateFrames()
/hardware/interfaces/automotive/sv/1.0/default/tests/fuzzer/
DAutomotiveSvV1_0Fuzzer.cpp121 SvFramesDesc frames; in invoke2dSessionAPI() local
122 frames.timestampNs = mFuzzedDataProvider.ConsumeIntegral<uint64_t>(); in invoke2dSessionAPI()
123 frames.sequenceId = mFuzzedDataProvider.ConsumeIntegral<uint32_t>(); in invoke2dSessionAPI()
126 frames.svBuffers.resize(numSvBuffers); in invoke2dSessionAPI()
128 frames.svBuffers[i].viewId = in invoke2dSessionAPI()
130 frames.svBuffers[i].hardwareBuffer.nativeHandle = new native_handle_t(); in invoke2dSessionAPI()
131 frames.svBuffers[i].hardwareBuffer.description[0] = in invoke2dSessionAPI()
133 frames.svBuffers[i].hardwareBuffer.description[1] = in invoke2dSessionAPI()
136 surroundView2dSession->doneWithFrames(frames); in invoke2dSessionAPI()
138 delete frames.svBuffers[i].hardwareBuffer.nativeHandle; in invoke2dSessionAPI()
[all …]
/hardware/libhardware/modules/audio_remote_submix/tests/
Dremote_submix_tests.cpp246 uint64_t frames; in TEST_F() local
248 EXPECT_EQ(0, streamOut->get_presentation_position(streamOut, &frames, &timestamp)); in TEST_F()
249 EXPECT_EQ(uint64_t{0}, frames); in TEST_F()
250 uint64_t prevFrames = frames; in TEST_F()
253 EXPECT_EQ(0, streamOut->get_presentation_position(streamOut, &frames, &timestamp)); in TEST_F()
254 EXPECT_LE(prevFrames, frames); in TEST_F()
255 prevFrames = frames; in TEST_F()
264 uint32_t frames; in TEST_F() local
265 EXPECT_EQ(0, streamOut->get_render_position(streamOut, &frames)); in TEST_F()
266 EXPECT_EQ(0U, frames); in TEST_F()
[all …]
/hardware/interfaces/automotive/sv/1.0/vts/functional/
DVtsHalSurroundViewV1_0TargetTest.cpp208 SvFramesDesc frames = handler->getLastReceivedFrames(); in TEST_P() local
209 EXPECT_EQ(frames.svBuffers.size(), 1); in TEST_P()
211 SvBuffer svBuffer2d = frames.svBuffers[0]; in TEST_P()
275 SvFramesDesc frames = handler->getLastReceivedFrames(); in TEST_P() local
276 EXPECT_EQ(frames.svBuffers.size(), 1); in TEST_P()
277 SvBuffer svBuffer2d = frames.svBuffers[0]; in TEST_P()
315 SvFramesDesc frames = handler->getLastReceivedFrames(); in TEST_P() local
316 EXPECT_EQ(frames.svBuffers.size(), 1); in TEST_P()
317 SvBuffer svBuffer2d = frames.svBuffers[0]; in TEST_P()
373 SvFramesDesc frames = handler->getLastReceivedFrames(); in TEST_P() local
[all …]
/hardware/qcom/audio/legacy/libalsa-intf/
Darec.c185 long frames; in record_file() local
269 frames = bufsize / 2; in record_file()
271 frames = bufsize / 8; in record_file()
273 frames = bufsize / 12; in record_file()
275 frames = bufsize / 4; in record_file()
277 x.frames = frames; in record_file()
300 fprintf(stderr, "Arec:avail 1 = %d frames = %ld\n",avail, frames); in record_file()
307 if (x.frames > avail) in record_file()
308 frames = avail; in record_file()
324 x.frames -= frames; in record_file()
[all …]
Dalsa_pcm.c468 long frames) in mmap_transfer() argument
478 frames = frames * channels *2 ; in mmap_transfer()
480 while (frames-- > 0) { in mmap_transfer()
489 long frames) in mmap_transfer_capture() argument
502 frames = frames * channels *2 ; in mmap_transfer_capture()
504 while (frames-- > 0) { in mmap_transfer_capture()
524 long frames; in pcm_write_mmap() local
528 frames = (pcm->flags & PCM_MONO) ? (count / 2) : (count / 4); in pcm_write_mmap()
539 pcm->sync_ptr->c.control.appl_ptr += frames; in pcm_write_mmap()
580 x.frames = (count / (channels * 2)) ; in pcm_write_nmmap()
[all …]
/hardware/interfaces/audio/2.0/
DIStreamIn.hal78 uint64_t frames;
101 * @param framesCount the number of frames in a buffer.
123 * Return the amount of input frames lost in the audio driver since the last
131 * @return framesLost the number of input audio frames lost.
136 * Return a recent count of the number of audio frames received and the
142 * @return frames the total frame count received. This must be as early in
143 * the capture pipeline as possible. In general, frames
145 * @return time is the clock monotonic time when frames was measured. In
150 generates (Result retval, uint64_t frames, uint64_t time);
DIStreamOut.hal71 uint64_t frames; // getPresentationPosition.
95 * @param framesCount the number of frames in a buffer.
116 * Return the number of audio frames written by the audio DSP to DAC since
120 * @return dspFrames number of audio frames written.
234 * Return a recent count of the number of audio frames presented to an
235 * external observer. This excludes frames which have been written but are
242 * Example: assume that N+M frames have been presented, where M is a 'small'
248 * @return frames count of presented audio frames.
252 generates (Result retval, uint64_t frames, TimeSpec timeStamp);
/hardware/interfaces/automotive/sv/1.0/
DISurroundViewStream.hal22 * This interface is to be implemented by client to receive callback for output frames and events.
26 * Receives callback of surround view 2d/3d frames.
28 * @param svFramesDesc Frames descriptor containing the output frames.
/hardware/interfaces/audio/4.0/
DIStreamIn.hal78 uint64_t frames;
107 * @param framesCount the number of frames in a buffer.
129 * Return the amount of input frames lost in the audio driver since the last
137 * @return framesLost the number of input audio frames lost.
142 * Return a recent count of the number of audio frames received and the
148 * @return frames the total frame count received. This must be as early in
149 * the capture pipeline as possible. In general, frames
151 * @return time is the clock monotonic time when frames was measured. In
156 generates (Result retval, uint64_t frames, uint64_t time);
DIStreamOut.hal71 uint64_t frames; // getPresentationPosition.
101 * @param framesCount the number of frames in a buffer.
122 * Return the number of audio frames written by the audio DSP to DAC since
127 * @return dspFrames number of audio frames written.
246 * Return a recent count of the number of audio frames presented to an
247 * external observer. This excludes frames which have been written but are
254 * Example: assume that N+M frames have been presented, where M is a 'small'
262 * @return frames count of presented audio frames.
266 generates (Result retval, uint64_t frames, TimeSpec timeStamp);
/hardware/interfaces/automotive/evs/1.1/
DIEvsCameraStream.hal29 * Receives calls from the HAL each time video frames is ready for inspection.
38 * A camera device will deliver the same number of frames as number of
41 * frames as many as number of backing physical camera devices.
43 * @param buffer Buffer descriptors of delivered image frames.
DIEvsUltrasonicsArray.hal40 * Up to this many data frames may be held concurrently by the client of IEvsUltrasonicsArray.
41 * If this many frames have been delivered to the receiver without being returned
42 * by doneWithFrame, the stream must skip frames until a buffer is returned for reuse.
65 * Requests to stop the delivery of the ultrasonic array data frames.
67 * Because delivery is asynchronous, frames may continue to arrive for
/hardware/qcom/audio/legacy/alsa_sound/
DAudioUsbALSA.cpp383 long frames; in RecordingThreadEntry() local
432 frames = (musbRecordingHandle->flags & PCM_MONO) ? (bufsize / 2) : (bufsize / 4); in RecordingThreadEntry()
433 x.frames = (musbRecordingHandle->flags & PCM_MONO) ? (bufsize / 2) : (bufsize / 4); in RecordingThreadEntry()
492 if (x.frames > avail) in RecordingThreadEntry()
493 frames = avail; in RecordingThreadEntry()
537 musbRecordingHandle->sync_ptr->c.control.appl_ptr += frames; in RecordingThreadEntry()
549 mproxyRecordingHandle->sync_ptr->c.control.appl_ptr += frames; in RecordingThreadEntry()
772 long frames; in PlaybackThreadEntry() local
839 frames = (mproxyPlaybackHandle->flags & PCM_MONO) ? (proxyPeriod / 2) : (proxyPeriod / 4); in PlaybackThreadEntry()
840 x.frames = (mproxyPlaybackHandle->flags & PCM_MONO) ? (proxyPeriod / 2) : (proxyPeriod / 4); in PlaybackThreadEntry()
[all …]
/hardware/google/gfxstream/codegen/vulkan/vulkan-docs-next/proposals/
DVK_EXT_frame_boundary.adoc9 debuggers) to group queue submissions per frames in non-trivial scenarios,
22 tools where the frames start and end.
44 ### Overlapping frames
47 different frames is being submitted in an interleaved way, a scenario we call
48 here "`overlapping frames`".
64 the erroneous grouping of queue submissions of different frames as the work of a
72 command buffers may be used concurrently by several frames, so that is not a
74 satisfactory since they cannot handle overlapping frames.
84 We want applications to be able to group queue submissions by frames. To this
144 // frameID is necessary to disambiguate overlapping frames
[all …]
/hardware/interfaces/audio/5.0/
DIStreamIn.hal78 uint64_t frames;
107 * @param framesCount the number of frames in a buffer.
129 * Return the amount of input frames lost in the audio driver since the last
137 * @return framesLost the number of input audio frames lost.
142 * Return a recent count of the number of audio frames received and the
148 * @return frames the total frame count received. This must be as early in
149 * the capture pipeline as possible. In general, frames
151 * @return time is the clock monotonic time when frames was measured. In
156 generates (Result retval, uint64_t frames, uint64_t time);
/hardware/interfaces/audio/6.0/
DIStreamIn.hal78 uint64_t frames;
107 * @param framesCount the number of frames in a buffer.
129 * Return the amount of input frames lost in the audio driver since the last
137 * @return framesLost the number of input audio frames lost.
142 * Return a recent count of the number of audio frames received and the
148 * @return frames the total frame count received. This must be as early in
149 * the capture pipeline as possible. In general, frames
151 * @return time is the clock monotonic time when frames was measured. In
156 generates (Result retval, uint64_t frames, uint64_t time);
/hardware/interfaces/automotive/evs/1.0/
DIEvsCamera.hal38 * Up to this many frames may be held concurrently by the client of IEvsCamera.
39 * If this many frames have been delivered to the receiver without being returned
40 * by doneWithFrame, the stream must skip frames until a buffer is returned for reuse.
68 * as one), and if the supply is exhausted, no further frames may be
76 * Stop the delivery of EVS camera frames.
78 * Because delivery is asynchronous, frames may continue to arrive for
/hardware/google/gfxstream/guest/mesa/src/util/tests/
Du_debug_stack_test.cpp53 func_c(struct debug_stack_frame *frames) in func_c() argument
55 debug_backtrace_capture(frames, 0, 16); in func_c()
/hardware/google/gchips/include/libaudio/audiohal/
Daudio_proxy_interface.h96 int proxy_get_render_position(void *proxy_stream, uint32_t *frames);
97 int proxy_get_presen_position(void *proxy_stream, uint64_t *frames, struct timespec *timestamp);
112 int __proxy_get_capture_position(void *proxy_stream, int64_t *frames, int64_t *time);
/hardware/interfaces/audio/7.0/
DIStreamIn.hal93 uint64_t frames;
116 * @param framesCount the number of frames in a buffer.
140 * Return the amount of input frames lost in the audio driver since the last
148 * @return framesLost the number of input audio frames lost.
153 * Return a recent count of the number of audio frames received and the
160 * @return frames the total frame count received. This must be as early in
161 * the capture pipeline as possible. In general, frames
163 * @return time is the clock monotonic time when frames was measured. In
168 generates (Result retval, uint64_t frames, uint64_t time);
/hardware/interfaces/audio/aidl/default/alsa/
DStreamAlsa.cpp127 alsa::resetTransferredFrames(mAlsaDeviceProxies[0], position->frames); in refinePosition()
129 if (int ret = proxy_get_capture_position(mAlsaDeviceProxies[0].get(), &position->frames, in refinePosition()
144 position->frames = static_cast<int64_t>(hwFrames); in refinePosition()
/hardware/interfaces/input/classifier/1.0/vts/functional/
DVtsHalInputClassifierV1_0TargetTest.cpp56 event.frames = {}; in getSimpleMotionEvent()
132 event.frames = {frame}; in TEST_P()
161 event.frames = {frame1, frame2}; in TEST_P()
/hardware/interfaces/audio/core/all-versions/vts/functional/7.0/
DAudioPrimaryHidlHalTest.cpp585 uint64_t frames = framesInitial; in waitForPresentationPositionAdvance() local
587 frames <= framesInitial && !writer.hasError() && in waitForPresentationPositionAdvance()
590 ASSERT_OK(stream->getPresentationPosition(returnIn(res, frames, ts))); in waitForPresentationPositionAdvance()
595 EXPECT_GT(frames, framesInitial); in waitForPresentationPositionAdvance()
597 if (lastPosition) *lastPosition = frames; in waitForPresentationPositionAdvance()
645 uint64_t frames = std::numeric_limits<uint64_t>::max(); in TEST_P() local
646 ASSERT_NO_FATAL_FAILURE(waitForPresentationPositionAdvance(writer, &frames)); in TEST_P()
647 EXPECT_GT(frames, framesInitial); in TEST_P()
745 uint64_t frames = framesInitial; in waitForCapturePositionAdvance() local
747 frames <= framesInitial && !reader.hasError() && in waitForCapturePositionAdvance()
[all …]

123456