Home
last modified time | relevance | path

Searched refs:frame (Results 1 – 25 of 280) sorted by relevance

12345678910>>...12

/hardware/google/aemu/host-common/
DMediaTexturePool.cpp73 TextureFrame frame{textures[2 * i], textures[2 * i + 1]}; in getTextureFrame() local
74 H264_DPRINT("allocated Y %d UV %d", frame.Ytex, frame.UVtex); in getTextureFrame()
75 m_Frame_to_PoolHandle[TexFrame{frame.Ytex, frame.UVtex}] = ph; in getTextureFrame()
76 ph->push_back(frame); in getTextureFrame()
79 TextureFrame frame = ph->front(); in getTextureFrame() local
81 H264_DPRINT("done %s %d ret Y %d UV %d", __func__, __LINE__, frame.Ytex, in getTextureFrame()
82 frame.UVtex); in getTextureFrame()
83 return frame; in getTextureFrame()
86 void MediaTexturePool::saveDecodedFrameToTexture(TextureFrame frame, in saveDecodedFrameToTexture() argument
90 (int)frame.Ytex, (int)frame.UVtex); in saveDecodedFrameToTexture()
[all …]
DH264NaluParser.cpp85 bool H264NaluParser::checkSpsFrame(const uint8_t* frame, size_t szBytes) { in checkSpsFrame() argument
87 H264NaluParser::getFrameNaluType(frame, szBytes, NULL); in checkSpsFrame()
95 bool H264NaluParser::checkIFrame(const uint8_t* frame, size_t szBytes) { in checkIFrame() argument
97 H264NaluParser::getFrameNaluType(frame, szBytes, NULL); in checkIFrame()
105 bool H264NaluParser::checkPpsFrame(const uint8_t* frame, size_t szBytes) { in checkPpsFrame() argument
107 H264NaluParser::getFrameNaluType(frame, szBytes, NULL); in checkPpsFrame()
115 H264NaluParser::H264NaluType H264NaluParser::getFrameNaluType(const uint8_t* frame, size_t szBytes,… in getFrameNaluType() argument
121 if (frame[0] != 0 || frame[1] != 0) { in getFrameNaluType()
127 if (frame[2] == 0 && frame[3] == 1) { in getFrameNaluType()
133 uint8_t naluType = 0x1f & frame[4]; in getFrameNaluType()
[all …]
DMediaHostRenderer.cpp61 void MediaHostRenderer::saveDecodedFrameToTexture(TextureFrame frame, in saveDecodedFrameToTexture() argument
64 mTexturePool.saveDecodedFrameToTexture(frame, privData, func); in saveDecodedFrameToTexture()
94 TextureFrame frame) { in renderToHostColorBufferWithTextures() argument
100 mTexturePool.putTextureFrame(frame); in renderToHostColorBufferWithTextures()
103 if (frame.Ytex <= 0 || frame.UVtex <= 0) { in renderToHostColorBufferWithTextures()
104 H264_DPRINT("ERROR: invalid tex ids: Ytex %d UVtex %d", (int)frame.Ytex, in renderToHostColorBufferWithTextures()
105 (int)frame.UVtex); in renderToHostColorBufferWithTextures()
109 uint32_t textures[2] = {frame.Ytex, frame.UVtex}; in renderToHostColorBufferWithTextures()
114 frame.Ytex = textures[0]; in renderToHostColorBufferWithTextures()
115 frame.UVtex = textures[1]; in renderToHostColorBufferWithTextures()
[all …]
DMediaSnapshotState.cpp42 bool MediaSnapshotState::savePacket(const uint8_t* frame, in savePacket() argument
49 vec.assign(frame, frame + size); in savePacket()
61 FrameInfo frame{ in saveDecodedFrame() local
64 savedFrames.push_back(std::move(frame)); in saveDecodedFrame()
73 FrameInfo frame{std::vector<uint8_t>{}, in saveDecodedFrame() local
79 savedFrames.push_back(std::move(frame)); in saveDecodedFrame()
102 const FrameInfo& frame) const { in saveFrameInfo()
104 saveVec(stream, frame.data); in saveFrameInfo()
106 saveVec(stream, frame.texture); in saveFrameInfo()
107 stream->putBe32(frame.width); in saveFrameInfo()
[all …]
DMediaSnapshotHelper.cpp35 void MediaSnapshotHelper::savePacket(const uint8_t* frame, in savePacket() argument
39 return saveH264Packet(frame, szBytes, inputPts); in savePacket()
41 return saveVPXPacket(frame, szBytes, inputPts); in savePacket()
72 void MediaSnapshotHelper::saveH264Packet(const uint8_t* frame, in saveH264Packet() argument
78 v.assign(frame, frame + szBytes); in saveH264Packet()
79 bool hasSps = H264NaluParser::checkSpsFrame(frame, szBytes); in saveH264Packet()
89 bool hasPps = H264NaluParser::checkPpsFrame(frame, szBytes); in saveH264Packet()
95 bool isIFrame = H264NaluParser::checkIFrame(frame, szBytes); in saveH264Packet()
DMediaVideoToolBoxVideoHelper.cpp96 MediaSnapshotState::FrameInfo frame; in extractFrameInfo() local
97 bool success = mFfmpegVideoHelper->receiveFrame(&frame); in extractFrameInfo()
99 mColorAspects = frame.color; in extractFrameInfo()
105 void MediaVideoToolBoxVideoHelper::decode(const uint8_t* frame, in decode() argument
108 VTB_DPRINT("%s(frame=%p, sz=%zu)", __func__, frame, szBytes); in decode()
111 const bool parseOk = parseInputFrames(frame, szBytes); in decode()
115 VTB_DPRINT("Failed to parse frame=%p, sz=%zu, give up.", frame, szBytes); in decode()
147 mFfmpegVideoHelper->decode(frame, szBytes, inputPts); in decode()
185 mFfmpegVideoHelper->decode(frame, szBytes, inputPts); in decode()
252 bool MediaVideoToolBoxVideoHelper::parseInputFrames(const uint8_t* frame, in parseInputFrames() argument
[all …]
DMediaH264DecoderCuvid.cpp212 const uint8_t* frame = param.pData; in decodeFrame() local
219 v.assign(frame, frame + szBytes); in decodeFrame()
220 bool hasSps = H264NaluParser::checkSpsFrame(frame, szBytes); in decodeFrame()
225 bool hasPps = H264NaluParser::checkPpsFrame(frame, szBytes); in decodeFrame()
231 bool isIFrame = H264NaluParser::checkIFrame(frame, szBytes); in decodeFrame()
242 decodeFrameInternal(param.pConsumedBytes, param.pDecoderErrorCode, frame, in decodeFrame()
248 const uint8_t* frame, in decodeFrameInternal() argument
252 H264_DPRINT("%s(frame=%p, sz=%zu)", __func__, frame, szBytes); in decodeFrameInternal()
256 packet.payload = frame; in decodeFrameInternal()
260 if (!frame || szBytes == 0) { in decodeFrameInternal()
[all …]
/hardware/libhardware/modules/camera/3_4/
Drequest_tracker_test.cpp41 uint32_t frame, std::vector<camera3_stream_t*> streams) { in GenerateCaptureRequest() argument
46 request->frame_number = frame; in GenerateCaptureRequest()
56 void AddRequest(uint32_t frame, in AddRequest() argument
60 GenerateCaptureRequest(frame, streams); in AddRequest()
63 EXPECT_FALSE(dut_->InFlight(frame)); in AddRequest()
67 EXPECT_TRUE(dut_->InFlight(frame)); in AddRequest()
78 uint32_t frame = 34; in TEST_F() local
79 EXPECT_FALSE(dut_->InFlight(frame)); in TEST_F()
80 AddRequest(frame, {&stream1_}); in TEST_F()
87 uint32_t frame = 42; in TEST_F() local
[all …]
/hardware/google/gfxstream/codegen/vulkan/vulkan-docs-next/proposals/
DVK_EXT_frame_boundary.adoc10 typically when `vkQueuePresentKHR` is not a relevant frame boundary delimiter.
17 commands, and need to know where a frame starts and ends. In general,
18 `vkQueuePresentKHR` is a natural frame boundary delimiter: a frame is made by
20 `vkQueuePresentKHR` is not a suitable frame boundary delimiter. The
24 Note: here, "`frame`" is understood as "`a unit of workload that spans one or more
25 queue submissions`". The notion of frame is application-dependent. In graphical
26 applications, a frame is typically the work needed to render into the image that
27 is eventually presented. In a compute-only application, a frame could be a set
30 There are a number of cases where `vkQueuePresentKHR` is not a suitable frame
46 A graphics application may pipeline its frame preparation such that work for
[all …]
/hardware/qcom/camera/msm8998/QCamera2/stack/mm-lib2d-interface/src/
Dmm_lib2d.c210 p_frame->frame[0].plane_cnt = 1; in lib2d_fill_img_frame()
211 p_frame->frame[0].plane[0].plane_type = PLANE_ARGB; in lib2d_fill_img_frame()
212 p_frame->frame[0].plane[0].addr = rgb_buffer->buffer; in lib2d_fill_img_frame()
213 p_frame->frame[0].plane[0].stride = rgb_buffer->stride; in lib2d_fill_img_frame()
214 p_frame->frame[0].plane[0].length = (rgb_buffer->stride * in lib2d_fill_img_frame()
216 p_frame->frame[0].plane[0].fd = rgb_buffer->fd; in lib2d_fill_img_frame()
217 p_frame->frame[0].plane[0].height = rgb_buffer->height; in lib2d_fill_img_frame()
218 p_frame->frame[0].plane[0].width = rgb_buffer->width; in lib2d_fill_img_frame()
219 p_frame->frame[0].plane[0].offset = 0; in lib2d_fill_img_frame()
220 p_frame->frame[0].plane[0].scanline = rgb_buffer->height; in lib2d_fill_img_frame()
[all …]
/hardware/qcom/camera/msm8998/QCamera2/HAL/
DQCamera2HWICallbacks.cpp146 mm_camera_super_buf_t* frame = in zsl_channel_cb() local
148 if (frame == NULL) { in zsl_channel_cb()
153 *frame = *recvd_frame; in zsl_channel_cb()
192 for (uint32_t i = 0; i < frame->num_bufs; i++) { in zsl_channel_cb()
193 QCameraStream *pStream = pChannel->getStreamByHandle(frame->bufs[i]->stream_id); in zsl_channel_cb()
196 pMetaFrame = frame->bufs[i]; //find the metadata in zsl_channel_cb()
233 for (uint32_t i = 0; i < frame->num_bufs; i++) { in zsl_channel_cb()
234 pStream = pChannel->getStreamByHandle(frame->bufs[i]->stream_id); in zsl_channel_cb()
237 pMetaFrame = frame->bufs[i]; in zsl_channel_cb()
253 for (uint32_t i = 0; i < frame->num_bufs; i++) { in zsl_channel_cb()
[all …]
/hardware/google/aemu/host-common/include/host-common/
DMediaH264DecoderPlugin.h107 bool savePacket(const uint8_t* frame, size_t size, uint64_t pts = 0) {
113 vec.assign(frame, frame+size);
124 FrameInfo frame{data, width, height, xcolor, pts};
125 savedDecodedFrame = std::move(frame);
137 void saveFrameInfo(base::Stream* stream, const FrameInfo& frame) const { in saveFrameInfo()
138 saveVec(stream, frame.data); in saveFrameInfo()
139 stream->putBe32(frame.width); in saveFrameInfo()
140 stream->putBe32(frame.height); in saveFrameInfo()
141 saveColor(stream, frame.color); in saveFrameInfo()
142 stream->putBe64(frame.pts); in saveFrameInfo()
[all …]
DH264NaluParser.h91 static bool checkSpsFrame(const uint8_t* frame, size_t szBytes);
92 static bool checkPpsFrame(const uint8_t* frame, size_t szBytes);
93 static bool checkIFrame(const uint8_t* frame, size_t szBytes);
99 static H264NaluType getFrameNaluType(const uint8_t* frame, size_t szBytes, uint8_t** data = nullptr…
103 static const uint8_t* getNextStartCodeHeader(const uint8_t* frame, size_t szBytes);
DMediaHostRenderer.h36 void saveDecodedFrameToTexture(MediaTexturePool::TextureFrame frame,
42 void putTextureFrame(MediaTexturePool::TextureFrame frame) { in putTextureFrame() argument
43 mTexturePool.putTextureFrame(frame); in putTextureFrame()
67 MediaTexturePool::TextureFrame frame,
DMediaSnapshotState.h66 void saveDecodedFrame(FrameInfo frame) { in saveDecodedFrame() argument
67 savedFrames.push_back(std::move(frame)); in saveDecodedFrame()
96 bool savePacket(const uint8_t* frame, size_t size, uint64_t pts = 0);
101 void saveFrameInfo(base::Stream* stream, const FrameInfo& frame) const;
106 void loadFrameInfo(base::Stream* stream, FrameInfo& frame);
DMediaTexturePool.h41 void saveDecodedFrameToTexture(TextureFrame frame,
47 void putTextureFrame(TextureFrame frame);
60 void deleteTextures(TextureFrame frame);
/hardware/interfaces/automotive/can/1.0/default/
DCanBus.cpp43 struct canfd_frame frame = {}; in send() local
44 frame.can_id = message.id; in send()
45 if (message.isExtendedId) frame.can_id |= CAN_EFF_FLAG; in send()
46 if (message.remoteTransmissionRequest) frame.can_id |= CAN_RTR_FLAG; in send()
47 frame.len = message.payload.size(); in send()
48 memcpy(frame.data, message.payload.data(), message.payload.size()); in send()
50 if (!mSocket->send(frame)) return Result::TRANSMISSION_FAILURE; in send()
276 static ErrorEvent parseErrorFrame(const struct canfd_frame& frame) { in parseErrorFrame() argument
278 if ((frame.can_id & (CAN_ERR_BUSERROR | CAN_ERR_BUSOFF)) != 0) { in parseErrorFrame()
281 if ((frame.data[1] & CAN_ERR_CRTL_TX_OVERFLOW) != 0) { in parseErrorFrame()
[all …]
DCanSocket.cpp67 bool CanSocket::send(const struct canfd_frame& frame) { in send() argument
68 const auto res = write(mSocket.get(), &frame, CAN_MTU); in send()
109 struct canfd_frame frame; in readerThread() local
110 const auto nbytes = read(mSocket.get(), &frame, CAN_MTU); in readerThread()
137 mReadCallback(frame, ts); in readerThread()
/hardware/google/gfxstream/guest/mesa/src/util/
Du_debug_stack_android.cpp77 const backtrace_frame_data_t* frame = bt->GetFrame(i); in debug_backtrace_capture() local
78 if (frame) { in debug_backtrace_capture()
79 backtrace[i].procname = intern_symbol(frame->func_name.c_str()); in debug_backtrace_capture()
80 backtrace[i].start_ip = frame->pc; in debug_backtrace_capture()
81 backtrace[i].off = frame->func_offset; in debug_backtrace_capture()
82 backtrace[i].map = intern_symbol(frame->map.Name().c_str()); in debug_backtrace_capture()
83 backtrace[i].map_off = frame->rel_pc; in debug_backtrace_capture()
/hardware/interfaces/automotive/evs/1.0/
DIEvsCameraStream.hal21 * Implemented on client side to receive asynchronous video frame deliveries.
26 * Receives calls from the HAL each time a video frame is ready for inspection.
30 * some time as the pipeline drains. Each frame must still be returned.
31 * When the last frame in the stream has been delivered, a NULL bufferHandle
32 * must be delivered, signifying the end of the stream. No further frame
35 * @param buffer a buffer descriptor of a delivered image frame.
/hardware/google/camera/devices/EmulatedCamera/hwl/
DJpegCompressor.cpp351 size_t JpegCompressor::CompressYUV420Frame(YUV420Frame frame) { in CompressYUV420Frame() argument
380 dmgr.buffer = static_cast<JOCTET*>(frame.output_buffer); in CompressYUV420Frame()
381 dmgr.buffer_size = frame.output_buffer_size; in CompressYUV420Frame()
408 cinfo->image_width = frame.width; in CompressYUV420Frame()
409 cinfo->image_height = frame.height; in CompressYUV420Frame()
444 if ((frame.app1_buffer != nullptr) && (frame.app1_buffer_size > 0)) { in CompressYUV420Frame()
446 static_cast<const JOCTET*>(frame.app1_buffer), in CompressYUV420Frame()
447 frame.app1_buffer_size); in CompressYUV420Frame()
452 switch (frame.color_space) { in CompressYUV420Frame()
482 uint8_t* py = static_cast<uint8_t*>(frame.yuv_planes.img_y); in CompressYUV420Frame()
[all …]
/hardware/qcom/camera/msm8998/QCamera2/stack/mm-camera-test/src/
Dmm_qcamera_video.c41 mm_camera_buf_def_t *frame; in mm_app_metadata_notify_cb() local
48 frame = bufs->bufs[0]; in mm_app_metadata_notify_cb()
79 frame = bufs->bufs[i]; in mm_app_metadata_notify_cb()
92 memcpy(pme->metadata, frame->buffer, sizeof(metadata_buffer_t)); in mm_app_metadata_notify_cb()
94 pMetadata = (metadata_buffer_t *)frame->buffer; in mm_app_metadata_notify_cb()
108 pme->user_metadata_cb(frame); in mm_app_metadata_notify_cb()
113 frame)) { in mm_app_metadata_notify_cb()
116 mm_app_cache_ops((mm_camera_app_meminfo_t *)frame->mem_info, in mm_app_metadata_notify_cb()
244 mm_camera_buf_def_t *frame = NULL; in mm_app_preview_notify_cb() local
252 frame = bufs->bufs[0]; in mm_app_preview_notify_cb()
[all …]
Dmm_qcamera_rdi.c40 static void mm_app_rdi_dump_frame(mm_camera_buf_def_t *frame, in mm_app_rdi_dump_frame() argument
49 if (frame != NULL) { in mm_app_rdi_dump_frame()
56 for (i = 0; i < frame->planes_buf.num_planes; i++) { in mm_app_rdi_dump_frame()
58 (uint8_t *)frame->buffer + frame->planes_buf.planes[i].data_offset, in mm_app_rdi_dump_frame()
72 mm_camera_buf_def_t *frame = bufs->bufs[0]; in mm_app_rdi_notify_cb() local
76 frame->frame_len, frame->frame_idx, frame->stream_id); in mm_app_rdi_notify_cb()
78 mm_app_rdi_dump_frame(frame, file_name, "raw", frame->frame_idx); in mm_app_rdi_notify_cb()
82 frame)) { in mm_app_rdi_notify_cb()
85 mm_app_cache_ops((mm_camera_app_meminfo_t *)frame->mem_info, in mm_app_rdi_notify_cb()
/hardware/interfaces/input/classifier/1.0/vts/functional/
DVtsHalInputClassifierV1_0TargetTest.cpp127 VideoFrame frame; in TEST_P() local
128 frame.data = {1, 2, 3, 4}; in TEST_P()
129 frame.height = 2; in TEST_P()
130 frame.width = 2; in TEST_P()
131 frame.timestamp = event.eventTime; in TEST_P()
132 event.frames = {frame}; in TEST_P()
/hardware/qcom/camera/msm8998/QCamera2/util/
DQCameraHALPP.cpp233 mm_camera_super_buf_t *frame = pData->frame; in releaseData() local
234 if (frame) { in releaseData()
238 m_pQCameraPostProc->releaseSuperBuf(frame); in releaseData()
240 free(frame); in releaseData()
241 frame = NULL; in releaseData()

12345678910>>...12