/hardware/google/aemu/host-common/ |
D | MediaTexturePool.cpp | 73 TextureFrame frame{textures[2 * i], textures[2 * i + 1]}; in getTextureFrame() local 74 H264_DPRINT("allocated Y %d UV %d", frame.Ytex, frame.UVtex); in getTextureFrame() 75 m_Frame_to_PoolHandle[TexFrame{frame.Ytex, frame.UVtex}] = ph; in getTextureFrame() 76 ph->push_back(frame); in getTextureFrame() 79 TextureFrame frame = ph->front(); in getTextureFrame() local 81 H264_DPRINT("done %s %d ret Y %d UV %d", __func__, __LINE__, frame.Ytex, in getTextureFrame() 82 frame.UVtex); in getTextureFrame() 83 return frame; in getTextureFrame() 86 void MediaTexturePool::saveDecodedFrameToTexture(TextureFrame frame, in saveDecodedFrameToTexture() argument 90 (int)frame.Ytex, (int)frame.UVtex); in saveDecodedFrameToTexture() [all …]
|
D | H264NaluParser.cpp | 85 bool H264NaluParser::checkSpsFrame(const uint8_t* frame, size_t szBytes) { in checkSpsFrame() argument 87 H264NaluParser::getFrameNaluType(frame, szBytes, NULL); in checkSpsFrame() 95 bool H264NaluParser::checkIFrame(const uint8_t* frame, size_t szBytes) { in checkIFrame() argument 97 H264NaluParser::getFrameNaluType(frame, szBytes, NULL); in checkIFrame() 105 bool H264NaluParser::checkPpsFrame(const uint8_t* frame, size_t szBytes) { in checkPpsFrame() argument 107 H264NaluParser::getFrameNaluType(frame, szBytes, NULL); in checkPpsFrame() 115 H264NaluParser::H264NaluType H264NaluParser::getFrameNaluType(const uint8_t* frame, size_t szBytes,… in getFrameNaluType() argument 121 if (frame[0] != 0 || frame[1] != 0) { in getFrameNaluType() 127 if (frame[2] == 0 && frame[3] == 1) { in getFrameNaluType() 133 uint8_t naluType = 0x1f & frame[4]; in getFrameNaluType() [all …]
|
D | MediaHostRenderer.cpp | 61 void MediaHostRenderer::saveDecodedFrameToTexture(TextureFrame frame, in saveDecodedFrameToTexture() argument 64 mTexturePool.saveDecodedFrameToTexture(frame, privData, func); in saveDecodedFrameToTexture() 94 TextureFrame frame) { in renderToHostColorBufferWithTextures() argument 100 mTexturePool.putTextureFrame(frame); in renderToHostColorBufferWithTextures() 103 if (frame.Ytex <= 0 || frame.UVtex <= 0) { in renderToHostColorBufferWithTextures() 104 H264_DPRINT("ERROR: invalid tex ids: Ytex %d UVtex %d", (int)frame.Ytex, in renderToHostColorBufferWithTextures() 105 (int)frame.UVtex); in renderToHostColorBufferWithTextures() 109 uint32_t textures[2] = {frame.Ytex, frame.UVtex}; in renderToHostColorBufferWithTextures() 114 frame.Ytex = textures[0]; in renderToHostColorBufferWithTextures() 115 frame.UVtex = textures[1]; in renderToHostColorBufferWithTextures() [all …]
|
D | MediaSnapshotState.cpp | 42 bool MediaSnapshotState::savePacket(const uint8_t* frame, in savePacket() argument 49 vec.assign(frame, frame + size); in savePacket() 61 FrameInfo frame{ in saveDecodedFrame() local 64 savedFrames.push_back(std::move(frame)); in saveDecodedFrame() 73 FrameInfo frame{std::vector<uint8_t>{}, in saveDecodedFrame() local 79 savedFrames.push_back(std::move(frame)); in saveDecodedFrame() 102 const FrameInfo& frame) const { in saveFrameInfo() 104 saveVec(stream, frame.data); in saveFrameInfo() 106 saveVec(stream, frame.texture); in saveFrameInfo() 107 stream->putBe32(frame.width); in saveFrameInfo() [all …]
|
D | MediaSnapshotHelper.cpp | 35 void MediaSnapshotHelper::savePacket(const uint8_t* frame, in savePacket() argument 39 return saveH264Packet(frame, szBytes, inputPts); in savePacket() 41 return saveVPXPacket(frame, szBytes, inputPts); in savePacket() 72 void MediaSnapshotHelper::saveH264Packet(const uint8_t* frame, in saveH264Packet() argument 78 v.assign(frame, frame + szBytes); in saveH264Packet() 79 bool hasSps = H264NaluParser::checkSpsFrame(frame, szBytes); in saveH264Packet() 89 bool hasPps = H264NaluParser::checkPpsFrame(frame, szBytes); in saveH264Packet() 95 bool isIFrame = H264NaluParser::checkIFrame(frame, szBytes); in saveH264Packet()
|
D | MediaVideoToolBoxVideoHelper.cpp | 96 MediaSnapshotState::FrameInfo frame; in extractFrameInfo() local 97 bool success = mFfmpegVideoHelper->receiveFrame(&frame); in extractFrameInfo() 99 mColorAspects = frame.color; in extractFrameInfo() 105 void MediaVideoToolBoxVideoHelper::decode(const uint8_t* frame, in decode() argument 108 VTB_DPRINT("%s(frame=%p, sz=%zu)", __func__, frame, szBytes); in decode() 111 const bool parseOk = parseInputFrames(frame, szBytes); in decode() 115 VTB_DPRINT("Failed to parse frame=%p, sz=%zu, give up.", frame, szBytes); in decode() 147 mFfmpegVideoHelper->decode(frame, szBytes, inputPts); in decode() 185 mFfmpegVideoHelper->decode(frame, szBytes, inputPts); in decode() 252 bool MediaVideoToolBoxVideoHelper::parseInputFrames(const uint8_t* frame, in parseInputFrames() argument [all …]
|
D | MediaH264DecoderCuvid.cpp | 212 const uint8_t* frame = param.pData; in decodeFrame() local 219 v.assign(frame, frame + szBytes); in decodeFrame() 220 bool hasSps = H264NaluParser::checkSpsFrame(frame, szBytes); in decodeFrame() 225 bool hasPps = H264NaluParser::checkPpsFrame(frame, szBytes); in decodeFrame() 231 bool isIFrame = H264NaluParser::checkIFrame(frame, szBytes); in decodeFrame() 242 decodeFrameInternal(param.pConsumedBytes, param.pDecoderErrorCode, frame, in decodeFrame() 248 const uint8_t* frame, in decodeFrameInternal() argument 252 H264_DPRINT("%s(frame=%p, sz=%zu)", __func__, frame, szBytes); in decodeFrameInternal() 256 packet.payload = frame; in decodeFrameInternal() 260 if (!frame || szBytes == 0) { in decodeFrameInternal() [all …]
|
/hardware/libhardware/modules/camera/3_4/ |
D | request_tracker_test.cpp | 41 uint32_t frame, std::vector<camera3_stream_t*> streams) { in GenerateCaptureRequest() argument 46 request->frame_number = frame; in GenerateCaptureRequest() 56 void AddRequest(uint32_t frame, in AddRequest() argument 60 GenerateCaptureRequest(frame, streams); in AddRequest() 63 EXPECT_FALSE(dut_->InFlight(frame)); in AddRequest() 67 EXPECT_TRUE(dut_->InFlight(frame)); in AddRequest() 78 uint32_t frame = 34; in TEST_F() local 79 EXPECT_FALSE(dut_->InFlight(frame)); in TEST_F() 80 AddRequest(frame, {&stream1_}); in TEST_F() 87 uint32_t frame = 42; in TEST_F() local [all …]
|
/hardware/google/gfxstream/codegen/vulkan/vulkan-docs-next/proposals/ |
D | VK_EXT_frame_boundary.adoc | 10 typically when `vkQueuePresentKHR` is not a relevant frame boundary delimiter. 17 commands, and need to know where a frame starts and ends. In general, 18 `vkQueuePresentKHR` is a natural frame boundary delimiter: a frame is made by 20 `vkQueuePresentKHR` is not a suitable frame boundary delimiter. The 24 Note: here, "`frame`" is understood as "`a unit of workload that spans one or more 25 queue submissions`". The notion of frame is application-dependent. In graphical 26 applications, a frame is typically the work needed to render into the image that 27 is eventually presented. In a compute-only application, a frame could be a set 30 There are a number of cases where `vkQueuePresentKHR` is not a suitable frame 46 A graphics application may pipeline its frame preparation such that work for [all …]
|
/hardware/qcom/camera/msm8998/QCamera2/stack/mm-lib2d-interface/src/ |
D | mm_lib2d.c | 210 p_frame->frame[0].plane_cnt = 1; in lib2d_fill_img_frame() 211 p_frame->frame[0].plane[0].plane_type = PLANE_ARGB; in lib2d_fill_img_frame() 212 p_frame->frame[0].plane[0].addr = rgb_buffer->buffer; in lib2d_fill_img_frame() 213 p_frame->frame[0].plane[0].stride = rgb_buffer->stride; in lib2d_fill_img_frame() 214 p_frame->frame[0].plane[0].length = (rgb_buffer->stride * in lib2d_fill_img_frame() 216 p_frame->frame[0].plane[0].fd = rgb_buffer->fd; in lib2d_fill_img_frame() 217 p_frame->frame[0].plane[0].height = rgb_buffer->height; in lib2d_fill_img_frame() 218 p_frame->frame[0].plane[0].width = rgb_buffer->width; in lib2d_fill_img_frame() 219 p_frame->frame[0].plane[0].offset = 0; in lib2d_fill_img_frame() 220 p_frame->frame[0].plane[0].scanline = rgb_buffer->height; in lib2d_fill_img_frame() [all …]
|
/hardware/qcom/camera/msm8998/QCamera2/HAL/ |
D | QCamera2HWICallbacks.cpp | 146 mm_camera_super_buf_t* frame = in zsl_channel_cb() local 148 if (frame == NULL) { in zsl_channel_cb() 153 *frame = *recvd_frame; in zsl_channel_cb() 192 for (uint32_t i = 0; i < frame->num_bufs; i++) { in zsl_channel_cb() 193 QCameraStream *pStream = pChannel->getStreamByHandle(frame->bufs[i]->stream_id); in zsl_channel_cb() 196 pMetaFrame = frame->bufs[i]; //find the metadata in zsl_channel_cb() 233 for (uint32_t i = 0; i < frame->num_bufs; i++) { in zsl_channel_cb() 234 pStream = pChannel->getStreamByHandle(frame->bufs[i]->stream_id); in zsl_channel_cb() 237 pMetaFrame = frame->bufs[i]; in zsl_channel_cb() 253 for (uint32_t i = 0; i < frame->num_bufs; i++) { in zsl_channel_cb() [all …]
|
/hardware/google/aemu/host-common/include/host-common/ |
D | MediaH264DecoderPlugin.h | 107 bool savePacket(const uint8_t* frame, size_t size, uint64_t pts = 0) { 113 vec.assign(frame, frame+size); 124 FrameInfo frame{data, width, height, xcolor, pts}; 125 savedDecodedFrame = std::move(frame); 137 void saveFrameInfo(base::Stream* stream, const FrameInfo& frame) const { in saveFrameInfo() 138 saveVec(stream, frame.data); in saveFrameInfo() 139 stream->putBe32(frame.width); in saveFrameInfo() 140 stream->putBe32(frame.height); in saveFrameInfo() 141 saveColor(stream, frame.color); in saveFrameInfo() 142 stream->putBe64(frame.pts); in saveFrameInfo() [all …]
|
D | H264NaluParser.h | 91 static bool checkSpsFrame(const uint8_t* frame, size_t szBytes); 92 static bool checkPpsFrame(const uint8_t* frame, size_t szBytes); 93 static bool checkIFrame(const uint8_t* frame, size_t szBytes); 99 static H264NaluType getFrameNaluType(const uint8_t* frame, size_t szBytes, uint8_t** data = nullptr… 103 static const uint8_t* getNextStartCodeHeader(const uint8_t* frame, size_t szBytes);
|
D | MediaHostRenderer.h | 36 void saveDecodedFrameToTexture(MediaTexturePool::TextureFrame frame, 42 void putTextureFrame(MediaTexturePool::TextureFrame frame) { in putTextureFrame() argument 43 mTexturePool.putTextureFrame(frame); in putTextureFrame() 67 MediaTexturePool::TextureFrame frame,
|
D | MediaSnapshotState.h | 66 void saveDecodedFrame(FrameInfo frame) { in saveDecodedFrame() argument 67 savedFrames.push_back(std::move(frame)); in saveDecodedFrame() 96 bool savePacket(const uint8_t* frame, size_t size, uint64_t pts = 0); 101 void saveFrameInfo(base::Stream* stream, const FrameInfo& frame) const; 106 void loadFrameInfo(base::Stream* stream, FrameInfo& frame);
|
D | MediaTexturePool.h | 41 void saveDecodedFrameToTexture(TextureFrame frame, 47 void putTextureFrame(TextureFrame frame); 60 void deleteTextures(TextureFrame frame);
|
/hardware/interfaces/automotive/can/1.0/default/ |
D | CanBus.cpp | 43 struct canfd_frame frame = {}; in send() local 44 frame.can_id = message.id; in send() 45 if (message.isExtendedId) frame.can_id |= CAN_EFF_FLAG; in send() 46 if (message.remoteTransmissionRequest) frame.can_id |= CAN_RTR_FLAG; in send() 47 frame.len = message.payload.size(); in send() 48 memcpy(frame.data, message.payload.data(), message.payload.size()); in send() 50 if (!mSocket->send(frame)) return Result::TRANSMISSION_FAILURE; in send() 276 static ErrorEvent parseErrorFrame(const struct canfd_frame& frame) { in parseErrorFrame() argument 278 if ((frame.can_id & (CAN_ERR_BUSERROR | CAN_ERR_BUSOFF)) != 0) { in parseErrorFrame() 281 if ((frame.data[1] & CAN_ERR_CRTL_TX_OVERFLOW) != 0) { in parseErrorFrame() [all …]
|
D | CanSocket.cpp | 67 bool CanSocket::send(const struct canfd_frame& frame) { in send() argument 68 const auto res = write(mSocket.get(), &frame, CAN_MTU); in send() 109 struct canfd_frame frame; in readerThread() local 110 const auto nbytes = read(mSocket.get(), &frame, CAN_MTU); in readerThread() 137 mReadCallback(frame, ts); in readerThread()
|
/hardware/google/gfxstream/guest/mesa/src/util/ |
D | u_debug_stack_android.cpp | 77 const backtrace_frame_data_t* frame = bt->GetFrame(i); in debug_backtrace_capture() local 78 if (frame) { in debug_backtrace_capture() 79 backtrace[i].procname = intern_symbol(frame->func_name.c_str()); in debug_backtrace_capture() 80 backtrace[i].start_ip = frame->pc; in debug_backtrace_capture() 81 backtrace[i].off = frame->func_offset; in debug_backtrace_capture() 82 backtrace[i].map = intern_symbol(frame->map.Name().c_str()); in debug_backtrace_capture() 83 backtrace[i].map_off = frame->rel_pc; in debug_backtrace_capture()
|
/hardware/interfaces/automotive/evs/1.0/ |
D | IEvsCameraStream.hal | 21 * Implemented on client side to receive asynchronous video frame deliveries. 26 * Receives calls from the HAL each time a video frame is ready for inspection. 30 * some time as the pipeline drains. Each frame must still be returned. 31 * When the last frame in the stream has been delivered, a NULL bufferHandle 32 * must be delivered, signifying the end of the stream. No further frame 35 * @param buffer a buffer descriptor of a delivered image frame.
|
/hardware/google/camera/devices/EmulatedCamera/hwl/ |
D | JpegCompressor.cpp | 351 size_t JpegCompressor::CompressYUV420Frame(YUV420Frame frame) { in CompressYUV420Frame() argument 380 dmgr.buffer = static_cast<JOCTET*>(frame.output_buffer); in CompressYUV420Frame() 381 dmgr.buffer_size = frame.output_buffer_size; in CompressYUV420Frame() 408 cinfo->image_width = frame.width; in CompressYUV420Frame() 409 cinfo->image_height = frame.height; in CompressYUV420Frame() 444 if ((frame.app1_buffer != nullptr) && (frame.app1_buffer_size > 0)) { in CompressYUV420Frame() 446 static_cast<const JOCTET*>(frame.app1_buffer), in CompressYUV420Frame() 447 frame.app1_buffer_size); in CompressYUV420Frame() 452 switch (frame.color_space) { in CompressYUV420Frame() 482 uint8_t* py = static_cast<uint8_t*>(frame.yuv_planes.img_y); in CompressYUV420Frame() [all …]
|
/hardware/qcom/camera/msm8998/QCamera2/stack/mm-camera-test/src/ |
D | mm_qcamera_video.c | 41 mm_camera_buf_def_t *frame; in mm_app_metadata_notify_cb() local 48 frame = bufs->bufs[0]; in mm_app_metadata_notify_cb() 79 frame = bufs->bufs[i]; in mm_app_metadata_notify_cb() 92 memcpy(pme->metadata, frame->buffer, sizeof(metadata_buffer_t)); in mm_app_metadata_notify_cb() 94 pMetadata = (metadata_buffer_t *)frame->buffer; in mm_app_metadata_notify_cb() 108 pme->user_metadata_cb(frame); in mm_app_metadata_notify_cb() 113 frame)) { in mm_app_metadata_notify_cb() 116 mm_app_cache_ops((mm_camera_app_meminfo_t *)frame->mem_info, in mm_app_metadata_notify_cb() 244 mm_camera_buf_def_t *frame = NULL; in mm_app_preview_notify_cb() local 252 frame = bufs->bufs[0]; in mm_app_preview_notify_cb() [all …]
|
D | mm_qcamera_rdi.c | 40 static void mm_app_rdi_dump_frame(mm_camera_buf_def_t *frame, in mm_app_rdi_dump_frame() argument 49 if (frame != NULL) { in mm_app_rdi_dump_frame() 56 for (i = 0; i < frame->planes_buf.num_planes; i++) { in mm_app_rdi_dump_frame() 58 (uint8_t *)frame->buffer + frame->planes_buf.planes[i].data_offset, in mm_app_rdi_dump_frame() 72 mm_camera_buf_def_t *frame = bufs->bufs[0]; in mm_app_rdi_notify_cb() local 76 frame->frame_len, frame->frame_idx, frame->stream_id); in mm_app_rdi_notify_cb() 78 mm_app_rdi_dump_frame(frame, file_name, "raw", frame->frame_idx); in mm_app_rdi_notify_cb() 82 frame)) { in mm_app_rdi_notify_cb() 85 mm_app_cache_ops((mm_camera_app_meminfo_t *)frame->mem_info, in mm_app_rdi_notify_cb()
|
/hardware/interfaces/input/classifier/1.0/vts/functional/ |
D | VtsHalInputClassifierV1_0TargetTest.cpp | 127 VideoFrame frame; in TEST_P() local 128 frame.data = {1, 2, 3, 4}; in TEST_P() 129 frame.height = 2; in TEST_P() 130 frame.width = 2; in TEST_P() 131 frame.timestamp = event.eventTime; in TEST_P() 132 event.frames = {frame}; in TEST_P()
|
/hardware/qcom/camera/msm8998/QCamera2/util/ |
D | QCameraHALPP.cpp | 233 mm_camera_super_buf_t *frame = pData->frame; in releaseData() local 234 if (frame) { in releaseData() 238 m_pQCameraPostProc->releaseSuperBuf(frame); in releaseData() 240 free(frame); in releaseData() 241 frame = NULL; in releaseData()
|