/cts/tests/tests/security/res/raw/ |
D | drm_uaf.dm | 9 ��W��cg[�<�(�@b�i�X�����+�|�d���t�=_�<�{����/���?���������f�p�<}B���i�[��H�}uj8���A… 11 ~�*�{P���C!�+ڊ�����<�%�^�,?�*Ȣ���'�0%��~��)���Y�<�>�.�_�_ Y]���f���8p�]�h�P<�"��\��[��4��… 12 !��(����F ��h���ꥆ��@.�<��l��6�h�k�υ��fx���H ����<��9�_/*_�?���bQ�ձ�~��oI�a��f�<�y�\�a���Z… 19 _�~_`�S����Sō�䐗�Gu�a�<����J~J�|V`{�C��E�sBbeP<{{� 23 …<}���r0ߥ�:c/�͆�X/�'V��rAx <�)���������*��q��m�?0q��0GP<qw�� �h�t��_���*��.QV�}�m=Z�<�%;… 24 �7OtT�[^n6�=��K6�]�<�,�6�|P�h.�T��ݞ]<���0PC�yEa<���zC��*�r���m�wc!R�XP_<����(��Ӏș1I�… 25 �4��-?�k��w�K�4��+��b�%���<@<�%�Հ��,�Ʊ����)���)�<_��<����v~�<�r#O�d�!����6z�<�,����… 28 ��_,y�Wa2�5���]�@1G>-.��<�%9�'��YZc_��,/�����j$����2p<��@g�E?-��t���z�,2z9�F�]�<�#)���… 32 …ʕ��t�`��������3`<y',0&�Z��U��<��\A�_�@VQ��<���zh+�J�a�x_#щ@-f5�^�*p<�"��8���ԛ ��kW;qbR… 35 …_��/��θٰt�U�V(!�P<� @,8?��t[<���S#�̊>���<� �^~��+����ǁ����'`��!��[�@<�/:>�?���B��…
|
/cts/tests/tests/renderscript/src/android/renderscript/cts/ |
D | struct_field.rscript | 63 // Arguments are named outerFieldName"_"index"_x", outerFieldName"_"index"_y", 64 // and outerFieldName"_"index"_f". 66 ARGS_INNERONE_(outerFieldName ## _ ## index) 70 // Arguments are named outerFieldName"_"index"_z", outerFieldName"_"index"_innerOne_x", etc. 72 …int8_t outerFieldName ## _ ## index ## _z, ARGS_INNERONE_(outerFieldName ## _ ## index ## _innerOn… 112 // Compares "outer"outerFieldName"["index"].x" to outerFieldName"_"index"_x", etc. 114 CHECK_INNERONE_(outer.outerFieldName[index], outerFieldName ## _ ## index) 118 // Compares "outer."outerFieldName"["index"].z" to outerFieldName"_"index"_z", 124 _RS_ASSERT(outer.outerFieldName[index].z == outerFieldName ## _ ## index ## _z); \ 125 …CHECK_INNERONE_(outer.outerFieldName[index].innerOne, outerFieldName ## _ ## index ## _innerOne); \
|
/cts/apps/CameraITS/tests/scene8/ |
D | test_ae_awb_regions.py | 101 aruco_corners, aruco_ids, _ = opencv_processing_utils.find_aruco_markers( 349 _, yellow, _, dark, _, light, _, blue = ( 353 _, blue, _, light, _, dark, _, yellow = (
|
/cts/apps/CameraITS/tests/ |
D | tutorial.py | 190 yimg, _, _ = image_processing_utils.convert_capture_to_planes(cap) 191 hist, _ = numpy.histogram(yimg*255, 256, (0, 256)) 205 r, _, _ = image_processing_utils.convert_capture_to_planes(cap)
|
/cts/apps/CameraITS/utils/ |
D | lighting_control_utils.py | 125 _, _, _ = select.select([sys.stdin], [], [], KEYBOARD_ENTRY_WAIT_TIME)
|
D | video_processing_utils.py | 157 output, _ = p.communicate() 203 _ = subprocess.call(cmd, 270 _ = subprocess.call(cmd, 276 [_ for _ in os.listdir(log_path) if (_.endswith(img_format) 277 and ffmpeg_image_name in _)])
|
D | its_session_utils.py | 447 data, _ = self.__read_response_from_socket() 486 data, _ = self.__read_response_from_socket() 509 data, _ = self.__read_response_from_socket() 567 data, _ = self.__read_response_from_socket() 574 data, _ = self.__read_response_from_socket() 598 data, _ = self.__read_response_from_socket() 613 data, _ = self.__read_response_from_socket() 635 data, _ = self.__read_response_from_socket() 652 data, _ = self.__read_response_from_socket() 683 data, _ = self.__read_response_from_socket() [all …]
|
D | noise_model_utils.py | 110 scale_a, _, offset_a, offset_b = zip(*noise_model) 173 height, width, _ = img.shape 212 height, width, _ = mean_img.shape 459 iso_ae, exp_ae, _, _, _ = cam.do_3a( 589 measured_models = [[] for _ in range(num_planes)] 591 samples = [[] for _ in range(num_planes)] 595 stats_per_plane = [[] for _ in range(num_planes)] 596 for _, means, vars_ in iso_to_stats_dict[iso]: 609 slope, intercept, rvalue, _, _ = scipy.stats.linregress(mean_var_pairs) 702 coeffs, _ = scipy.optimize.curve_fit(f, (gains, means), vars_ / (gains))
|
D | low_light_utils.py | 71 _, labels, centers = cv2.kmeans(data, _NUM_CLUSTERS, None, k_means_criteria, 90 contours, _ = cv2.findContours(mask, cv2.RETR_EXTERNAL, 139 contours, _ = cv2.findContours(eroded, cv2.RETR_EXTERNAL, 294 luminance_values = [luminance for _, luminance in results[:6]] 309 luminance_values = [luminance for _, luminance in results[:6]] 323 luminance_values = [luminance for _, luminance in results] 348 luminance_values = [luminance for _, luminance in results]
|
D | opencv_processing_utils.py | 121 y, _, _ = cv2.split(cv2.cvtColor(img, cv2.COLOR_RGB2YUV)) 123 y, _, _ = cv2.split(cv2.cvtColor(img, cv2.COLOR_BGR2YUV)) 137 _, img_bw = cv2.threshold(numpy.uint8(img_gray), 0, 255, 145 for path, _, files in os.walk(cv2_directory): 181 _, contours, _ = cv2.findContours(img, cv2.RETR_TREE, 184 contours, _ = cv2.findContours(img, cv2.RETR_TREE, cv2.CHAIN_APPROX_SIMPLE) 379 _, opt_val, _, top_left_scaled = cv2.minMaxLoc(result) 796 _, (width, height), angle = rect 824 _, (width, height), angle = cv2.minAreaRect(square) 876 for _, xy1 in enumerate(coordinates_list): [all …]
|
/cts/apps/CameraITS/tests/scene1_2/ |
D | test_raw_burst_sensitivity.py | 88 sens_min, _ = props['android.sensor.info.sensitivityRange'] 92 sens_ae, exp_ae, _, _, _ = cam.do_3a(do_af=False, get_results=True) 117 _, var_image = image_processing_utils.unpack_rawstats_capture(cap)
|
D | test_raw_sensitivity.py | 77 sens_min, _ = props['android.sensor.info.sensitivityRange'] 83 s_ae, e_ae, _, _, _ = cam.do_3a(do_af=False, get_results=True)
|
D | test_post_raw_sensitivity_boost.py | 108 for i, _ in enumerate(_COLORS): 183 sens_prev, _ = settings[step - 1] 186 for ch, _ in enumerate(_COLORS): 200 for ch, _ in enumerate(_COLORS):
|
/cts/apps/CameraITS/tests/scene_extensions/scene_hdr/ |
D | test_hdr_extension.py | 129 qr_code, _, _ = qr_detector.detectAndDecode(tile) 135 qr_code, _ = qr_detector.detect(tile) 144 qr_code, _, _ = qr_detector.detectAndDecode(qr_tile) 155 qr_code, _ = qr_detector.detect(qr_tile) 208 y_plane, _, _ = image_processing_utils.convert_capture_to_planes(cap)
|
/cts/apps/CameraITS/tests/scene3/ |
D | test_flip_mirror.py | 60 y, _, _ = image_processing_utils.convert_capture_to_planes(cap, props) 108 _, opt_val, _, _ = cv2.minMaxLoc(correlation)
|
D | test_edge_enhancement.py | 90 y, _, _ = image_processing_utils.convert_capture_to_planes(cap) 142 s, e, _, _, fd = cam.do_3a(get_results=True, mono_camera=mono_camera)
|
D | test_reprocess_edge_enhancement.py | 109 y, _, _ = image_processing_utils.convert_capture_to_planes(caps[n]) 177 s, e, _, _, fd = cam.do_3a(get_results=True, mono_camera=mono_camera)
|
/cts/apps/CameraITS/tests/inprog/rolling_shutter_skew/ |
D | test_rolling_shutter_skew.py | 136 (frame_h, _, _) = frames[0].shape 189 s, e, _, _, _ = cam.do_3a(get_results=True, do_af=False) 324 (frame_h, frame_w, _) = frame.shape 331 _, _, cluster_w, _ = find_cluster_bounding_rect(largest_cluster, 376 _, thresh = cv2.threshold(red_img, 0, 255, cv2.THRESH_BINARY + 392 _, contours, _ = cv2.findContours( 395 contours, _ = cv2.findContours(
|
/cts/tests/input/src/android/input/cts/ |
D | TwoWindowsActivity.kt | 82 leftView.setOnTouchListener { _: View, event: MotionEvent -> in launchTwoWindows() method 84 leftView.setOnHoverListener { _, event -> in launchTwoWindows() method 95 rightView.setOnTouchListener{ _: View, event: MotionEvent -> in launchTwoWindows() method 97 rightView.setOnHoverListener { _, event -> in launchTwoWindows() method
|
/cts/apps/CameraITS/tests/sensor_fusion/ |
D | test_multi_camera_frame_sync.py | 115 cam0_angles = [i for i, _ in frame_pair_angles] 116 cam1_angles = [j for _, j in frame_pair_angles] 163 sens, exp, _, _, _ = cam.do_3a(get_results=True, do_af=False) 225 angles = [i for i, _ in frame_pair_angles]
|
/cts/common/device-side/bedstead/modules/multiuser/src/test/kotlin/com/android/bedstead/multiuser/ |
D | MultiUserAnnotationExecutorTest.kt | 38 users().createUser().create().use { _ -> in <lambda>() method 39 users().createUser().create().use { _ -> } in <lambda>() method
|
/cts/tests/inputmethod/src/android/view/inputmethod/cts/ |
D | SpellCheckerTest.kt | 133 val (_, editText) = startTestActivity() in <lambda>() constant 175 val (_, editText) = startTestActivity() in <lambda>() constant 205 val (_, editText) = startTestActivity() in <lambda>() constant 230 val (_, editText) = startTestActivity() in <lambda>() constant 280 MockImeSession.create(context).use { _ -> in <lambda>() method 328 MockImeSession.create(context).use { _ -> in <lambda>() method 363 MockImeSession.create(context).use { _ -> in <lambda>() method 529 MockSpellCheckerClient.create(context, configuration).use { _ -> in <lambda>() method 530 val (_, editText) = startTestActivity() in <lambda>() constant 556 MockImeSession.create(context).use { _ -> in <lambda>() method [all …]
|
/cts/apps/CameraITS/tests/scene1_1/ |
D | test_dng_noise_model.py | 79 sens_min, _ = props['android.sensor.info.sensitivityRange'] 82 s_ae, e_ae, _, _, _ = cam.do_3a(get_results=True, do_af=False)
|
D | test_linearity.py | 130 line, residuals, _, _, _ = np.polyfit(
|
/cts/tests/tests/companion/core/src/android/companion/cts/core/ |
D | TransportsListenerTest.kt | 51 val listener = Consumer<List<AssociationInfo>> { _: List<AssociationInfo> -> } in <lambda>() method 82 val listener = BiConsumer<Int, ByteArray> { _: Int, _: ByteArray -> } in <lambda>() method
|