1 /*
2 * Copyright (C) 2008 The Android Open Source Project
3 * All rights reserved.
4 *
5 * Redistribution and use in source and binary forms, with or without
6 * modification, are permitted provided that the following conditions
7 * are met:
8 * * Redistributions of source code must retain the above copyright
9 * notice, this list of conditions and the following disclaimer.
10 * * Redistributions in binary form must reproduce the above copyright
11 * notice, this list of conditions and the following disclaimer in
12 * the documentation and/or other materials provided with the
13 * distribution.
14 *
15 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
16 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
17 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS
18 * FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE
19 * COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
20 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING,
21 * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS
22 * OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED
23 * AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
24 * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT
25 * OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
26 * SUCH DAMAGE.
27 */
28
29 #include <android/api-level.h>
30 #include <elf.h>
31 #include <errno.h>
32 #include <malloc.h>
33 #include <stddef.h>
34 #include <stdint.h>
35 #include <stdio.h>
36 #include <stdlib.h>
37 #include <sys/auxv.h>
38 #include <sys/mman.h>
39
40 #include "async_safe/log.h"
41 #include "heap_tagging.h"
42 #include "libc_init_common.h"
43 #include "platform/bionic/macros.h"
44 #include "platform/bionic/mte.h"
45 #include "platform/bionic/page.h"
46 #include "platform/bionic/reserved_signals.h"
47 #include "private/KernelArgumentBlock.h"
48 #include "private/bionic_asm.h"
49 #include "private/bionic_asm_note.h"
50 #include "private/bionic_call_ifunc_resolver.h"
51 #include "private/bionic_elf_tls.h"
52 #include "private/bionic_globals.h"
53 #include "private/bionic_tls.h"
54 #include "private/elf_note.h"
55 #include "pthread_internal.h"
56 #include "sys/system_properties.h"
57 #include "sysprop_helpers.h"
58
59 #if __has_feature(hwaddress_sanitizer)
60 #include <sanitizer/hwasan_interface.h>
61 #endif
62
63 // Leave the variable uninitialized for the sake of the dynamic loader, which
64 // links in this file. The loader will initialize this variable before
65 // relocating itself.
66 #if defined(__i386__)
67 __LIBC_HIDDEN__ void* __libc_sysinfo;
68 #endif
69
70 extern "C" int __cxa_atexit(void (*)(void *), void *, void *);
71 extern "C" const char* __gnu_basename(const char* path);
72
call_array(init_func_t ** list,size_t count,int argc,char * argv[],char * envp[])73 static void call_array(init_func_t** list, size_t count, int argc, char* argv[], char* envp[]) {
74 while (count-- > 0) {
75 init_func_t* function = *list++;
76 (*function)(argc, argv, envp);
77 }
78 }
79
call_fini_array(void * arg)80 static void call_fini_array(void* arg) {
81 structors_array_t* structors = reinterpret_cast<structors_array_t*>(arg);
82 fini_func_t** array = structors->fini_array;
83 size_t count = structors->fini_array_count;
84 // Now call each destructor in reverse order.
85 while (count-- > 0) {
86 fini_func_t* function = array[count];
87 (*function)();
88 }
89 }
90
91 #if defined(__arm__) || defined(__i386__) // Legacy architectures used REL...
92 extern __LIBC_HIDDEN__ __attribute__((weak)) ElfW(Rel) __rel_iplt_start[], __rel_iplt_end[];
93
call_ifunc_resolvers()94 static void call_ifunc_resolvers() {
95 for (ElfW(Rel)* r = __rel_iplt_start; r != __rel_iplt_end; ++r) {
96 ElfW(Addr)* offset = reinterpret_cast<ElfW(Addr)*>(r->r_offset);
97 ElfW(Addr) resolver = *offset;
98 *offset = __bionic_call_ifunc_resolver(resolver);
99 }
100 }
101 #else // ...but modern architectures use RELA instead.
102 extern __LIBC_HIDDEN__ __attribute__((weak)) ElfW(Rela) __rela_iplt_start[], __rela_iplt_end[];
103
call_ifunc_resolvers()104 static void call_ifunc_resolvers() {
105 for (ElfW(Rela)* r = __rela_iplt_start; r != __rela_iplt_end; ++r) {
106 ElfW(Addr)* offset = reinterpret_cast<ElfW(Addr)*>(r->r_offset);
107 ElfW(Addr) resolver = r->r_addend;
108 *offset = __bionic_call_ifunc_resolver(resolver);
109 }
110 }
111 #endif
112
apply_gnu_relro()113 static void apply_gnu_relro() {
114 ElfW(Phdr)* phdr_start = reinterpret_cast<ElfW(Phdr)*>(getauxval(AT_PHDR));
115 unsigned long int phdr_ct = getauxval(AT_PHNUM);
116
117 for (ElfW(Phdr)* phdr = phdr_start; phdr < (phdr_start + phdr_ct); phdr++) {
118 if (phdr->p_type != PT_GNU_RELRO) {
119 continue;
120 }
121
122 ElfW(Addr) seg_page_start = page_start(phdr->p_vaddr);
123 ElfW(Addr) seg_page_end = page_end(phdr->p_vaddr + phdr->p_memsz);
124
125 // Check return value here? What do we do if we fail?
126 mprotect(reinterpret_cast<void*>(seg_page_start), seg_page_end - seg_page_start, PROT_READ);
127 }
128 }
129
layout_static_tls(KernelArgumentBlock & args)130 static void layout_static_tls(KernelArgumentBlock& args) {
131 StaticTlsLayout& layout = __libc_shared_globals()->static_tls_layout;
132 layout.reserve_bionic_tls();
133
134 const char* progname = args.argv[0];
135 ElfW(Phdr)* phdr_start = reinterpret_cast<ElfW(Phdr)*>(getauxval(AT_PHDR));
136 size_t phdr_ct = getauxval(AT_PHNUM);
137
138 static TlsModule mod;
139 TlsModules& modules = __libc_shared_globals()->tls_modules;
140 if (__bionic_get_tls_segment(phdr_start, phdr_ct, 0, &mod.segment)) {
141 if (!__bionic_check_tls_align(mod.segment.aligned_size.align.value)) {
142 async_safe_fatal("error: TLS segment alignment in \"%s\" is not a power of 2: %zu\n",
143 progname, mod.segment.aligned_size.align.value);
144 }
145 mod.static_offset = layout.reserve_exe_segment_and_tcb(&mod.segment, progname);
146 mod.first_generation = kTlsGenerationFirst;
147
148 modules.module_count = 1;
149 modules.static_module_count = 1;
150 modules.module_table = &mod;
151 } else {
152 layout.reserve_exe_segment_and_tcb(nullptr, progname);
153 }
154 // Enable the fast path in __tls_get_addr.
155 __libc_tls_generation_copy = modules.generation;
156
157 layout.finish_layout();
158 }
159
160 #ifdef __aarch64__
__get_memtag_level_from_note(const ElfW (Phdr)* phdr_start,size_t phdr_ct,const ElfW (Addr)load_bias,bool * stack)161 static HeapTaggingLevel __get_memtag_level_from_note(const ElfW(Phdr) * phdr_start, size_t phdr_ct,
162 const ElfW(Addr) load_bias, bool* stack) {
163 const ElfW(Nhdr) * note;
164 const char* desc;
165 if (!__find_elf_note(NT_ANDROID_TYPE_MEMTAG, "Android", phdr_start, phdr_ct, ¬e, &desc,
166 load_bias)) {
167 return M_HEAP_TAGGING_LEVEL_TBI;
168 }
169
170 // Previously (in Android 12), if the note was != 4 bytes, we check-failed
171 // here. Let's be more permissive to allow future expansion.
172 if (note->n_descsz < 4) {
173 async_safe_fatal("unrecognized android.memtag note: n_descsz = %d, expected >= 4",
174 note->n_descsz);
175 }
176
177 // `desc` is always aligned due to ELF requirements, enforced in __find_elf_note().
178 ElfW(Word) note_val = *reinterpret_cast<const ElfW(Word)*>(desc);
179 *stack = (note_val & NT_MEMTAG_STACK) != 0;
180
181 // Warning: In Android 12, any value outside of bits [0..3] resulted in a check-fail.
182 if (!(note_val & (NT_MEMTAG_HEAP | NT_MEMTAG_STACK))) {
183 async_safe_format_log(ANDROID_LOG_INFO, "libc",
184 "unrecognised memtag note_val did not specificy heap or stack: %u",
185 note_val);
186 return M_HEAP_TAGGING_LEVEL_TBI;
187 }
188
189 unsigned mode = note_val & NT_MEMTAG_LEVEL_MASK;
190 switch (mode) {
191 case NT_MEMTAG_LEVEL_NONE:
192 // Note, previously (in Android 12), NT_MEMTAG_LEVEL_NONE was
193 // NT_MEMTAG_LEVEL_DEFAULT, which implied SYNC mode. This was never used
194 // by anyone, but we note it (heh) here for posterity, in case the zero
195 // level becomes meaningful, and binaries with this note can be executed
196 // on Android 12 devices.
197 return M_HEAP_TAGGING_LEVEL_TBI;
198 case NT_MEMTAG_LEVEL_ASYNC:
199 return M_HEAP_TAGGING_LEVEL_ASYNC;
200 case NT_MEMTAG_LEVEL_SYNC:
201 default:
202 // We allow future extensions to specify mode 3 (currently unused), with
203 // the idea that it might be used for ASYMM mode or something else. On
204 // this version of Android, it falls back to SYNC mode.
205 return M_HEAP_TAGGING_LEVEL_SYNC;
206 }
207 }
208
209 // Returns true if there's an environment setting (either sysprop or env var)
210 // that should overwrite the ELF note, and places the equivalent heap tagging
211 // level into *level.
get_environment_memtag_setting(HeapTaggingLevel * level)212 static bool get_environment_memtag_setting(HeapTaggingLevel* level) {
213 static const char kMemtagPrognameSyspropPrefix[] = "arm64.memtag.process.";
214 static const char kMemtagGlobalSysprop[] = "persist.arm64.memtag.default";
215 static const char kMemtagOverrideSyspropPrefix[] =
216 "persist.device_config.memory_safety_native.mode_override.process.";
217
218 const char* progname = __libc_shared_globals()->init_progname;
219 if (progname == nullptr) return false;
220
221 const char* basename = __gnu_basename(progname);
222
223 char options_str[PROP_VALUE_MAX];
224 char sysprop_name[512];
225 async_safe_format_buffer(sysprop_name, sizeof(sysprop_name), "%s%s", kMemtagPrognameSyspropPrefix,
226 basename);
227 char remote_sysprop_name[512];
228 async_safe_format_buffer(remote_sysprop_name, sizeof(remote_sysprop_name), "%s%s",
229 kMemtagOverrideSyspropPrefix, basename);
230 const char* sys_prop_names[] = {sysprop_name, remote_sysprop_name, kMemtagGlobalSysprop};
231
232 if (!get_config_from_env_or_sysprops("MEMTAG_OPTIONS", sys_prop_names, arraysize(sys_prop_names),
233 options_str, sizeof(options_str))) {
234 return false;
235 }
236
237 if (strcmp("sync", options_str) == 0) {
238 *level = M_HEAP_TAGGING_LEVEL_SYNC;
239 } else if (strcmp("async", options_str) == 0) {
240 *level = M_HEAP_TAGGING_LEVEL_ASYNC;
241 } else if (strcmp("off", options_str) == 0) {
242 *level = M_HEAP_TAGGING_LEVEL_TBI;
243 } else {
244 async_safe_format_log(
245 ANDROID_LOG_ERROR, "libc",
246 "unrecognized memtag level: \"%s\" (options are \"sync\", \"async\", or \"off\").",
247 options_str);
248 return false;
249 }
250
251 return true;
252 }
253
254 // Returns the initial heap tagging level. Note: This function will never return
255 // M_HEAP_TAGGING_LEVEL_NONE, if MTE isn't enabled for this process we enable
256 // M_HEAP_TAGGING_LEVEL_TBI.
__get_tagging_level(const memtag_dynamic_entries_t * memtag_dynamic_entries,const void * phdr_start,size_t phdr_ct,uintptr_t load_bias,bool * stack)257 static HeapTaggingLevel __get_tagging_level(const memtag_dynamic_entries_t* memtag_dynamic_entries,
258 const void* phdr_start, size_t phdr_ct,
259 uintptr_t load_bias, bool* stack) {
260 HeapTaggingLevel level = M_HEAP_TAGGING_LEVEL_TBI;
261
262 // If the dynamic entries exist, use those. Otherwise, fall back to the old
263 // Android note, which is still used for fully static executables. When
264 // -fsanitize=memtag* is used in newer toolchains, currently both the dynamic
265 // entries and the old note are created, but we'd expect to move to just the
266 // dynamic entries for dynamically linked executables in the future. In
267 // addition, there's still some cleanup of the build system (that uses a
268 // manually-constructed note) needed. For more information about the dynamic
269 // entries, see:
270 // https://github.com/ARM-software/abi-aa/blob/main/memtagabielf64/memtagabielf64.rst#dynamic-section
271 if (memtag_dynamic_entries && memtag_dynamic_entries->has_memtag_mode) {
272 switch (memtag_dynamic_entries->memtag_mode) {
273 case 0:
274 level = M_HEAP_TAGGING_LEVEL_SYNC;
275 break;
276 case 1:
277 level = M_HEAP_TAGGING_LEVEL_ASYNC;
278 break;
279 default:
280 async_safe_format_log(ANDROID_LOG_INFO, "libc",
281 "unrecognised DT_AARCH64_MEMTAG_MODE value: %u",
282 memtag_dynamic_entries->memtag_mode);
283 }
284 *stack = memtag_dynamic_entries->memtag_stack;
285 } else {
286 level = __get_memtag_level_from_note(reinterpret_cast<const ElfW(Phdr)*>(phdr_start), phdr_ct,
287 load_bias, stack);
288 }
289
290 // We can't short-circuit the environment override, as `stack` is still inherited from the
291 // binary's settings.
292 if (get_environment_memtag_setting(&level)) {
293 if (level == M_HEAP_TAGGING_LEVEL_NONE || level == M_HEAP_TAGGING_LEVEL_TBI) {
294 *stack = false;
295 }
296 }
297 return level;
298 }
299
__get_memtag_upgrade_secs()300 static int64_t __get_memtag_upgrade_secs() {
301 char* env = getenv("BIONIC_MEMTAG_UPGRADE_SECS");
302 if (!env) return 0;
303 int64_t timed_upgrade = 0;
304 static const char kAppProcessName[] = "app_process64";
305 const char* progname = __libc_shared_globals()->init_progname;
306 progname = progname ? __gnu_basename(progname) : nullptr;
307 // disable timed upgrade for zygote, as the thread spawned will violate the requirement
308 // that it be single-threaded.
309 if (!progname || strncmp(progname, kAppProcessName, sizeof(kAppProcessName)) != 0) {
310 char* endptr;
311 timed_upgrade = strtoll(env, &endptr, 10);
312 if (*endptr != '\0' || timed_upgrade < 0) {
313 async_safe_format_log(ANDROID_LOG_ERROR, "libc",
314 "Invalid value for BIONIC_MEMTAG_UPGRADE_SECS: %s", env);
315 timed_upgrade = 0;
316 }
317 }
318 // Make sure that this does not get passed to potential processes inheriting
319 // this environment.
320 unsetenv("BIONIC_MEMTAG_UPGRADE_SECS");
321 return timed_upgrade;
322 }
323
324 // Figure out the desired memory tagging mode (sync/async, heap/globals/stack) for this executable.
325 // This function is called from the linker before the main executable is relocated.
__libc_init_mte(const memtag_dynamic_entries_t * memtag_dynamic_entries,const void * phdr_start,size_t phdr_ct,uintptr_t load_bias,void * stack_top)326 __attribute__((no_sanitize("hwaddress", "memtag"))) void __libc_init_mte(
327 const memtag_dynamic_entries_t* memtag_dynamic_entries, const void* phdr_start, size_t phdr_ct,
328 uintptr_t load_bias, void* stack_top) {
329 bool memtag_stack = false;
330 HeapTaggingLevel level =
331 __get_tagging_level(memtag_dynamic_entries, phdr_start, phdr_ct, load_bias, &memtag_stack);
332 // This is used by the linker (in linker.cpp) to communicate than any library linked by this
333 // executable enables memtag-stack.
334 if (__libc_shared_globals()->initial_memtag_stack) {
335 if (!memtag_stack) {
336 async_safe_format_log(ANDROID_LOG_INFO, "libc", "enabling PROT_MTE as requested by linker");
337 }
338 memtag_stack = true;
339 }
340 if (int64_t timed_upgrade = __get_memtag_upgrade_secs()) {
341 if (level == M_HEAP_TAGGING_LEVEL_ASYNC) {
342 async_safe_format_log(ANDROID_LOG_INFO, "libc",
343 "Attempting timed MTE upgrade from async to sync.");
344 __libc_shared_globals()->heap_tagging_upgrade_timer_sec = timed_upgrade;
345 level = M_HEAP_TAGGING_LEVEL_SYNC;
346 } else if (level != M_HEAP_TAGGING_LEVEL_SYNC) {
347 async_safe_format_log(
348 ANDROID_LOG_ERROR, "libc",
349 "Requested timed MTE upgrade from invalid %s to sync. Ignoring.",
350 DescribeTaggingLevel(level));
351 }
352 }
353 if (level == M_HEAP_TAGGING_LEVEL_SYNC || level == M_HEAP_TAGGING_LEVEL_ASYNC) {
354 unsigned long prctl_arg = PR_TAGGED_ADDR_ENABLE | PR_MTE_TAG_SET_NONZERO;
355 prctl_arg |= (level == M_HEAP_TAGGING_LEVEL_SYNC) ? PR_MTE_TCF_SYNC : PR_MTE_TCF_ASYNC;
356
357 // When entering ASYNC mode, specify that we want to allow upgrading to SYNC by OR'ing in the
358 // SYNC flag. But if the kernel doesn't support specifying multiple TCF modes, fall back to
359 // specifying a single mode.
360 if (prctl(PR_SET_TAGGED_ADDR_CTRL, prctl_arg | PR_MTE_TCF_SYNC, 0, 0, 0) == 0 ||
361 prctl(PR_SET_TAGGED_ADDR_CTRL, prctl_arg, 0, 0, 0) == 0) {
362 __libc_shared_globals()->initial_heap_tagging_level = level;
363 __libc_shared_globals()->initial_memtag_stack = memtag_stack;
364
365 if (memtag_stack) {
366 void* pg_start =
367 reinterpret_cast<void*>(page_start(reinterpret_cast<uintptr_t>(stack_top)));
368 if (mprotect(pg_start, page_size(), PROT_READ | PROT_WRITE | PROT_MTE | PROT_GROWSDOWN)) {
369 async_safe_fatal("error: failed to set PROT_MTE on main thread stack: %m");
370 }
371 }
372
373 return;
374 }
375 }
376
377 // MTE was either not enabled, or wasn't supported on this device. Try and use
378 // TBI.
379 if (prctl(PR_SET_TAGGED_ADDR_CTRL, PR_TAGGED_ADDR_ENABLE, 0, 0, 0) == 0) {
380 __libc_shared_globals()->initial_heap_tagging_level = M_HEAP_TAGGING_LEVEL_TBI;
381 }
382 // We did not enable MTE, so we do not need to arm the upgrade timer.
383 __libc_shared_globals()->heap_tagging_upgrade_timer_sec = 0;
384 // We also didn't enable memtag_stack.
385 __libc_shared_globals()->initial_memtag_stack = false;
386 }
387 #else // __aarch64__
__libc_init_mte(const memtag_dynamic_entries_t *,const void *,size_t,uintptr_t,void *)388 void __libc_init_mte(const memtag_dynamic_entries_t*, const void*, size_t, uintptr_t, void*) {}
389 #endif // __aarch64__
390
__libc_init_profiling_handlers()391 void __libc_init_profiling_handlers() {
392 // The dynamic variant of this function is more interesting, but this
393 // at least ensures that static binaries aren't killed by the kernel's
394 // default disposition for these two real-time signals that would have
395 // handlers installed if this was a dynamic binary.
396 signal(BIONIC_SIGNAL_PROFILER, SIG_IGN);
397 signal(BIONIC_SIGNAL_ART_PROFILER, SIG_IGN);
398 }
399
__real_libc_init(void * raw_args,void (* onexit)(void)__unused,int (* slingshot)(int,char **,char **),structors_array_t const * const structors,bionic_tcb * temp_tcb)400 __attribute__((no_sanitize("memtag"))) __noreturn static void __real_libc_init(
401 void* raw_args, void (*onexit)(void) __unused, int (*slingshot)(int, char**, char**),
402 structors_array_t const* const structors, bionic_tcb* temp_tcb) {
403 BIONIC_STOP_UNWIND;
404
405 // Initialize TLS early so system calls and errno work.
406 KernelArgumentBlock args(raw_args);
407 __libc_init_main_thread_early(args, temp_tcb);
408 __libc_init_main_thread_late();
409 __libc_init_globals();
410 __libc_shared_globals()->init_progname = args.argv[0];
411 __libc_init_AT_SECURE(args.envp);
412 layout_static_tls(args);
413 __libc_init_main_thread_final();
414 __libc_init_common();
415 __libc_init_mte(/*memtag_dynamic_entries=*/nullptr,
416 reinterpret_cast<ElfW(Phdr)*>(getauxval(AT_PHDR)), getauxval(AT_PHNUM),
417 /*load_bias = */ 0, /*stack_top = */ raw_args);
418 __libc_init_scudo();
419 __libc_init_profiling_handlers();
420 __libc_init_fork_handler();
421
422 call_ifunc_resolvers();
423 apply_gnu_relro();
424
425 // Several Linux ABIs don't pass the onexit pointer, and the ones that
426 // do never use it. Therefore, we ignore it.
427
428 call_array(structors->preinit_array, structors->preinit_array_count, args.argc, args.argv,
429 args.envp);
430 call_array(structors->init_array, structors->init_array_count, args.argc, args.argv, args.envp);
431
432 // The executable may have its own destructors listed in its .fini_array
433 // so we need to ensure that these are called when the program exits
434 // normally.
435 if (structors->fini_array_count > 0) {
436 __cxa_atexit(call_fini_array, const_cast<structors_array_t*>(structors), nullptr);
437 }
438
439 __libc_init_mte_late();
440
441 exit(slingshot(args.argc, args.argv, args.envp));
442 }
443
444 extern "C" void __hwasan_init_static();
445
446 // This __libc_init() is only used for static executables, and is called from crtbegin.c.
447 //
448 // The 'structors' parameter contains pointers to various initializer
449 // arrays that must be run before the program's 'main' routine is launched.
__libc_init(void * raw_args,void (* onexit)(void)__unused,int (* slingshot)(int,char **,char **),structors_array_t const * const structors)450 __attribute__((no_sanitize("hwaddress", "memtag"))) __noreturn void __libc_init(
451 void* raw_args, void (*onexit)(void) __unused, int (*slingshot)(int, char**, char**),
452 structors_array_t const* const structors) {
453 bionic_tcb temp_tcb = {};
454 #if __has_feature(hwaddress_sanitizer)
455 // Install main thread TLS early. It will be initialized later in __libc_init_main_thread. For now
456 // all we need is access to TLS_SLOT_SANITIZER.
457 __set_tls(&temp_tcb.tls_slot(0));
458 // Initialize HWASan enough to run instrumented code. This sets up TLS_SLOT_SANITIZER, among other
459 // things.
460 __hwasan_init_static();
461 // We are ready to run HWASan-instrumented code, proceed with libc initialization...
462 #endif
463 __real_libc_init(raw_args, onexit, slingshot, structors, &temp_tcb);
464 }
465
466 static int g_target_sdk_version{__ANDROID_API__};
467
android_get_application_target_sdk_version()468 extern "C" int android_get_application_target_sdk_version() {
469 return g_target_sdk_version;
470 }
471
android_set_application_target_sdk_version(int target)472 extern "C" void android_set_application_target_sdk_version(int target) {
473 g_target_sdk_version = target;
474 __libc_set_target_sdk_version(target);
475 }
476
477 // This function is called in the dynamic linker before ifunc resolvers have run, so this file is
478 // compiled with -ffreestanding to avoid implicit string.h function calls. (It shouldn't strictly
479 // be necessary, though.)
__libc_shared_globals()480 __LIBC_HIDDEN__ libc_shared_globals* __libc_shared_globals() {
481 static libc_shared_globals globals;
482 return &globals;
483 }
484