elf: Always call destructors in reverse constructor order (bug 30785)

The current implementation of dlclose (and process exit) re-sorts the
link maps before calling ELF destructors.  Destructor order is not the
reverse of the constructor order as a result: The second sort takes
relocation dependencies into account, and other differences can result
from ambiguous inputs, such as cycles.  (The force_first handling in
_dl_sort_maps is not effective for dlclose.)  After the changes in
this commit, there is still a required difference due to
dlopen/dlclose ordering by the application, but the previous
discrepancies went beyond that.

A new global (namespace-spanning) list of link maps,
_dl_init_called_list, is updated right before ELF constructors are
called from _dl_init.

In dl_close_worker, the maps variable, an on-stack variable length
array, is eliminated.  (VLAs are problematic, and dlclose should not
call malloc because it cannot readily deal with malloc failure.)
Marking still-used objects uses the namespace list directly, with
next and next_idx replacing the done_index variable.

After marking, _dl_init_called_list is used to call the destructors
of now-unused maps in reverse destructor order.  These destructors
can call dlopen.  Previously, new objects do not have l_map_used set.
This had to change: There is no copy of the link map list anymore,
so processing would cover newly opened (and unmarked) mappings,
unloading them.  Now, _dl_init (indirectly) sets l_map_used, too.
(dlclose is handled by the existing reentrancy guard.)

After _dl_init_called_list traversal, two more loops follow.  The
processing order changes to the original link map order in the
namespace.  Previously, dependency order was used.  The difference
should not matter because relocation dependencies could already
reorder link maps in the old code.

The changes to _dl_fini remove the sorting step and replace it with
a traversal of _dl_init_called_list.  The l_direct_opencount
decrement outside the loader lock is removed because it appears
incorrect: the counter manipulation could race with other dynamic
loader operations.

tst-audit23 needs adjustments to the changes in LA_ACT_DELETE
notifications.  The new approach for checking la_activity should
make it clearer that la_activty calls come in pairs around namespace
updates.

The dependency sorting test cases need updates because the destructor
order is always the opposite order of constructor order, even with
relocation dependencies or cycles present.

There is a future cleanup opportunity to remove the now-constant
force_first and for_fini arguments from the _dl_sort_maps function.

Fixes commit 1df71d32fe ("elf: Implement
force_first handling in _dl_sort_maps_dfs (bug 28937)").

Reviewed-by: DJ Delorie <dj@redhat.com>
(cherry picked from commit 6985865bc3)
This commit is contained in:
Florian Weimer 2023-09-08 12:32:14 +02:00
parent 7ae211a01b
commit a3189f66a5
8 changed files with 184 additions and 171 deletions

7
NEWS
View File

@ -4,6 +4,13 @@ See the end for copying conditions.
Please send GNU C library bug reports via <https://sourceware.org/bugzilla/> Please send GNU C library bug reports via <https://sourceware.org/bugzilla/>
using `glibc' in the "product" field. using `glibc' in the "product" field.
Version 2.38.1
The following bugs are resolved with this release:
[30785] Always call destructors in reverse constructor order
Version 2.38 Version 2.38

View File

@ -138,30 +138,31 @@ _dl_close_worker (struct link_map *map, bool force)
bool any_tls = false; bool any_tls = false;
const unsigned int nloaded = ns->_ns_nloaded; const unsigned int nloaded = ns->_ns_nloaded;
struct link_map *maps[nloaded];
/* Run over the list and assign indexes to the link maps and enter /* Run over the list and assign indexes to the link maps. */
them into the MAPS array. */
int idx = 0; int idx = 0;
for (struct link_map *l = ns->_ns_loaded; l != NULL; l = l->l_next) for (struct link_map *l = ns->_ns_loaded; l != NULL; l = l->l_next)
{ {
l->l_map_used = 0; l->l_map_used = 0;
l->l_map_done = 0; l->l_map_done = 0;
l->l_idx = idx; l->l_idx = idx;
maps[idx] = l;
++idx; ++idx;
} }
assert (idx == nloaded); assert (idx == nloaded);
/* Keep track of the lowest index link map we have covered already. */ /* Keep marking link maps until no new link maps are found. */
int done_index = -1; for (struct link_map *l = ns->_ns_loaded; l != NULL; )
while (++done_index < nloaded)
{ {
struct link_map *l = maps[done_index]; /* next is reset to earlier link maps for remarking. */
struct link_map *next = l->l_next;
int next_idx = l->l_idx + 1; /* next->l_idx, but covers next == NULL. */
if (l->l_map_done) if (l->l_map_done)
/* Already handled. */ {
continue; /* Already handled. */
l = next;
continue;
}
/* Check whether this object is still used. */ /* Check whether this object is still used. */
if (l->l_type == lt_loaded if (l->l_type == lt_loaded
@ -171,7 +172,10 @@ _dl_close_worker (struct link_map *map, bool force)
acquire is sufficient and correct. */ acquire is sufficient and correct. */
&& atomic_load_acquire (&l->l_tls_dtor_count) == 0 && atomic_load_acquire (&l->l_tls_dtor_count) == 0
&& !l->l_map_used) && !l->l_map_used)
continue; {
l = next;
continue;
}
/* We need this object and we handle it now. */ /* We need this object and we handle it now. */
l->l_map_used = 1; l->l_map_used = 1;
@ -198,8 +202,11 @@ _dl_close_worker (struct link_map *map, bool force)
already processed it, then we need to go back already processed it, then we need to go back
and process again from that point forward to and process again from that point forward to
ensure we keep all of its dependencies also. */ ensure we keep all of its dependencies also. */
if ((*lp)->l_idx - 1 < done_index) if ((*lp)->l_idx < next_idx)
done_index = (*lp)->l_idx - 1; {
next = *lp;
next_idx = next->l_idx;
}
} }
} }
@ -219,44 +226,65 @@ _dl_close_worker (struct link_map *map, bool force)
if (!jmap->l_map_used) if (!jmap->l_map_used)
{ {
jmap->l_map_used = 1; jmap->l_map_used = 1;
if (jmap->l_idx - 1 < done_index) if (jmap->l_idx < next_idx)
done_index = jmap->l_idx - 1; {
next = jmap;
next_idx = next->l_idx;
}
} }
} }
} }
l = next;
} }
/* Sort the entries. We can skip looking for the binary itself which is /* Call the destructors in reverse constructor order, and remove the
at the front of the search list for the main namespace. */ closed link maps from the list. */
_dl_sort_maps (maps, nloaded, (nsid == LM_ID_BASE), true); for (struct link_map **init_called_head = &_dl_init_called_list;
*init_called_head != NULL; )
/* Call all termination functions at once. */
bool unload_any = false;
bool scope_mem_left = false;
unsigned int unload_global = 0;
unsigned int first_loaded = ~0;
for (unsigned int i = 0; i < nloaded; ++i)
{ {
struct link_map *imap = maps[i]; struct link_map *imap = *init_called_head;
/* All elements must be in the same namespace. */ /* _dl_init_called_list is global, to produce a global odering.
assert (imap->l_ns == nsid); Ignore the other namespaces (and link maps that are still used). */
if (imap->l_ns != nsid || imap->l_map_used)
if (!imap->l_map_used) init_called_head = &imap->l_init_called_next;
else
{ {
assert (imap->l_type == lt_loaded && !imap->l_nodelete_active); assert (imap->l_type == lt_loaded && !imap->l_nodelete_active);
/* Call its termination function. Do not do it for /* _dl_init_called_list is updated at the same time as
half-cooked objects. Temporarily disable exception l_init_called. */
handling, so that errors are fatal. */ assert (imap->l_init_called);
if (imap->l_init_called)
if (imap->l_info[DT_FINI_ARRAY] != NULL
|| imap->l_info[DT_FINI] != NULL)
_dl_catch_exception (NULL, _dl_call_fini, imap); _dl_catch_exception (NULL, _dl_call_fini, imap);
#ifdef SHARED #ifdef SHARED
/* Auditing checkpoint: we remove an object. */ /* Auditing checkpoint: we remove an object. */
_dl_audit_objclose (imap); _dl_audit_objclose (imap);
#endif #endif
/* Unlink this link map. */
*init_called_head = imap->l_init_called_next;
}
}
bool unload_any = false;
bool scope_mem_left = false;
unsigned int unload_global = 0;
/* For skipping un-unloadable link maps in the second loop. */
struct link_map *first_loaded = ns->_ns_loaded;
/* Iterate over the namespace to find objects to unload. Some
unloadable objects may not be on _dl_init_called_list due to
dlopen failure. */
for (struct link_map *imap = first_loaded; imap != NULL; imap = imap->l_next)
{
if (!imap->l_map_used)
{
/* This object must not be used anymore. */ /* This object must not be used anymore. */
imap->l_removed = 1; imap->l_removed = 1;
@ -267,8 +295,8 @@ _dl_close_worker (struct link_map *map, bool force)
++unload_global; ++unload_global;
/* Remember where the first dynamically loaded object is. */ /* Remember where the first dynamically loaded object is. */
if (i < first_loaded) if (first_loaded == NULL)
first_loaded = i; first_loaded = imap;
} }
/* Else imap->l_map_used. */ /* Else imap->l_map_used. */
else if (imap->l_type == lt_loaded) else if (imap->l_type == lt_loaded)
@ -404,8 +432,8 @@ _dl_close_worker (struct link_map *map, bool force)
imap->l_loader = NULL; imap->l_loader = NULL;
/* Remember where the first dynamically loaded object is. */ /* Remember where the first dynamically loaded object is. */
if (i < first_loaded) if (first_loaded == NULL)
first_loaded = i; first_loaded = imap;
} }
} }
@ -476,10 +504,11 @@ _dl_close_worker (struct link_map *map, bool force)
/* Check each element of the search list to see if all references to /* Check each element of the search list to see if all references to
it are gone. */ it are gone. */
for (unsigned int i = first_loaded; i < nloaded; ++i) for (struct link_map *imap = first_loaded; imap != NULL; )
{ {
struct link_map *imap = maps[i]; if (imap->l_map_used)
if (!imap->l_map_used) imap = imap->l_next;
else
{ {
assert (imap->l_type == lt_loaded); assert (imap->l_type == lt_loaded);
@ -690,7 +719,9 @@ _dl_close_worker (struct link_map *map, bool force)
if (imap == GL(dl_initfirst)) if (imap == GL(dl_initfirst))
GL(dl_initfirst) = NULL; GL(dl_initfirst) = NULL;
struct link_map *next = imap->l_next;
free (imap); free (imap);
imap = next;
} }
} }

View File

@ -24,116 +24,68 @@
void void
_dl_fini (void) _dl_fini (void)
{ {
/* Lots of fun ahead. We have to call the destructors for all still /* Call destructors strictly in the reverse order of constructors.
loaded objects, in all namespaces. The problem is that the ELF This causes fewer surprises than some arbitrary reordering based
specification now demands that dependencies between the modules on new (relocation) dependencies. None of the objects are
are taken into account. I.e., the destructor for a module is unmapped, so applications can deal with this if their DSOs remain
called before the ones for any of its dependencies. in a consistent state after destructors have run. */
To make things more complicated, we cannot simply use the reverse /* Protect against concurrent loads and unloads. */
order of the constructors. Since the user might have loaded objects __rtld_lock_lock_recursive (GL(dl_load_lock));
using `dlopen' there are possibly several other modules with its
dependencies to be taken into account. Therefore we have to start
determining the order of the modules once again from the beginning. */
/* We run the destructors of the main namespaces last. As for the /* Ignore objects which are opened during shutdown. */
other namespaces, we pick run the destructors in them in reverse struct link_map *local_init_called_list = _dl_init_called_list;
order of the namespace ID. */
for (struct link_map *l = local_init_called_list; l != NULL;
l = l->l_init_called_next)
/* Bump l_direct_opencount of all objects so that they
are not dlclose()ed from underneath us. */
++l->l_direct_opencount;
/* After this point, everything linked from local_init_called_list
cannot be unloaded because of the reference counter update. */
__rtld_lock_unlock_recursive (GL(dl_load_lock));
/* Perform two passes: One for non-audit modules, one for audit
modules. This way, audit modules receive unload notifications
for non-audit objects, and the destructors for audit modules
still run. */
#ifdef SHARED #ifdef SHARED
int do_audit = 0; int last_pass = GLRO(dl_naudit) > 0;
again: Lmid_t last_ns = -1;
for (int do_audit = 0; do_audit <= last_pass; ++do_audit)
#endif #endif
for (Lmid_t ns = GL(dl_nns) - 1; ns >= 0; --ns) for (struct link_map *l = local_init_called_list; l != NULL;
{ l = l->l_init_called_next)
/* Protect against concurrent loads and unloads. */ {
__rtld_lock_lock_recursive (GL(dl_load_lock)); #ifdef SHARED
if (GL(dl_ns)[l->l_ns]._ns_loaded->l_auditing != do_audit)
continue;
unsigned int nloaded = GL(dl_ns)[ns]._ns_nloaded; /* Avoid back-to-back calls of _dl_audit_activity_nsid for the
/* No need to do anything for empty namespaces or those used for same namespace. */
auditing DSOs. */ if (last_ns != l->l_ns)
if (nloaded == 0 {
#ifdef SHARED if (last_ns >= 0)
|| GL(dl_ns)[ns]._ns_loaded->l_auditing != do_audit _dl_audit_activity_nsid (last_ns, LA_ACT_CONSISTENT);
#endif _dl_audit_activity_nsid (l->l_ns, LA_ACT_DELETE);
) last_ns = l->l_ns;
__rtld_lock_unlock_recursive (GL(dl_load_lock)); }
else
{
#ifdef SHARED
_dl_audit_activity_nsid (ns, LA_ACT_DELETE);
#endif #endif
/* Now we can allocate an array to hold all the pointers and /* There is no need to re-enable exceptions because _dl_fini
copy the pointers in. */ is not called from a context where exceptions are caught. */
struct link_map *maps[nloaded]; _dl_call_fini (l);
unsigned int i;
struct link_map *l;
assert (nloaded != 0 || GL(dl_ns)[ns]._ns_loaded == NULL);
for (l = GL(dl_ns)[ns]._ns_loaded, i = 0; l != NULL; l = l->l_next)
/* Do not handle ld.so in secondary namespaces. */
if (l == l->l_real)
{
assert (i < nloaded);
maps[i] = l;
l->l_idx = i;
++i;
/* Bump l_direct_opencount of all objects so that they
are not dlclose()ed from underneath us. */
++l->l_direct_opencount;
}
assert (ns != LM_ID_BASE || i == nloaded);
assert (ns == LM_ID_BASE || i == nloaded || i == nloaded - 1);
unsigned int nmaps = i;
/* Now we have to do the sorting. We can skip looking for the
binary itself which is at the front of the search list for
the main namespace. */
_dl_sort_maps (maps, nmaps, (ns == LM_ID_BASE), true);
/* We do not rely on the linked list of loaded object anymore
from this point on. We have our own list here (maps). The
various members of this list cannot vanish since the open
count is too high and will be decremented in this loop. So
we release the lock so that some code which might be called
from a destructor can directly or indirectly access the
lock. */
__rtld_lock_unlock_recursive (GL(dl_load_lock));
/* 'maps' now contains the objects in the right order. Now
call the destructors. We have to process this array from
the front. */
for (i = 0; i < nmaps; ++i)
{
struct link_map *l = maps[i];
if (l->l_init_called)
{
_dl_call_fini (l);
#ifdef SHARED #ifdef SHARED
/* Auditing checkpoint: another object closed. */ /* Auditing checkpoint: another object closed. */
_dl_audit_objclose (l); _dl_audit_objclose (l);
#endif #endif
} }
/* Correct the previous increment. */
--l->l_direct_opencount;
}
#ifdef SHARED #ifdef SHARED
_dl_audit_activity_nsid (ns, LA_ACT_CONSISTENT); if (last_ns >= 0)
#endif _dl_audit_activity_nsid (last_ns, LA_ACT_CONSISTENT);
}
}
#ifdef SHARED
if (! do_audit && GLRO(dl_naudit) > 0)
{
do_audit = 1;
goto again;
}
if (__glibc_unlikely (GLRO(dl_debug_mask) & DL_DEBUG_STATISTICS)) if (__glibc_unlikely (GLRO(dl_debug_mask) & DL_DEBUG_STATISTICS))
_dl_debug_printf ("\nruntime linker statistics:\n" _dl_debug_printf ("\nruntime linker statistics:\n"

View File

@ -21,6 +21,7 @@
#include <ldsodefs.h> #include <ldsodefs.h>
#include <elf-initfini.h> #include <elf-initfini.h>
struct link_map *_dl_init_called_list;
static void static void
call_init (struct link_map *l, int argc, char **argv, char **env) call_init (struct link_map *l, int argc, char **argv, char **env)
@ -42,6 +43,21 @@ call_init (struct link_map *l, int argc, char **argv, char **env)
dependency. */ dependency. */
l->l_init_called = 1; l->l_init_called = 1;
/* Help an already-running dlclose: The just-loaded object must not
be removed during the current pass. (No effect if no dlclose in
progress.) */
l->l_map_used = 1;
/* Record execution before starting any initializers. This way, if
the initializers themselves call dlopen, their ELF destructors
will eventually be run before this object is destructed, matching
that their ELF constructors have run before this object was
constructed. _dl_fini uses this list for audit callbacks, so
register objects on the list even if they do not have a
constructor. */
l->l_init_called_next = _dl_init_called_list;
_dl_init_called_list = l;
/* Check for object which constructors we do not run here. */ /* Check for object which constructors we do not run here. */
if (__builtin_expect (l->l_name[0], 'a') == '\0' if (__builtin_expect (l->l_name[0], 'a') == '\0'
&& l->l_type == lt_executable) && l->l_type == lt_executable)

View File

@ -53,21 +53,14 @@ tst-dso-ordering10: {}->a->b->c;soname({})=c
output: b>a>{}<a<b output: b>a>{}<a<b
# Complex example from Bugzilla #15311, under-linked and with circular # Complex example from Bugzilla #15311, under-linked and with circular
# relocation(dynamic) dependencies. While this is technically unspecified, the # relocation(dynamic) dependencies. For both sorting algorithms, the
# presumed reasonable practical behavior is for the destructor order to respect # destruction order is the reverse of the construction order, and
# the static DT_NEEDED links (here this means the a->b->c->d order). # relocation dependencies are not taken into account.
# The older dynamic_sort=1 algorithm does not achieve this, while the DFS-based
# dynamic_sort=2 algorithm does, although it is still arguable whether going
# beyond spec to do this is the right thing to do.
# The below expected outputs are what the two algorithms currently produce
# respectively, for regression testing purposes.
tst-bz15311: {+a;+e;+f;+g;+d;%d;-d;-g;-f;-e;-a};a->b->c->d;d=>[ba];c=>a;b=>e=>a;c=>f=>b;d=>g=>c tst-bz15311: {+a;+e;+f;+g;+d;%d;-d;-g;-f;-e;-a};a->b->c->d;d=>[ba];c=>a;b=>e=>a;c=>f=>b;d=>g=>c
output(glibc.rtld.dynamic_sort=1): {+a[d>c>b>a>];+e[e>];+f[f>];+g[g>];+d[];%d(b(e(a()))a()g(c(a()f(b(e(a()))))));-d[];-g[];-f[];-e[];-a[<a<c<d<g<f<b<e];} output: {+a[d>c>b>a>];+e[e>];+f[f>];+g[g>];+d[];%d(b(e(a()))a()g(c(a()f(b(e(a()))))));-d[];-g[];-f[];-e[];-a[<g<f<e<a<b<c<d];}
output(glibc.rtld.dynamic_sort=2): {+a[d>c>b>a>];+e[e>];+f[f>];+g[g>];+d[];%d(b(e(a()))a()g(c(a()f(b(e(a()))))));-d[];-g[];-f[];-e[];-a[<g<f<a<b<c<d<e];}
# Test that even in the presence of dependency loops involving dlopen'ed # Test that even in the presence of dependency loops involving dlopen'ed
# object, that object is initialized last (and not unloaded prematurely). # object, that object is initialized last (and not unloaded prematurely).
# Final destructor order is indeterminate due to the cycle. # Final destructor order is the opposite of constructor order.
tst-bz28937: {+a;+b;-b;+c;%c};a->a1;a->a2;a2->a;b->b1;c->a1;c=>a1 tst-bz28937: {+a;+b;-b;+c;%c};a->a1;a->a2;a2->a;b->b1;c->a1;c=>a1
output(glibc.rtld.dynamic_sort=1): {+a[a2>a1>a>];+b[b1>b>];-b[<b<b1];+c[c>];%c(a1());}<a<a2<c<a1 output: {+a[a2>a1>a>];+b[b1>b>];-b[<b<b1];+c[c>];%c(a1());}<c<a<a1<a2
output(glibc.rtld.dynamic_sort=2): {+a[a2>a1>a>];+b[b1>b>];-b[<b<b1];+c[c>];%c(a1());}<a2<a<c<a1

View File

@ -98,6 +98,8 @@ do_test (int argc, char *argv[])
char *lname; char *lname;
uintptr_t laddr; uintptr_t laddr;
Lmid_t lmid; Lmid_t lmid;
uintptr_t cookie;
uintptr_t namespace;
bool closed; bool closed;
} objs[max_objs] = { [0 ... max_objs-1] = { .closed = false } }; } objs[max_objs] = { [0 ... max_objs-1] = { .closed = false } };
size_t nobjs = 0; size_t nobjs = 0;
@ -117,6 +119,9 @@ do_test (int argc, char *argv[])
size_t buffer_length = 0; size_t buffer_length = 0;
while (xgetline (&buffer, &buffer_length, out)) while (xgetline (&buffer, &buffer_length, out))
{ {
*strchrnul (buffer, '\n') = '\0';
printf ("info: subprocess output: %s\n", buffer);
if (startswith (buffer, "la_activity: ")) if (startswith (buffer, "la_activity: "))
{ {
uintptr_t cookie; uintptr_t cookie;
@ -125,29 +130,26 @@ do_test (int argc, char *argv[])
&cookie); &cookie);
TEST_COMPARE (r, 2); TEST_COMPARE (r, 2);
/* The cookie identifies the object at the head of the link map,
so we only add a new namespace if it changes from the previous
one. This works since dlmopen is the last in the test body. */
if (cookie != last_act_cookie && last_act_cookie != -1)
TEST_COMPARE (last_act, LA_ACT_CONSISTENT);
if (this_act == LA_ACT_ADD && acts[nacts] != cookie) if (this_act == LA_ACT_ADD && acts[nacts] != cookie)
{ {
/* The cookie identifies the object at the head of the
link map, so we only add a new namespace if it
changes from the previous one. This works since
dlmopen is the last in the test body. */
if (cookie != last_act_cookie && last_act_cookie != -1)
TEST_COMPARE (last_act, LA_ACT_CONSISTENT);
acts[nacts++] = cookie; acts[nacts++] = cookie;
last_act_cookie = cookie; last_act_cookie = cookie;
} }
/* The LA_ACT_DELETE is called in the reverse order of LA_ACT_ADD /* LA_ACT_DELETE is called multiple times for each
at program termination (if the tests adds a dlclose or a library namespace, depending on destruction order. */
with extra dependencies this will need to be adapted). */
else if (this_act == LA_ACT_DELETE) else if (this_act == LA_ACT_DELETE)
{ last_act_cookie = cookie;
last_act_cookie = acts[--nacts];
TEST_COMPARE (acts[nacts], cookie);
acts[nacts] = 0;
}
else if (this_act == LA_ACT_CONSISTENT) else if (this_act == LA_ACT_CONSISTENT)
{ {
TEST_COMPARE (cookie, last_act_cookie); TEST_COMPARE (cookie, last_act_cookie);
last_act_cookie = -1;
/* LA_ACT_DELETE must always be followed by an la_objclose. */ /* LA_ACT_DELETE must always be followed by an la_objclose. */
if (last_act == LA_ACT_DELETE) if (last_act == LA_ACT_DELETE)
@ -179,6 +181,8 @@ do_test (int argc, char *argv[])
objs[nobjs].lname = lname; objs[nobjs].lname = lname;
objs[nobjs].laddr = laddr; objs[nobjs].laddr = laddr;
objs[nobjs].lmid = lmid; objs[nobjs].lmid = lmid;
objs[nobjs].cookie = cookie;
objs[nobjs].namespace = last_act_cookie;
objs[nobjs].closed = false; objs[nobjs].closed = false;
nobjs++; nobjs++;
@ -201,6 +205,12 @@ do_test (int argc, char *argv[])
if (strcmp (lname, objs[i].lname) == 0 && lmid == objs[i].lmid) if (strcmp (lname, objs[i].lname) == 0 && lmid == objs[i].lmid)
{ {
TEST_COMPARE (objs[i].closed, false); TEST_COMPARE (objs[i].closed, false);
TEST_COMPARE (objs[i].cookie, cookie);
if (objs[i].namespace == -1)
/* No LA_ACT_ADD before the first la_objopen call. */
TEST_COMPARE (acts[0], last_act_cookie);
else
TEST_COMPARE (objs[i].namespace, last_act_cookie);
objs[i].closed = true; objs[i].closed = true;
break; break;
} }
@ -209,11 +219,7 @@ do_test (int argc, char *argv[])
/* la_objclose should be called after la_activity(LA_ACT_DELETE) for /* la_objclose should be called after la_activity(LA_ACT_DELETE) for
the closed object's namespace. */ the closed object's namespace. */
TEST_COMPARE (last_act, LA_ACT_DELETE); TEST_COMPARE (last_act, LA_ACT_DELETE);
if (!seen_first_objclose) seen_first_objclose = true;
{
TEST_COMPARE (last_act_cookie, cookie);
seen_first_objclose = true;
}
} }
} }

View File

@ -278,6 +278,10 @@ struct link_map
/* List of object in order of the init and fini calls. */ /* List of object in order of the init and fini calls. */
struct link_map **l_initfini; struct link_map **l_initfini;
/* Linked list of objects in reverse ELF constructor execution
order. Head of list is stored in _dl_init_called_list. */
struct link_map *l_init_called_next;
/* List of the dependencies introduced through symbol binding. */ /* List of the dependencies introduced through symbol binding. */
struct link_map_reldeps struct link_map_reldeps
{ {

View File

@ -1037,6 +1037,10 @@ extern int _dl_check_map_versions (struct link_map *map, int verbose,
extern void _dl_init (struct link_map *main_map, int argc, char **argv, extern void _dl_init (struct link_map *main_map, int argc, char **argv,
char **env) attribute_hidden; char **env) attribute_hidden;
/* List of ELF objects in reverse order of their constructor
invocation. */
extern struct link_map *_dl_init_called_list attribute_hidden;
/* Call the finalizer functions of all shared objects whose /* Call the finalizer functions of all shared objects whose
initializer functions have completed. */ initializer functions have completed. */
extern void _dl_fini (void) attribute_hidden; extern void _dl_fini (void) attribute_hidden;