mirror of
https://sourceware.org/git/glibc.git
synced 2024-11-23 05:20:06 +00:00
265 lines
9.0 KiB
C
265 lines
9.0 KiB
C
/* Map in a shared object's segments. NaCl version.
|
|
Copyright (C) 2015-2016 Free Software Foundation, Inc.
|
|
This file is part of the GNU C Library.
|
|
|
|
The GNU C Library is free software; you can redistribute it and/or
|
|
modify it under the terms of the GNU Lesser General Public
|
|
License as published by the Free Software Foundation; either
|
|
version 2.1 of the License, or (at your option) any later version.
|
|
|
|
The GNU C Library is distributed in the hope that it will be useful,
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
Lesser General Public License for more details.
|
|
|
|
You should have received a copy of the GNU Lesser General Public
|
|
License along with the GNU C Library; if not, see
|
|
<http://www.gnu.org/licenses/>. */
|
|
|
|
#include <assert.h>
|
|
#include <dl-load.h>
|
|
#include <errno.h>
|
|
#include <stdbool.h>
|
|
#include <unistd.h>
|
|
#include <libc-internal.h>
|
|
|
|
|
|
/* This is basically pread, but with iteration after short reads. */
|
|
static bool
|
|
read_in_data (int fd, void *data, size_t len, off_t pos)
|
|
{
|
|
if (__glibc_unlikely (__lseek (fd, pos, SEEK_SET) == (off_t) -1))
|
|
return true;
|
|
while (len > 0)
|
|
{
|
|
ssize_t n = __read (fd, data, len);
|
|
if (__glibc_unlikely (n < 0))
|
|
return true;
|
|
if (__glibc_unlikely (n == 0))
|
|
{
|
|
errno = EFTYPE;
|
|
return true;
|
|
}
|
|
data += n;
|
|
len -= n;
|
|
}
|
|
return false;
|
|
}
|
|
|
|
static const char *
|
|
_dl_map_segments (struct link_map *l, int fd,
|
|
const ElfW(Ehdr) *header, int type,
|
|
const struct loadcmd loadcmds[], size_t nloadcmds,
|
|
const size_t maplength, bool has_holes,
|
|
struct link_map *loader)
|
|
{
|
|
if (__glibc_likely (type == ET_DYN))
|
|
{
|
|
/* This is a position-independent shared object. Let the system
|
|
choose where to place it.
|
|
|
|
As a refinement, sometimes we have an address that we would
|
|
prefer to map such objects at; but this is only a preference,
|
|
the OS can do whatever it likes. */
|
|
ElfW(Addr) mappref
|
|
= (ELF_PREFERRED_ADDRESS (loader, maplength,
|
|
loadcmds[0].mapstart & GLRO(dl_use_load_bias))
|
|
- MAP_BASE_ADDR (l));
|
|
|
|
uintptr_t mapstart;
|
|
if (__glibc_likely (loadcmds[0].prot & PROT_EXEC))
|
|
{
|
|
/* When there is a code segment, we must use the
|
|
allocate_code_data interface to choose a location. */
|
|
|
|
uintptr_t code_size = loadcmds[0].allocend - loadcmds[0].mapstart;
|
|
uintptr_t data_offset;
|
|
size_t data_size;
|
|
|
|
if (__glibc_likely (nloadcmds > 1))
|
|
{
|
|
data_offset = loadcmds[1].mapstart - loadcmds[0].mapstart;
|
|
data_size = ALIGN_UP (maplength - data_offset,
|
|
GLRO(dl_pagesize));
|
|
}
|
|
else
|
|
{
|
|
data_offset = 0;
|
|
data_size = 0;
|
|
}
|
|
|
|
int error = __nacl_irt_code_data_alloc.allocate_code_data
|
|
(mappref, code_size, data_offset, data_size, &mapstart);
|
|
if (__glibc_unlikely (error))
|
|
{
|
|
errno = error;
|
|
return DL_MAP_SEGMENTS_ERROR_MAP_SEGMENT;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
/* With no code pages involved, plain mmap works fine. */
|
|
void *mapped = __mmap ((void *) mappref, maplength,
|
|
PROT_NONE, MAP_ANON, -1, 0);
|
|
if (__glibc_unlikely (mapped == MAP_FAILED))
|
|
return DL_MAP_SEGMENTS_ERROR_MAP_SEGMENT;
|
|
mapstart = (uintptr_t) mapped;
|
|
}
|
|
|
|
l->l_addr = mapstart - loadcmds[0].mapstart;
|
|
}
|
|
|
|
/* Remember which part of the address space this object uses. */
|
|
l->l_map_start = loadcmds[0].mapstart + l->l_addr;
|
|
l->l_map_end = l->l_map_start + maplength;
|
|
l->l_contiguous = !has_holes;
|
|
|
|
/* Now actually map (or read) in each segment. */
|
|
for (const struct loadcmd *c = loadcmds; c < &loadcmds[nloadcmds]; ++c)
|
|
if (__glibc_likely (c->mapend > c->mapstart))
|
|
{
|
|
/* Unlike POSIX mmap, NaCl's mmap does not reliably handle COW
|
|
faults in the remainder of the final partial page. So to get
|
|
the expected behavior for the unaligned boundary between data
|
|
and bss, it's necessary to allocate the final partial page of
|
|
data as anonymous memory rather than mapping it from the file. */
|
|
|
|
size_t maplen = c->mapend - c->mapstart;
|
|
if (c->mapend > c->dataend && c->allocend > c->dataend)
|
|
maplen = (c->dataend & -GLRO(dl_pagesize)) - c->mapstart;
|
|
|
|
/* Map the segment contents from the file. */
|
|
if (__glibc_unlikely (__mmap ((void *) (l->l_addr + c->mapstart),
|
|
maplen, c->prot,
|
|
MAP_FIXED|MAP_COPY|MAP_FILE,
|
|
fd, c->mapoff)
|
|
== MAP_FAILED))
|
|
{
|
|
switch (errno)
|
|
{
|
|
case EINVAL:
|
|
case ENOTSUP:
|
|
case ENOSYS:
|
|
break;
|
|
default:
|
|
return DL_MAP_SEGMENTS_ERROR_MAP_SEGMENT;
|
|
}
|
|
|
|
/* No mmap support for this file. */
|
|
if (c->prot & PROT_EXEC)
|
|
{
|
|
/* Read the data into a temporary buffer. */
|
|
const size_t len = c->mapend - c->mapstart;
|
|
void *data = __mmap (NULL, len, PROT_READ | PROT_WRITE,
|
|
MAP_ANON|MAP_PRIVATE, -1, 0);
|
|
if (__glibc_unlikely (data == MAP_FAILED))
|
|
return DL_MAP_SEGMENTS_ERROR_MAP_ZERO_FILL;
|
|
if (read_in_data (fd, data, len, c->mapoff))
|
|
return DL_MAP_SEGMENTS_ERROR_MAP_SEGMENT;
|
|
/* Now validate and install the code. */
|
|
int error = __nacl_irt_dyncode.dyncode_create
|
|
((void *) (l->l_addr + c->mapstart), data, len);
|
|
__munmap (data, len);
|
|
if (__glibc_unlikely (error))
|
|
{
|
|
errno = error;
|
|
return DL_MAP_SEGMENTS_ERROR_MAP_SEGMENT;
|
|
}
|
|
if (__glibc_unlikely (type != ET_DYN))
|
|
{
|
|
/* A successful PROT_EXEC mmap would have implicitly
|
|
updated the bookkeeping so that a future
|
|
allocate_code_data call would know that this range
|
|
of the address space is already occupied. That
|
|
doesn't happen implicitly with dyncode_create, so
|
|
it's necessary to do an explicit call to update the
|
|
bookkeeping. */
|
|
uintptr_t allocated_address;
|
|
error = __nacl_irt_code_data_alloc.allocate_code_data
|
|
(l->l_addr + c->mapstart, len, 0, 0, &allocated_address);
|
|
if (__glibc_unlikely (error))
|
|
{
|
|
errno = error;
|
|
return DL_MAP_SEGMENTS_ERROR_MAP_SEGMENT;
|
|
}
|
|
if (__glibc_unlikely
|
|
(allocated_address != l->l_addr + c->mapstart))
|
|
{
|
|
/* This is not a very helpful error for this case,
|
|
but there isn't really anything better to use. */
|
|
errno = ENOMEM;
|
|
return DL_MAP_SEGMENTS_ERROR_MAP_SEGMENT;
|
|
}
|
|
}
|
|
}
|
|
else
|
|
{
|
|
/* Allocate the pages. */
|
|
if (__mmap ((void *) (l->l_addr + c->mapstart),
|
|
c->mapend - c->mapstart, c->prot | PROT_WRITE,
|
|
MAP_FIXED|MAP_ANON|MAP_PRIVATE, -1, 0)
|
|
== MAP_FAILED)
|
|
return DL_MAP_SEGMENTS_ERROR_MAP_ZERO_FILL;
|
|
/* Now read in the data. */
|
|
if (read_in_data (fd, (void *) (l->l_addr + c->mapstart),
|
|
c->dataend - c->mapstart, c->mapoff))
|
|
return DL_MAP_SEGMENTS_ERROR_MAP_SEGMENT;
|
|
/* Now that we've filled the pages, reset the page
|
|
protections to what they should be. */
|
|
if (!(c->prot & PROT_WRITE)
|
|
&& __mprotect ((void *) (l->l_addr + c->mapstart),
|
|
c->mapend - c->mapstart, c->prot) < 0)
|
|
return DL_MAP_SEGMENTS_ERROR_MPROTECT;
|
|
}
|
|
}
|
|
else if (c->allocend > c->dataend)
|
|
{
|
|
/* Extra zero pages should appear at the end of this segment,
|
|
after the data mapped from the file. */
|
|
|
|
uintptr_t allocend = c->mapend;
|
|
if (c->mapend > c->dataend)
|
|
{
|
|
/* The final data page was partial. So we didn't map it in.
|
|
Instead, we must allocate an anonymous page to fill. */
|
|
if (c->prot & PROT_WRITE)
|
|
/* Do the whole allocation right here. */
|
|
allocend = c->allocend;
|
|
if (__mmap ((void *) (l->l_addr + c->mapstart + maplen),
|
|
allocend - (c->mapstart + maplen), c->prot,
|
|
MAP_FIXED|MAP_ANON|MAP_PRIVATE, -1, 0)
|
|
== MAP_FAILED)
|
|
return DL_MAP_SEGMENTS_ERROR_MAP_ZERO_FILL;
|
|
if (read_in_data (fd,
|
|
(void *) (l->l_addr + c->mapstart + maplen),
|
|
c->dataend & (GLRO(dl_pagesize) - 1),
|
|
c->mapoff + maplen))
|
|
return DL_MAP_SEGMENTS_ERROR_MAP_SEGMENT;
|
|
/* Now that we've filled the page, reset its
|
|
protections to what they should be. */
|
|
if (!(c->prot & PROT_WRITE)
|
|
&& __mprotect ((void *) (l->l_addr + c->mapstart + maplen),
|
|
c->mapend - (c->mapstart + maplen),
|
|
c->prot) < 0)
|
|
return DL_MAP_SEGMENTS_ERROR_MPROTECT;
|
|
}
|
|
|
|
/* Now allocate the pure zero-fill pages. */
|
|
if (allocend < c->allocend
|
|
&& (__mmap ((void *) (l->l_addr + c->mapstart + allocend),
|
|
c->allocend - (c->mapstart + allocend), c->prot,
|
|
MAP_FIXED|MAP_ANON|MAP_PRIVATE, -1, 0)
|
|
== MAP_FAILED))
|
|
return DL_MAP_SEGMENTS_ERROR_MAP_ZERO_FILL;
|
|
}
|
|
|
|
_dl_postprocess_loadcmd (l, header, c);
|
|
}
|
|
|
|
/* Notify ELF_PREFERRED_ADDRESS that we have to load this one
|
|
fixed. */
|
|
ELF_FIXED_ADDRESS (loader, c->mapstart);
|
|
|
|
return NULL;
|
|
}
|