aboutsummaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorSzabolcs Nagy <szabolcs.nagy@arm.com>2022-10-21 12:38:26 +0100
committerSzabolcs Nagy <szabolcs.nagy@arm.com>2022-10-27 14:46:53 +0100
commit28bd4285ad0f8ce6026583e9cc0a612018d461ba (patch)
tree7a3aa851f83addae07337267feba755a07cade39
parent5eedf66625842a56c0ed7e16a1f79fda4b52b425 (diff)
downloadglibc-28bd4285ad0f8ce6026583e9cc0a612018d461ba.zip
glibc-28bd4285ad0f8ce6026583e9cc0a612018d461ba.tar.gz
glibc-28bd4285ad0f8ce6026583e9cc0a612018d461ba.tar.bz2
cheri: elf: Fix segment mapping permissions
Ensure mmap returns pointers with RWX permission covering all segments. These pointers later get restricted to RX and RW permission.
-rw-r--r--elf/dl-map-segments.h17
1 files changed, 13 insertions, 4 deletions
diff --git a/elf/dl-map-segments.h b/elf/dl-map-segments.h
index ce51585..e25fb57 100644
--- a/elf/dl-map-segments.h
+++ b/elf/dl-map-segments.h
@@ -22,6 +22,13 @@
# include <cheri_perms.h>
#endif
+/* Allow RWX mprotect later, on CHERI this means RWX capability permission. */
+#ifdef PROT_MAX
+# define PROT_MAX_RWX PROT_MAX (PROT_READ | PROT_WRITE | PROT_EXEC)
+#else
+# define PROT_MAX_RWX 0
+#endif
+
/* Map a segment and align it properly. */
static __always_inline elfptr_t
@@ -29,7 +36,8 @@ _dl_map_segment (const struct loadcmd *c, ElfW(Addr) mappref,
const size_t maplength, int fd)
{
if (__glibc_likely (c->mapalign <= GLRO(dl_pagesize)))
- return (elfptr_t) __mmap ((void *) mappref, maplength, c->prot,
+ return (elfptr_t) __mmap ((void *) mappref, maplength,
+ c->prot|PROT_MAX_RWX,
MAP_COPY|MAP_FILE, fd, c->mapoff);
/* If the segment alignment > the page size, allocate enough space to
@@ -38,7 +46,7 @@ _dl_map_segment (const struct loadcmd *c, ElfW(Addr) mappref,
? (maplength + c->mapalign)
: (2 * c->mapalign));
elfptr_t map_start = (elfptr_t) __mmap ((void *) mappref, maplen,
- PROT_NONE,
+ PROT_NONE|PROT_MAX_RWX,
MAP_ANONYMOUS|MAP_PRIVATE,
-1, 0);
if (__glibc_unlikely ((void *) map_start == MAP_FAILED))
@@ -46,7 +54,7 @@ _dl_map_segment (const struct loadcmd *c, ElfW(Addr) mappref,
elfptr_t map_start_aligned = ALIGN_UP (map_start, c->mapalign);
map_start_aligned = (elfptr_t) __mmap ((void *) map_start_aligned,
- maplength, c->prot,
+ maplength, c->prot|PROT_MAX_RWX,
MAP_COPY|MAP_FILE|MAP_FIXED,
fd, c->mapoff);
if (__glibc_unlikely ((void *) map_start_aligned == MAP_FAILED))
@@ -137,7 +145,8 @@ _dl_map_segments (struct link_map *l, int fd,
else
{
/* Need a single capability to cover all load segments. */
- void *p = __mmap ((void *) c->mapstart, maplength, c->prot,
+ void *p = __mmap ((void *) c->mapstart, maplength,
+ c->prot|PROT_MAX_RWX,
MAP_FIXED|MAP_COPY|MAP_FILE,
fd, c->mapoff);
if (p == MAP_FAILED)