aboutsummaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorSzabolcs Nagy <szabolcs.nagy@arm.com>2023-01-05 18:12:36 +0000
committerSzabolcs Nagy <szabolcs.nagy@arm.com>2023-01-06 10:03:51 +0000
commit807b692012ae921ac388cb2ea852645ee2ce4027 (patch)
treeb0078cd754eb708dd2dfe14f0f06c56bcf610418
parentbbce0c75f35f55b9f5e43dd4fa01a82cc27bf020 (diff)
downloadglibc-807b692012ae921ac388cb2ea852645ee2ce4027.zip
glibc-807b692012ae921ac388cb2ea852645ee2ce4027.tar.gz
glibc-807b692012ae921ac388cb2ea852645ee2ce4027.tar.bz2
cheri: malloc: Use memcpy for obstack internal copy
Internally obstack used a backward word at a time copy. This can be an aliasing violation and likely slower than memcpy on most targets. The code is a badly written memcpy with performance issues such as the h->object_base indirection cannot be hoisted out of the loop. On CHERI architecture this is needed for copying pointers correctly.
-rw-r--r--malloc/obstack.c31
1 files changed, 3 insertions, 28 deletions
diff --git a/malloc/obstack.c b/malloc/obstack.c
index aad6a34..2056a5f 100644
--- a/malloc/obstack.c
+++ b/malloc/obstack.c
@@ -57,6 +57,7 @@
# include <stdint.h>
+# include <string.h>
/* Determine default alignment. */
union fooround
@@ -79,15 +80,6 @@ enum
DEFAULT_ROUNDING = sizeof (union fooround)
};
-/* When we copy a long block of data, this is the unit to do it with.
- On some machines, copying successive ints does not work;
- in such a case, redefine COPYING_UNIT to 'long' (if that works)
- or 'char' as a last resort. */
-# ifndef COPYING_UNIT
-# define COPYING_UNIT int
-# endif
-
-
/* The functions allocating more room by calling 'obstack_chunk_alloc'
jump to the handler pointed to by 'obstack_alloc_failed_handler'.
This can be set to a user defined function which should either
@@ -273,25 +265,8 @@ _obstack_newchunk (struct obstack *h, int length)
object_base =
__PTR_ALIGN ((char *) new_chunk, new_chunk->contents, h->alignment_mask);
- /* Move the existing object to the new chunk.
- Word at a time is fast and is safe if the object
- is sufficiently aligned. */
- if (h->alignment_mask + 1 >= DEFAULT_ALIGNMENT)
- {
- for (i = obj_size / sizeof (COPYING_UNIT) - 1;
- i >= 0; i--)
- ((COPYING_UNIT *) object_base)[i]
- = ((COPYING_UNIT *) h->object_base)[i];
- /* We used to copy the odd few remaining bytes as one extra COPYING_UNIT,
- but that can cross a page boundary on a machine
- which does not do strict alignment for COPYING_UNITS. */
- already = obj_size / sizeof (COPYING_UNIT) * sizeof (COPYING_UNIT);
- }
- else
- already = 0;
- /* Copy remaining bytes one by one. */
- for (i = already; i < obj_size; i++)
- object_base[i] = h->object_base[i];
+ /* Move the existing object to the new chunk. */
+ memcpy (object_base, h->object_base, obj_size);
/* If the object just copied was the only data in OLD_CHUNK,
free that chunk and remove it from the chain.