The framework handles the following mem* functions:
memcpy
memmove
memset
bzero
bcmp
memcmp
These functions can be built out of a set of lower-level operations:
block
: operates on a block of SIZE
bytes.tail
: operates on the last SIZE
bytes of the buffer (e.g., [dst + count - SIZE, dst + count]
)head_tail
: operates on the first and last SIZE
bytes. This is the same as calling block
and tail
.loop_and_tail
: calls block
in a loop to consume as much as possible of the count
bytes and handle the remaining bytes with a tail
operation.As an illustration, let's take the example of a trivial memset
implementation:
extern "C" void memset(const char* dst, int value, size_t count) { if (count == 0) return; if (count == 1) return Memset<1>::block(dst, value); if (count == 2) return Memset<2>::block(dst, value); if (count == 3) return Memset<3>::block(dst, value); if (count <= 8) return Memset<4>::head_tail(dst, value, count); // Note that 0 to 4 bytes are written twice. if (count <= 16) return Memset<8>::head_tail(dst, value, count); // Same here. return Memset<16>::loop_and_tail(dst, value, count); }
Now let's have a look into the Memset
structure:
template <size_t Size> struct Memset { static constexpr size_t SIZE = Size; LIBC_INLINE static void block(Ptr dst, uint8_t value) { // Implement me } LIBC_INLINE static void tail(Ptr dst, uint8_t value, size_t count) { block(dst + count - SIZE, value); } LIBC_INLINE static void head_tail(Ptr dst, uint8_t value, size_t count) { block(dst, value); tail(dst, value, count); } LIBC_INLINE static void loop_and_tail(Ptr dst, uint8_t value, size_t count) { size_t offset = 0; do { block(dst + offset, value); offset += SIZE; } while (offset < count - SIZE); tail(dst, value, count); } };
As you can see, the tail
, head_tail
and loop_and_tail
are higher order functions that build on each others. Only block
really needs to be implemented. In earlier designs we were implementing these higher order functions with templated functions but it appears that it is more readable to have the implementation explicitly stated. This design is useful because it provides customization points. For instance, for bcmp
on aarch64
we can provide a better implementation of head_tail
using vector reduction intrinsics.
We can have several specializations of the Memset
structure. Depending on the target requirements we can use one or several scopes for the same implementation.
In the following example we use the generic
implementation for the small sizes but use the x86
implementation for the loop.
extern "C" void memset(const char* dst, int value, size_t count) { if (count == 0) return; if (count == 1) return generic::Memset<1>::block(dst, value); if (count == 2) return generic::Memset<2>::block(dst, value); if (count == 3) return generic::Memset<3>::block(dst, value); if (count <= 8) return generic::Memset<4>::head_tail(dst, value, count); if (count <= 16) return generic::Memset<8>::head_tail(dst, value, count); return x86::Memset<16>::loop_and_tail(dst, value, count); }
builtin
scopeUltimately we would like the compiler to provide the code for the block
function. For this we rely on dedicated builtins available in Clang (e.g., __builtin_memset_inline
)
generic
scopeIn this scope we define pure C++ implementations using native integral types and clang vector extensions.
Then comes implementations that are using specific architectures or microarchitectures features (e.g., rep;movsb
for x86
or dc zva
for aarch64
).
The purpose here is to rely on builtins as much as possible and fallback to asm volatile
as a last resort.