summaryrefslogtreecommitdiff
path: root/src/mixed_arena.h
diff options
context:
space:
mode:
authorAlon Zakai <alonzakai@gmail.com>2018-11-13 16:36:29 -0800
committerGitHub <noreply@github.com>2018-11-13 16:36:29 -0800
commitb67917ee7392c4d49402cb5e7320e663208390ef (patch)
treebfc32d7f2bbd007e10dbc6a7c6fa352673552741 /src/mixed_arena.h
parent6b99d143a32263478b7d525886b0bea46cbbdcaa (diff)
downloadbinaryen-b67917ee7392c4d49402cb5e7320e663208390ef.tar.gz
binaryen-b67917ee7392c4d49402cb5e7320e663208390ef.tar.bz2
binaryen-b67917ee7392c4d49402cb5e7320e663208390ef.zip
Fix alignment in MixedAllocator (#1740)
Necessary for simd, as we add a type with alignment >8. We were just broken on that before this PR.
Diffstat (limited to 'src/mixed_arena.h')
-rw-r--r--src/mixed_arena.h47
1 files changed, 29 insertions, 18 deletions
diff --git a/src/mixed_arena.h b/src/mixed_arena.h
index 90203842e..10b84c236 100644
--- a/src/mixed_arena.h
+++ b/src/mixed_arena.h
@@ -23,6 +23,7 @@
#include <memory>
#include <mutex>
#include <thread>
+#include <type_traits>
#include <vector>
//
@@ -58,9 +59,17 @@
struct MixedArena {
// fast bump allocation
- std::vector<char*> chunks;
- size_t chunkSize = 32768;
- size_t index; // in last chunk
+
+ static const size_t CHUNK_SIZE = 32768;
+ static const size_t MAX_ALIGN = 16; // allow 128bit SIMD
+
+ typedef std::aligned_storage<CHUNK_SIZE, MAX_ALIGN>::type Chunk;
+
+ // Each pointer in chunks is to an array of Chunk structs; typically 1,
+ // but possibly more.
+ std::vector<Chunk*> chunks;
+
+ size_t index = 0; // in last chunk
std::thread::id threadId;
@@ -74,7 +83,8 @@ struct MixedArena {
next.store(nullptr);
}
- void* allocSpace(size_t size) {
+ // Allocate an amount of space with a guaranteed alignment
+ void* allocSpace(size_t size, size_t align) {
// the bump allocator data should not be modified by multiple threads at once.
auto myId = std::this_thread::get_id();
if (myId != threadId) {
@@ -104,32 +114,33 @@ struct MixedArena {
curr = seen;
}
if (allocated) delete allocated;
- return curr->allocSpace(size);
- }
- size = (size + 7) & (-8); // same alignment as malloc TODO optimize?
- bool mustAllocate = false;
- while (chunkSize <= size) {
- chunkSize *= 2;
- mustAllocate = true;
+ return curr->allocSpace(size, align);
}
- if (chunks.size() == 0 || index + size >= chunkSize || mustAllocate) {
- chunks.push_back(new char[chunkSize]);
+ // First, move the current index in the last chunk to an aligned position.
+ index = (index + align - 1) & (-align);
+ if (index + size > CHUNK_SIZE || chunks.size() == 0) {
+ // Allocate a new chunk.
+ auto numChunks = (size + CHUNK_SIZE - 1) / CHUNK_SIZE;
+ assert(size <= numChunks * CHUNK_SIZE);
+ chunks.push_back(new Chunk[numChunks]);
index = 0;
}
- auto* ret = chunks.back() + index;
- index += size;
+ uint8_t* ret = static_cast<uint8_t*>(static_cast<void*>(chunks.back()));
+ ret += index;
+ index += size; // TODO: if we allocated more than 1 chunk, reuse the remainder, right now we allocate another next time
return static_cast<void*>(ret);
}
template<class T>
T* alloc() {
- auto* ret = static_cast<T*>(allocSpace(sizeof(T)));
+ static_assert(alignof(T) <= MAX_ALIGN, "maximum alignment not large enough");
+ auto* ret = static_cast<T*>(allocSpace(sizeof(T), alignof(T)));
new (ret) T(*this); // allocated objects receive the allocator, so they can allocate more later if necessary
return ret;
}
void clear() {
- for (char* chunk : chunks) {
+ for (auto* chunk : chunks) {
delete[] chunk;
}
chunks.clear();
@@ -324,7 +335,7 @@ public:
void allocate(size_t size) {
this->allocatedElements = size;
- this->data = static_cast<T*>(allocator.allocSpace(sizeof(T) * this->allocatedElements));
+ this->data = static_cast<T*>(allocator.allocSpace(sizeof(T) * this->allocatedElements, alignof(T)));
}
};