summaryrefslogtreecommitdiff
path: root/src
diff options
context:
space:
mode:
authorAlon Zakai <alonzakai@gmail.com>2016-11-30 15:02:01 -0800
committerGitHub <noreply@github.com>2016-11-30 15:02:01 -0800
commit36be3e0151dd7357e47b2d8f432bdd706a30466c (patch)
tree99a40bc874dbe786ae14eaec0e78bdcfc0ea0c86 /src
parent7d2d9ec9f2f936d0a53b7dc60089456a0654d29c (diff)
downloadbinaryen-36be3e0151dd7357e47b2d8f432bdd706a30466c.tar.gz
binaryen-36be3e0151dd7357e47b2d8f432bdd706a30466c.tar.bz2
binaryen-36be3e0151dd7357e47b2d8f432bdd706a30466c.zip
Fix regression from #850 (#851)
* fix regression from #850 - it is not always safe to fold added offsets into load/store offsets, as the add wraps but offset does not * small refactoring to simplify asm2wasm pass invocation
Diffstat (limited to 'src')
-rw-r--r--src/asm2wasm.h8
-rw-r--r--src/passes/OptimizeInstructions.cpp28
-rw-r--r--src/passes/PostEmscripten.cpp57
3 files changed, 37 insertions, 56 deletions
diff --git a/src/asm2wasm.h b/src/asm2wasm.h
index 9cf8cf75a..2a6bcd0e7 100644
--- a/src/asm2wasm.h
+++ b/src/asm2wasm.h
@@ -859,13 +859,6 @@ void Asm2WasmBuilder::processAsm(Ref ast) {
if (runOptimizationPasses) {
optimizingBuilder->finish();
- PassRunner passRunner(&wasm);
- if (debug) {
- passRunner.setDebug(true);
- passRunner.setValidateGlobally(false);
- }
- passRunner.add("post-emscripten");
- passRunner.run();
}
// second pass. first, function imports
@@ -993,6 +986,7 @@ void Asm2WasmBuilder::processAsm(Ref ast) {
passRunner.add("vacuum");
passRunner.add("remove-unused-brs");
passRunner.add("optimize-instructions");
+ passRunner.add("post-emscripten");
}
passRunner.run();
diff --git a/src/passes/OptimizeInstructions.cpp b/src/passes/OptimizeInstructions.cpp
index 37207ffc7..04216f1ee 100644
--- a/src/passes/OptimizeInstructions.cpp
+++ b/src/passes/OptimizeInstructions.cpp
@@ -436,33 +436,7 @@ private:
// fold constant factors into the offset
void optimizeMemoryAccess(Expression*& ptr, Address& offset) {
- while (1) {
- auto* add = ptr->dynCast<Binary>();
- if (!add) break;
- if (add->op != AddInt32) break;
- auto* left = add->left->dynCast<Const>();
- auto* right = add->right->dynCast<Const>();
- // note: in optimized code, we shouldn't see an add of two constants, so don't worry about that much
- // (precompute would optimize that)
- if (left) {
- auto value = left->value.geti32();
- if (value >= 0) {
- offset = offset + value;
- ptr = add->right;
- continue;
- }
- }
- if (right) {
- auto value = right->value.geti32();
- if (value >= 0) {
- offset = offset + value;
- ptr = add->left;
- continue;
- }
- }
- break;
- }
- // finally, ptr may be a const, but it isn't worth folding that in (we still have a const); in fact,
+ // ptr may be a const, but it isn't worth folding that in (we still have a const); in fact,
// it's better to do the opposite for gzip purposes as well as for readability.
auto* last = ptr->dynCast<Const>();
if (last) {
diff --git a/src/passes/PostEmscripten.cpp b/src/passes/PostEmscripten.cpp
index e46270c50..9a28efb1c 100644
--- a/src/passes/PostEmscripten.cpp
+++ b/src/passes/PostEmscripten.cpp
@@ -44,36 +44,49 @@ struct PostEmscripten : public WalkerPass<PostWalker<PostEmscripten, Visitor<Pos
// it's 0 or an unused section of memory that was reserved for mapped globlas).
// Thus it is ok to optimize such small constants into Load offsets.
- template<typename T>
- void visitMemoryOp(T *curr) {
- if (curr->offset) return;
- Expression* ptr = curr->ptr;
- auto add = ptr->dynCast<Binary>();
- if (!add || add->op != AddInt32) return;
- assert(add->type == i32);
- auto c = add->right->dynCast<Const>();
- if (!c) {
- c = add->left->dynCast<Const>();
- if (c) {
- // if one is a const, it's ok to swap
- add->left = add->right;
- add->right = c;
+ #define SAFE_MAX 1024
+
+ void optimizeMemoryAccess(Expression*& ptr, Address& offset) {
+ while (1) {
+ auto* add = ptr->dynCast<Binary>();
+ if (!add) break;
+ if (add->op != AddInt32) break;
+ auto* left = add->left->dynCast<Const>();
+ auto* right = add->right->dynCast<Const>();
+ // note: in optimized code, we shouldn't see an add of two constants, so don't worry about that much
+ // (precompute would optimize that)
+ if (left) {
+ auto value = left->value.geti32();
+ if (value >= 0 && value < SAFE_MAX) {
+ offset = offset + value;
+ ptr = add->right;
+ continue;
+ }
+ }
+ if (right) {
+ auto value = right->value.geti32();
+ if (value >= 0 && value < SAFE_MAX) {
+ offset = offset + value;
+ ptr = add->left;
+ continue;
+ }
}
+ break;
}
- if (!c) return;
- auto value = c->value.geti32();
- if (value >= 0 && value < 1024) {
- // foldable, by the above logic
- curr->ptr = add->left;
- curr->offset = value;
+ // finally ptr may be a const, but it isn't worth folding that in (we still have a const); in fact,
+ // it's better to do the opposite for gzip purposes as well as for readability.
+ auto* last = ptr->dynCast<Const>();
+ if (last) {
+ last->value = Literal(int32_t(last->value.geti32() + offset));
+ offset = 0;
}
}
void visitLoad(Load* curr) {
- visitMemoryOp(curr);
+ optimizeMemoryAccess(curr->ptr, curr->offset);
}
void visitStore(Store* curr) {
- visitMemoryOp(curr);
+ optimizeMemoryAccess(curr->ptr, curr->offset);
}
};