summaryrefslogtreecommitdiff
path: root/src/literal.h
diff options
context:
space:
mode:
authorThomas Lively <tlively@google.com>2024-03-22 16:56:33 -0700
committerGitHub <noreply@github.com>2024-03-22 23:56:33 +0000
commitb3fea30f84fef3ff7aa77775e00b83ba62d997cc (patch)
tree53494a466d8e56d34d849d14927817a22f843748 /src/literal.h
parentd3414c3deaebe7ba35731a8c20d7fa5f5a833ca3 (diff)
downloadbinaryen-b3fea30f84fef3ff7aa77775e00b83ba62d997cc.tar.gz
binaryen-b3fea30f84fef3ff7aa77775e00b83ba62d997cc.tar.bz2
binaryen-b3fea30f84fef3ff7aa77775e00b83ba62d997cc.zip
[Strings] Represent string values as WTF-16 internally (#6418)
WTF-16, i.e. arbitrary sequences of 16-bit values, is the encoding of Java and JavaScript strings, and using the same encoding makes the interpretation of string operations trivial, even when accounting for non-ascii characters. Specifically, use little-endian WTF-16. Re-encode string constants from WTF-8 to WTF-16 in the parsers, then back to WTF-8 in the writers. Update the constructor for string `Literal`s to interpret the string as WTF-16 and store a sequence of WTF-16 code units, i.e. 16-bit integers. Update `Builder::makeConstantExpression` accordingly to convert from the new `Literal` string representation back to a WTF-16 string. Update the interpreter to remove the logic for detecting non-ascii characters and bailing out. The naive implementations of all the string operations are correct now that our string encoding matches the JS string encoding.
Diffstat (limited to 'src/literal.h')
-rw-r--r--src/literal.h2
1 files changed, 1 insertions, 1 deletions
diff --git a/src/literal.h b/src/literal.h
index 971cc8b3e..a4017f6ec 100644
--- a/src/literal.h
+++ b/src/literal.h
@@ -85,7 +85,7 @@ public:
assert(type.isSignature());
}
explicit Literal(std::shared_ptr<GCData> gcData, HeapType type);
- explicit Literal(std::string string);
+ explicit Literal(std::string_view string);
Literal(const Literal& other);
Literal& operator=(const Literal& other);
~Literal();