-
Notifications
You must be signed in to change notification settings - Fork 161
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Implement some functions in AssemblyScript/WebAssembly #26
Conversation
You could try speedup this a little use You could try this approach: const threshold = 100; // adjust this
function utf8ToUtf16String(bytes: Uint8Array) {
if (bytes.length > threshold) {
return wasmUtf8ToUtf16(...)
} else {
return jsUtf8ToUtf16(...)
}
} |
@MaxGraey Thanks for your advice! As I attached my benchmark script to the description, I am using a large target string like
(only decode() uses AssemblyScript ver.) My environment: NodeJS v12.2.0 on macOS 10.14.4 (Mojave) Do you have any idea? |
Try to decrease iteration count from |
Tried: diff --git a/benchmark/string.ts b/benchmark/string.ts
index ebf0314..b47613d 100644
--- a/benchmark/string.ts
+++ b/benchmark/string.ts
@@ -1,6 +1,6 @@
import { encode, decode } from "../src";
-const data = "Hello, 🌏\n".repeat(1000);
+const data = "Hello, 🌏\n".repeat(10_000);
// warm up
const encoded = encode(data);
@@ -9,13 +9,13 @@ decode(encoded);
// run
console.time("encode");
-for (let i = 0; i < 10000; i++) {
+for (let i = 0; i < 1000; i++) {
encode(data);
}
console.timeEnd("encode");
console.time("decode");
-for (let i = 0; i < 10000; i++) {
+for (let i = 0; i < 1000; i++) {
decode(encoded);
}
console.timeEnd("decode"); But little changes:
Note: |
Try to avoid using I think if you decrease |
assembly/index.ts
Outdated
} else if ((byte1 & 0xf0) === 0xe0) { | ||
// 3 bytes | ||
let byte2: u16 = load<u16>(inputOffset++) & 0x3f; | ||
let byte3: u16 = load<u16>(inputOffset++) & 0x3f; |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Also it seems this should be:
let byte2: u16 = load<u8>(inputOffset++) & 0x3f;
let byte3: u16 = load<u8>(inputOffset++) & 0x3f;
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Right: eea2f04
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
other hint which could potentially improve speed is using u32
instead u16
for every temporary variable.
Hmm? As far as I know, |
Right, but better use |
Interesting! I thought there's no difference on Now WASM ver. is faster than JS ver. for a large string! dfe06c3
|
now wasm faster on ~60%. It's pretty good result. In comparison with JS 1.5-3x speedup usually achieved |
Yep! not "a little bit", but 1.5x faster. Sounds great enough to proceed to merge this PR. |
Codecov Report
@@ Coverage Diff @@
## master #26 +/- ##
==========================================
+ Coverage 95.15% 96.37% +1.21%
==========================================
Files 13 14 +1
Lines 681 744 +63
Branches 151 156 +5
==========================================
+ Hits 648 717 +69
+ Misses 13 12 -1
+ Partials 20 15 -5
Continue to review full report at Codecov.
|
because it increses the size of bundle +5KiB (+base64-js dep)
I've implemented $ WASM=never npx ts-node benchmark/string.ts && echo && WASM=force npx ts-node benchmark/string.ts
WASM_AVAILABLE=false
encode / decode ascii data.length=40000 encoded.byteLength=40003
encode ascii: 542.870ms
decode ascii: 868.292ms
encode / decode emoji data.length=40000 encoded.byteLength=80005
encode emoji: 653.479ms
decode emoji: 892.014ms
WASM_AVAILABLE=true
encode / decode ascii data.length=40000 encoded.byteLength=40003
encode ascii: 528.937ms
decode ascii: 520.596ms
encode / decode emoji data.length=40000 encoded.byteLength=80005
encode emoji: 555.948ms
decode emoji: 595.192ms I think the overhead of preparation that converts string to an array of utf16 units (i.e. Do you have any idea about it? |
setMemoryStr(inputU16BePtr, inputByteLength, str, strLength); | ||
|
||
const maxOutputHeaderSize = 1 + 4; // headByte + u32 | ||
const outputPtr: pointer = wm.malloc(maxOutputHeaderSize + strLength * 4); |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I guess better allocate this inside utf8EncodeUint16Array
on AS side. This reduce overhead for one js <-> wasm call. But you should retreave this pointer somehow anyway and this required do other call like getOutputPointer
. Hmm
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Tried to reduce a malloc()
call there but the performance score didn't get better.
diff --git a/assembly/utf8EncodeUint16Array.ts b/assembly/utf8EncodeUint16Array.ts
index 195991f..536311c 100644
--- a/assembly/utf8EncodeUint16Array.ts
+++ b/assembly/utf8EncodeUint16Array.ts
@@ -29,9 +29,15 @@ function storeStringHeader(outputPtr: usize, utf8ByteLength: usize): usize {
// outputPtr: u8*
// inputPtr: u16*
// It adds MessagePack str head bytes to the output
-export function utf8EncodeUint16Array(outputPtr: usize, inputPtr: usize, inputLength: usize): usize {
+export function utf8EncodeUint16Array(inputPtr: usize, inputLength: usize): usize {
+ const maxOutputHeaderSize = sizeof<u8>() + sizeof<u32>();
+
+ // outputPtr: [u32 outputBufferSize][outputBuffer]
+ let outputPtr = memory.allocate(maxOutputHeaderSize + inputLength * 4);
+ let outputPtrStart = outputPtr + sizeof<u32>();
+
let utf8ByteLength = utf8CountUint16Array(inputPtr, inputLength);
- let strHeaderOffset = storeStringHeader(outputPtr, utf8ByteLength);
+ let strHeaderOffset = storeStringHeader(outputPtrStart, utf8ByteLength);
const u16s = sizeof<u16>();
let inputOffset = inputPtr;
@@ -76,5 +82,7 @@ export function utf8EncodeUint16Array(outputPtr: usize, inputPtr: usize, inputLe
store<u8>(outputOffset++, (value & 0x3f) | 0x80);
}
- return outputOffset - outputPtr;
+ let outputByteLength = outputOffset - outputPtrStart;
+ storeUint32BE(outputPtr, outputByteLength);
+ return outputPtr;
}
diff --git a/src/wasmFunctions.ts b/src/wasmFunctions.ts
index 5a3c44a..3a6ab4b 100644
--- a/src/wasmFunctions.ts
+++ b/src/wasmFunctions.ts
@@ -54,9 +54,11 @@ export function utf8EncodeWasm(str: string, output: Uint8Array): number {
const maxOutputHeaderSize = 1 + 4; // headByte + u32
const outputPtr: pointer = wm.malloc(maxOutputHeaderSize + strLength * 4);
try {
- const outputLength = wm.utf8EncodeUint16Array(outputPtr, inputU16BePtr, strLength);
- output.set(new Uint8Array(wm.memory.buffer, outputPtr, outputLength));
- return outputLength;
+ const outputPtr = wm.utf8EncodeUint16Array(inputU16BePtr, strLength);
+ // the first 4 bytes is the outputByteLength in big-endian
+ const outputByteLength = new DataView(wm.memory.buffer, outputPtr, 4).getUint32(0);
+ output.set(new Uint8Array(wm.memory.buffer, outputPtr + 4, outputByteLength));
+ return outputByteLength;
} finally {
wm.free(inputU16BePtr);
wm.free(outputPtr);
I guess calling malloc()
is not so heavy.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Hmm, in this case I recommend profile this better. Try manually measure (for example via console.time/console.timeEnd) for whole js utf8EncodeUint16Array
method and for wm.utf8EncodeUint16Array
call exclusively. Diff between times give you picture about which is price you pay for interop
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Tried to merge multiple malloc into a single one, but there were no effects.
Will try reference types https://github.com/WebAssembly/reference-types/blob/master/proposals/reference-types/Overview.md in a future.
Definitely you have interop overhead. Hard to tell how this improve. Try eliminate calls between JS and wasm. For example: use one shared buffer (with max size of both) for input and output and use inplace process if this possible |
Hmm! Will give it a try to reduce some js<->wasm calls. Thank you for your advice, anyway! I'll merge this PR in a few days and will send feedbacks to AssemblyScript. |
@gfx Sure! It will be great |
Currently, it is much slower than JS version 🤔Thanks to @MaxGraey, now it's 1.5x faster than JS ver. for large string chunks. Good enough to proceed to merge this PR.
The benchmark script I am using (
benchmark/string.ts
):