mirror of
https://github.com/bytecodealliance/wasm-micro-runtime.git
synced 2024-11-26 23:42:05 +00:00
8123 lines
293 KiB
C
8123 lines
293 KiB
C
/*
|
|
* Copyright (C) 2019 Intel Corporation. All rights reserved.
|
|
* SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
|
|
*/
|
|
|
|
#include "wasm_loader.h"
|
|
#include "bh_common.h"
|
|
#include "bh_log.h"
|
|
#include "wasm.h"
|
|
#include "wasm_opcode.h"
|
|
#include "wasm_runtime.h"
|
|
#include "../common/wasm_native.h"
|
|
#include "../common/wasm_memory.h"
|
|
#include "wasm_loader_common.h"
|
|
#if WASM_ENABLE_FAST_JIT != 0
|
|
#include "../fast-jit/jit_compiler.h"
|
|
#include "../fast-jit/jit_codecache.h"
|
|
#endif
|
|
#if WASM_ENABLE_JIT != 0
|
|
#include "../compilation/aot_llvm.h"
|
|
#endif
|
|
|
|
/* Read a value of given type from the address pointed to by the given
|
|
pointer and increase the pointer to the position just after the
|
|
value being read. */
|
|
#define TEMPLATE_READ_VALUE(Type, p) \
|
|
(p += sizeof(Type), *(Type *)(p - sizeof(Type)))
|
|
|
|
#if WASM_ENABLE_MEMORY64 != 0
|
|
static bool
|
|
has_module_memory64(WASMModule *module)
|
|
{
|
|
/* TODO: multi-memories for now assuming the memory idx type is consistent
|
|
* across multi-memories */
|
|
if (module->import_memory_count > 0)
|
|
return !!(module->import_memories[0].u.memory.flags & MEMORY64_FLAG);
|
|
else if (module->memory_count > 0)
|
|
return !!(module->memories[0].flags & MEMORY64_FLAG);
|
|
|
|
return false;
|
|
}
|
|
#endif
|
|
|
|
static void
|
|
set_error_buf(char *error_buf, uint32 error_buf_size, const char *string)
|
|
{
|
|
if (error_buf != NULL)
|
|
snprintf(error_buf, error_buf_size, "WASM module load failed: %s",
|
|
string);
|
|
}
|
|
|
|
#define CHECK_BUF(buf, buf_end, length) \
|
|
do { \
|
|
bh_assert(buf + length >= buf && buf + length <= buf_end); \
|
|
} while (0)
|
|
|
|
#define CHECK_BUF1(buf, buf_end, length) \
|
|
do { \
|
|
bh_assert(buf + length >= buf && buf + length <= buf_end); \
|
|
} while (0)
|
|
|
|
#define skip_leb(p) while (*p++ & 0x80)
|
|
#define skip_leb_int64(p, p_end) skip_leb(p)
|
|
#define skip_leb_uint32(p, p_end) skip_leb(p)
|
|
#define skip_leb_int32(p, p_end) skip_leb(p)
|
|
#define skip_leb_mem_offset(p, p_end) skip_leb(p)
|
|
|
|
static bool
|
|
is_32bit_type(uint8 type)
|
|
{
|
|
if (type == VALUE_TYPE_I32
|
|
|| type == VALUE_TYPE_F32
|
|
/* the operand stack is in polymorphic state */
|
|
|| type == VALUE_TYPE_ANY
|
|
#if WASM_ENABLE_REF_TYPES != 0
|
|
|| type == VALUE_TYPE_FUNCREF || type == VALUE_TYPE_EXTERNREF
|
|
#endif
|
|
)
|
|
return true;
|
|
return false;
|
|
}
|
|
|
|
static bool
|
|
is_64bit_type(uint8 type)
|
|
{
|
|
if (type == VALUE_TYPE_I64 || type == VALUE_TYPE_F64)
|
|
return true;
|
|
return false;
|
|
}
|
|
|
|
static bool
|
|
is_value_type(uint8 type)
|
|
{
|
|
if (type == VALUE_TYPE_I32 || type == VALUE_TYPE_I64
|
|
|| type == VALUE_TYPE_F32 || type == VALUE_TYPE_F64
|
|
#if WASM_ENABLE_REF_TYPES != 0
|
|
|| type == VALUE_TYPE_FUNCREF || type == VALUE_TYPE_EXTERNREF
|
|
#endif
|
|
)
|
|
return true;
|
|
return false;
|
|
}
|
|
|
|
static bool
|
|
is_byte_a_type(uint8 type)
|
|
{
|
|
return is_value_type(type) || (type == VALUE_TYPE_VOID);
|
|
}
|
|
|
|
static void
|
|
read_leb(uint8 **p_buf, const uint8 *buf_end, uint32 maxbits, bool sign,
|
|
uint64 *p_result, char *error_buf, uint32 error_buf_size)
|
|
{
|
|
const uint8 *buf = *p_buf;
|
|
uint64 result = 0;
|
|
uint32 shift = 0;
|
|
uint32 offset = 0, bcnt = 0;
|
|
uint64 byte;
|
|
|
|
while (true) {
|
|
bh_assert(bcnt + 1 <= (maxbits + 6) / 7);
|
|
CHECK_BUF(buf, buf_end, offset + 1);
|
|
byte = buf[offset];
|
|
offset += 1;
|
|
result |= ((byte & 0x7f) << shift);
|
|
shift += 7;
|
|
bcnt += 1;
|
|
if ((byte & 0x80) == 0) {
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (!sign && maxbits == 32 && shift >= maxbits) {
|
|
/* The top bits set represent values > 32 bits */
|
|
bh_assert(!(((uint8)byte) & 0xf0));
|
|
}
|
|
else if (sign && maxbits == 32) {
|
|
if (shift < maxbits) {
|
|
/* Sign extend, second-highest bit is the sign bit */
|
|
if ((uint8)byte & 0x40)
|
|
result |= (~((uint64)0)) << shift;
|
|
}
|
|
else {
|
|
/* The top bits should be a sign-extension of the sign bit */
|
|
bool sign_bit_set = ((uint8)byte) & 0x8;
|
|
int top_bits = ((uint8)byte) & 0xf0;
|
|
bh_assert(!((sign_bit_set && top_bits != 0x70)
|
|
|| (!sign_bit_set && top_bits != 0)));
|
|
(void)top_bits;
|
|
(void)sign_bit_set;
|
|
}
|
|
}
|
|
else if (sign && maxbits == 64) {
|
|
if (shift < maxbits) {
|
|
/* Sign extend, second-highest bit is the sign bit */
|
|
if ((uint8)byte & 0x40)
|
|
result |= (~((uint64)0)) << shift;
|
|
}
|
|
else {
|
|
/* The top bits should be a sign-extension of the sign bit */
|
|
bool sign_bit_set = ((uint8)byte) & 0x1;
|
|
int top_bits = ((uint8)byte) & 0xfe;
|
|
|
|
bh_assert(!((sign_bit_set && top_bits != 0x7e)
|
|
|| (!sign_bit_set && top_bits != 0)));
|
|
(void)top_bits;
|
|
(void)sign_bit_set;
|
|
}
|
|
}
|
|
|
|
*p_buf += offset;
|
|
*p_result = result;
|
|
}
|
|
|
|
#define read_uint8(p) TEMPLATE_READ_VALUE(uint8, p)
|
|
#define read_uint32(p) TEMPLATE_READ_VALUE(uint32, p)
|
|
#define read_bool(p) TEMPLATE_READ_VALUE(bool, p)
|
|
|
|
#define read_leb_int64(p, p_end, res) \
|
|
do { \
|
|
uint64 res64; \
|
|
read_leb((uint8 **)&p, p_end, 64, true, &res64, error_buf, \
|
|
error_buf_size); \
|
|
res = (int64)res64; \
|
|
} while (0)
|
|
|
|
#define read_leb_uint32(p, p_end, res) \
|
|
do { \
|
|
uint64 res64; \
|
|
read_leb((uint8 **)&p, p_end, 32, false, &res64, error_buf, \
|
|
error_buf_size); \
|
|
res = (uint32)res64; \
|
|
} while (0)
|
|
|
|
#define read_leb_int32(p, p_end, res) \
|
|
do { \
|
|
uint64 res64; \
|
|
read_leb((uint8 **)&p, p_end, 32, true, &res64, error_buf, \
|
|
error_buf_size); \
|
|
res = (int32)res64; \
|
|
} while (0)
|
|
|
|
#if WASM_ENABLE_MEMORY64 != 0
|
|
#define read_leb_mem_offset(p, p_end, res) \
|
|
do { \
|
|
uint64 res64; \
|
|
read_leb((uint8 **)&p, p_end, is_memory64 ? 64 : 32, false, &res64, \
|
|
error_buf, error_buf_size); \
|
|
res = (mem_offset_t)res64; \
|
|
} while (0)
|
|
#else
|
|
#define read_leb_mem_offset(p, p_end, res) read_leb_uint32(p, p_end, res)
|
|
#endif
|
|
|
|
static void *
|
|
loader_malloc(uint64 size, char *error_buf, uint32 error_buf_size)
|
|
{
|
|
void *mem;
|
|
|
|
if (size >= UINT32_MAX || !(mem = wasm_runtime_malloc((uint32)size))) {
|
|
set_error_buf(error_buf, error_buf_size, "allocate memory failed");
|
|
return NULL;
|
|
}
|
|
|
|
memset(mem, 0, (uint32)size);
|
|
return mem;
|
|
}
|
|
|
|
static void *
|
|
memory_realloc(void *mem_old, uint32 size_old, uint32 size_new, char *error_buf,
|
|
uint32 error_buf_size)
|
|
{
|
|
uint8 *mem_new;
|
|
bh_assert(size_new > size_old);
|
|
if ((mem_new = loader_malloc(size_new, error_buf, error_buf_size))) {
|
|
bh_memcpy_s(mem_new, size_new, mem_old, size_old);
|
|
memset(mem_new + size_old, 0, size_new - size_old);
|
|
wasm_runtime_free(mem_old);
|
|
}
|
|
return mem_new;
|
|
}
|
|
|
|
#define MEM_REALLOC(mem, size_old, size_new) \
|
|
do { \
|
|
void *mem_new = memory_realloc(mem, size_old, size_new, error_buf, \
|
|
error_buf_size); \
|
|
if (!mem_new) \
|
|
goto fail; \
|
|
mem = mem_new; \
|
|
} while (0)
|
|
|
|
static void
|
|
destroy_wasm_type(WASMFuncType *type)
|
|
{
|
|
if (type->ref_count > 1) {
|
|
/* The type is referenced by other types
|
|
of current wasm module */
|
|
type->ref_count--;
|
|
return;
|
|
}
|
|
|
|
#if WASM_ENABLE_FAST_JIT != 0 && WASM_ENABLE_JIT != 0 \
|
|
&& WASM_ENABLE_LAZY_JIT != 0
|
|
if (type->call_to_llvm_jit_from_fast_jit)
|
|
jit_code_cache_free(type->call_to_llvm_jit_from_fast_jit);
|
|
#endif
|
|
|
|
wasm_runtime_free(type);
|
|
}
|
|
|
|
static bool
|
|
check_function_index(const WASMModule *module, uint32 function_index,
|
|
char *error_buf, uint32 error_buf_size)
|
|
{
|
|
return (function_index
|
|
< module->import_function_count + module->function_count);
|
|
}
|
|
|
|
typedef struct InitValue {
|
|
uint8 type;
|
|
uint8 flag;
|
|
WASMValue value;
|
|
} InitValue;
|
|
|
|
typedef struct ConstExprContext {
|
|
uint32 sp;
|
|
uint32 size;
|
|
WASMModule *module;
|
|
InitValue *stack;
|
|
InitValue data[WASM_CONST_EXPR_STACK_SIZE];
|
|
} ConstExprContext;
|
|
|
|
static void
|
|
init_const_expr_stack(ConstExprContext *ctx, WASMModule *module)
|
|
{
|
|
ctx->sp = 0;
|
|
ctx->module = module;
|
|
ctx->stack = ctx->data;
|
|
ctx->size = WASM_CONST_EXPR_STACK_SIZE;
|
|
}
|
|
|
|
static bool
|
|
push_const_expr_stack(ConstExprContext *ctx, uint8 flag, uint8 type,
|
|
WASMValue *value, char *error_buf, uint32 error_buf_size)
|
|
{
|
|
InitValue *cur_value;
|
|
|
|
if (ctx->sp >= ctx->size) {
|
|
if (ctx->stack != ctx->data) {
|
|
MEM_REALLOC(ctx->stack, ctx->size * sizeof(InitValue),
|
|
(ctx->size + 4) * sizeof(InitValue));
|
|
}
|
|
else {
|
|
if (!(ctx->stack =
|
|
loader_malloc((ctx->size + 4) * (uint64)sizeof(InitValue),
|
|
error_buf, error_buf_size))) {
|
|
return false;
|
|
}
|
|
}
|
|
ctx->size += 4;
|
|
}
|
|
|
|
cur_value = &ctx->stack[ctx->sp++];
|
|
cur_value->type = type;
|
|
cur_value->flag = flag;
|
|
cur_value->value = *value;
|
|
|
|
return true;
|
|
fail:
|
|
return false;
|
|
}
|
|
|
|
static bool
|
|
pop_const_expr_stack(ConstExprContext *ctx, uint8 *p_flag, uint8 type,
|
|
WASMValue *p_value, char *error_buf, uint32 error_buf_size)
|
|
{
|
|
InitValue *cur_value;
|
|
|
|
if (ctx->sp == 0) {
|
|
return false;
|
|
}
|
|
|
|
cur_value = &ctx->stack[--ctx->sp];
|
|
|
|
if (cur_value->type != type) {
|
|
return false;
|
|
}
|
|
|
|
if (p_flag)
|
|
*p_flag = cur_value->flag;
|
|
if (p_value)
|
|
*p_value = cur_value->value;
|
|
|
|
return true;
|
|
}
|
|
|
|
static void
|
|
destroy_const_expr_stack(ConstExprContext *ctx)
|
|
{
|
|
if (ctx->stack != ctx->data) {
|
|
wasm_runtime_free(ctx->stack);
|
|
}
|
|
}
|
|
|
|
static bool
|
|
load_init_expr(WASMModule *module, const uint8 **p_buf, const uint8 *buf_end,
|
|
InitializerExpression *init_expr, uint8 type, char *error_buf,
|
|
uint32 error_buf_size)
|
|
{
|
|
const uint8 *p = *p_buf, *p_end = buf_end;
|
|
uint8 flag, *p_float;
|
|
uint32 i;
|
|
ConstExprContext const_expr_ctx = { 0 };
|
|
WASMValue cur_value = { 0 };
|
|
|
|
init_const_expr_stack(&const_expr_ctx, module);
|
|
|
|
CHECK_BUF(p, p_end, 1);
|
|
flag = read_uint8(p);
|
|
|
|
while (flag != WASM_OP_END) {
|
|
switch (flag) {
|
|
/* i32.const */
|
|
case INIT_EXPR_TYPE_I32_CONST:
|
|
read_leb_int32(p, p_end, cur_value.i32);
|
|
|
|
if (!push_const_expr_stack(&const_expr_ctx, flag,
|
|
VALUE_TYPE_I32, &cur_value,
|
|
error_buf, error_buf_size)) {
|
|
bh_assert(0);
|
|
}
|
|
break;
|
|
/* i64.const */
|
|
case INIT_EXPR_TYPE_I64_CONST:
|
|
read_leb_int64(p, p_end, cur_value.i64);
|
|
|
|
if (!push_const_expr_stack(&const_expr_ctx, flag,
|
|
VALUE_TYPE_I64, &cur_value,
|
|
error_buf, error_buf_size)) {
|
|
bh_assert(0);
|
|
}
|
|
break;
|
|
/* f32.const */
|
|
case INIT_EXPR_TYPE_F32_CONST:
|
|
CHECK_BUF(p, p_end, 4);
|
|
p_float = (uint8 *)&cur_value.f32;
|
|
for (i = 0; i < sizeof(float32); i++)
|
|
*p_float++ = *p++;
|
|
|
|
if (!push_const_expr_stack(&const_expr_ctx, flag,
|
|
VALUE_TYPE_F32, &cur_value,
|
|
error_buf, error_buf_size)) {
|
|
bh_assert(0);
|
|
}
|
|
break;
|
|
/* f64.const */
|
|
case INIT_EXPR_TYPE_F64_CONST:
|
|
CHECK_BUF(p, p_end, 8);
|
|
p_float = (uint8 *)&cur_value.f64;
|
|
for (i = 0; i < sizeof(float64); i++)
|
|
*p_float++ = *p++;
|
|
|
|
if (!push_const_expr_stack(&const_expr_ctx, flag,
|
|
VALUE_TYPE_F64, &cur_value,
|
|
error_buf, error_buf_size)) {
|
|
bh_assert(0);
|
|
}
|
|
break;
|
|
|
|
#if WASM_ENABLE_REF_TYPES != 0
|
|
/* ref.func */
|
|
case INIT_EXPR_TYPE_FUNCREF_CONST:
|
|
{
|
|
uint32 func_idx;
|
|
read_leb_uint32(p, p_end, func_idx);
|
|
cur_value.ref_index = func_idx;
|
|
if (!check_function_index(module, func_idx, error_buf,
|
|
error_buf_size)) {
|
|
bh_assert(0);
|
|
}
|
|
|
|
if (!push_const_expr_stack(&const_expr_ctx, flag,
|
|
VALUE_TYPE_FUNCREF, &cur_value,
|
|
error_buf, error_buf_size)) {
|
|
bh_assert(0);
|
|
}
|
|
break;
|
|
}
|
|
|
|
/* ref.null */
|
|
case INIT_EXPR_TYPE_REFNULL_CONST:
|
|
{
|
|
uint8 type1;
|
|
|
|
CHECK_BUF(p, p_end, 1);
|
|
type1 = read_uint8(p);
|
|
|
|
cur_value.ref_index = UINT32_MAX;
|
|
if (!push_const_expr_stack(&const_expr_ctx, flag, type1,
|
|
&cur_value, error_buf,
|
|
error_buf_size)) {
|
|
bh_assert(0);
|
|
}
|
|
break;
|
|
}
|
|
#endif /* end of WASM_ENABLE_REF_TYPES != 0 */
|
|
|
|
/* get_global */
|
|
case INIT_EXPR_TYPE_GET_GLOBAL:
|
|
{
|
|
uint32 global_idx;
|
|
uint8 global_type;
|
|
|
|
read_leb_uint32(p, p_end, cur_value.global_index);
|
|
global_idx = cur_value.global_index;
|
|
|
|
bh_assert(global_idx < module->import_global_count);
|
|
bh_assert(!module->import_globals[global_idx]
|
|
.u.global.type.is_mutable);
|
|
|
|
if (global_idx < module->import_global_count) {
|
|
global_type = module->import_globals[global_idx]
|
|
.u.global.type.val_type;
|
|
}
|
|
else {
|
|
global_type =
|
|
module
|
|
->globals[global_idx - module->import_global_count]
|
|
.type.val_type;
|
|
}
|
|
|
|
if (!push_const_expr_stack(&const_expr_ctx, flag, global_type,
|
|
&cur_value, error_buf,
|
|
error_buf_size))
|
|
bh_assert(0);
|
|
|
|
break;
|
|
}
|
|
default:
|
|
{
|
|
bh_assert(0);
|
|
}
|
|
}
|
|
|
|
CHECK_BUF(p, p_end, 1);
|
|
flag = read_uint8(p);
|
|
}
|
|
|
|
/* There should be only one value left on the init value stack */
|
|
if (!pop_const_expr_stack(&const_expr_ctx, &flag, type, &cur_value,
|
|
error_buf, error_buf_size)) {
|
|
bh_assert(0);
|
|
}
|
|
|
|
bh_assert(const_expr_ctx.sp == 0);
|
|
|
|
init_expr->init_expr_type = flag;
|
|
init_expr->u = cur_value;
|
|
|
|
*p_buf = p;
|
|
destroy_const_expr_stack(&const_expr_ctx);
|
|
return true;
|
|
}
|
|
|
|
static bool
|
|
load_type_section(const uint8 *buf, const uint8 *buf_end, WASMModule *module,
|
|
char *error_buf, uint32 error_buf_size)
|
|
{
|
|
const uint8 *p = buf, *p_end = buf_end, *p_org;
|
|
uint32 type_count, param_count, result_count, i, j;
|
|
uint32 param_cell_num, ret_cell_num;
|
|
uint64 total_size;
|
|
uint8 flag;
|
|
WASMFuncType *type;
|
|
|
|
read_leb_uint32(p, p_end, type_count);
|
|
|
|
if (type_count) {
|
|
module->type_count = type_count;
|
|
total_size = sizeof(WASMFuncType *) * (uint64)type_count;
|
|
if (!(module->types =
|
|
loader_malloc(total_size, error_buf, error_buf_size))) {
|
|
return false;
|
|
}
|
|
|
|
for (i = 0; i < type_count; i++) {
|
|
CHECK_BUF(p, p_end, 1);
|
|
flag = read_uint8(p);
|
|
bh_assert(flag == 0x60);
|
|
|
|
read_leb_uint32(p, p_end, param_count);
|
|
|
|
/* Resolve param count and result count firstly */
|
|
p_org = p;
|
|
CHECK_BUF(p, p_end, param_count);
|
|
p += param_count;
|
|
read_leb_uint32(p, p_end, result_count);
|
|
CHECK_BUF(p, p_end, result_count);
|
|
p = p_org;
|
|
|
|
bh_assert(param_count <= UINT16_MAX && result_count <= UINT16_MAX);
|
|
|
|
total_size = offsetof(WASMFuncType, types)
|
|
+ sizeof(uint8) * (uint64)(param_count + result_count);
|
|
if (!(type = module->types[i] =
|
|
loader_malloc(total_size, error_buf, error_buf_size))) {
|
|
return false;
|
|
}
|
|
|
|
/* Resolve param types and result types */
|
|
type->ref_count = 1;
|
|
type->param_count = (uint16)param_count;
|
|
type->result_count = (uint16)result_count;
|
|
for (j = 0; j < param_count; j++) {
|
|
CHECK_BUF(p, p_end, 1);
|
|
type->types[j] = read_uint8(p);
|
|
}
|
|
read_leb_uint32(p, p_end, result_count);
|
|
for (j = 0; j < result_count; j++) {
|
|
CHECK_BUF(p, p_end, 1);
|
|
type->types[param_count + j] = read_uint8(p);
|
|
}
|
|
for (j = 0; j < param_count + result_count; j++) {
|
|
bh_assert(is_value_type(type->types[j]));
|
|
}
|
|
|
|
param_cell_num = wasm_get_cell_num(type->types, param_count);
|
|
ret_cell_num =
|
|
wasm_get_cell_num(type->types + param_count, result_count);
|
|
bh_assert(param_cell_num <= UINT16_MAX
|
|
&& ret_cell_num <= UINT16_MAX);
|
|
type->param_cell_num = (uint16)param_cell_num;
|
|
type->ret_cell_num = (uint16)ret_cell_num;
|
|
|
|
#if WASM_ENABLE_QUICK_AOT_ENTRY != 0
|
|
type->quick_aot_entry = wasm_native_lookup_quick_aot_entry(type);
|
|
#endif
|
|
|
|
/* If there is already a same type created, use it instead */
|
|
for (j = 0; j < i; ++j) {
|
|
if (wasm_type_equal(type, module->types[j], module->types, i)) {
|
|
bh_assert(module->types[j]->ref_count != UINT16_MAX);
|
|
destroy_wasm_type(type);
|
|
module->types[i] = module->types[j];
|
|
module->types[j]->ref_count++;
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
bh_assert(p == p_end);
|
|
LOG_VERBOSE("Load type section success.\n");
|
|
(void)flag;
|
|
return true;
|
|
}
|
|
|
|
static void
|
|
adjust_table_max_size(uint32 init_size, uint32 max_size_flag, uint32 *max_size)
|
|
{
|
|
uint32 default_max_size = init_size * 2 > WASM_TABLE_MAX_SIZE
|
|
? init_size * 2
|
|
: WASM_TABLE_MAX_SIZE;
|
|
|
|
if (max_size_flag) {
|
|
/* module defines the table limitation */
|
|
bh_assert(init_size <= *max_size);
|
|
|
|
if (init_size < *max_size) {
|
|
*max_size =
|
|
*max_size < default_max_size ? *max_size : default_max_size;
|
|
}
|
|
}
|
|
else {
|
|
/* partial defined table limitation, gives a default value */
|
|
*max_size = default_max_size;
|
|
}
|
|
}
|
|
|
|
static bool
|
|
load_function_import(const uint8 **p_buf, const uint8 *buf_end,
|
|
const WASMModule *parent_module,
|
|
const char *sub_module_name, const char *function_name,
|
|
WASMFunctionImport *function, char *error_buf,
|
|
uint32 error_buf_size)
|
|
{
|
|
const uint8 *p = *p_buf, *p_end = buf_end;
|
|
uint32 declare_type_index = 0;
|
|
WASMFuncType *declare_func_type = NULL;
|
|
WASMFunction *linked_func = NULL;
|
|
const char *linked_signature = NULL;
|
|
void *linked_attachment = NULL;
|
|
bool linked_call_conv_raw = false;
|
|
|
|
read_leb_uint32(p, p_end, declare_type_index);
|
|
*p_buf = p;
|
|
|
|
bh_assert(declare_type_index < parent_module->type_count);
|
|
|
|
declare_func_type = parent_module->types[declare_type_index];
|
|
|
|
/* check built-in modules */
|
|
linked_func = wasm_native_resolve_symbol(
|
|
sub_module_name, function_name, declare_func_type, &linked_signature,
|
|
&linked_attachment, &linked_call_conv_raw);
|
|
|
|
function->module_name = (char *)sub_module_name;
|
|
function->field_name = (char *)function_name;
|
|
function->func_type = declare_func_type;
|
|
function->func_ptr_linked = linked_func;
|
|
function->signature = linked_signature;
|
|
function->attachment = linked_attachment;
|
|
function->call_conv_raw = linked_call_conv_raw;
|
|
return true;
|
|
}
|
|
|
|
static bool
|
|
load_table_import(const uint8 **p_buf, const uint8 *buf_end,
|
|
WASMModule *parent_module, const char *sub_module_name,
|
|
const char *table_name, WASMTableImport *table,
|
|
char *error_buf, uint32 error_buf_size)
|
|
{
|
|
const uint8 *p = *p_buf, *p_end = buf_end;
|
|
uint32 declare_elem_type = 0, declare_max_size_flag = 0,
|
|
declare_init_size = 0, declare_max_size = 0;
|
|
|
|
CHECK_BUF(p, p_end, 1);
|
|
/* 0x70 or 0x6F */
|
|
declare_elem_type = read_uint8(p);
|
|
bh_assert(VALUE_TYPE_FUNCREF == declare_elem_type
|
|
#if WASM_ENABLE_REF_TYPES != 0
|
|
|| VALUE_TYPE_EXTERNREF == declare_elem_type
|
|
#endif
|
|
);
|
|
|
|
read_leb_uint32(p, p_end, declare_max_size_flag);
|
|
read_leb_uint32(p, p_end, declare_init_size);
|
|
if (declare_max_size_flag & 1) {
|
|
read_leb_uint32(p, p_end, declare_max_size);
|
|
bh_assert(table->init_size <= table->max_size);
|
|
}
|
|
|
|
adjust_table_max_size(declare_init_size, declare_max_size_flag,
|
|
&declare_max_size);
|
|
*p_buf = p;
|
|
|
|
bh_assert(
|
|
!((declare_max_size_flag & 1) && declare_init_size > declare_max_size));
|
|
|
|
/* now we believe all declaration are ok */
|
|
table->elem_type = declare_elem_type;
|
|
table->init_size = declare_init_size;
|
|
table->flags = declare_max_size_flag;
|
|
table->max_size = declare_max_size;
|
|
return true;
|
|
}
|
|
|
|
static bool
|
|
load_memory_import(const uint8 **p_buf, const uint8 *buf_end,
|
|
WASMModule *parent_module, const char *sub_module_name,
|
|
const char *memory_name, WASMMemoryImport *memory,
|
|
char *error_buf, uint32 error_buf_size)
|
|
{
|
|
const uint8 *p = *p_buf, *p_end = buf_end;
|
|
#if WASM_ENABLE_APP_FRAMEWORK != 0
|
|
uint32 pool_size = wasm_runtime_memory_pool_size();
|
|
uint32 max_page_count = pool_size * APP_MEMORY_MAX_GLOBAL_HEAP_PERCENT
|
|
/ DEFAULT_NUM_BYTES_PER_PAGE;
|
|
#else
|
|
uint32 max_page_count;
|
|
#endif /* WASM_ENABLE_APP_FRAMEWORK */
|
|
uint32 mem_flag = 0;
|
|
bool is_memory64 = false;
|
|
uint32 declare_init_page_count = 0;
|
|
uint32 declare_max_page_count = 0;
|
|
|
|
read_leb_uint32(p, p_end, mem_flag);
|
|
if (!wasm_memory_check_flags(mem_flag, error_buf, error_buf_size, false)) {
|
|
return false;
|
|
}
|
|
|
|
#if WASM_ENABLE_APP_FRAMEWORK == 0
|
|
is_memory64 = mem_flag & MEMORY64_FLAG;
|
|
max_page_count = is_memory64 ? DEFAULT_MEM64_MAX_PAGES : DEFAULT_MAX_PAGES;
|
|
#endif
|
|
|
|
read_leb_uint32(p, p_end, declare_init_page_count);
|
|
bh_assert(declare_init_page_count <= max_page_count);
|
|
|
|
if (mem_flag & MAX_PAGE_COUNT_FLAG) {
|
|
read_leb_uint32(p, p_end, declare_max_page_count);
|
|
bh_assert(declare_init_page_count <= declare_max_page_count);
|
|
bh_assert(declare_max_page_count <= max_page_count);
|
|
if (declare_max_page_count > max_page_count) {
|
|
declare_max_page_count = max_page_count;
|
|
}
|
|
}
|
|
else {
|
|
/* Limit the maximum memory size to max_page_count */
|
|
declare_max_page_count = max_page_count;
|
|
}
|
|
|
|
/* now we believe all declaration are ok */
|
|
memory->flags = mem_flag;
|
|
memory->init_page_count = declare_init_page_count;
|
|
memory->max_page_count = declare_max_page_count;
|
|
memory->num_bytes_per_page = DEFAULT_NUM_BYTES_PER_PAGE;
|
|
|
|
*p_buf = p;
|
|
return true;
|
|
}
|
|
|
|
static bool
|
|
load_global_import(const uint8 **p_buf, const uint8 *buf_end,
|
|
const WASMModule *parent_module, char *sub_module_name,
|
|
char *global_name, WASMGlobalImport *global, char *error_buf,
|
|
uint32 error_buf_size)
|
|
{
|
|
const uint8 *p = *p_buf, *p_end = buf_end;
|
|
uint8 declare_type = 0;
|
|
uint8 declare_mutable = 0;
|
|
bool is_mutable = false;
|
|
bool ret = false;
|
|
|
|
CHECK_BUF(p, p_end, 2);
|
|
declare_type = read_uint8(p);
|
|
declare_mutable = read_uint8(p);
|
|
*p_buf = p;
|
|
|
|
bh_assert(declare_mutable < 2);
|
|
|
|
is_mutable = declare_mutable & 1 ? true : false;
|
|
|
|
#if WASM_ENABLE_LIBC_BUILTIN != 0
|
|
/* check built-in modules */
|
|
ret = wasm_native_lookup_libc_builtin_global(sub_module_name, global_name,
|
|
global);
|
|
if (ret) {
|
|
bh_assert(global->type.val_type == declare_type
|
|
&& global->type.is_mutable != declare_mutable);
|
|
}
|
|
#endif /* WASM_ENABLE_LIBC_BUILTIN */
|
|
|
|
global->is_linked = ret;
|
|
global->module_name = sub_module_name;
|
|
global->field_name = global_name;
|
|
global->type.val_type = declare_type;
|
|
global->type.is_mutable = is_mutable;
|
|
(void)p_end;
|
|
return true;
|
|
}
|
|
|
|
static bool
|
|
load_table(const uint8 **p_buf, const uint8 *buf_end, WASMTable *table,
|
|
char *error_buf, uint32 error_buf_size)
|
|
{
|
|
const uint8 *p = *p_buf, *p_end = buf_end, *p_org;
|
|
|
|
CHECK_BUF(p, p_end, 1);
|
|
/* 0x70 or 0x6F */
|
|
table->elem_type = read_uint8(p);
|
|
bh_assert((VALUE_TYPE_FUNCREF == table->elem_type)
|
|
#if WASM_ENABLE_REF_TYPES != 0
|
|
|| VALUE_TYPE_EXTERNREF == table->elem_type
|
|
#endif
|
|
);
|
|
|
|
p_org = p;
|
|
read_leb_uint32(p, p_end, table->flags);
|
|
bh_assert(p - p_org <= 1);
|
|
bh_assert(table->flags <= 1);
|
|
(void)p_org;
|
|
|
|
read_leb_uint32(p, p_end, table->init_size);
|
|
if (table->flags == 1) {
|
|
read_leb_uint32(p, p_end, table->max_size);
|
|
bh_assert(table->init_size <= table->max_size);
|
|
}
|
|
|
|
adjust_table_max_size(table->init_size, table->flags, &table->max_size);
|
|
|
|
*p_buf = p;
|
|
return true;
|
|
}
|
|
|
|
static bool
|
|
load_memory(const uint8 **p_buf, const uint8 *buf_end, WASMMemory *memory,
|
|
char *error_buf, uint32 error_buf_size)
|
|
{
|
|
const uint8 *p = *p_buf, *p_end = buf_end, *p_org;
|
|
#if WASM_ENABLE_APP_FRAMEWORK != 0
|
|
uint32 pool_size = wasm_runtime_memory_pool_size();
|
|
uint32 max_page_count = pool_size * APP_MEMORY_MAX_GLOBAL_HEAP_PERCENT
|
|
/ DEFAULT_NUM_BYTES_PER_PAGE;
|
|
#else
|
|
uint32 max_page_count;
|
|
bool is_memory64 = false;
|
|
#endif
|
|
|
|
p_org = p;
|
|
read_leb_uint32(p, p_end, memory->flags);
|
|
bh_assert(p - p_org <= 1);
|
|
(void)p_org;
|
|
if (!wasm_memory_check_flags(memory->flags, error_buf, error_buf_size,
|
|
false)) {
|
|
return false;
|
|
}
|
|
|
|
#if WASM_ENABLE_APP_FRAMEWORK == 0
|
|
is_memory64 = memory->flags & MEMORY64_FLAG;
|
|
max_page_count = is_memory64 ? DEFAULT_MEM64_MAX_PAGES : DEFAULT_MAX_PAGES;
|
|
#endif
|
|
|
|
read_leb_uint32(p, p_end, memory->init_page_count);
|
|
bh_assert(memory->init_page_count <= max_page_count);
|
|
|
|
if (memory->flags & 1) {
|
|
read_leb_uint32(p, p_end, memory->max_page_count);
|
|
bh_assert(memory->init_page_count <= memory->max_page_count);
|
|
bh_assert(memory->max_page_count <= max_page_count);
|
|
if (memory->max_page_count > max_page_count)
|
|
memory->max_page_count = max_page_count;
|
|
}
|
|
else {
|
|
/* Limit the maximum memory size to max_page_count */
|
|
memory->max_page_count = max_page_count;
|
|
}
|
|
|
|
memory->num_bytes_per_page = DEFAULT_NUM_BYTES_PER_PAGE;
|
|
|
|
*p_buf = p;
|
|
return true;
|
|
}
|
|
|
|
static bool
|
|
load_import_section(const uint8 *buf, const uint8 *buf_end, WASMModule *module,
|
|
bool is_load_from_file_buf, char *error_buf,
|
|
uint32 error_buf_size)
|
|
{
|
|
const uint8 *p = buf, *p_end = buf_end, *p_old;
|
|
uint32 import_count, name_len, type_index, i, u32, flags;
|
|
uint64 total_size;
|
|
WASMImport *import;
|
|
WASMImport *import_functions = NULL, *import_tables = NULL;
|
|
WASMImport *import_memories = NULL, *import_globals = NULL;
|
|
char *sub_module_name, *field_name;
|
|
uint8 u8, kind;
|
|
|
|
read_leb_uint32(p, p_end, import_count);
|
|
|
|
if (import_count) {
|
|
module->import_count = import_count;
|
|
total_size = sizeof(WASMImport) * (uint64)import_count;
|
|
if (!(module->imports =
|
|
loader_malloc(total_size, error_buf, error_buf_size))) {
|
|
return false;
|
|
}
|
|
|
|
p_old = p;
|
|
|
|
/* Scan firstly to get import count of each type */
|
|
for (i = 0; i < import_count; i++) {
|
|
/* module name */
|
|
read_leb_uint32(p, p_end, name_len);
|
|
CHECK_BUF(p, p_end, name_len);
|
|
p += name_len;
|
|
|
|
/* field name */
|
|
read_leb_uint32(p, p_end, name_len);
|
|
CHECK_BUF(p, p_end, name_len);
|
|
p += name_len;
|
|
|
|
CHECK_BUF(p, p_end, 1);
|
|
/* 0x00/0x01/0x02/0x03 */
|
|
kind = read_uint8(p);
|
|
|
|
switch (kind) {
|
|
case IMPORT_KIND_FUNC: /* import function */
|
|
read_leb_uint32(p, p_end, type_index);
|
|
module->import_function_count++;
|
|
break;
|
|
|
|
case IMPORT_KIND_TABLE: /* import table */
|
|
CHECK_BUF(p, p_end, 1);
|
|
/* 0x70 */
|
|
u8 = read_uint8(p);
|
|
read_leb_uint32(p, p_end, flags);
|
|
read_leb_uint32(p, p_end, u32);
|
|
if (flags & 1)
|
|
read_leb_uint32(p, p_end, u32);
|
|
module->import_table_count++;
|
|
#if WASM_ENABLE_REF_TYPES == 0
|
|
bh_assert(module->import_table_count <= 1);
|
|
#endif
|
|
break;
|
|
|
|
case IMPORT_KIND_MEMORY: /* import memory */
|
|
read_leb_uint32(p, p_end, flags);
|
|
read_leb_uint32(p, p_end, u32);
|
|
if (flags & 1)
|
|
read_leb_uint32(p, p_end, u32);
|
|
module->import_memory_count++;
|
|
bh_assert(module->import_memory_count <= 1);
|
|
break;
|
|
|
|
case IMPORT_KIND_GLOBAL: /* import global */
|
|
CHECK_BUF(p, p_end, 2);
|
|
p += 2;
|
|
module->import_global_count++;
|
|
break;
|
|
|
|
default:
|
|
bh_assert(0);
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (module->import_function_count)
|
|
import_functions = module->import_functions = module->imports;
|
|
if (module->import_table_count)
|
|
import_tables = module->import_tables =
|
|
module->imports + module->import_function_count;
|
|
if (module->import_memory_count)
|
|
import_memories = module->import_memories =
|
|
module->imports + module->import_function_count
|
|
+ module->import_table_count;
|
|
if (module->import_global_count)
|
|
import_globals = module->import_globals =
|
|
module->imports + module->import_function_count
|
|
+ module->import_table_count + module->import_memory_count;
|
|
|
|
p = p_old;
|
|
|
|
/* Scan again to resolve the data */
|
|
for (i = 0; i < import_count; i++) {
|
|
WASMModule *sub_module = NULL;
|
|
|
|
/* load module name */
|
|
read_leb_uint32(p, p_end, name_len);
|
|
CHECK_BUF(p, p_end, name_len);
|
|
if (!(sub_module_name = wasm_const_str_list_insert(
|
|
p, name_len, module, is_load_from_file_buf, error_buf,
|
|
error_buf_size))) {
|
|
return false;
|
|
}
|
|
p += name_len;
|
|
|
|
/* load field name */
|
|
read_leb_uint32(p, p_end, name_len);
|
|
CHECK_BUF(p, p_end, name_len);
|
|
if (!(field_name = wasm_const_str_list_insert(
|
|
p, name_len, module, is_load_from_file_buf, error_buf,
|
|
error_buf_size))) {
|
|
return false;
|
|
}
|
|
p += name_len;
|
|
|
|
CHECK_BUF(p, p_end, 1);
|
|
/* 0x00/0x01/0x02/0x03 */
|
|
kind = read_uint8(p);
|
|
|
|
LOG_DEBUG("import #%d: (%s, %s), kind: %d", i, sub_module_name,
|
|
field_name, kind);
|
|
switch (kind) {
|
|
case IMPORT_KIND_FUNC: /* import function */
|
|
bh_assert(import_functions);
|
|
import = import_functions++;
|
|
if (!load_function_import(
|
|
&p, p_end, module, sub_module_name, field_name,
|
|
&import->u.function, error_buf, error_buf_size)) {
|
|
return false;
|
|
}
|
|
break;
|
|
|
|
case IMPORT_KIND_TABLE: /* import table */
|
|
bh_assert(import_tables);
|
|
import = import_tables++;
|
|
if (!load_table_import(&p, p_end, module, sub_module_name,
|
|
field_name, &import->u.table,
|
|
error_buf, error_buf_size)) {
|
|
LOG_DEBUG("can not import such a table (%s,%s)",
|
|
sub_module_name, field_name);
|
|
return false;
|
|
}
|
|
break;
|
|
|
|
case IMPORT_KIND_MEMORY: /* import memory */
|
|
bh_assert(import_memories);
|
|
import = import_memories++;
|
|
if (!load_memory_import(&p, p_end, module, sub_module_name,
|
|
field_name, &import->u.memory,
|
|
error_buf, error_buf_size)) {
|
|
return false;
|
|
}
|
|
break;
|
|
|
|
case IMPORT_KIND_GLOBAL: /* import global */
|
|
bh_assert(import_globals);
|
|
import = import_globals++;
|
|
if (!load_global_import(&p, p_end, module, sub_module_name,
|
|
field_name, &import->u.global,
|
|
error_buf, error_buf_size)) {
|
|
return false;
|
|
}
|
|
break;
|
|
|
|
default:
|
|
bh_assert(0);
|
|
import = NULL;
|
|
break;
|
|
}
|
|
import->kind = kind;
|
|
import->u.names.module_name = sub_module_name;
|
|
import->u.names.field_name = field_name;
|
|
(void)sub_module;
|
|
}
|
|
|
|
#if WASM_ENABLE_LIBC_WASI != 0
|
|
import = module->import_functions;
|
|
for (i = 0; i < module->import_function_count; i++, import++) {
|
|
if (!strcmp(import->u.names.module_name, "wasi_unstable")
|
|
|| !strcmp(import->u.names.module_name,
|
|
"wasi_snapshot_preview1")) {
|
|
module->import_wasi_api = true;
|
|
break;
|
|
}
|
|
}
|
|
#endif
|
|
}
|
|
|
|
bh_assert(p == p_end);
|
|
|
|
LOG_VERBOSE("Load import section success.\n");
|
|
(void)u8;
|
|
(void)u32;
|
|
(void)type_index;
|
|
return true;
|
|
}
|
|
|
|
static bool
|
|
init_function_local_offsets(WASMFunction *func, char *error_buf,
|
|
uint32 error_buf_size)
|
|
{
|
|
WASMFuncType *param_type = func->func_type;
|
|
uint32 param_count = param_type->param_count;
|
|
uint8 *param_types = param_type->types;
|
|
uint32 local_count = func->local_count;
|
|
uint8 *local_types = func->local_types;
|
|
uint32 i, local_offset = 0;
|
|
uint64 total_size = sizeof(uint16) * ((uint64)param_count + local_count);
|
|
|
|
if (total_size > 0
|
|
&& !(func->local_offsets =
|
|
loader_malloc(total_size, error_buf, error_buf_size))) {
|
|
return false;
|
|
}
|
|
|
|
for (i = 0; i < param_count; i++) {
|
|
func->local_offsets[i] = (uint16)local_offset;
|
|
local_offset += wasm_value_type_cell_num(param_types[i]);
|
|
}
|
|
|
|
for (i = 0; i < local_count; i++) {
|
|
func->local_offsets[param_count + i] = (uint16)local_offset;
|
|
local_offset += wasm_value_type_cell_num(local_types[i]);
|
|
}
|
|
|
|
bh_assert(local_offset == func->param_cell_num + func->local_cell_num);
|
|
return true;
|
|
}
|
|
|
|
static bool
|
|
load_function_section(const uint8 *buf, const uint8 *buf_end,
|
|
const uint8 *buf_code, const uint8 *buf_code_end,
|
|
WASMModule *module, char *error_buf,
|
|
uint32 error_buf_size)
|
|
{
|
|
const uint8 *p = buf, *p_end = buf_end;
|
|
const uint8 *p_code = buf_code, *p_code_end, *p_code_save;
|
|
uint32 func_count;
|
|
uint64 total_size;
|
|
uint32 code_count = 0, code_size, type_index, i, j, k, local_type_index;
|
|
uint32 local_count, local_set_count, sub_local_count, local_cell_num;
|
|
uint8 type;
|
|
WASMFunction *func;
|
|
|
|
read_leb_uint32(p, p_end, func_count);
|
|
|
|
if (buf_code)
|
|
read_leb_uint32(p_code, buf_code_end, code_count);
|
|
|
|
bh_assert(func_count == code_count);
|
|
|
|
if (func_count) {
|
|
module->function_count = func_count;
|
|
total_size = sizeof(WASMFunction *) * (uint64)func_count;
|
|
if (!(module->functions =
|
|
loader_malloc(total_size, error_buf, error_buf_size))) {
|
|
return false;
|
|
}
|
|
|
|
for (i = 0; i < func_count; i++) {
|
|
/* Resolve function type */
|
|
read_leb_uint32(p, p_end, type_index);
|
|
bh_assert(type_index < module->type_count);
|
|
|
|
#if (WASM_ENABLE_WAMR_COMPILER != 0) || (WASM_ENABLE_JIT != 0)
|
|
type_index = wasm_get_smallest_type_idx(
|
|
module->types, module->type_count, type_index);
|
|
#endif
|
|
|
|
read_leb_uint32(p_code, buf_code_end, code_size);
|
|
bh_assert(code_size > 0 && p_code + code_size <= buf_code_end);
|
|
|
|
/* Resolve local set count */
|
|
p_code_end = p_code + code_size;
|
|
local_count = 0;
|
|
read_leb_uint32(p_code, buf_code_end, local_set_count);
|
|
p_code_save = p_code;
|
|
|
|
/* Calculate total local count */
|
|
for (j = 0; j < local_set_count; j++) {
|
|
read_leb_uint32(p_code, buf_code_end, sub_local_count);
|
|
bh_assert(sub_local_count <= UINT32_MAX - local_count);
|
|
|
|
CHECK_BUF(p_code, buf_code_end, 1);
|
|
/* 0x7F/0x7E/0x7D/0x7C */
|
|
type = read_uint8(p_code);
|
|
local_count += sub_local_count;
|
|
}
|
|
|
|
/* Alloc memory, layout: function structure + local types */
|
|
code_size = (uint32)(p_code_end - p_code);
|
|
|
|
total_size = sizeof(WASMFunction) + (uint64)local_count;
|
|
if (!(func = module->functions[i] =
|
|
loader_malloc(total_size, error_buf, error_buf_size))) {
|
|
return false;
|
|
}
|
|
|
|
/* Set function type, local count, code size and code body */
|
|
func->func_type = module->types[type_index];
|
|
func->local_count = local_count;
|
|
if (local_count > 0)
|
|
func->local_types = (uint8 *)func + sizeof(WASMFunction);
|
|
func->code_size = code_size;
|
|
/*
|
|
* we shall make a copy of code body [p_code, p_code + code_size]
|
|
* when we are worrying about inappropriate releasing behaviour.
|
|
* all code bodies are actually in a buffer which user allocates in
|
|
* his embedding environment and we don't have power on them.
|
|
* it will be like:
|
|
* code_body_cp = malloc(code_size);
|
|
* memcpy(code_body_cp, p_code, code_size);
|
|
* func->code = code_body_cp;
|
|
*/
|
|
func->code = (uint8 *)p_code;
|
|
|
|
/* Load each local type */
|
|
p_code = p_code_save;
|
|
local_type_index = 0;
|
|
for (j = 0; j < local_set_count; j++) {
|
|
read_leb_uint32(p_code, buf_code_end, sub_local_count);
|
|
/* Note: sub_local_count is allowed to be 0 */
|
|
bh_assert(local_type_index <= UINT32_MAX - sub_local_count
|
|
&& local_type_index + sub_local_count <= local_count);
|
|
|
|
CHECK_BUF(p_code, buf_code_end, 1);
|
|
/* 0x7F/0x7E/0x7D/0x7C */
|
|
type = read_uint8(p_code);
|
|
bh_assert(is_value_type(type));
|
|
for (k = 0; k < sub_local_count; k++) {
|
|
func->local_types[local_type_index++] = type;
|
|
}
|
|
}
|
|
|
|
func->param_cell_num = func->func_type->param_cell_num;
|
|
func->ret_cell_num = func->func_type->ret_cell_num;
|
|
local_cell_num =
|
|
wasm_get_cell_num(func->local_types, func->local_count);
|
|
bh_assert(local_cell_num <= UINT16_MAX);
|
|
|
|
func->local_cell_num = (uint16)local_cell_num;
|
|
|
|
if (!init_function_local_offsets(func, error_buf, error_buf_size))
|
|
return false;
|
|
|
|
p_code = p_code_end;
|
|
}
|
|
}
|
|
|
|
bh_assert(p == p_end);
|
|
LOG_VERBOSE("Load function section success.\n");
|
|
(void)code_count;
|
|
return true;
|
|
}
|
|
|
|
static bool
|
|
load_table_section(const uint8 *buf, const uint8 *buf_end, WASMModule *module,
|
|
char *error_buf, uint32 error_buf_size)
|
|
{
|
|
const uint8 *p = buf, *p_end = buf_end;
|
|
uint32 table_count, i;
|
|
uint64 total_size;
|
|
WASMTable *table;
|
|
|
|
read_leb_uint32(p, p_end, table_count);
|
|
#if WASM_ENABLE_REF_TYPES == 0
|
|
bh_assert(module->import_table_count + table_count <= 1);
|
|
#endif
|
|
|
|
if (table_count) {
|
|
module->table_count = table_count;
|
|
total_size = sizeof(WASMTable) * (uint64)table_count;
|
|
if (!(module->tables =
|
|
loader_malloc(total_size, error_buf, error_buf_size))) {
|
|
return false;
|
|
}
|
|
|
|
/* load each table */
|
|
table = module->tables;
|
|
for (i = 0; i < table_count; i++, table++)
|
|
if (!load_table(&p, p_end, table, error_buf, error_buf_size))
|
|
return false;
|
|
}
|
|
|
|
bh_assert(p == p_end);
|
|
LOG_VERBOSE("Load table section success.\n");
|
|
return true;
|
|
}
|
|
|
|
static bool
|
|
load_memory_section(const uint8 *buf, const uint8 *buf_end, WASMModule *module,
|
|
char *error_buf, uint32 error_buf_size)
|
|
{
|
|
const uint8 *p = buf, *p_end = buf_end;
|
|
uint32 memory_count, i;
|
|
uint64 total_size;
|
|
WASMMemory *memory;
|
|
|
|
read_leb_uint32(p, p_end, memory_count);
|
|
bh_assert(module->import_memory_count + memory_count <= 1);
|
|
|
|
if (memory_count) {
|
|
module->memory_count = memory_count;
|
|
total_size = sizeof(WASMMemory) * (uint64)memory_count;
|
|
if (!(module->memories =
|
|
loader_malloc(total_size, error_buf, error_buf_size))) {
|
|
return false;
|
|
}
|
|
|
|
/* load each memory */
|
|
memory = module->memories;
|
|
for (i = 0; i < memory_count; i++, memory++)
|
|
if (!load_memory(&p, p_end, memory, error_buf, error_buf_size))
|
|
return false;
|
|
}
|
|
|
|
bh_assert(p == p_end);
|
|
LOG_VERBOSE("Load memory section success.\n");
|
|
return true;
|
|
}
|
|
|
|
static bool
|
|
load_global_section(const uint8 *buf, const uint8 *buf_end, WASMModule *module,
|
|
char *error_buf, uint32 error_buf_size)
|
|
{
|
|
const uint8 *p = buf, *p_end = buf_end;
|
|
uint32 global_count, i;
|
|
uint64 total_size;
|
|
WASMGlobal *global;
|
|
uint8 mutable;
|
|
|
|
read_leb_uint32(p, p_end, global_count);
|
|
|
|
module->global_count = 0;
|
|
if (global_count) {
|
|
total_size = sizeof(WASMGlobal) * (uint64)global_count;
|
|
if (!(module->globals =
|
|
loader_malloc(total_size, error_buf, error_buf_size))) {
|
|
return false;
|
|
}
|
|
|
|
global = module->globals;
|
|
|
|
for (i = 0; i < global_count; i++, global++) {
|
|
CHECK_BUF(p, p_end, 2);
|
|
global->type.val_type = read_uint8(p);
|
|
mutable = read_uint8(p);
|
|
bh_assert(mutable < 2);
|
|
global->type.is_mutable = mutable ? true : false;
|
|
|
|
/* initialize expression */
|
|
if (!load_init_expr(module, &p, p_end, &(global->init_expr),
|
|
global->type.val_type, error_buf,
|
|
error_buf_size))
|
|
return false;
|
|
|
|
if (INIT_EXPR_TYPE_GET_GLOBAL == global->init_expr.init_expr_type) {
|
|
/**
|
|
* Currently, constant expressions occurring as initializers
|
|
* of globals are further constrained in that contained
|
|
* global.get instructions are
|
|
* only allowed to refer to imported globals.
|
|
*/
|
|
uint32 target_global_index = global->init_expr.u.global_index;
|
|
bh_assert(target_global_index < module->import_global_count);
|
|
(void)target_global_index;
|
|
}
|
|
else if (INIT_EXPR_TYPE_FUNCREF_CONST
|
|
== global->init_expr.init_expr_type) {
|
|
bh_assert(global->init_expr.u.ref_index
|
|
< module->import_function_count
|
|
+ module->function_count);
|
|
}
|
|
|
|
module->global_count++;
|
|
}
|
|
bh_assert(module->global_count == global_count);
|
|
}
|
|
|
|
bh_assert(p == p_end);
|
|
LOG_VERBOSE("Load global section success.\n");
|
|
return true;
|
|
}
|
|
|
|
static bool
|
|
load_export_section(const uint8 *buf, const uint8 *buf_end, WASMModule *module,
|
|
bool is_load_from_file_buf, char *error_buf,
|
|
uint32 error_buf_size)
|
|
{
|
|
const uint8 *p = buf, *p_end = buf_end;
|
|
uint32 export_count, i, j, index;
|
|
uint64 total_size;
|
|
uint32 str_len;
|
|
WASMExport *export;
|
|
const char *name;
|
|
|
|
read_leb_uint32(p, p_end, export_count);
|
|
|
|
if (export_count) {
|
|
module->export_count = export_count;
|
|
total_size = sizeof(WASMExport) * (uint64)export_count;
|
|
if (!(module->exports =
|
|
loader_malloc(total_size, error_buf, error_buf_size))) {
|
|
return false;
|
|
}
|
|
|
|
export = module->exports;
|
|
for (i = 0; i < export_count; i++, export ++) {
|
|
read_leb_uint32(p, p_end, str_len);
|
|
CHECK_BUF(p, p_end, str_len);
|
|
|
|
for (j = 0; j < i; j++) {
|
|
name = module->exports[j].name;
|
|
bh_assert(!(strlen(name) == str_len
|
|
&& memcmp(name, p, str_len) == 0));
|
|
}
|
|
|
|
if (!(export->name = wasm_const_str_list_insert(
|
|
p, str_len, module, is_load_from_file_buf, error_buf,
|
|
error_buf_size))) {
|
|
return false;
|
|
}
|
|
|
|
p += str_len;
|
|
CHECK_BUF(p, p_end, 1);
|
|
export->kind = read_uint8(p);
|
|
read_leb_uint32(p, p_end, index);
|
|
export->index = index;
|
|
|
|
switch (export->kind) {
|
|
/* function index */
|
|
case EXPORT_KIND_FUNC:
|
|
bh_assert(index < module->function_count
|
|
+ module->import_function_count);
|
|
break;
|
|
/* table index */
|
|
case EXPORT_KIND_TABLE:
|
|
bh_assert(index < module->table_count
|
|
+ module->import_table_count);
|
|
break;
|
|
/* memory index */
|
|
case EXPORT_KIND_MEMORY:
|
|
bh_assert(index < module->memory_count
|
|
+ module->import_memory_count);
|
|
break;
|
|
/* global index */
|
|
case EXPORT_KIND_GLOBAL:
|
|
bh_assert(index < module->global_count
|
|
+ module->import_global_count);
|
|
break;
|
|
default:
|
|
bh_assert(0);
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
bh_assert(p == p_end);
|
|
LOG_VERBOSE("Load export section success.\n");
|
|
(void)name;
|
|
return true;
|
|
}
|
|
|
|
static bool
|
|
check_table_index(const WASMModule *module, uint32 table_index, char *error_buf,
|
|
uint32 error_buf_size)
|
|
{
|
|
#if WASM_ENABLE_REF_TYPES == 0
|
|
if (table_index != 0) {
|
|
return false;
|
|
}
|
|
#endif
|
|
|
|
if (table_index >= module->import_table_count + module->table_count) {
|
|
return false;
|
|
}
|
|
return true;
|
|
}
|
|
|
|
static bool
|
|
load_table_index(const uint8 **p_buf, const uint8 *buf_end, WASMModule *module,
|
|
uint32 *p_table_index, char *error_buf, uint32 error_buf_size)
|
|
{
|
|
const uint8 *p = *p_buf, *p_end = buf_end;
|
|
uint32 table_index;
|
|
|
|
read_leb_uint32(p, p_end, table_index);
|
|
if (!check_table_index(module, table_index, error_buf, error_buf_size)) {
|
|
return false;
|
|
}
|
|
|
|
*p_table_index = table_index;
|
|
*p_buf = p;
|
|
return true;
|
|
}
|
|
|
|
#if WASM_ENABLE_REF_TYPES != 0
|
|
static bool
|
|
load_elem_type(const uint8 **p_buf, const uint8 *buf_end, uint32 *p_elem_type,
|
|
bool elemkind_zero, char *error_buf, uint32 error_buf_size)
|
|
{
|
|
const uint8 *p = *p_buf, *p_end = buf_end;
|
|
uint8 elem_type;
|
|
|
|
CHECK_BUF(p, p_end, 1);
|
|
elem_type = read_uint8(p);
|
|
if ((elemkind_zero && elem_type != 0)
|
|
|| (!elemkind_zero && elem_type != VALUE_TYPE_FUNCREF
|
|
&& elem_type != VALUE_TYPE_EXTERNREF)) {
|
|
set_error_buf(error_buf, error_buf_size, "invalid reference type");
|
|
return false;
|
|
}
|
|
|
|
if (elemkind_zero)
|
|
*p_elem_type = VALUE_TYPE_FUNCREF;
|
|
else
|
|
*p_elem_type = elem_type;
|
|
*p_buf = p;
|
|
|
|
(void)p_end;
|
|
return true;
|
|
}
|
|
#endif /* WASM_ENABLE_REF_TYPES != 0*/
|
|
|
|
static bool
|
|
load_func_index_vec(const uint8 **p_buf, const uint8 *buf_end,
|
|
WASMModule *module, WASMTableSeg *table_segment,
|
|
char *error_buf, uint32 error_buf_size)
|
|
{
|
|
const uint8 *p = *p_buf, *p_end = buf_end;
|
|
uint32 function_count, function_index = 0, i;
|
|
uint64 total_size;
|
|
|
|
read_leb_uint32(p, p_end, function_count);
|
|
table_segment->value_count = function_count;
|
|
total_size = sizeof(InitializerExpression) * (uint64)function_count;
|
|
if (total_size > 0
|
|
&& !(table_segment->init_values =
|
|
(InitializerExpression *)loader_malloc(total_size, error_buf,
|
|
error_buf_size))) {
|
|
return false;
|
|
}
|
|
|
|
for (i = 0; i < function_count; i++) {
|
|
InitializerExpression *init_expr = &table_segment->init_values[i];
|
|
|
|
read_leb_uint32(p, p_end, function_index);
|
|
if (!check_function_index(module, function_index, error_buf,
|
|
error_buf_size)) {
|
|
return false;
|
|
}
|
|
|
|
init_expr->init_expr_type = INIT_EXPR_TYPE_FUNCREF_CONST;
|
|
init_expr->u.ref_index = function_index;
|
|
}
|
|
|
|
*p_buf = p;
|
|
return true;
|
|
}
|
|
|
|
#if WASM_ENABLE_REF_TYPES != 0
|
|
static bool
|
|
load_init_expr_vec(const uint8 **p_buf, const uint8 *buf_end,
|
|
WASMModule *module, WASMTableSeg *table_segment,
|
|
char *error_buf, uint32 error_buf_size)
|
|
{
|
|
const uint8 *p = *p_buf, *p_end = buf_end;
|
|
uint32 ref_count, i;
|
|
uint64 total_size;
|
|
|
|
read_leb_uint32(p, p_end, ref_count);
|
|
table_segment->value_count = ref_count;
|
|
total_size = sizeof(InitializerExpression) * (uint64)ref_count;
|
|
if (total_size > 0
|
|
&& !(table_segment->init_values =
|
|
(InitializerExpression *)loader_malloc(total_size, error_buf,
|
|
error_buf_size))) {
|
|
return false;
|
|
}
|
|
|
|
for (i = 0; i < ref_count; i++) {
|
|
InitializerExpression *init_expr = &table_segment->init_values[i];
|
|
|
|
if (!load_init_expr(module, &p, p_end, init_expr,
|
|
table_segment->elem_type, error_buf,
|
|
error_buf_size))
|
|
return false;
|
|
|
|
bh_assert(
|
|
(init_expr->init_expr_type == INIT_EXPR_TYPE_GET_GLOBAL)
|
|
|| (init_expr->init_expr_type == INIT_EXPR_TYPE_REFNULL_CONST)
|
|
|| (init_expr->init_expr_type == INIT_EXPR_TYPE_FUNCREF_CONST));
|
|
}
|
|
|
|
*p_buf = p;
|
|
return true;
|
|
}
|
|
#endif /* end of WASM_ENABLE_REF_TYPES != 0 */
|
|
|
|
static bool
|
|
load_table_segment_section(const uint8 *buf, const uint8 *buf_end,
|
|
WASMModule *module, char *error_buf,
|
|
uint32 error_buf_size)
|
|
{
|
|
const uint8 *p = buf, *p_end = buf_end;
|
|
uint32 table_segment_count, i, table_index, function_count;
|
|
uint64 total_size;
|
|
WASMTableSeg *table_segment;
|
|
|
|
read_leb_uint32(p, p_end, table_segment_count);
|
|
|
|
if (table_segment_count) {
|
|
module->table_seg_count = table_segment_count;
|
|
total_size = sizeof(WASMTableSeg) * (uint64)table_segment_count;
|
|
if (!(module->table_segments =
|
|
loader_malloc(total_size, error_buf, error_buf_size))) {
|
|
return false;
|
|
}
|
|
|
|
table_segment = module->table_segments;
|
|
for (i = 0; i < table_segment_count; i++, table_segment++) {
|
|
bh_assert(p < p_end);
|
|
|
|
#if WASM_ENABLE_REF_TYPES != 0
|
|
read_leb_uint32(p, p_end, table_segment->mode);
|
|
/* last three bits */
|
|
table_segment->mode = table_segment->mode & 0x07;
|
|
switch (table_segment->mode) {
|
|
/* elemkind/elemtype + active */
|
|
case 0:
|
|
case 4:
|
|
table_segment->elem_type = VALUE_TYPE_FUNCREF;
|
|
table_segment->table_index = 0;
|
|
|
|
if (!check_table_index(module, table_segment->table_index,
|
|
error_buf, error_buf_size))
|
|
return false;
|
|
|
|
if (!load_init_expr(
|
|
module, &p, p_end, &table_segment->base_offset,
|
|
VALUE_TYPE_I32, error_buf, error_buf_size))
|
|
return false;
|
|
|
|
if (table_segment->mode == 0) {
|
|
/* vec(funcidx) */
|
|
if (!load_func_index_vec(&p, p_end, module,
|
|
table_segment, error_buf,
|
|
error_buf_size))
|
|
return false;
|
|
}
|
|
else {
|
|
/* vec(expr) */
|
|
if (!load_init_expr_vec(&p, p_end, module,
|
|
table_segment, error_buf,
|
|
error_buf_size))
|
|
return false;
|
|
}
|
|
break;
|
|
/* elemkind + passive/declarative */
|
|
case 1:
|
|
case 3:
|
|
if (!load_elem_type(&p, p_end, &table_segment->elem_type,
|
|
true, error_buf, error_buf_size))
|
|
return false;
|
|
/* vec(funcidx) */
|
|
if (!load_func_index_vec(&p, p_end, module, table_segment,
|
|
error_buf, error_buf_size))
|
|
return false;
|
|
break;
|
|
/* elemkind/elemtype + table_idx + active */
|
|
case 2:
|
|
case 6:
|
|
if (!load_table_index(&p, p_end, module,
|
|
&table_segment->table_index,
|
|
error_buf, error_buf_size))
|
|
return false;
|
|
if (!load_init_expr(
|
|
module, &p, p_end, &table_segment->base_offset,
|
|
VALUE_TYPE_I32, error_buf, error_buf_size))
|
|
return false;
|
|
if (!load_elem_type(&p, p_end, &table_segment->elem_type,
|
|
table_segment->mode == 2 ? true : false,
|
|
error_buf, error_buf_size))
|
|
return false;
|
|
if (table_segment->mode == 2) {
|
|
/* vec(funcidx) */
|
|
if (!load_func_index_vec(&p, p_end, module,
|
|
table_segment, error_buf,
|
|
error_buf_size))
|
|
return false;
|
|
}
|
|
else {
|
|
/* vec(expr) */
|
|
if (!load_init_expr_vec(&p, p_end, module,
|
|
table_segment, error_buf,
|
|
error_buf_size))
|
|
return false;
|
|
}
|
|
break;
|
|
case 5:
|
|
case 7:
|
|
if (!load_elem_type(&p, p_end, &table_segment->elem_type,
|
|
false, error_buf, error_buf_size))
|
|
return false;
|
|
/* vec(expr) */
|
|
if (!load_init_expr_vec(&p, p_end, module, table_segment,
|
|
error_buf, error_buf_size))
|
|
return false;
|
|
break;
|
|
default:
|
|
return false;
|
|
}
|
|
#else
|
|
/*
|
|
* like: 00 41 05 0b 04 00 01 00 01
|
|
* for: (elem 0 (offset (i32.const 5)) $f1 $f2 $f1 $f2)
|
|
*/
|
|
if (!load_table_index(&p, p_end, module,
|
|
&table_segment->table_index, error_buf,
|
|
error_buf_size))
|
|
return false;
|
|
if (!load_init_expr(module, &p, p_end, &table_segment->base_offset,
|
|
VALUE_TYPE_I32, error_buf, error_buf_size))
|
|
return false;
|
|
if (!load_func_index_vec(&p, p_end, module, table_segment,
|
|
error_buf, error_buf_size))
|
|
return false;
|
|
#endif /* WASM_ENABLE_REF_TYPES != 0 */
|
|
}
|
|
}
|
|
|
|
(void)table_index;
|
|
(void)function_count;
|
|
bh_assert(p == p_end);
|
|
LOG_VERBOSE("Load table segment section success.\n");
|
|
return true;
|
|
}
|
|
|
|
static bool
|
|
load_data_segment_section(const uint8 *buf, const uint8 *buf_end,
|
|
WASMModule *module, bool clone_data_seg,
|
|
char *error_buf, uint32 error_buf_size)
|
|
{
|
|
const uint8 *p = buf, *p_end = buf_end;
|
|
uint32 data_seg_count, i, mem_index, data_seg_len;
|
|
uint64 total_size;
|
|
WASMDataSeg *dataseg;
|
|
InitializerExpression init_expr;
|
|
#if WASM_ENABLE_BULK_MEMORY != 0
|
|
bool is_passive = false;
|
|
uint32 mem_flag;
|
|
#endif
|
|
uint8 mem_offset_type = VALUE_TYPE_I32;
|
|
|
|
read_leb_uint32(p, p_end, data_seg_count);
|
|
|
|
#if WASM_ENABLE_BULK_MEMORY != 0
|
|
bh_assert(module->data_seg_count1 == 0
|
|
|| data_seg_count == module->data_seg_count1);
|
|
#endif
|
|
|
|
if (data_seg_count) {
|
|
module->data_seg_count = data_seg_count;
|
|
total_size = sizeof(WASMDataSeg *) * (uint64)data_seg_count;
|
|
if (!(module->data_segments =
|
|
loader_malloc(total_size, error_buf, error_buf_size))) {
|
|
return false;
|
|
}
|
|
|
|
for (i = 0; i < data_seg_count; i++) {
|
|
read_leb_uint32(p, p_end, mem_index);
|
|
#if WASM_ENABLE_BULK_MEMORY != 0
|
|
is_passive = false;
|
|
mem_flag = mem_index & 0x03;
|
|
switch (mem_flag) {
|
|
case 0x01:
|
|
is_passive = true;
|
|
break;
|
|
case 0x00:
|
|
/* no memory index, treat index as 0 */
|
|
mem_index = 0;
|
|
goto check_mem_index;
|
|
case 0x02:
|
|
/* read following memory index */
|
|
read_leb_uint32(p, p_end, mem_index);
|
|
check_mem_index:
|
|
bh_assert(mem_index < module->import_memory_count
|
|
+ module->memory_count);
|
|
break;
|
|
case 0x03:
|
|
default:
|
|
bh_assert(0);
|
|
break;
|
|
}
|
|
#else
|
|
bh_assert(mem_index
|
|
< module->import_memory_count + module->memory_count);
|
|
#endif /* WASM_ENABLE_BULK_MEMORY */
|
|
|
|
#if WASM_ENABLE_BULK_MEMORY != 0
|
|
if (!is_passive)
|
|
#endif /* WASM_ENABLE_BULK_MEMORY */
|
|
{
|
|
#if WASM_ENABLE_MEMORY64 != 0
|
|
/* This memory_flag is from memory instead of data segment */
|
|
uint8 memory_flag;
|
|
if (module->import_memory_count > 0) {
|
|
memory_flag =
|
|
module->import_memories[mem_index].u.memory.flags;
|
|
}
|
|
else {
|
|
memory_flag =
|
|
module
|
|
->memories[mem_index - module->import_memory_count]
|
|
.flags;
|
|
}
|
|
mem_offset_type = memory_flag & MEMORY64_FLAG ? VALUE_TYPE_I64
|
|
: VALUE_TYPE_I32;
|
|
#else
|
|
mem_offset_type = VALUE_TYPE_I32;
|
|
#endif /* WASM_ENABLE_MEMORY64 */
|
|
}
|
|
|
|
#if WASM_ENABLE_BULK_MEMORY != 0
|
|
if (!is_passive)
|
|
#endif
|
|
if (!load_init_expr(module, &p, p_end, &init_expr,
|
|
mem_offset_type, error_buf, error_buf_size))
|
|
return false;
|
|
|
|
read_leb_uint32(p, p_end, data_seg_len);
|
|
|
|
if (!(dataseg = module->data_segments[i] = loader_malloc(
|
|
sizeof(WASMDataSeg), error_buf, error_buf_size))) {
|
|
return false;
|
|
}
|
|
|
|
#if WASM_ENABLE_BULK_MEMORY != 0
|
|
dataseg->is_passive = is_passive;
|
|
if (!is_passive)
|
|
#endif
|
|
{
|
|
bh_memcpy_s(&dataseg->base_offset,
|
|
sizeof(InitializerExpression), &init_expr,
|
|
sizeof(InitializerExpression));
|
|
|
|
dataseg->memory_index = mem_index;
|
|
}
|
|
|
|
dataseg->data_length = data_seg_len;
|
|
CHECK_BUF(p, p_end, data_seg_len);
|
|
if (clone_data_seg) {
|
|
if (!(dataseg->data = loader_malloc(
|
|
dataseg->data_length, error_buf, error_buf_size))) {
|
|
return false;
|
|
}
|
|
|
|
bh_memcpy_s(dataseg->data, dataseg->data_length, p,
|
|
data_seg_len);
|
|
}
|
|
else {
|
|
dataseg->data = (uint8 *)p;
|
|
}
|
|
dataseg->is_data_cloned = clone_data_seg;
|
|
p += data_seg_len;
|
|
}
|
|
}
|
|
|
|
bh_assert(p == p_end);
|
|
LOG_VERBOSE("Load data segment section success.\n");
|
|
return true;
|
|
}
|
|
|
|
#if WASM_ENABLE_BULK_MEMORY != 0
|
|
static bool
|
|
load_datacount_section(const uint8 *buf, const uint8 *buf_end,
|
|
WASMModule *module, char *error_buf,
|
|
uint32 error_buf_size)
|
|
{
|
|
const uint8 *p = buf, *p_end = buf_end;
|
|
uint32 data_seg_count1 = 0;
|
|
|
|
read_leb_uint32(p, p_end, data_seg_count1);
|
|
module->data_seg_count1 = data_seg_count1;
|
|
|
|
bh_assert(p == p_end);
|
|
LOG_VERBOSE("Load datacount section success.\n");
|
|
return true;
|
|
}
|
|
#endif
|
|
|
|
static bool
|
|
load_code_section(const uint8 *buf, const uint8 *buf_end, const uint8 *buf_func,
|
|
const uint8 *buf_func_end, WASMModule *module,
|
|
char *error_buf, uint32 error_buf_size)
|
|
{
|
|
const uint8 *p = buf, *p_end = buf_end;
|
|
const uint8 *p_func = buf_func;
|
|
uint32 func_count = 0, code_count;
|
|
|
|
/* code has been loaded in function section, so pass it here, just check
|
|
* whether function and code section have inconsistent lengths */
|
|
read_leb_uint32(p, p_end, code_count);
|
|
|
|
if (buf_func)
|
|
read_leb_uint32(p_func, buf_func_end, func_count);
|
|
|
|
bh_assert(func_count == code_count);
|
|
LOG_VERBOSE("Load code segment section success.\n");
|
|
(void)code_count;
|
|
(void)func_count;
|
|
return true;
|
|
}
|
|
|
|
static bool
|
|
load_start_section(const uint8 *buf, const uint8 *buf_end, WASMModule *module,
|
|
char *error_buf, uint32 error_buf_size)
|
|
{
|
|
const uint8 *p = buf, *p_end = buf_end;
|
|
WASMFuncType *type;
|
|
uint32 start_function;
|
|
|
|
read_leb_uint32(p, p_end, start_function);
|
|
|
|
bh_assert(start_function
|
|
< module->function_count + module->import_function_count);
|
|
|
|
if (start_function < module->import_function_count)
|
|
type = module->import_functions[start_function].u.function.func_type;
|
|
else
|
|
type = module->functions[start_function - module->import_function_count]
|
|
->func_type;
|
|
|
|
bh_assert(type->param_count == 0 && type->result_count == 0);
|
|
|
|
module->start_function = start_function;
|
|
|
|
bh_assert(p == p_end);
|
|
LOG_VERBOSE("Load start section success.\n");
|
|
(void)type;
|
|
return true;
|
|
}
|
|
|
|
#if WASM_ENABLE_CUSTOM_NAME_SECTION != 0
|
|
static bool
|
|
handle_name_section(const uint8 *buf, const uint8 *buf_end, WASMModule *module,
|
|
bool is_load_from_file_buf, char *error_buf,
|
|
uint32 error_buf_size)
|
|
{
|
|
const uint8 *p = buf, *p_end = buf_end;
|
|
uint32 name_type, subsection_size;
|
|
uint32 previous_name_type = 0;
|
|
uint32 num_func_name;
|
|
uint32 func_index;
|
|
uint32 previous_func_index = ~0U;
|
|
uint32 func_name_len;
|
|
uint32 name_index;
|
|
int i = 0;
|
|
|
|
bh_assert(p < p_end);
|
|
|
|
while (p < p_end) {
|
|
read_leb_uint32(p, p_end, name_type);
|
|
if (i != 0) {
|
|
bh_assert(name_type > previous_name_type);
|
|
}
|
|
previous_name_type = name_type;
|
|
read_leb_uint32(p, p_end, subsection_size);
|
|
CHECK_BUF(p, p_end, subsection_size);
|
|
switch (name_type) {
|
|
case SUB_SECTION_TYPE_FUNC:
|
|
if (subsection_size) {
|
|
read_leb_uint32(p, p_end, num_func_name);
|
|
for (name_index = 0; name_index < num_func_name;
|
|
name_index++) {
|
|
read_leb_uint32(p, p_end, func_index);
|
|
bh_assert(func_index > previous_func_index);
|
|
previous_func_index = func_index;
|
|
read_leb_uint32(p, p_end, func_name_len);
|
|
CHECK_BUF(p, p_end, func_name_len);
|
|
/* Skip the import functions */
|
|
if (func_index >= module->import_function_count) {
|
|
func_index -= module->import_function_count;
|
|
bh_assert(func_index < module->function_count);
|
|
if (!(module->functions[func_index]->field_name =
|
|
wasm_const_str_list_insert(
|
|
p, func_name_len, module,
|
|
is_load_from_file_buf, error_buf,
|
|
error_buf_size))) {
|
|
return false;
|
|
}
|
|
}
|
|
p += func_name_len;
|
|
}
|
|
}
|
|
break;
|
|
case SUB_SECTION_TYPE_MODULE: /* TODO: Parse for module subsection
|
|
*/
|
|
case SUB_SECTION_TYPE_LOCAL: /* TODO: Parse for local subsection */
|
|
default:
|
|
p = p + subsection_size;
|
|
break;
|
|
}
|
|
i++;
|
|
}
|
|
|
|
(void)previous_name_type;
|
|
(void)previous_func_index;
|
|
return true;
|
|
}
|
|
#endif
|
|
|
|
static bool
|
|
load_user_section(const uint8 *buf, const uint8 *buf_end, WASMModule *module,
|
|
bool is_load_from_file_buf, char *error_buf,
|
|
uint32 error_buf_size)
|
|
{
|
|
const uint8 *p = buf, *p_end = buf_end;
|
|
uint32 name_len;
|
|
|
|
bh_assert(p < p_end);
|
|
|
|
read_leb_uint32(p, p_end, name_len);
|
|
|
|
bh_assert(name_len > 0 && p + name_len <= p_end);
|
|
|
|
#if WASM_ENABLE_CUSTOM_NAME_SECTION != 0
|
|
if (name_len == 4 && memcmp(p, "name", 4) == 0) {
|
|
p += name_len;
|
|
handle_name_section(p, p_end, module, is_load_from_file_buf, error_buf,
|
|
error_buf_size);
|
|
}
|
|
#endif
|
|
LOG_VERBOSE("Load custom section success.\n");
|
|
(void)name_len;
|
|
return true;
|
|
}
|
|
|
|
static void
|
|
calculate_global_data_offset(WASMModule *module)
|
|
{
|
|
uint32 i, data_offset;
|
|
|
|
data_offset = 0;
|
|
for (i = 0; i < module->import_global_count; i++) {
|
|
WASMGlobalImport *import_global =
|
|
&((module->import_globals + i)->u.global);
|
|
#if WASM_ENABLE_FAST_JIT != 0
|
|
import_global->data_offset = data_offset;
|
|
#endif
|
|
data_offset += wasm_value_type_size(import_global->type.val_type);
|
|
}
|
|
|
|
for (i = 0; i < module->global_count; i++) {
|
|
WASMGlobal *global = module->globals + i;
|
|
#if WASM_ENABLE_FAST_JIT != 0
|
|
global->data_offset = data_offset;
|
|
#endif
|
|
data_offset += wasm_value_type_size(global->type.val_type);
|
|
}
|
|
|
|
module->global_data_size = data_offset;
|
|
}
|
|
|
|
#if WASM_ENABLE_FAST_JIT != 0
|
|
static bool
|
|
init_fast_jit_functions(WASMModule *module, char *error_buf,
|
|
uint32 error_buf_size)
|
|
{
|
|
#if WASM_ENABLE_LAZY_JIT != 0
|
|
JitGlobals *jit_globals = jit_compiler_get_jit_globals();
|
|
#endif
|
|
uint32 i;
|
|
|
|
if (!module->function_count)
|
|
return true;
|
|
|
|
if (!(module->fast_jit_func_ptrs =
|
|
loader_malloc(sizeof(void *) * module->function_count, error_buf,
|
|
error_buf_size))) {
|
|
return false;
|
|
}
|
|
|
|
#if WASM_ENABLE_LAZY_JIT != 0
|
|
for (i = 0; i < module->function_count; i++) {
|
|
module->fast_jit_func_ptrs[i] =
|
|
jit_globals->compile_fast_jit_and_then_call;
|
|
}
|
|
#endif
|
|
|
|
for (i = 0; i < WASM_ORC_JIT_BACKEND_THREAD_NUM; i++) {
|
|
if (os_mutex_init(&module->fast_jit_thread_locks[i]) != 0) {
|
|
set_error_buf(error_buf, error_buf_size,
|
|
"init fast jit thread lock failed");
|
|
return false;
|
|
}
|
|
module->fast_jit_thread_locks_inited[i] = true;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
#endif /* end of WASM_ENABLE_FAST_JIT != 0 */
|
|
|
|
#if WASM_ENABLE_JIT != 0
|
|
static bool
|
|
init_llvm_jit_functions_stage1(WASMModule *module, char *error_buf,
|
|
uint32 error_buf_size)
|
|
{
|
|
LLVMJITOptions llvm_jit_options = wasm_runtime_get_llvm_jit_options();
|
|
AOTCompOption option = { 0 };
|
|
char *aot_last_error;
|
|
uint64 size;
|
|
bool gc_enabled = false; /* GC hasn't been enabled in mini loader */
|
|
|
|
if (module->function_count == 0)
|
|
return true;
|
|
|
|
#if WASM_ENABLE_FAST_JIT != 0 && WASM_ENABLE_LAZY_JIT != 0
|
|
if (os_mutex_init(&module->tierup_wait_lock) != 0) {
|
|
set_error_buf(error_buf, error_buf_size, "init jit tierup lock failed");
|
|
return false;
|
|
}
|
|
if (os_cond_init(&module->tierup_wait_cond) != 0) {
|
|
set_error_buf(error_buf, error_buf_size, "init jit tierup cond failed");
|
|
os_mutex_destroy(&module->tierup_wait_lock);
|
|
return false;
|
|
}
|
|
module->tierup_wait_lock_inited = true;
|
|
#endif
|
|
|
|
size = sizeof(void *) * (uint64)module->function_count
|
|
+ sizeof(bool) * (uint64)module->function_count;
|
|
if (!(module->func_ptrs = loader_malloc(size, error_buf, error_buf_size))) {
|
|
return false;
|
|
}
|
|
module->func_ptrs_compiled =
|
|
(bool *)((uint8 *)module->func_ptrs
|
|
+ sizeof(void *) * module->function_count);
|
|
|
|
module->comp_data = aot_create_comp_data(module, NULL, gc_enabled);
|
|
if (!module->comp_data) {
|
|
aot_last_error = aot_get_last_error();
|
|
bh_assert(aot_last_error != NULL);
|
|
set_error_buf(error_buf, error_buf_size, aot_last_error);
|
|
return false;
|
|
}
|
|
|
|
option.is_jit_mode = true;
|
|
option.opt_level = llvm_jit_options->opt_level;
|
|
option.size_level = llvm_jit_options->size_level;
|
|
option.segue_flags = llvm_jit_options->segue_flags;
|
|
option.quick_invoke_c_api_import =
|
|
llvm_jit_options->quick_invoke_c_api_import;
|
|
|
|
#if WASM_ENABLE_BULK_MEMORY != 0
|
|
option.enable_bulk_memory = true;
|
|
#endif
|
|
#if WASM_ENABLE_THREAD_MGR != 0
|
|
option.enable_thread_mgr = true;
|
|
#endif
|
|
#if WASM_ENABLE_TAIL_CALL != 0
|
|
option.enable_tail_call = true;
|
|
#endif
|
|
#if WASM_ENABLE_SIMD != 0
|
|
option.enable_simd = true;
|
|
#endif
|
|
#if WASM_ENABLE_REF_TYPES != 0
|
|
option.enable_ref_types = true;
|
|
#endif
|
|
option.enable_aux_stack_check = true;
|
|
#if WASM_ENABLE_PERF_PROFILING != 0 || WASM_ENABLE_DUMP_CALL_STACK != 0 \
|
|
|| WASM_ENABLE_AOT_STACK_FRAME != 0
|
|
option.enable_aux_stack_frame = true;
|
|
#endif
|
|
#if WASM_ENABLE_PERF_PROFILING != 0
|
|
option.enable_perf_profiling = true;
|
|
#endif
|
|
#if WASM_ENABLE_MEMORY_PROFILING != 0
|
|
option.enable_memory_profiling = true;
|
|
option.enable_stack_estimation = true;
|
|
#endif
|
|
|
|
module->comp_ctx = aot_create_comp_context(module->comp_data, &option);
|
|
if (!module->comp_ctx) {
|
|
aot_last_error = aot_get_last_error();
|
|
bh_assert(aot_last_error != NULL);
|
|
set_error_buf(error_buf, error_buf_size, aot_last_error);
|
|
return false;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
static bool
|
|
init_llvm_jit_functions_stage2(WASMModule *module, char *error_buf,
|
|
uint32 error_buf_size)
|
|
{
|
|
char *aot_last_error;
|
|
uint32 i;
|
|
|
|
if (module->function_count == 0)
|
|
return true;
|
|
|
|
if (!aot_compile_wasm(module->comp_ctx)) {
|
|
aot_last_error = aot_get_last_error();
|
|
bh_assert(aot_last_error != NULL);
|
|
set_error_buf(error_buf, error_buf_size, aot_last_error);
|
|
return false;
|
|
}
|
|
|
|
#if WASM_ENABLE_FAST_JIT != 0 && WASM_ENABLE_LAZY_JIT != 0
|
|
if (module->orcjit_stop_compiling)
|
|
return false;
|
|
#endif
|
|
|
|
bh_print_time("Begin to lookup llvm jit functions");
|
|
|
|
for (i = 0; i < module->function_count; i++) {
|
|
LLVMOrcJITTargetAddress func_addr = 0;
|
|
LLVMErrorRef error;
|
|
char func_name[48];
|
|
|
|
snprintf(func_name, sizeof(func_name), "%s%d", AOT_FUNC_PREFIX, i);
|
|
error = LLVMOrcLLLazyJITLookup(module->comp_ctx->orc_jit, &func_addr,
|
|
func_name);
|
|
if (error != LLVMErrorSuccess) {
|
|
char *err_msg = LLVMGetErrorMessage(error);
|
|
char buf[96];
|
|
snprintf(buf, sizeof(buf),
|
|
"failed to compile llvm jit function: %s", err_msg);
|
|
set_error_buf(error_buf, error_buf_size, buf);
|
|
LLVMDisposeErrorMessage(err_msg);
|
|
return false;
|
|
}
|
|
|
|
/**
|
|
* No need to lock the func_ptr[func_idx] here as it is basic
|
|
* data type, the load/store for it can be finished by one cpu
|
|
* instruction, and there can be only one cpu instruction
|
|
* loading/storing at the same time.
|
|
*/
|
|
module->func_ptrs[i] = (void *)func_addr;
|
|
|
|
#if WASM_ENABLE_FAST_JIT != 0 && WASM_ENABLE_LAZY_JIT != 0
|
|
module->functions[i]->llvm_jit_func_ptr = (void *)func_addr;
|
|
|
|
if (module->orcjit_stop_compiling)
|
|
return false;
|
|
#endif
|
|
}
|
|
|
|
bh_print_time("End lookup llvm jit functions");
|
|
|
|
return true;
|
|
}
|
|
#endif /* end of WASM_ENABLE_JIT != 0 */
|
|
|
|
#if WASM_ENABLE_FAST_JIT != 0 && WASM_ENABLE_JIT != 0 \
|
|
&& WASM_ENABLE_LAZY_JIT != 0
|
|
static void *
|
|
init_llvm_jit_functions_stage2_callback(void *arg)
|
|
{
|
|
WASMModule *module = (WASMModule *)arg;
|
|
char error_buf[128];
|
|
uint32 error_buf_size = (uint32)sizeof(error_buf);
|
|
|
|
if (!init_llvm_jit_functions_stage2(module, error_buf, error_buf_size)) {
|
|
module->orcjit_stop_compiling = true;
|
|
return NULL;
|
|
}
|
|
|
|
os_mutex_lock(&module->tierup_wait_lock);
|
|
module->llvm_jit_inited = true;
|
|
os_cond_broadcast(&module->tierup_wait_cond);
|
|
os_mutex_unlock(&module->tierup_wait_lock);
|
|
|
|
return NULL;
|
|
}
|
|
#endif
|
|
|
|
#if WASM_ENABLE_FAST_JIT != 0 || WASM_ENABLE_JIT != 0
|
|
/* The callback function to compile jit functions */
|
|
static void *
|
|
orcjit_thread_callback(void *arg)
|
|
{
|
|
OrcJitThreadArg *thread_arg = (OrcJitThreadArg *)arg;
|
|
#if WASM_ENABLE_JIT != 0
|
|
AOTCompContext *comp_ctx = thread_arg->comp_ctx;
|
|
#endif
|
|
WASMModule *module = thread_arg->module;
|
|
uint32 group_idx = thread_arg->group_idx;
|
|
uint32 group_stride = WASM_ORC_JIT_BACKEND_THREAD_NUM;
|
|
uint32 func_count = module->function_count;
|
|
uint32 i;
|
|
|
|
#if WASM_ENABLE_FAST_JIT != 0
|
|
/* Compile fast jit functions of this group */
|
|
for (i = group_idx; i < func_count; i += group_stride) {
|
|
if (!jit_compiler_compile(module, i + module->import_function_count)) {
|
|
LOG_ERROR("failed to compile fast jit function %u\n", i);
|
|
break;
|
|
}
|
|
|
|
if (module->orcjit_stop_compiling) {
|
|
return NULL;
|
|
}
|
|
}
|
|
#if WASM_ENABLE_JIT != 0 && WASM_ENABLE_LAZY_JIT != 0
|
|
os_mutex_lock(&module->tierup_wait_lock);
|
|
module->fast_jit_ready_groups++;
|
|
os_mutex_unlock(&module->tierup_wait_lock);
|
|
#endif
|
|
#endif
|
|
|
|
#if WASM_ENABLE_FAST_JIT != 0 && WASM_ENABLE_JIT != 0 \
|
|
&& WASM_ENABLE_LAZY_JIT != 0
|
|
/* For JIT tier-up, set each llvm jit func to call_to_fast_jit */
|
|
for (i = group_idx; i < func_count;
|
|
i += group_stride * WASM_ORC_JIT_COMPILE_THREAD_NUM) {
|
|
uint32 j;
|
|
|
|
for (j = 0; j < WASM_ORC_JIT_COMPILE_THREAD_NUM; j++) {
|
|
if (i + j * group_stride < func_count) {
|
|
if (!jit_compiler_set_call_to_fast_jit(
|
|
module,
|
|
i + j * group_stride + module->import_function_count)) {
|
|
LOG_ERROR(
|
|
"failed to compile call_to_fast_jit for func %u\n",
|
|
i + j * group_stride + module->import_function_count);
|
|
module->orcjit_stop_compiling = true;
|
|
return NULL;
|
|
}
|
|
}
|
|
if (module->orcjit_stop_compiling) {
|
|
return NULL;
|
|
}
|
|
}
|
|
}
|
|
|
|
/* Wait until init_llvm_jit_functions_stage2 finishes and all
|
|
fast jit functions are compiled */
|
|
os_mutex_lock(&module->tierup_wait_lock);
|
|
while (!(module->llvm_jit_inited && module->enable_llvm_jit_compilation
|
|
&& module->fast_jit_ready_groups >= group_stride)) {
|
|
os_cond_reltimedwait(&module->tierup_wait_cond,
|
|
&module->tierup_wait_lock, 10000);
|
|
if (module->orcjit_stop_compiling) {
|
|
/* init_llvm_jit_functions_stage2 failed */
|
|
os_mutex_unlock(&module->tierup_wait_lock);
|
|
return NULL;
|
|
}
|
|
}
|
|
os_mutex_unlock(&module->tierup_wait_lock);
|
|
#endif
|
|
|
|
#if WASM_ENABLE_JIT != 0
|
|
/* Compile llvm jit functions of this group */
|
|
for (i = group_idx; i < func_count;
|
|
i += group_stride * WASM_ORC_JIT_COMPILE_THREAD_NUM) {
|
|
LLVMOrcJITTargetAddress func_addr = 0;
|
|
LLVMErrorRef error;
|
|
char func_name[48];
|
|
typedef void (*F)(void);
|
|
union {
|
|
F f;
|
|
void *v;
|
|
} u;
|
|
uint32 j;
|
|
|
|
snprintf(func_name, sizeof(func_name), "%s%d%s", AOT_FUNC_PREFIX, i,
|
|
"_wrapper");
|
|
LOG_DEBUG("compile llvm jit func %s", func_name);
|
|
error =
|
|
LLVMOrcLLLazyJITLookup(comp_ctx->orc_jit, &func_addr, func_name);
|
|
if (error != LLVMErrorSuccess) {
|
|
char *err_msg = LLVMGetErrorMessage(error);
|
|
LOG_ERROR("failed to compile llvm jit function %u: %s", i, err_msg);
|
|
LLVMDisposeErrorMessage(err_msg);
|
|
break;
|
|
}
|
|
|
|
/* Call the jit wrapper function to trigger its compilation, so as
|
|
to compile the actual jit functions, since we add the latter to
|
|
function list in the PartitionFunction callback */
|
|
u.v = (void *)func_addr;
|
|
u.f();
|
|
|
|
for (j = 0; j < WASM_ORC_JIT_COMPILE_THREAD_NUM; j++) {
|
|
if (i + j * group_stride < func_count) {
|
|
module->func_ptrs_compiled[i + j * group_stride] = true;
|
|
#if WASM_ENABLE_FAST_JIT != 0 && WASM_ENABLE_LAZY_JIT != 0
|
|
snprintf(func_name, sizeof(func_name), "%s%d", AOT_FUNC_PREFIX,
|
|
i + j * group_stride);
|
|
error = LLVMOrcLLLazyJITLookup(comp_ctx->orc_jit, &func_addr,
|
|
func_name);
|
|
if (error != LLVMErrorSuccess) {
|
|
char *err_msg = LLVMGetErrorMessage(error);
|
|
LOG_ERROR("failed to compile llvm jit function %u: %s", i,
|
|
err_msg);
|
|
LLVMDisposeErrorMessage(err_msg);
|
|
/* Ignore current llvm jit func, as its func ptr is
|
|
previous set to call_to_fast_jit, which also works */
|
|
continue;
|
|
}
|
|
|
|
jit_compiler_set_llvm_jit_func_ptr(
|
|
module,
|
|
i + j * group_stride + module->import_function_count,
|
|
(void *)func_addr);
|
|
|
|
/* Try to switch to call this llvm jit funtion instead of
|
|
fast jit function from fast jit jitted code */
|
|
jit_compiler_set_call_to_llvm_jit(
|
|
module,
|
|
i + j * group_stride + module->import_function_count);
|
|
#endif
|
|
}
|
|
}
|
|
|
|
if (module->orcjit_stop_compiling) {
|
|
break;
|
|
}
|
|
}
|
|
#endif
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static void
|
|
orcjit_stop_compile_threads(WASMModule *module)
|
|
{
|
|
uint32 i, thread_num = (uint32)(sizeof(module->orcjit_thread_args)
|
|
/ sizeof(OrcJitThreadArg));
|
|
|
|
module->orcjit_stop_compiling = true;
|
|
for (i = 0; i < thread_num; i++) {
|
|
if (module->orcjit_threads[i])
|
|
os_thread_join(module->orcjit_threads[i], NULL);
|
|
}
|
|
}
|
|
|
|
static bool
|
|
compile_jit_functions(WASMModule *module, char *error_buf,
|
|
uint32 error_buf_size)
|
|
{
|
|
uint32 thread_num =
|
|
(uint32)(sizeof(module->orcjit_thread_args) / sizeof(OrcJitThreadArg));
|
|
uint32 i, j;
|
|
|
|
bh_print_time("Begin to compile jit functions");
|
|
|
|
/* Create threads to compile the jit functions */
|
|
for (i = 0; i < thread_num && i < module->function_count; i++) {
|
|
#if WASM_ENABLE_JIT != 0
|
|
module->orcjit_thread_args[i].comp_ctx = module->comp_ctx;
|
|
#endif
|
|
module->orcjit_thread_args[i].module = module;
|
|
module->orcjit_thread_args[i].group_idx = i;
|
|
|
|
if (os_thread_create(&module->orcjit_threads[i], orcjit_thread_callback,
|
|
(void *)&module->orcjit_thread_args[i],
|
|
APP_THREAD_STACK_SIZE_DEFAULT)
|
|
!= 0) {
|
|
set_error_buf(error_buf, error_buf_size,
|
|
"create orcjit compile thread failed");
|
|
/* Terminate the threads created */
|
|
module->orcjit_stop_compiling = true;
|
|
for (j = 0; j < i; j++) {
|
|
os_thread_join(module->orcjit_threads[j], NULL);
|
|
}
|
|
return false;
|
|
}
|
|
}
|
|
|
|
#if WASM_ENABLE_LAZY_JIT == 0
|
|
/* Wait until all jit functions are compiled for eager mode */
|
|
for (i = 0; i < thread_num; i++) {
|
|
if (module->orcjit_threads[i])
|
|
os_thread_join(module->orcjit_threads[i], NULL);
|
|
}
|
|
|
|
#if WASM_ENABLE_FAST_JIT != 0
|
|
/* Ensure all the fast-jit functions are compiled */
|
|
for (i = 0; i < module->function_count; i++) {
|
|
if (!jit_compiler_is_compiled(module,
|
|
i + module->import_function_count)) {
|
|
set_error_buf(error_buf, error_buf_size,
|
|
"failed to compile fast jit function");
|
|
return false;
|
|
}
|
|
}
|
|
#endif
|
|
|
|
#if WASM_ENABLE_JIT != 0
|
|
/* Ensure all the llvm-jit functions are compiled */
|
|
for (i = 0; i < module->function_count; i++) {
|
|
if (!module->func_ptrs_compiled[i]) {
|
|
set_error_buf(error_buf, error_buf_size,
|
|
"failed to compile llvm jit function");
|
|
return false;
|
|
}
|
|
}
|
|
#endif
|
|
#endif /* end of WASM_ENABLE_LAZY_JIT == 0 */
|
|
|
|
bh_print_time("End compile jit functions");
|
|
|
|
return true;
|
|
}
|
|
#endif /* end of WASM_ENABLE_FAST_JIT != 0 || WASM_ENABLE_JIT != 0 */
|
|
|
|
#if WASM_ENABLE_REF_TYPES != 0
|
|
static bool
|
|
get_table_elem_type(const WASMModule *module, uint32 table_idx,
|
|
uint8 *p_elem_type, char *error_buf, uint32 error_buf_size)
|
|
{
|
|
if (!check_table_index(module, table_idx, error_buf, error_buf_size)) {
|
|
return false;
|
|
}
|
|
|
|
if (p_elem_type) {
|
|
if (table_idx < module->import_table_count)
|
|
*p_elem_type = module->import_tables[table_idx].u.table.elem_type;
|
|
else
|
|
*p_elem_type =
|
|
module->tables[module->import_table_count + table_idx]
|
|
.elem_type;
|
|
}
|
|
return true;
|
|
}
|
|
|
|
static bool
|
|
get_table_seg_elem_type(const WASMModule *module, uint32 table_seg_idx,
|
|
uint8 *p_elem_type, char *error_buf,
|
|
uint32 error_buf_size)
|
|
{
|
|
if (table_seg_idx >= module->table_seg_count) {
|
|
return false;
|
|
}
|
|
|
|
if (p_elem_type) {
|
|
*p_elem_type = module->table_segments[table_seg_idx].elem_type;
|
|
}
|
|
return true;
|
|
}
|
|
#endif
|
|
|
|
static bool
|
|
wasm_loader_prepare_bytecode(WASMModule *module, WASMFunction *func,
|
|
uint32 cur_func_idx, char *error_buf,
|
|
uint32 error_buf_size);
|
|
|
|
#if WASM_ENABLE_FAST_INTERP != 0 && WASM_ENABLE_LABELS_AS_VALUES != 0
|
|
void **
|
|
wasm_interp_get_handle_table();
|
|
|
|
static void **handle_table;
|
|
#endif
|
|
|
|
static bool
|
|
load_from_sections(WASMModule *module, WASMSection *sections,
|
|
bool is_load_from_file_buf, bool wasm_binary_freeable,
|
|
char *error_buf, uint32 error_buf_size)
|
|
{
|
|
WASMExport *export;
|
|
WASMSection *section = sections;
|
|
const uint8 *buf, *buf_end, *buf_code = NULL, *buf_code_end = NULL,
|
|
*buf_func = NULL, *buf_func_end = NULL;
|
|
WASMGlobal *aux_data_end_global = NULL, *aux_heap_base_global = NULL;
|
|
WASMGlobal *aux_stack_top_global = NULL, *global;
|
|
uint64 aux_data_end = (uint64)-1LL, aux_heap_base = (uint64)-1LL,
|
|
aux_stack_top = (uint64)-1LL;
|
|
uint32 global_index, func_index, i;
|
|
uint32 aux_data_end_global_index = (uint32)-1;
|
|
uint32 aux_heap_base_global_index = (uint32)-1;
|
|
WASMFuncType *func_type;
|
|
uint8 malloc_free_io_type = VALUE_TYPE_I32;
|
|
bool reuse_const_strings = is_load_from_file_buf && !wasm_binary_freeable;
|
|
bool clone_data_seg = is_load_from_file_buf && wasm_binary_freeable;
|
|
|
|
/* Find code and function sections if have */
|
|
while (section) {
|
|
if (section->section_type == SECTION_TYPE_CODE) {
|
|
buf_code = section->section_body;
|
|
buf_code_end = buf_code + section->section_body_size;
|
|
}
|
|
else if (section->section_type == SECTION_TYPE_FUNC) {
|
|
buf_func = section->section_body;
|
|
buf_func_end = buf_func + section->section_body_size;
|
|
}
|
|
section = section->next;
|
|
}
|
|
|
|
section = sections;
|
|
while (section) {
|
|
buf = section->section_body;
|
|
buf_end = buf + section->section_body_size;
|
|
LOG_DEBUG("load section, type: %d", section->section_type);
|
|
switch (section->section_type) {
|
|
case SECTION_TYPE_USER:
|
|
/* unsupported user section, ignore it. */
|
|
if (!load_user_section(buf, buf_end, module,
|
|
reuse_const_strings, error_buf,
|
|
error_buf_size))
|
|
return false;
|
|
break;
|
|
case SECTION_TYPE_TYPE:
|
|
if (!load_type_section(buf, buf_end, module, error_buf,
|
|
error_buf_size))
|
|
return false;
|
|
break;
|
|
case SECTION_TYPE_IMPORT:
|
|
if (!load_import_section(buf, buf_end, module,
|
|
reuse_const_strings, error_buf,
|
|
error_buf_size))
|
|
return false;
|
|
break;
|
|
case SECTION_TYPE_FUNC:
|
|
if (!load_function_section(buf, buf_end, buf_code, buf_code_end,
|
|
module, error_buf, error_buf_size))
|
|
return false;
|
|
break;
|
|
case SECTION_TYPE_TABLE:
|
|
if (!load_table_section(buf, buf_end, module, error_buf,
|
|
error_buf_size))
|
|
return false;
|
|
break;
|
|
case SECTION_TYPE_MEMORY:
|
|
if (!load_memory_section(buf, buf_end, module, error_buf,
|
|
error_buf_size))
|
|
return false;
|
|
break;
|
|
case SECTION_TYPE_GLOBAL:
|
|
if (!load_global_section(buf, buf_end, module, error_buf,
|
|
error_buf_size))
|
|
return false;
|
|
break;
|
|
case SECTION_TYPE_EXPORT:
|
|
if (!load_export_section(buf, buf_end, module,
|
|
reuse_const_strings, error_buf,
|
|
error_buf_size))
|
|
return false;
|
|
break;
|
|
case SECTION_TYPE_START:
|
|
if (!load_start_section(buf, buf_end, module, error_buf,
|
|
error_buf_size))
|
|
return false;
|
|
break;
|
|
case SECTION_TYPE_ELEM:
|
|
if (!load_table_segment_section(buf, buf_end, module, error_buf,
|
|
error_buf_size))
|
|
return false;
|
|
break;
|
|
case SECTION_TYPE_CODE:
|
|
if (!load_code_section(buf, buf_end, buf_func, buf_func_end,
|
|
module, error_buf, error_buf_size))
|
|
return false;
|
|
break;
|
|
case SECTION_TYPE_DATA:
|
|
if (!load_data_segment_section(buf, buf_end, module,
|
|
clone_data_seg, error_buf,
|
|
error_buf_size))
|
|
return false;
|
|
break;
|
|
#if WASM_ENABLE_BULK_MEMORY != 0
|
|
case SECTION_TYPE_DATACOUNT:
|
|
if (!load_datacount_section(buf, buf_end, module, error_buf,
|
|
error_buf_size))
|
|
return false;
|
|
break;
|
|
#endif
|
|
default:
|
|
set_error_buf(error_buf, error_buf_size, "invalid section id");
|
|
return false;
|
|
}
|
|
|
|
section = section->next;
|
|
}
|
|
|
|
module->aux_data_end_global_index = (uint32)-1;
|
|
module->aux_heap_base_global_index = (uint32)-1;
|
|
module->aux_stack_top_global_index = (uint32)-1;
|
|
|
|
/* Resolve auxiliary data/stack/heap info and reset memory info */
|
|
export = module->exports;
|
|
for (i = 0; i < module->export_count; i++, export ++) {
|
|
if (export->kind == EXPORT_KIND_GLOBAL) {
|
|
if (!strcmp(export->name, "__heap_base")) {
|
|
global_index = export->index - module->import_global_count;
|
|
global = module->globals + global_index;
|
|
if (global->type.val_type == VALUE_TYPE_I32
|
|
&& !global->type.is_mutable
|
|
&& global->init_expr.init_expr_type
|
|
== INIT_EXPR_TYPE_I32_CONST) {
|
|
aux_heap_base_global = global;
|
|
aux_heap_base = (uint64)(uint32)global->init_expr.u.i32;
|
|
aux_heap_base_global_index = export->index;
|
|
LOG_VERBOSE("Found aux __heap_base global, value: %" PRIu64,
|
|
aux_heap_base);
|
|
}
|
|
}
|
|
else if (!strcmp(export->name, "__data_end")) {
|
|
global_index = export->index - module->import_global_count;
|
|
global = module->globals + global_index;
|
|
if (global->type.val_type == VALUE_TYPE_I32
|
|
&& !global->type.is_mutable
|
|
&& global->init_expr.init_expr_type
|
|
== INIT_EXPR_TYPE_I32_CONST) {
|
|
aux_data_end_global = global;
|
|
aux_data_end = (uint64)(uint32)global->init_expr.u.i32;
|
|
aux_data_end_global_index = export->index;
|
|
LOG_VERBOSE("Found aux __data_end global, value: %" PRIu64,
|
|
aux_data_end);
|
|
aux_data_end = align_uint64(aux_data_end, 16);
|
|
}
|
|
}
|
|
|
|
/* For module compiled with -pthread option, the global is:
|
|
[0] stack_top <-- 0
|
|
[1] tls_pointer
|
|
[2] tls_size
|
|
[3] data_end <-- 3
|
|
[4] global_base
|
|
[5] heap_base <-- 5
|
|
[6] dso_handle
|
|
|
|
For module compiled without -pthread option:
|
|
[0] stack_top <-- 0
|
|
[1] data_end <-- 1
|
|
[2] global_base
|
|
[3] heap_base <-- 3
|
|
[4] dso_handle
|
|
*/
|
|
if (aux_data_end_global && aux_heap_base_global
|
|
&& aux_data_end <= aux_heap_base) {
|
|
module->aux_data_end_global_index = aux_data_end_global_index;
|
|
module->aux_data_end = aux_data_end;
|
|
module->aux_heap_base_global_index = aux_heap_base_global_index;
|
|
module->aux_heap_base = aux_heap_base;
|
|
|
|
/* Resolve aux stack top global */
|
|
for (global_index = 0; global_index < module->global_count;
|
|
global_index++) {
|
|
global = module->globals + global_index;
|
|
if (global->type.is_mutable /* heap_base and data_end is
|
|
not mutable */
|
|
&& global->type.val_type == VALUE_TYPE_I32
|
|
&& global->init_expr.init_expr_type
|
|
== INIT_EXPR_TYPE_I32_CONST
|
|
&& (uint64)(uint32)global->init_expr.u.i32
|
|
<= aux_heap_base) {
|
|
aux_stack_top_global = global;
|
|
aux_stack_top = (uint64)(uint32)global->init_expr.u.i32;
|
|
module->aux_stack_top_global_index =
|
|
module->import_global_count + global_index;
|
|
module->aux_stack_bottom = aux_stack_top;
|
|
module->aux_stack_size =
|
|
aux_stack_top > aux_data_end
|
|
? (uint32)(aux_stack_top - aux_data_end)
|
|
: (uint32)aux_stack_top;
|
|
LOG_VERBOSE(
|
|
"Found aux stack top global, value: %" PRIu64 ", "
|
|
"global index: %d, stack size: %d",
|
|
aux_stack_top, global_index,
|
|
module->aux_stack_size);
|
|
break;
|
|
}
|
|
}
|
|
if (!aux_stack_top_global) {
|
|
/* Auxiliary stack global isn't found, it must be unused
|
|
in the wasm app, as if it is used, the global must be
|
|
defined. Here we set it to __heap_base global and set
|
|
its size to 0. */
|
|
aux_stack_top_global = aux_heap_base_global;
|
|
aux_stack_top = aux_heap_base;
|
|
module->aux_stack_top_global_index =
|
|
module->aux_heap_base_global_index;
|
|
module->aux_stack_bottom = aux_stack_top;
|
|
module->aux_stack_size = 0;
|
|
}
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
module->malloc_function = (uint32)-1;
|
|
module->free_function = (uint32)-1;
|
|
module->retain_function = (uint32)-1;
|
|
|
|
/* Resolve malloc/free function exported by wasm module */
|
|
#if WASM_ENABLE_MEMORY64 != 0
|
|
if (has_module_memory64(module))
|
|
malloc_free_io_type = VALUE_TYPE_I64;
|
|
#endif
|
|
export = module->exports;
|
|
for (i = 0; i < module->export_count; i++, export ++) {
|
|
if (export->kind == EXPORT_KIND_FUNC) {
|
|
if (!strcmp(export->name, "malloc")
|
|
&& export->index >= module->import_function_count) {
|
|
func_index = export->index - module->import_function_count;
|
|
func_type = module->functions[func_index]->func_type;
|
|
if (func_type->param_count == 1 && func_type->result_count == 1
|
|
&& func_type->types[0] == malloc_free_io_type
|
|
&& func_type->types[1] == malloc_free_io_type) {
|
|
bh_assert(module->malloc_function == (uint32)-1);
|
|
module->malloc_function = export->index;
|
|
LOG_VERBOSE("Found malloc function, name: %s, index: %u",
|
|
export->name, export->index);
|
|
}
|
|
}
|
|
else if (!strcmp(export->name, "__new")
|
|
&& export->index >= module->import_function_count) {
|
|
/* __new && __pin for AssemblyScript */
|
|
func_index = export->index - module->import_function_count;
|
|
func_type = module->functions[func_index]->func_type;
|
|
if (func_type->param_count == 2 && func_type->result_count == 1
|
|
&& func_type->types[0] == malloc_free_io_type
|
|
&& func_type->types[1] == VALUE_TYPE_I32
|
|
&& func_type->types[2] == malloc_free_io_type) {
|
|
uint32 j;
|
|
WASMExport *export_tmp;
|
|
|
|
bh_assert(module->malloc_function == (uint32)-1);
|
|
module->malloc_function = export->index;
|
|
LOG_VERBOSE("Found malloc function, name: %s, index: %u",
|
|
export->name, export->index);
|
|
|
|
/* resolve retain function.
|
|
If not found, reset malloc function index */
|
|
export_tmp = module->exports;
|
|
for (j = 0; j < module->export_count; j++, export_tmp++) {
|
|
if ((export_tmp->kind == EXPORT_KIND_FUNC)
|
|
&& (!strcmp(export_tmp->name, "__retain")
|
|
|| !strcmp(export_tmp->name, "__pin"))
|
|
&& (export_tmp->index
|
|
>= module->import_function_count)) {
|
|
func_index = export_tmp->index
|
|
- module->import_function_count;
|
|
func_type =
|
|
module->functions[func_index]->func_type;
|
|
if (func_type->param_count == 1
|
|
&& func_type->result_count == 1
|
|
&& func_type->types[0] == malloc_free_io_type
|
|
&& func_type->types[1] == malloc_free_io_type) {
|
|
bh_assert(module->retain_function
|
|
== (uint32)-1);
|
|
module->retain_function = export_tmp->index;
|
|
LOG_VERBOSE("Found retain function, name: %s, "
|
|
"index: %u",
|
|
export_tmp->name,
|
|
export_tmp->index);
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
if (j == module->export_count) {
|
|
module->malloc_function = (uint32)-1;
|
|
LOG_VERBOSE("Can't find retain function,"
|
|
"reset malloc function index to -1");
|
|
}
|
|
}
|
|
}
|
|
else if (((!strcmp(export->name, "free"))
|
|
|| (!strcmp(export->name, "__release"))
|
|
|| (!strcmp(export->name, "__unpin")))
|
|
&& export->index >= module->import_function_count) {
|
|
func_index = export->index - module->import_function_count;
|
|
func_type = module->functions[func_index]->func_type;
|
|
if (func_type->param_count == 1 && func_type->result_count == 0
|
|
&& func_type->types[0] == malloc_free_io_type) {
|
|
bh_assert(module->free_function == (uint32)-1);
|
|
module->free_function = export->index;
|
|
LOG_VERBOSE("Found free function, name: %s, index: %u",
|
|
export->name, export->index);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
#if WASM_ENABLE_FAST_INTERP != 0 && WASM_ENABLE_LABELS_AS_VALUES != 0
|
|
handle_table = wasm_interp_get_handle_table();
|
|
#endif
|
|
|
|
for (i = 0; i < module->function_count; i++) {
|
|
WASMFunction *func = module->functions[i];
|
|
if (!wasm_loader_prepare_bytecode(module, func, i, error_buf,
|
|
error_buf_size)) {
|
|
return false;
|
|
}
|
|
|
|
if (i == module->function_count - 1) {
|
|
bh_assert(func->code + func->code_size == buf_code_end);
|
|
}
|
|
}
|
|
|
|
if (!module->possible_memory_grow) {
|
|
WASMMemoryImport *memory_import;
|
|
WASMMemory *memory;
|
|
|
|
if (aux_data_end_global && aux_heap_base_global
|
|
&& aux_stack_top_global) {
|
|
uint64 init_memory_size;
|
|
uint64 shrunk_memory_size = align_uint64(aux_heap_base, 8);
|
|
|
|
/* Only resize(shrunk) the memory size if num_bytes_per_page is in
|
|
* valid range of uint32 */
|
|
if (shrunk_memory_size <= UINT32_MAX) {
|
|
if (module->import_memory_count) {
|
|
memory_import = &module->import_memories[0].u.memory;
|
|
init_memory_size = (uint64)memory_import->num_bytes_per_page
|
|
* memory_import->init_page_count;
|
|
if (shrunk_memory_size <= init_memory_size) {
|
|
/* Reset memory info to decrease memory usage */
|
|
memory_import->num_bytes_per_page = shrunk_memory_size;
|
|
memory_import->init_page_count = 1;
|
|
LOG_VERBOSE("Shrink import memory size to %" PRIu64,
|
|
shrunk_memory_size);
|
|
}
|
|
}
|
|
|
|
if (module->memory_count) {
|
|
memory = &module->memories[0];
|
|
init_memory_size = (uint64)memory->num_bytes_per_page
|
|
* memory->init_page_count;
|
|
if (shrunk_memory_size <= init_memory_size) {
|
|
/* Reset memory info to decrease memory usage */
|
|
memory->num_bytes_per_page = shrunk_memory_size;
|
|
memory->init_page_count = 1;
|
|
LOG_VERBOSE("Shrink memory size to %" PRIu64,
|
|
shrunk_memory_size);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
if (module->import_memory_count) {
|
|
memory_import = &module->import_memories[0].u.memory;
|
|
if (memory_import->init_page_count < DEFAULT_MAX_PAGES) {
|
|
memory_import->num_bytes_per_page *=
|
|
memory_import->init_page_count;
|
|
if (memory_import->init_page_count > 0)
|
|
memory_import->init_page_count =
|
|
memory_import->max_page_count = 1;
|
|
else
|
|
memory_import->init_page_count =
|
|
memory_import->max_page_count = 0;
|
|
}
|
|
}
|
|
|
|
if (module->memory_count) {
|
|
memory = &module->memories[0];
|
|
if (memory->init_page_count < DEFAULT_MAX_PAGES) {
|
|
memory->num_bytes_per_page *= memory->init_page_count;
|
|
if (memory->init_page_count > 0)
|
|
memory->init_page_count = memory->max_page_count = 1;
|
|
else
|
|
memory->init_page_count = memory->max_page_count = 0;
|
|
}
|
|
}
|
|
}
|
|
|
|
calculate_global_data_offset(module);
|
|
|
|
#if WASM_ENABLE_FAST_JIT != 0
|
|
if (!init_fast_jit_functions(module, error_buf, error_buf_size)) {
|
|
return false;
|
|
}
|
|
#endif
|
|
|
|
#if WASM_ENABLE_JIT != 0
|
|
if (!init_llvm_jit_functions_stage1(module, error_buf, error_buf_size)) {
|
|
return false;
|
|
}
|
|
#if !(WASM_ENABLE_FAST_JIT != 0 && WASM_ENABLE_LAZY_JIT != 0)
|
|
if (!init_llvm_jit_functions_stage2(module, error_buf, error_buf_size)) {
|
|
return false;
|
|
}
|
|
#else
|
|
/* Run aot_compile_wasm in a backend thread, so as not to block the main
|
|
thread fast jit execution, since applying llvm optimizations in
|
|
aot_compile_wasm may cost a lot of time.
|
|
Create thread with enough native stack to apply llvm optimizations */
|
|
if (os_thread_create(&module->llvm_jit_init_thread,
|
|
init_llvm_jit_functions_stage2_callback,
|
|
(void *)module, APP_THREAD_STACK_SIZE_DEFAULT * 8)
|
|
!= 0) {
|
|
set_error_buf(error_buf, error_buf_size,
|
|
"create orcjit compile thread failed");
|
|
return false;
|
|
}
|
|
#endif
|
|
#endif
|
|
|
|
#if WASM_ENABLE_FAST_JIT != 0 || WASM_ENABLE_JIT != 0
|
|
/* Create threads to compile the jit functions */
|
|
if (!compile_jit_functions(module, error_buf, error_buf_size)) {
|
|
return false;
|
|
}
|
|
#endif
|
|
|
|
#if WASM_ENABLE_MEMORY_TRACING != 0
|
|
wasm_runtime_dump_module_mem_consumption(module);
|
|
#endif
|
|
return true;
|
|
}
|
|
|
|
static WASMModule *
|
|
create_module(char *name, char *error_buf, uint32 error_buf_size)
|
|
{
|
|
WASMModule *module =
|
|
loader_malloc(sizeof(WASMModule), error_buf, error_buf_size);
|
|
bh_list_status ret;
|
|
|
|
if (!module) {
|
|
return NULL;
|
|
}
|
|
|
|
module->module_type = Wasm_Module_Bytecode;
|
|
|
|
/* Set start_function to -1, means no start function */
|
|
module->start_function = (uint32)-1;
|
|
|
|
module->name = name;
|
|
module->is_binary_freeable = false;
|
|
|
|
#if WASM_ENABLE_FAST_INTERP == 0
|
|
module->br_table_cache_list = &module->br_table_cache_list_head;
|
|
ret = bh_list_init(module->br_table_cache_list);
|
|
bh_assert(ret == BH_LIST_SUCCESS);
|
|
#endif
|
|
|
|
#if WASM_ENABLE_FAST_JIT != 0 && WASM_ENABLE_JIT != 0 \
|
|
&& WASM_ENABLE_LAZY_JIT != 0
|
|
if (os_mutex_init(&module->instance_list_lock) != 0) {
|
|
set_error_buf(error_buf, error_buf_size,
|
|
"init instance list lock failed");
|
|
wasm_runtime_free(module);
|
|
return NULL;
|
|
}
|
|
#endif
|
|
|
|
(void)ret;
|
|
return module;
|
|
}
|
|
|
|
WASMModule *
|
|
wasm_loader_load_from_sections(WASMSection *section_list, char *error_buf,
|
|
uint32 error_buf_size)
|
|
{
|
|
WASMModule *module = create_module("", error_buf, error_buf_size);
|
|
if (!module)
|
|
return NULL;
|
|
|
|
if (!load_from_sections(module, section_list, false, true, error_buf,
|
|
error_buf_size)) {
|
|
wasm_loader_unload(module);
|
|
return NULL;
|
|
}
|
|
|
|
LOG_VERBOSE("Load module from sections success.\n");
|
|
return module;
|
|
}
|
|
|
|
static void
|
|
destroy_sections(WASMSection *section_list)
|
|
{
|
|
WASMSection *section = section_list, *next;
|
|
while (section) {
|
|
next = section->next;
|
|
wasm_runtime_free(section);
|
|
section = next;
|
|
}
|
|
}
|
|
|
|
/* clang-format off */
|
|
static uint8 section_ids[] = {
|
|
SECTION_TYPE_USER,
|
|
SECTION_TYPE_TYPE,
|
|
SECTION_TYPE_IMPORT,
|
|
SECTION_TYPE_FUNC,
|
|
SECTION_TYPE_TABLE,
|
|
SECTION_TYPE_MEMORY,
|
|
SECTION_TYPE_GLOBAL,
|
|
SECTION_TYPE_EXPORT,
|
|
SECTION_TYPE_START,
|
|
SECTION_TYPE_ELEM,
|
|
#if WASM_ENABLE_BULK_MEMORY != 0
|
|
SECTION_TYPE_DATACOUNT,
|
|
#endif
|
|
SECTION_TYPE_CODE,
|
|
SECTION_TYPE_DATA
|
|
};
|
|
/* clang-format on */
|
|
|
|
static uint8
|
|
get_section_index(uint8 section_type)
|
|
{
|
|
uint8 max_id = sizeof(section_ids) / sizeof(uint8);
|
|
|
|
for (uint8 i = 0; i < max_id; i++) {
|
|
if (section_type == section_ids[i])
|
|
return i;
|
|
}
|
|
|
|
return (uint8)-1;
|
|
}
|
|
|
|
static bool
|
|
create_sections(const uint8 *buf, uint32 size, WASMSection **p_section_list,
|
|
char *error_buf, uint32 error_buf_size)
|
|
{
|
|
WASMSection *section_list_end = NULL, *section;
|
|
const uint8 *p = buf, *p_end = buf + size /*, *section_body*/;
|
|
uint8 section_type, section_index, last_section_index = (uint8)-1;
|
|
uint32 section_size;
|
|
|
|
bh_assert(!*p_section_list);
|
|
|
|
p += 8;
|
|
while (p < p_end) {
|
|
CHECK_BUF(p, p_end, 1);
|
|
section_type = read_uint8(p);
|
|
section_index = get_section_index(section_type);
|
|
if (section_index != (uint8)-1) {
|
|
if (section_type != SECTION_TYPE_USER) {
|
|
/* Custom sections may be inserted at any place,
|
|
while other sections must occur at most once
|
|
and in prescribed order. */
|
|
bh_assert(last_section_index == (uint8)-1
|
|
|| last_section_index < section_index);
|
|
last_section_index = section_index;
|
|
}
|
|
read_leb_uint32(p, p_end, section_size);
|
|
CHECK_BUF1(p, p_end, section_size);
|
|
|
|
if (!(section = loader_malloc(sizeof(WASMSection), error_buf,
|
|
error_buf_size))) {
|
|
return false;
|
|
}
|
|
|
|
section->section_type = section_type;
|
|
section->section_body = (uint8 *)p;
|
|
section->section_body_size = section_size;
|
|
|
|
if (!*p_section_list)
|
|
*p_section_list = section_list_end = section;
|
|
else {
|
|
section_list_end->next = section;
|
|
section_list_end = section;
|
|
}
|
|
|
|
p += section_size;
|
|
}
|
|
else {
|
|
bh_assert(0);
|
|
}
|
|
}
|
|
|
|
(void)last_section_index;
|
|
return true;
|
|
}
|
|
|
|
static void
|
|
exchange32(uint8 *p_data)
|
|
{
|
|
uint8 value = *p_data;
|
|
*p_data = *(p_data + 3);
|
|
*(p_data + 3) = value;
|
|
|
|
value = *(p_data + 1);
|
|
*(p_data + 1) = *(p_data + 2);
|
|
*(p_data + 2) = value;
|
|
}
|
|
|
|
static union {
|
|
int a;
|
|
char b;
|
|
} __ue = { .a = 1 };
|
|
|
|
#define is_little_endian() (__ue.b == 1)
|
|
|
|
static bool
|
|
load(const uint8 *buf, uint32 size, WASMModule *module,
|
|
bool wasm_binary_freeable, char *error_buf, uint32 error_buf_size)
|
|
{
|
|
const uint8 *buf_end = buf + size;
|
|
const uint8 *p = buf, *p_end = buf_end;
|
|
uint32 magic_number, version;
|
|
WASMSection *section_list = NULL;
|
|
|
|
CHECK_BUF1(p, p_end, sizeof(uint32));
|
|
magic_number = read_uint32(p);
|
|
if (!is_little_endian())
|
|
exchange32((uint8 *)&magic_number);
|
|
|
|
bh_assert(magic_number == WASM_MAGIC_NUMBER);
|
|
|
|
CHECK_BUF1(p, p_end, sizeof(uint32));
|
|
version = read_uint32(p);
|
|
if (!is_little_endian())
|
|
exchange32((uint8 *)&version);
|
|
|
|
if (version != WASM_CURRENT_VERSION) {
|
|
set_error_buf(error_buf, error_buf_size, "unknown binary version");
|
|
return false;
|
|
}
|
|
|
|
if (!create_sections(buf, size, §ion_list, error_buf, error_buf_size)
|
|
|| !load_from_sections(module, section_list, true, wasm_binary_freeable,
|
|
error_buf, error_buf_size)) {
|
|
destroy_sections(section_list);
|
|
return false;
|
|
}
|
|
|
|
destroy_sections(section_list);
|
|
(void)p_end;
|
|
return true;
|
|
}
|
|
|
|
WASMModule *
|
|
wasm_loader_load(uint8 *buf, uint32 size, const LoadArgs *args, char *error_buf,
|
|
uint32 error_buf_size)
|
|
{
|
|
WASMModule *module = create_module(args->name, error_buf, error_buf_size);
|
|
if (!module) {
|
|
return NULL;
|
|
}
|
|
|
|
#if WASM_ENABLE_FAST_JIT != 0 || WASM_ENABLE_DUMP_CALL_STACK != 0 \
|
|
|| WASM_ENABLE_JIT != 0
|
|
module->load_addr = (uint8 *)buf;
|
|
module->load_size = size;
|
|
#endif
|
|
|
|
if (!load(buf, size, module, args->wasm_binary_freeable, error_buf,
|
|
error_buf_size)) {
|
|
goto fail;
|
|
}
|
|
|
|
LOG_VERBOSE("Load module success.\n");
|
|
return module;
|
|
|
|
fail:
|
|
wasm_loader_unload(module);
|
|
return NULL;
|
|
}
|
|
|
|
void
|
|
wasm_loader_unload(WASMModule *module)
|
|
{
|
|
uint32 i;
|
|
|
|
if (!module)
|
|
return;
|
|
|
|
#if WASM_ENABLE_FAST_JIT != 0 && WASM_ENABLE_JIT != 0 \
|
|
&& WASM_ENABLE_LAZY_JIT != 0
|
|
module->orcjit_stop_compiling = true;
|
|
if (module->llvm_jit_init_thread)
|
|
os_thread_join(module->llvm_jit_init_thread, NULL);
|
|
#endif
|
|
|
|
#if WASM_ENABLE_FAST_JIT != 0 || WASM_ENABLE_JIT != 0
|
|
/* Stop Fast/LLVM JIT compilation firstly to avoid accessing
|
|
module internal data after they were freed */
|
|
orcjit_stop_compile_threads(module);
|
|
#endif
|
|
|
|
#if WASM_ENABLE_JIT != 0
|
|
if (module->func_ptrs)
|
|
wasm_runtime_free(module->func_ptrs);
|
|
if (module->comp_ctx)
|
|
aot_destroy_comp_context(module->comp_ctx);
|
|
if (module->comp_data)
|
|
aot_destroy_comp_data(module->comp_data);
|
|
#endif
|
|
|
|
#if WASM_ENABLE_FAST_JIT != 0 && WASM_ENABLE_JIT != 0 \
|
|
&& WASM_ENABLE_LAZY_JIT != 0
|
|
if (module->tierup_wait_lock_inited) {
|
|
os_mutex_destroy(&module->tierup_wait_lock);
|
|
os_cond_destroy(&module->tierup_wait_cond);
|
|
}
|
|
#endif
|
|
|
|
if (module->types) {
|
|
for (i = 0; i < module->type_count; i++) {
|
|
if (module->types[i])
|
|
destroy_wasm_type(module->types[i]);
|
|
}
|
|
wasm_runtime_free(module->types);
|
|
}
|
|
|
|
if (module->imports)
|
|
wasm_runtime_free(module->imports);
|
|
|
|
if (module->functions) {
|
|
for (i = 0; i < module->function_count; i++) {
|
|
if (module->functions[i]) {
|
|
if (module->functions[i]->local_offsets)
|
|
wasm_runtime_free(module->functions[i]->local_offsets);
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
if (module->functions[i]->code_compiled)
|
|
wasm_runtime_free(module->functions[i]->code_compiled);
|
|
if (module->functions[i]->consts)
|
|
wasm_runtime_free(module->functions[i]->consts);
|
|
#endif
|
|
#if WASM_ENABLE_FAST_JIT != 0
|
|
if (module->functions[i]->fast_jit_jitted_code) {
|
|
jit_code_cache_free(
|
|
module->functions[i]->fast_jit_jitted_code);
|
|
}
|
|
#if WASM_ENABLE_JIT != 0 && WASM_ENABLE_LAZY_JIT != 0
|
|
if (module->functions[i]->call_to_fast_jit_from_llvm_jit) {
|
|
jit_code_cache_free(
|
|
module->functions[i]->call_to_fast_jit_from_llvm_jit);
|
|
}
|
|
#endif
|
|
#endif
|
|
wasm_runtime_free(module->functions[i]);
|
|
}
|
|
}
|
|
wasm_runtime_free(module->functions);
|
|
}
|
|
|
|
if (module->tables)
|
|
wasm_runtime_free(module->tables);
|
|
|
|
if (module->memories)
|
|
wasm_runtime_free(module->memories);
|
|
|
|
if (module->globals)
|
|
wasm_runtime_free(module->globals);
|
|
|
|
if (module->exports)
|
|
wasm_runtime_free(module->exports);
|
|
|
|
if (module->table_segments) {
|
|
for (i = 0; i < module->table_seg_count; i++) {
|
|
if (module->table_segments[i].init_values)
|
|
wasm_runtime_free(module->table_segments[i].init_values);
|
|
}
|
|
wasm_runtime_free(module->table_segments);
|
|
}
|
|
|
|
if (module->data_segments) {
|
|
for (i = 0; i < module->data_seg_count; i++) {
|
|
if (module->data_segments[i]) {
|
|
if (module->data_segments[i]->is_data_cloned)
|
|
wasm_runtime_free(module->data_segments[i]->data);
|
|
wasm_runtime_free(module->data_segments[i]);
|
|
}
|
|
}
|
|
wasm_runtime_free(module->data_segments);
|
|
}
|
|
|
|
if (module->const_str_list) {
|
|
StringNode *node = module->const_str_list, *node_next;
|
|
while (node) {
|
|
node_next = node->next;
|
|
wasm_runtime_free(node);
|
|
node = node_next;
|
|
}
|
|
}
|
|
|
|
#if WASM_ENABLE_FAST_INTERP == 0
|
|
if (module->br_table_cache_list) {
|
|
BrTableCache *node = bh_list_first_elem(module->br_table_cache_list);
|
|
BrTableCache *node_next;
|
|
while (node) {
|
|
node_next = bh_list_elem_next(node);
|
|
wasm_runtime_free(node);
|
|
node = node_next;
|
|
}
|
|
}
|
|
#endif
|
|
|
|
#if WASM_ENABLE_FAST_JIT != 0 && WASM_ENABLE_JIT != 0 \
|
|
&& WASM_ENABLE_LAZY_JIT != 0
|
|
os_mutex_destroy(&module->instance_list_lock);
|
|
#endif
|
|
|
|
#if WASM_ENABLE_FAST_JIT != 0
|
|
if (module->fast_jit_func_ptrs) {
|
|
wasm_runtime_free(module->fast_jit_func_ptrs);
|
|
}
|
|
|
|
for (i = 0; i < WASM_ORC_JIT_BACKEND_THREAD_NUM; i++) {
|
|
if (module->fast_jit_thread_locks_inited[i]) {
|
|
os_mutex_destroy(&module->fast_jit_thread_locks[i]);
|
|
}
|
|
}
|
|
#endif
|
|
|
|
wasm_runtime_free(module);
|
|
}
|
|
|
|
bool
|
|
wasm_loader_find_block_addr(WASMExecEnv *exec_env, BlockAddr *block_addr_cache,
|
|
const uint8 *start_addr, const uint8 *code_end_addr,
|
|
uint8 label_type, uint8 **p_else_addr,
|
|
uint8 **p_end_addr)
|
|
{
|
|
const uint8 *p = start_addr, *p_end = code_end_addr;
|
|
uint8 *else_addr = NULL;
|
|
char error_buf[128];
|
|
uint32 block_nested_depth = 1, count, i, j, t;
|
|
uint32 error_buf_size = sizeof(error_buf);
|
|
uint8 opcode, u8;
|
|
BlockAddr block_stack[16] = { 0 }, *block;
|
|
|
|
i = ((uintptr_t)start_addr) & (uintptr_t)(BLOCK_ADDR_CACHE_SIZE - 1);
|
|
block = block_addr_cache + BLOCK_ADDR_CONFLICT_SIZE * i;
|
|
|
|
for (j = 0; j < BLOCK_ADDR_CONFLICT_SIZE; j++) {
|
|
if (block[j].start_addr == start_addr) {
|
|
/* Cache hit */
|
|
*p_else_addr = block[j].else_addr;
|
|
*p_end_addr = block[j].end_addr;
|
|
return true;
|
|
}
|
|
}
|
|
|
|
/* Cache unhit */
|
|
block_stack[0].start_addr = start_addr;
|
|
|
|
while (p < code_end_addr) {
|
|
opcode = *p++;
|
|
|
|
switch (opcode) {
|
|
case WASM_OP_UNREACHABLE:
|
|
case WASM_OP_NOP:
|
|
break;
|
|
|
|
case WASM_OP_BLOCK:
|
|
case WASM_OP_LOOP:
|
|
case WASM_OP_IF:
|
|
/* block result type: 0x40/0x7F/0x7E/0x7D/0x7C */
|
|
u8 = read_uint8(p);
|
|
if (block_nested_depth
|
|
< sizeof(block_stack) / sizeof(BlockAddr)) {
|
|
block_stack[block_nested_depth].start_addr = p;
|
|
block_stack[block_nested_depth].else_addr = NULL;
|
|
}
|
|
block_nested_depth++;
|
|
break;
|
|
|
|
case EXT_OP_BLOCK:
|
|
case EXT_OP_LOOP:
|
|
case EXT_OP_IF:
|
|
/* block type */
|
|
skip_leb_uint32(p, p_end);
|
|
if (block_nested_depth
|
|
< sizeof(block_stack) / sizeof(BlockAddr)) {
|
|
block_stack[block_nested_depth].start_addr = p;
|
|
block_stack[block_nested_depth].else_addr = NULL;
|
|
}
|
|
block_nested_depth++;
|
|
break;
|
|
|
|
case WASM_OP_ELSE:
|
|
if (label_type == LABEL_TYPE_IF && block_nested_depth == 1)
|
|
else_addr = (uint8 *)(p - 1);
|
|
if (block_nested_depth - 1
|
|
< sizeof(block_stack) / sizeof(BlockAddr))
|
|
block_stack[block_nested_depth - 1].else_addr =
|
|
(uint8 *)(p - 1);
|
|
break;
|
|
|
|
case WASM_OP_END:
|
|
if (block_nested_depth == 1) {
|
|
if (label_type == LABEL_TYPE_IF)
|
|
*p_else_addr = else_addr;
|
|
*p_end_addr = (uint8 *)(p - 1);
|
|
|
|
block_stack[0].end_addr = (uint8 *)(p - 1);
|
|
for (t = 0; t < sizeof(block_stack) / sizeof(BlockAddr);
|
|
t++) {
|
|
start_addr = block_stack[t].start_addr;
|
|
if (start_addr) {
|
|
i = ((uintptr_t)start_addr)
|
|
& (uintptr_t)(BLOCK_ADDR_CACHE_SIZE - 1);
|
|
block =
|
|
block_addr_cache + BLOCK_ADDR_CONFLICT_SIZE * i;
|
|
for (j = 0; j < BLOCK_ADDR_CONFLICT_SIZE; j++)
|
|
if (!block[j].start_addr)
|
|
break;
|
|
|
|
if (j == BLOCK_ADDR_CONFLICT_SIZE) {
|
|
memmove(block + 1, block,
|
|
(BLOCK_ADDR_CONFLICT_SIZE - 1)
|
|
* sizeof(BlockAddr));
|
|
j = 0;
|
|
}
|
|
block[j].start_addr = block_stack[t].start_addr;
|
|
block[j].else_addr = block_stack[t].else_addr;
|
|
block[j].end_addr = block_stack[t].end_addr;
|
|
}
|
|
else
|
|
break;
|
|
}
|
|
return true;
|
|
}
|
|
else {
|
|
block_nested_depth--;
|
|
if (block_nested_depth
|
|
< sizeof(block_stack) / sizeof(BlockAddr))
|
|
block_stack[block_nested_depth].end_addr =
|
|
(uint8 *)(p - 1);
|
|
}
|
|
break;
|
|
|
|
case WASM_OP_BR:
|
|
case WASM_OP_BR_IF:
|
|
skip_leb_uint32(p, p_end); /* labelidx */
|
|
break;
|
|
|
|
case WASM_OP_BR_TABLE:
|
|
read_leb_uint32(p, p_end, count); /* lable num */
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
for (i = 0; i <= count; i++) /* lableidxs */
|
|
skip_leb_uint32(p, p_end);
|
|
#else
|
|
p += count + 1;
|
|
while (*p == WASM_OP_NOP)
|
|
p++;
|
|
#endif
|
|
break;
|
|
|
|
#if WASM_ENABLE_FAST_INTERP == 0
|
|
case EXT_OP_BR_TABLE_CACHE:
|
|
read_leb_uint32(p, p_end, count); /* lable num */
|
|
while (*p == WASM_OP_NOP)
|
|
p++;
|
|
break;
|
|
#endif
|
|
|
|
case WASM_OP_RETURN:
|
|
break;
|
|
|
|
case WASM_OP_CALL:
|
|
#if WASM_ENABLE_TAIL_CALL != 0
|
|
case WASM_OP_RETURN_CALL:
|
|
#endif
|
|
skip_leb_uint32(p, p_end); /* funcidx */
|
|
break;
|
|
|
|
case WASM_OP_CALL_INDIRECT:
|
|
#if WASM_ENABLE_TAIL_CALL != 0
|
|
case WASM_OP_RETURN_CALL_INDIRECT:
|
|
#endif
|
|
skip_leb_uint32(p, p_end); /* typeidx */
|
|
CHECK_BUF(p, p_end, 1);
|
|
u8 = read_uint8(p); /* 0x00 */
|
|
break;
|
|
|
|
#if WASM_ENABLE_EXCE_HANDLING != 0
|
|
case WASM_OP_TRY:
|
|
case WASM_OP_CATCH:
|
|
case WASM_OP_THROW:
|
|
case WASM_OP_RETHROW:
|
|
case WASM_OP_DELEGATE:
|
|
case WASM_OP_CATCH_ALL:
|
|
/* TODO */
|
|
return false;
|
|
#endif
|
|
|
|
case WASM_OP_DROP:
|
|
case WASM_OP_SELECT:
|
|
case WASM_OP_DROP_64:
|
|
case WASM_OP_SELECT_64:
|
|
break;
|
|
#if WASM_ENABLE_REF_TYPES != 0
|
|
case WASM_OP_SELECT_T:
|
|
skip_leb_uint32(p, p_end); /* vec length */
|
|
CHECK_BUF(p, p_end, 1);
|
|
u8 = read_uint8(p); /* typeidx */
|
|
break;
|
|
case WASM_OP_TABLE_GET:
|
|
case WASM_OP_TABLE_SET:
|
|
skip_leb_uint32(p, p_end); /* table index */
|
|
break;
|
|
case WASM_OP_REF_NULL:
|
|
CHECK_BUF(p, p_end, 1);
|
|
u8 = read_uint8(p); /* type */
|
|
break;
|
|
case WASM_OP_REF_IS_NULL:
|
|
break;
|
|
case WASM_OP_REF_FUNC:
|
|
skip_leb_uint32(p, p_end); /* func index */
|
|
break;
|
|
#endif /* WASM_ENABLE_REF_TYPES */
|
|
case WASM_OP_GET_LOCAL:
|
|
case WASM_OP_SET_LOCAL:
|
|
case WASM_OP_TEE_LOCAL:
|
|
case WASM_OP_GET_GLOBAL:
|
|
case WASM_OP_SET_GLOBAL:
|
|
case WASM_OP_GET_GLOBAL_64:
|
|
case WASM_OP_SET_GLOBAL_64:
|
|
case WASM_OP_SET_GLOBAL_AUX_STACK:
|
|
skip_leb_uint32(p, p_end); /* localidx */
|
|
break;
|
|
|
|
case EXT_OP_GET_LOCAL_FAST:
|
|
case EXT_OP_SET_LOCAL_FAST:
|
|
case EXT_OP_TEE_LOCAL_FAST:
|
|
CHECK_BUF(p, p_end, 1);
|
|
p++;
|
|
break;
|
|
|
|
case WASM_OP_I32_LOAD:
|
|
case WASM_OP_I64_LOAD:
|
|
case WASM_OP_F32_LOAD:
|
|
case WASM_OP_F64_LOAD:
|
|
case WASM_OP_I32_LOAD8_S:
|
|
case WASM_OP_I32_LOAD8_U:
|
|
case WASM_OP_I32_LOAD16_S:
|
|
case WASM_OP_I32_LOAD16_U:
|
|
case WASM_OP_I64_LOAD8_S:
|
|
case WASM_OP_I64_LOAD8_U:
|
|
case WASM_OP_I64_LOAD16_S:
|
|
case WASM_OP_I64_LOAD16_U:
|
|
case WASM_OP_I64_LOAD32_S:
|
|
case WASM_OP_I64_LOAD32_U:
|
|
case WASM_OP_I32_STORE:
|
|
case WASM_OP_I64_STORE:
|
|
case WASM_OP_F32_STORE:
|
|
case WASM_OP_F64_STORE:
|
|
case WASM_OP_I32_STORE8:
|
|
case WASM_OP_I32_STORE16:
|
|
case WASM_OP_I64_STORE8:
|
|
case WASM_OP_I64_STORE16:
|
|
case WASM_OP_I64_STORE32:
|
|
skip_leb_uint32(p, p_end); /* align */
|
|
skip_leb_mem_offset(p, p_end); /* offset */
|
|
break;
|
|
|
|
case WASM_OP_MEMORY_SIZE:
|
|
case WASM_OP_MEMORY_GROW:
|
|
skip_leb_uint32(p, p_end); /* 0x00 */
|
|
break;
|
|
|
|
case WASM_OP_I32_CONST:
|
|
skip_leb_int32(p, p_end);
|
|
break;
|
|
case WASM_OP_I64_CONST:
|
|
skip_leb_int64(p, p_end);
|
|
break;
|
|
case WASM_OP_F32_CONST:
|
|
p += sizeof(float32);
|
|
break;
|
|
case WASM_OP_F64_CONST:
|
|
p += sizeof(float64);
|
|
break;
|
|
|
|
case WASM_OP_I32_EQZ:
|
|
case WASM_OP_I32_EQ:
|
|
case WASM_OP_I32_NE:
|
|
case WASM_OP_I32_LT_S:
|
|
case WASM_OP_I32_LT_U:
|
|
case WASM_OP_I32_GT_S:
|
|
case WASM_OP_I32_GT_U:
|
|
case WASM_OP_I32_LE_S:
|
|
case WASM_OP_I32_LE_U:
|
|
case WASM_OP_I32_GE_S:
|
|
case WASM_OP_I32_GE_U:
|
|
case WASM_OP_I64_EQZ:
|
|
case WASM_OP_I64_EQ:
|
|
case WASM_OP_I64_NE:
|
|
case WASM_OP_I64_LT_S:
|
|
case WASM_OP_I64_LT_U:
|
|
case WASM_OP_I64_GT_S:
|
|
case WASM_OP_I64_GT_U:
|
|
case WASM_OP_I64_LE_S:
|
|
case WASM_OP_I64_LE_U:
|
|
case WASM_OP_I64_GE_S:
|
|
case WASM_OP_I64_GE_U:
|
|
case WASM_OP_F32_EQ:
|
|
case WASM_OP_F32_NE:
|
|
case WASM_OP_F32_LT:
|
|
case WASM_OP_F32_GT:
|
|
case WASM_OP_F32_LE:
|
|
case WASM_OP_F32_GE:
|
|
case WASM_OP_F64_EQ:
|
|
case WASM_OP_F64_NE:
|
|
case WASM_OP_F64_LT:
|
|
case WASM_OP_F64_GT:
|
|
case WASM_OP_F64_LE:
|
|
case WASM_OP_F64_GE:
|
|
case WASM_OP_I32_CLZ:
|
|
case WASM_OP_I32_CTZ:
|
|
case WASM_OP_I32_POPCNT:
|
|
case WASM_OP_I32_ADD:
|
|
case WASM_OP_I32_SUB:
|
|
case WASM_OP_I32_MUL:
|
|
case WASM_OP_I32_DIV_S:
|
|
case WASM_OP_I32_DIV_U:
|
|
case WASM_OP_I32_REM_S:
|
|
case WASM_OP_I32_REM_U:
|
|
case WASM_OP_I32_AND:
|
|
case WASM_OP_I32_OR:
|
|
case WASM_OP_I32_XOR:
|
|
case WASM_OP_I32_SHL:
|
|
case WASM_OP_I32_SHR_S:
|
|
case WASM_OP_I32_SHR_U:
|
|
case WASM_OP_I32_ROTL:
|
|
case WASM_OP_I32_ROTR:
|
|
case WASM_OP_I64_CLZ:
|
|
case WASM_OP_I64_CTZ:
|
|
case WASM_OP_I64_POPCNT:
|
|
case WASM_OP_I64_ADD:
|
|
case WASM_OP_I64_SUB:
|
|
case WASM_OP_I64_MUL:
|
|
case WASM_OP_I64_DIV_S:
|
|
case WASM_OP_I64_DIV_U:
|
|
case WASM_OP_I64_REM_S:
|
|
case WASM_OP_I64_REM_U:
|
|
case WASM_OP_I64_AND:
|
|
case WASM_OP_I64_OR:
|
|
case WASM_OP_I64_XOR:
|
|
case WASM_OP_I64_SHL:
|
|
case WASM_OP_I64_SHR_S:
|
|
case WASM_OP_I64_SHR_U:
|
|
case WASM_OP_I64_ROTL:
|
|
case WASM_OP_I64_ROTR:
|
|
case WASM_OP_F32_ABS:
|
|
case WASM_OP_F32_NEG:
|
|
case WASM_OP_F32_CEIL:
|
|
case WASM_OP_F32_FLOOR:
|
|
case WASM_OP_F32_TRUNC:
|
|
case WASM_OP_F32_NEAREST:
|
|
case WASM_OP_F32_SQRT:
|
|
case WASM_OP_F32_ADD:
|
|
case WASM_OP_F32_SUB:
|
|
case WASM_OP_F32_MUL:
|
|
case WASM_OP_F32_DIV:
|
|
case WASM_OP_F32_MIN:
|
|
case WASM_OP_F32_MAX:
|
|
case WASM_OP_F32_COPYSIGN:
|
|
case WASM_OP_F64_ABS:
|
|
case WASM_OP_F64_NEG:
|
|
case WASM_OP_F64_CEIL:
|
|
case WASM_OP_F64_FLOOR:
|
|
case WASM_OP_F64_TRUNC:
|
|
case WASM_OP_F64_NEAREST:
|
|
case WASM_OP_F64_SQRT:
|
|
case WASM_OP_F64_ADD:
|
|
case WASM_OP_F64_SUB:
|
|
case WASM_OP_F64_MUL:
|
|
case WASM_OP_F64_DIV:
|
|
case WASM_OP_F64_MIN:
|
|
case WASM_OP_F64_MAX:
|
|
case WASM_OP_F64_COPYSIGN:
|
|
case WASM_OP_I32_WRAP_I64:
|
|
case WASM_OP_I32_TRUNC_S_F32:
|
|
case WASM_OP_I32_TRUNC_U_F32:
|
|
case WASM_OP_I32_TRUNC_S_F64:
|
|
case WASM_OP_I32_TRUNC_U_F64:
|
|
case WASM_OP_I64_EXTEND_S_I32:
|
|
case WASM_OP_I64_EXTEND_U_I32:
|
|
case WASM_OP_I64_TRUNC_S_F32:
|
|
case WASM_OP_I64_TRUNC_U_F32:
|
|
case WASM_OP_I64_TRUNC_S_F64:
|
|
case WASM_OP_I64_TRUNC_U_F64:
|
|
case WASM_OP_F32_CONVERT_S_I32:
|
|
case WASM_OP_F32_CONVERT_U_I32:
|
|
case WASM_OP_F32_CONVERT_S_I64:
|
|
case WASM_OP_F32_CONVERT_U_I64:
|
|
case WASM_OP_F32_DEMOTE_F64:
|
|
case WASM_OP_F64_CONVERT_S_I32:
|
|
case WASM_OP_F64_CONVERT_U_I32:
|
|
case WASM_OP_F64_CONVERT_S_I64:
|
|
case WASM_OP_F64_CONVERT_U_I64:
|
|
case WASM_OP_F64_PROMOTE_F32:
|
|
case WASM_OP_I32_REINTERPRET_F32:
|
|
case WASM_OP_I64_REINTERPRET_F64:
|
|
case WASM_OP_F32_REINTERPRET_I32:
|
|
case WASM_OP_F64_REINTERPRET_I64:
|
|
case WASM_OP_I32_EXTEND8_S:
|
|
case WASM_OP_I32_EXTEND16_S:
|
|
case WASM_OP_I64_EXTEND8_S:
|
|
case WASM_OP_I64_EXTEND16_S:
|
|
case WASM_OP_I64_EXTEND32_S:
|
|
break;
|
|
case WASM_OP_MISC_PREFIX:
|
|
{
|
|
uint32 opcode1;
|
|
|
|
read_leb_uint32(p, p_end, opcode1);
|
|
/* opcode1 was checked in wasm_loader_prepare_bytecode and
|
|
is no larger than UINT8_MAX */
|
|
opcode = (uint8)opcode1;
|
|
|
|
switch (opcode) {
|
|
case WASM_OP_I32_TRUNC_SAT_S_F32:
|
|
case WASM_OP_I32_TRUNC_SAT_U_F32:
|
|
case WASM_OP_I32_TRUNC_SAT_S_F64:
|
|
case WASM_OP_I32_TRUNC_SAT_U_F64:
|
|
case WASM_OP_I64_TRUNC_SAT_S_F32:
|
|
case WASM_OP_I64_TRUNC_SAT_U_F32:
|
|
case WASM_OP_I64_TRUNC_SAT_S_F64:
|
|
case WASM_OP_I64_TRUNC_SAT_U_F64:
|
|
break;
|
|
#if WASM_ENABLE_BULK_MEMORY != 0
|
|
case WASM_OP_MEMORY_INIT:
|
|
skip_leb_uint32(p, p_end);
|
|
/* skip memory idx */
|
|
p++;
|
|
break;
|
|
case WASM_OP_DATA_DROP:
|
|
skip_leb_uint32(p, p_end);
|
|
break;
|
|
case WASM_OP_MEMORY_COPY:
|
|
/* skip two memory idx */
|
|
p += 2;
|
|
break;
|
|
case WASM_OP_MEMORY_FILL:
|
|
/* skip memory idx */
|
|
p++;
|
|
break;
|
|
#endif
|
|
#if WASM_ENABLE_REF_TYPES != 0
|
|
case WASM_OP_TABLE_INIT:
|
|
case WASM_OP_TABLE_COPY:
|
|
/* tableidx */
|
|
skip_leb_uint32(p, p_end);
|
|
/* elemidx */
|
|
skip_leb_uint32(p, p_end);
|
|
break;
|
|
case WASM_OP_ELEM_DROP:
|
|
/* elemidx */
|
|
skip_leb_uint32(p, p_end);
|
|
break;
|
|
case WASM_OP_TABLE_SIZE:
|
|
case WASM_OP_TABLE_GROW:
|
|
case WASM_OP_TABLE_FILL:
|
|
skip_leb_uint32(p, p_end); /* table idx */
|
|
break;
|
|
#endif /* WASM_ENABLE_REF_TYPES */
|
|
default:
|
|
bh_assert(0);
|
|
break;
|
|
}
|
|
break;
|
|
}
|
|
|
|
#if WASM_ENABLE_SHARED_MEMORY != 0
|
|
case WASM_OP_ATOMIC_PREFIX:
|
|
{
|
|
/* TODO: memory64 offset type changes */
|
|
uint32 opcode1;
|
|
|
|
/* atomic_op (u32_leb) + memarg (2 u32_leb) */
|
|
read_leb_uint32(p, p_end, opcode1);
|
|
/* opcode1 was checked in wasm_loader_prepare_bytecode and
|
|
is no larger than UINT8_MAX */
|
|
opcode = (uint8)opcode1;
|
|
|
|
if (opcode != WASM_OP_ATOMIC_FENCE) {
|
|
skip_leb_uint32(p, p_end); /* align */
|
|
skip_leb_mem_offset(p, p_end); /* offset */
|
|
}
|
|
else {
|
|
/* atomic.fence doesn't have memarg */
|
|
p++;
|
|
}
|
|
break;
|
|
}
|
|
#endif
|
|
|
|
default:
|
|
bh_assert(0);
|
|
break;
|
|
}
|
|
}
|
|
|
|
(void)u8;
|
|
return false;
|
|
}
|
|
|
|
#define REF_I32 VALUE_TYPE_I32
|
|
#define REF_F32 VALUE_TYPE_F32
|
|
#define REF_I64_1 VALUE_TYPE_I64
|
|
#define REF_I64_2 VALUE_TYPE_I64
|
|
#define REF_F64_1 VALUE_TYPE_F64
|
|
#define REF_F64_2 VALUE_TYPE_F64
|
|
#define REF_ANY VALUE_TYPE_ANY
|
|
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
|
|
#if WASM_DEBUG_PREPROCESSOR != 0
|
|
#define LOG_OP(...) os_printf(__VA_ARGS__)
|
|
#else
|
|
#define LOG_OP(...) (void)0
|
|
#endif
|
|
|
|
#define PATCH_ELSE 0
|
|
#define PATCH_END 1
|
|
typedef struct BranchBlockPatch {
|
|
struct BranchBlockPatch *next;
|
|
uint8 patch_type;
|
|
uint8 *code_compiled;
|
|
} BranchBlockPatch;
|
|
#endif
|
|
|
|
typedef struct BranchBlock {
|
|
uint8 label_type;
|
|
BlockType block_type;
|
|
uint8 *start_addr;
|
|
uint8 *else_addr;
|
|
uint8 *end_addr;
|
|
uint32 stack_cell_num;
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
uint16 dynamic_offset;
|
|
uint8 *code_compiled;
|
|
BranchBlockPatch *patch_list;
|
|
/* This is used to save params frame_offset of of if block */
|
|
int16 *param_frame_offsets;
|
|
/* This is used to store available param num for if/else branch, so the else
|
|
* opcode can know how many parameters should be copied to the stack */
|
|
uint32 available_param_num;
|
|
/* This is used to recover dynamic offset for else branch */
|
|
uint16 start_dynamic_offset;
|
|
#endif
|
|
|
|
/* Indicate the operand stack is in polymorphic state.
|
|
* If the opcode is one of unreachable/br/br_table/return, stack is marked
|
|
* to polymorphic state until the block's 'end' opcode is processed.
|
|
* If stack is in polymorphic state and stack is empty, instruction can
|
|
* pop any type of value directly without decreasing stack top pointer
|
|
* and stack cell num. */
|
|
bool is_stack_polymorphic;
|
|
} BranchBlock;
|
|
|
|
typedef struct WASMLoaderContext {
|
|
/* frame ref stack */
|
|
uint8 *frame_ref;
|
|
uint8 *frame_ref_bottom;
|
|
uint8 *frame_ref_boundary;
|
|
uint32 frame_ref_size;
|
|
uint32 stack_cell_num;
|
|
uint32 max_stack_cell_num;
|
|
|
|
/* frame csp stack */
|
|
BranchBlock *frame_csp;
|
|
BranchBlock *frame_csp_bottom;
|
|
BranchBlock *frame_csp_boundary;
|
|
uint32 frame_csp_size;
|
|
uint32 csp_num;
|
|
uint32 max_csp_num;
|
|
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
/* frame offset stack */
|
|
int16 *frame_offset;
|
|
int16 *frame_offset_bottom;
|
|
int16 *frame_offset_boundary;
|
|
uint32 frame_offset_size;
|
|
int16 dynamic_offset;
|
|
int16 start_dynamic_offset;
|
|
int16 max_dynamic_offset;
|
|
|
|
/* preserved local offset */
|
|
int16 preserved_local_offset;
|
|
|
|
/* const buffer */
|
|
uint8 *const_buf;
|
|
uint16 num_const;
|
|
uint16 const_cell_num;
|
|
uint32 const_buf_size;
|
|
|
|
/* processed code */
|
|
uint8 *p_code_compiled;
|
|
uint8 *p_code_compiled_end;
|
|
uint32 code_compiled_size;
|
|
/* If the last opcode will be dropped, the peak memory usage will be larger
|
|
* than the final code_compiled_size, we record the peak size to ensure
|
|
* there will not be invalid memory access during second traverse */
|
|
uint32 code_compiled_peak_size;
|
|
#endif
|
|
} WASMLoaderContext;
|
|
|
|
typedef struct Const {
|
|
WASMValue value;
|
|
uint16 slot_index;
|
|
uint8 value_type;
|
|
} Const;
|
|
|
|
#define CHECK_CSP_PUSH() \
|
|
do { \
|
|
if (ctx->frame_csp >= ctx->frame_csp_boundary) { \
|
|
MEM_REALLOC( \
|
|
ctx->frame_csp_bottom, ctx->frame_csp_size, \
|
|
(uint32)(ctx->frame_csp_size + 8 * sizeof(BranchBlock))); \
|
|
ctx->frame_csp_size += (uint32)(8 * sizeof(BranchBlock)); \
|
|
ctx->frame_csp_boundary = \
|
|
ctx->frame_csp_bottom \
|
|
+ ctx->frame_csp_size / sizeof(BranchBlock); \
|
|
ctx->frame_csp = ctx->frame_csp_bottom + ctx->csp_num; \
|
|
} \
|
|
} while (0)
|
|
|
|
#define CHECK_CSP_POP() \
|
|
do { \
|
|
bh_assert(ctx->csp_num >= 1); \
|
|
} while (0)
|
|
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
static bool
|
|
check_offset_push(WASMLoaderContext *ctx, char *error_buf,
|
|
uint32 error_buf_size)
|
|
{
|
|
uint32 cell_num = (uint32)(ctx->frame_offset - ctx->frame_offset_bottom);
|
|
if (ctx->frame_offset >= ctx->frame_offset_boundary) {
|
|
MEM_REALLOC(ctx->frame_offset_bottom, ctx->frame_offset_size,
|
|
ctx->frame_offset_size + 16);
|
|
ctx->frame_offset_size += 16;
|
|
ctx->frame_offset_boundary =
|
|
ctx->frame_offset_bottom + ctx->frame_offset_size / sizeof(int16);
|
|
ctx->frame_offset = ctx->frame_offset_bottom + cell_num;
|
|
}
|
|
return true;
|
|
fail:
|
|
return false;
|
|
}
|
|
|
|
static bool
|
|
check_offset_pop(WASMLoaderContext *ctx, uint32 cells)
|
|
{
|
|
if (ctx->frame_offset - cells < ctx->frame_offset_bottom)
|
|
return false;
|
|
return true;
|
|
}
|
|
|
|
static void
|
|
free_label_patch_list(BranchBlock *frame_csp)
|
|
{
|
|
BranchBlockPatch *label_patch = frame_csp->patch_list;
|
|
BranchBlockPatch *next;
|
|
while (label_patch != NULL) {
|
|
next = label_patch->next;
|
|
wasm_runtime_free(label_patch);
|
|
label_patch = next;
|
|
}
|
|
frame_csp->patch_list = NULL;
|
|
}
|
|
|
|
static void
|
|
free_all_label_patch_lists(BranchBlock *frame_csp, uint32 csp_num)
|
|
{
|
|
BranchBlock *tmp_csp = frame_csp;
|
|
|
|
for (uint32 i = 0; i < csp_num; i++) {
|
|
free_label_patch_list(tmp_csp);
|
|
tmp_csp++;
|
|
}
|
|
}
|
|
|
|
#endif
|
|
|
|
static bool
|
|
check_stack_push(WASMLoaderContext *ctx, char *error_buf, uint32 error_buf_size)
|
|
{
|
|
if (ctx->frame_ref >= ctx->frame_ref_boundary) {
|
|
MEM_REALLOC(ctx->frame_ref_bottom, ctx->frame_ref_size,
|
|
ctx->frame_ref_size + 16);
|
|
ctx->frame_ref_size += 16;
|
|
ctx->frame_ref_boundary = ctx->frame_ref_bottom + ctx->frame_ref_size;
|
|
ctx->frame_ref = ctx->frame_ref_bottom + ctx->stack_cell_num;
|
|
}
|
|
return true;
|
|
fail:
|
|
return false;
|
|
}
|
|
|
|
static bool
|
|
check_stack_top_values(uint8 *frame_ref, int32 stack_cell_num, uint8 type,
|
|
char *error_buf, uint32 error_buf_size)
|
|
{
|
|
bh_assert(!((is_32bit_type(type) && stack_cell_num < 1)
|
|
|| (is_64bit_type(type) && stack_cell_num < 2)));
|
|
|
|
bh_assert(!(
|
|
(type == VALUE_TYPE_I32 && *(frame_ref - 1) != REF_I32)
|
|
|| (type == VALUE_TYPE_F32 && *(frame_ref - 1) != REF_F32)
|
|
|| (type == VALUE_TYPE_I64
|
|
&& (*(frame_ref - 2) != REF_I64_1 || *(frame_ref - 1) != REF_I64_2))
|
|
|| (type == VALUE_TYPE_F64
|
|
&& (*(frame_ref - 2) != REF_F64_1
|
|
|| *(frame_ref - 1) != REF_F64_2))));
|
|
return true;
|
|
}
|
|
|
|
static bool
|
|
check_stack_pop(WASMLoaderContext *ctx, uint8 type, char *error_buf,
|
|
uint32 error_buf_size)
|
|
{
|
|
int32 block_stack_cell_num =
|
|
(int32)(ctx->stack_cell_num - (ctx->frame_csp - 1)->stack_cell_num);
|
|
|
|
if (block_stack_cell_num > 0 && *(ctx->frame_ref - 1) == VALUE_TYPE_ANY) {
|
|
/* the stack top is a value of any type, return success */
|
|
return true;
|
|
}
|
|
|
|
if (!check_stack_top_values(ctx->frame_ref, block_stack_cell_num, type,
|
|
error_buf, error_buf_size))
|
|
return false;
|
|
|
|
return true;
|
|
}
|
|
|
|
static void
|
|
wasm_loader_ctx_destroy(WASMLoaderContext *ctx)
|
|
{
|
|
if (ctx) {
|
|
if (ctx->frame_ref_bottom)
|
|
wasm_runtime_free(ctx->frame_ref_bottom);
|
|
if (ctx->frame_csp_bottom) {
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
free_all_label_patch_lists(ctx->frame_csp_bottom, ctx->csp_num);
|
|
#endif
|
|
wasm_runtime_free(ctx->frame_csp_bottom);
|
|
}
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
if (ctx->frame_offset_bottom)
|
|
wasm_runtime_free(ctx->frame_offset_bottom);
|
|
if (ctx->const_buf)
|
|
wasm_runtime_free(ctx->const_buf);
|
|
#endif
|
|
wasm_runtime_free(ctx);
|
|
}
|
|
}
|
|
|
|
static WASMLoaderContext *
|
|
wasm_loader_ctx_init(WASMFunction *func, char *error_buf, uint32 error_buf_size)
|
|
{
|
|
WASMLoaderContext *loader_ctx =
|
|
loader_malloc(sizeof(WASMLoaderContext), error_buf, error_buf_size);
|
|
if (!loader_ctx)
|
|
return NULL;
|
|
|
|
loader_ctx->frame_ref_size = 32;
|
|
if (!(loader_ctx->frame_ref_bottom = loader_ctx->frame_ref = loader_malloc(
|
|
loader_ctx->frame_ref_size, error_buf, error_buf_size)))
|
|
goto fail;
|
|
loader_ctx->frame_ref_boundary = loader_ctx->frame_ref_bottom + 32;
|
|
|
|
loader_ctx->frame_csp_size = sizeof(BranchBlock) * 8;
|
|
if (!(loader_ctx->frame_csp_bottom = loader_ctx->frame_csp = loader_malloc(
|
|
loader_ctx->frame_csp_size, error_buf, error_buf_size)))
|
|
goto fail;
|
|
loader_ctx->frame_csp_boundary = loader_ctx->frame_csp_bottom + 8;
|
|
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
loader_ctx->frame_offset_size = sizeof(int16) * 32;
|
|
if (!(loader_ctx->frame_offset_bottom = loader_ctx->frame_offset =
|
|
loader_malloc(loader_ctx->frame_offset_size, error_buf,
|
|
error_buf_size)))
|
|
goto fail;
|
|
loader_ctx->frame_offset_boundary = loader_ctx->frame_offset_bottom + 32;
|
|
|
|
loader_ctx->num_const = 0;
|
|
loader_ctx->const_buf_size = sizeof(Const) * 8;
|
|
if (!(loader_ctx->const_buf = loader_malloc(loader_ctx->const_buf_size,
|
|
error_buf, error_buf_size)))
|
|
goto fail;
|
|
|
|
if (func->param_cell_num >= (int32)INT16_MAX - func->local_cell_num) {
|
|
set_error_buf(error_buf, error_buf_size,
|
|
"fast interpreter offset overflow");
|
|
goto fail;
|
|
}
|
|
|
|
loader_ctx->start_dynamic_offset = loader_ctx->dynamic_offset =
|
|
loader_ctx->max_dynamic_offset =
|
|
func->param_cell_num + func->local_cell_num;
|
|
#endif
|
|
return loader_ctx;
|
|
|
|
fail:
|
|
wasm_loader_ctx_destroy(loader_ctx);
|
|
return NULL;
|
|
}
|
|
|
|
static bool
|
|
wasm_loader_push_frame_ref(WASMLoaderContext *ctx, uint8 type, char *error_buf,
|
|
uint32 error_buf_size)
|
|
{
|
|
if (type == VALUE_TYPE_VOID)
|
|
return true;
|
|
|
|
if (!check_stack_push(ctx, error_buf, error_buf_size))
|
|
return false;
|
|
|
|
*ctx->frame_ref++ = type;
|
|
ctx->stack_cell_num++;
|
|
if (ctx->stack_cell_num > ctx->max_stack_cell_num)
|
|
ctx->max_stack_cell_num = ctx->stack_cell_num;
|
|
|
|
if (is_32bit_type(type))
|
|
return true;
|
|
|
|
if (!check_stack_push(ctx, error_buf, error_buf_size))
|
|
return false;
|
|
*ctx->frame_ref++ = type;
|
|
ctx->stack_cell_num++;
|
|
if (ctx->stack_cell_num > ctx->max_stack_cell_num) {
|
|
ctx->max_stack_cell_num = ctx->stack_cell_num;
|
|
bh_assert(ctx->max_stack_cell_num <= UINT16_MAX);
|
|
}
|
|
return true;
|
|
}
|
|
|
|
static bool
|
|
wasm_loader_pop_frame_ref(WASMLoaderContext *ctx, uint8 type, char *error_buf,
|
|
uint32 error_buf_size)
|
|
{
|
|
BranchBlock *cur_block = ctx->frame_csp - 1;
|
|
int32 available_stack_cell =
|
|
(int32)(ctx->stack_cell_num - cur_block->stack_cell_num);
|
|
|
|
/* Directly return success if current block is in stack
|
|
* polymorphic state while stack is empty. */
|
|
if (available_stack_cell <= 0 && cur_block->is_stack_polymorphic)
|
|
return true;
|
|
|
|
if (type == VALUE_TYPE_VOID)
|
|
return true;
|
|
|
|
if (!check_stack_pop(ctx, type, error_buf, error_buf_size))
|
|
return false;
|
|
|
|
ctx->frame_ref--;
|
|
ctx->stack_cell_num--;
|
|
|
|
if (is_32bit_type(type) || *ctx->frame_ref == VALUE_TYPE_ANY)
|
|
return true;
|
|
|
|
ctx->frame_ref--;
|
|
ctx->stack_cell_num--;
|
|
return true;
|
|
}
|
|
|
|
#if WASM_ENABLE_FAST_INTERP == 0
|
|
static bool
|
|
wasm_loader_push_pop_frame_ref(WASMLoaderContext *ctx, uint8 pop_cnt,
|
|
uint8 type_push, uint8 type_pop, char *error_buf,
|
|
uint32 error_buf_size)
|
|
{
|
|
for (int i = 0; i < pop_cnt; i++) {
|
|
if (!wasm_loader_pop_frame_ref(ctx, type_pop, error_buf,
|
|
error_buf_size))
|
|
return false;
|
|
}
|
|
if (!wasm_loader_push_frame_ref(ctx, type_push, error_buf, error_buf_size))
|
|
return false;
|
|
return true;
|
|
}
|
|
#endif
|
|
|
|
static bool
|
|
wasm_loader_push_frame_csp(WASMLoaderContext *ctx, uint8 label_type,
|
|
BlockType block_type, uint8 *start_addr,
|
|
char *error_buf, uint32 error_buf_size)
|
|
{
|
|
CHECK_CSP_PUSH();
|
|
memset(ctx->frame_csp, 0, sizeof(BranchBlock));
|
|
ctx->frame_csp->label_type = label_type;
|
|
ctx->frame_csp->block_type = block_type;
|
|
ctx->frame_csp->start_addr = start_addr;
|
|
ctx->frame_csp->stack_cell_num = ctx->stack_cell_num;
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
ctx->frame_csp->dynamic_offset = ctx->dynamic_offset;
|
|
ctx->frame_csp->patch_list = NULL;
|
|
#endif
|
|
ctx->frame_csp++;
|
|
ctx->csp_num++;
|
|
if (ctx->csp_num > ctx->max_csp_num) {
|
|
ctx->max_csp_num = ctx->csp_num;
|
|
bh_assert(ctx->max_csp_num <= UINT16_MAX);
|
|
}
|
|
return true;
|
|
fail:
|
|
return false;
|
|
}
|
|
|
|
static bool
|
|
wasm_loader_pop_frame_csp(WASMLoaderContext *ctx, char *error_buf,
|
|
uint32 error_buf_size)
|
|
{
|
|
CHECK_CSP_POP();
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
if ((ctx->frame_csp - 1)->param_frame_offsets)
|
|
wasm_runtime_free((ctx->frame_csp - 1)->param_frame_offsets);
|
|
#endif
|
|
ctx->frame_csp--;
|
|
ctx->csp_num--;
|
|
return true;
|
|
}
|
|
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
|
|
#if WASM_ENABLE_LABELS_AS_VALUES != 0
|
|
#if WASM_CPU_SUPPORTS_UNALIGNED_ADDR_ACCESS != 0
|
|
#define emit_label(opcode) \
|
|
do { \
|
|
wasm_loader_emit_ptr(loader_ctx, handle_table[opcode]); \
|
|
LOG_OP("\nemit_op [%02x]\t", opcode); \
|
|
} while (0)
|
|
#define skip_label() \
|
|
do { \
|
|
wasm_loader_emit_backspace(loader_ctx, sizeof(void *)); \
|
|
LOG_OP("\ndelete last op\n"); \
|
|
} while (0)
|
|
#else /* else of WASM_CPU_SUPPORTS_UNALIGNED_ADDR_ACCESS */
|
|
#if UINTPTR_MAX == UINT64_MAX
|
|
#define emit_label(opcode) \
|
|
do { \
|
|
int32 offset = \
|
|
(int32)((uint8 *)handle_table[opcode] - (uint8 *)handle_table[0]); \
|
|
/* emit int32 relative offset in 64-bit target */ \
|
|
wasm_loader_emit_uint32(loader_ctx, offset); \
|
|
LOG_OP("\nemit_op [%02x]\t", opcode); \
|
|
} while (0)
|
|
#else
|
|
#define emit_label(opcode) \
|
|
do { \
|
|
uint32 label_addr = (uint32)(uintptr_t)handle_table[opcode]; \
|
|
/* emit uint32 label address in 32-bit target */ \
|
|
wasm_loader_emit_uint32(loader_ctx, label_addr); \
|
|
LOG_OP("\nemit_op [%02x]\t", opcode); \
|
|
} while (0)
|
|
#endif
|
|
#define skip_label() \
|
|
do { \
|
|
wasm_loader_emit_backspace(loader_ctx, sizeof(int32)); \
|
|
LOG_OP("\ndelete last op\n"); \
|
|
} while (0)
|
|
#endif /* end of WASM_CPU_SUPPORTS_UNALIGNED_ADDR_ACCESS */
|
|
#else /* else of WASM_ENABLE_LABELS_AS_VALUES */
|
|
#define emit_label(opcode) \
|
|
do { \
|
|
wasm_loader_emit_uint8(loader_ctx, opcode); \
|
|
LOG_OP("\nemit_op [%02x]\t", opcode); \
|
|
} while (0)
|
|
#define skip_label() \
|
|
do { \
|
|
wasm_loader_emit_backspace(loader_ctx, sizeof(uint8)); \
|
|
LOG_OP("\ndelete last op\n"); \
|
|
} while (0)
|
|
#endif /* end of WASM_ENABLE_LABELS_AS_VALUES */
|
|
|
|
#define emit_empty_label_addr_and_frame_ip(type) \
|
|
do { \
|
|
if (!add_label_patch_to_list(loader_ctx->frame_csp - 1, type, \
|
|
loader_ctx->p_code_compiled, error_buf, \
|
|
error_buf_size)) \
|
|
goto fail; \
|
|
/* label address, to be patched */ \
|
|
wasm_loader_emit_ptr(loader_ctx, NULL); \
|
|
} while (0)
|
|
|
|
#define emit_br_info(frame_csp, is_br) \
|
|
do { \
|
|
if (!wasm_loader_emit_br_info(loader_ctx, frame_csp, is_br, error_buf, \
|
|
error_buf_size)) \
|
|
goto fail; \
|
|
} while (0)
|
|
|
|
#define LAST_OP_OUTPUT_I32() \
|
|
(last_op >= WASM_OP_I32_EQZ && last_op <= WASM_OP_I32_ROTR) \
|
|
|| (last_op == WASM_OP_I32_LOAD || last_op == WASM_OP_F32_LOAD) \
|
|
|| (last_op >= WASM_OP_I32_LOAD8_S && last_op <= WASM_OP_I32_LOAD16_U) \
|
|
|| (last_op >= WASM_OP_F32_ABS && last_op <= WASM_OP_F32_COPYSIGN) \
|
|
|| (last_op >= WASM_OP_I32_WRAP_I64 \
|
|
&& last_op <= WASM_OP_I32_TRUNC_U_F64) \
|
|
|| (last_op >= WASM_OP_F32_CONVERT_S_I32 \
|
|
&& last_op <= WASM_OP_F32_DEMOTE_F64) \
|
|
|| (last_op == WASM_OP_I32_REINTERPRET_F32) \
|
|
|| (last_op == WASM_OP_F32_REINTERPRET_I32) \
|
|
|| (last_op == EXT_OP_COPY_STACK_TOP)
|
|
|
|
#define LAST_OP_OUTPUT_I64() \
|
|
(last_op >= WASM_OP_I64_CLZ && last_op <= WASM_OP_I64_ROTR) \
|
|
|| (last_op >= WASM_OP_F64_ABS && last_op <= WASM_OP_F64_COPYSIGN) \
|
|
|| (last_op == WASM_OP_I64_LOAD || last_op == WASM_OP_F64_LOAD) \
|
|
|| (last_op >= WASM_OP_I64_LOAD8_S && last_op <= WASM_OP_I64_LOAD32_U) \
|
|
|| (last_op >= WASM_OP_I64_EXTEND_S_I32 \
|
|
&& last_op <= WASM_OP_I64_TRUNC_U_F64) \
|
|
|| (last_op >= WASM_OP_F64_CONVERT_S_I32 \
|
|
&& last_op <= WASM_OP_F64_PROMOTE_F32) \
|
|
|| (last_op == WASM_OP_I64_REINTERPRET_F64) \
|
|
|| (last_op == WASM_OP_F64_REINTERPRET_I64) \
|
|
|| (last_op == EXT_OP_COPY_STACK_TOP_I64)
|
|
|
|
#define GET_CONST_OFFSET(type, val) \
|
|
do { \
|
|
if (!(wasm_loader_get_const_offset(loader_ctx, type, &val, \
|
|
&operand_offset, error_buf, \
|
|
error_buf_size))) \
|
|
goto fail; \
|
|
} while (0)
|
|
|
|
#define GET_CONST_F32_OFFSET(type, fval) \
|
|
do { \
|
|
if (!(wasm_loader_get_const_offset(loader_ctx, type, &fval, \
|
|
&operand_offset, error_buf, \
|
|
error_buf_size))) \
|
|
goto fail; \
|
|
} while (0)
|
|
|
|
#define GET_CONST_F64_OFFSET(type, fval) \
|
|
do { \
|
|
if (!(wasm_loader_get_const_offset(loader_ctx, type, &fval, \
|
|
&operand_offset, error_buf, \
|
|
error_buf_size))) \
|
|
goto fail; \
|
|
} while (0)
|
|
|
|
#define emit_operand(ctx, offset) \
|
|
do { \
|
|
wasm_loader_emit_int16(ctx, offset); \
|
|
LOG_OP("%d\t", offset); \
|
|
} while (0)
|
|
|
|
#define emit_byte(ctx, byte) \
|
|
do { \
|
|
wasm_loader_emit_uint8(ctx, byte); \
|
|
LOG_OP("%d\t", byte); \
|
|
} while (0)
|
|
|
|
#define emit_uint32(ctx, value) \
|
|
do { \
|
|
wasm_loader_emit_uint32(ctx, value); \
|
|
LOG_OP("%d\t", value); \
|
|
} while (0)
|
|
|
|
#define emit_uint64(ctx, value) \
|
|
do { \
|
|
wasm_loader_emit_const(ctx, &value, false); \
|
|
LOG_OP("%lld\t", value); \
|
|
} while (0)
|
|
|
|
#define emit_float32(ctx, value) \
|
|
do { \
|
|
wasm_loader_emit_const(ctx, &value, true); \
|
|
LOG_OP("%f\t", value); \
|
|
} while (0)
|
|
|
|
#define emit_float64(ctx, value) \
|
|
do { \
|
|
wasm_loader_emit_const(ctx, &value, false); \
|
|
LOG_OP("%f\t", value); \
|
|
} while (0)
|
|
|
|
static bool
|
|
wasm_loader_ctx_reinit(WASMLoaderContext *ctx)
|
|
{
|
|
if (!(ctx->p_code_compiled =
|
|
loader_malloc(ctx->code_compiled_peak_size, NULL, 0)))
|
|
return false;
|
|
ctx->p_code_compiled_end =
|
|
ctx->p_code_compiled + ctx->code_compiled_peak_size;
|
|
|
|
/* clean up frame ref */
|
|
memset(ctx->frame_ref_bottom, 0, ctx->frame_ref_size);
|
|
ctx->frame_ref = ctx->frame_ref_bottom;
|
|
ctx->stack_cell_num = 0;
|
|
|
|
/* clean up frame csp */
|
|
memset(ctx->frame_csp_bottom, 0, ctx->frame_csp_size);
|
|
ctx->frame_csp = ctx->frame_csp_bottom;
|
|
ctx->csp_num = 0;
|
|
ctx->max_csp_num = 0;
|
|
|
|
/* clean up frame offset */
|
|
memset(ctx->frame_offset_bottom, 0, ctx->frame_offset_size);
|
|
ctx->frame_offset = ctx->frame_offset_bottom;
|
|
ctx->dynamic_offset = ctx->start_dynamic_offset;
|
|
|
|
/* init preserved local offsets */
|
|
ctx->preserved_local_offset = ctx->max_dynamic_offset;
|
|
|
|
/* const buf is reserved */
|
|
return true;
|
|
}
|
|
|
|
static void
|
|
increase_compiled_code_space(WASMLoaderContext *ctx, int32 size)
|
|
{
|
|
ctx->code_compiled_size += size;
|
|
if (ctx->code_compiled_size >= ctx->code_compiled_peak_size) {
|
|
ctx->code_compiled_peak_size = ctx->code_compiled_size;
|
|
}
|
|
}
|
|
|
|
static void
|
|
wasm_loader_emit_const(WASMLoaderContext *ctx, void *value, bool is_32_bit)
|
|
{
|
|
uint32 size = is_32_bit ? sizeof(uint32) : sizeof(uint64);
|
|
|
|
if (ctx->p_code_compiled) {
|
|
#if WASM_CPU_SUPPORTS_UNALIGNED_ADDR_ACCESS == 0
|
|
bh_assert(((uintptr_t)ctx->p_code_compiled & 1) == 0);
|
|
#endif
|
|
bh_memcpy_s(ctx->p_code_compiled,
|
|
(uint32)(ctx->p_code_compiled_end - ctx->p_code_compiled),
|
|
value, size);
|
|
ctx->p_code_compiled += size;
|
|
}
|
|
else {
|
|
#if WASM_CPU_SUPPORTS_UNALIGNED_ADDR_ACCESS == 0
|
|
bh_assert((ctx->code_compiled_size & 1) == 0);
|
|
#endif
|
|
increase_compiled_code_space(ctx, size);
|
|
}
|
|
}
|
|
|
|
static void
|
|
wasm_loader_emit_uint32(WASMLoaderContext *ctx, uint32 value)
|
|
{
|
|
if (ctx->p_code_compiled) {
|
|
#if WASM_CPU_SUPPORTS_UNALIGNED_ADDR_ACCESS == 0
|
|
bh_assert(((uintptr_t)ctx->p_code_compiled & 1) == 0);
|
|
#endif
|
|
STORE_U32(ctx->p_code_compiled, value);
|
|
ctx->p_code_compiled += sizeof(uint32);
|
|
}
|
|
else {
|
|
#if WASM_CPU_SUPPORTS_UNALIGNED_ADDR_ACCESS == 0
|
|
bh_assert((ctx->code_compiled_size & 1) == 0);
|
|
#endif
|
|
increase_compiled_code_space(ctx, sizeof(uint32));
|
|
}
|
|
}
|
|
|
|
static void
|
|
wasm_loader_emit_int16(WASMLoaderContext *ctx, int16 value)
|
|
{
|
|
if (ctx->p_code_compiled) {
|
|
#if WASM_CPU_SUPPORTS_UNALIGNED_ADDR_ACCESS == 0
|
|
bh_assert(((uintptr_t)ctx->p_code_compiled & 1) == 0);
|
|
#endif
|
|
STORE_U16(ctx->p_code_compiled, (uint16)value);
|
|
ctx->p_code_compiled += sizeof(int16);
|
|
}
|
|
else {
|
|
#if WASM_CPU_SUPPORTS_UNALIGNED_ADDR_ACCESS == 0
|
|
bh_assert((ctx->code_compiled_size & 1) == 0);
|
|
#endif
|
|
increase_compiled_code_space(ctx, sizeof(uint16));
|
|
}
|
|
}
|
|
|
|
static void
|
|
wasm_loader_emit_uint8(WASMLoaderContext *ctx, uint8 value)
|
|
{
|
|
if (ctx->p_code_compiled) {
|
|
*(ctx->p_code_compiled) = value;
|
|
ctx->p_code_compiled += sizeof(uint8);
|
|
#if WASM_CPU_SUPPORTS_UNALIGNED_ADDR_ACCESS == 0
|
|
ctx->p_code_compiled++;
|
|
bh_assert(((uintptr_t)ctx->p_code_compiled & 1) == 0);
|
|
#endif
|
|
}
|
|
else {
|
|
increase_compiled_code_space(ctx, sizeof(uint8));
|
|
#if WASM_CPU_SUPPORTS_UNALIGNED_ADDR_ACCESS == 0
|
|
increase_compiled_code_space(ctx, sizeof(uint8));
|
|
bh_assert((ctx->code_compiled_size & 1) == 0);
|
|
#endif
|
|
}
|
|
}
|
|
|
|
static void
|
|
wasm_loader_emit_ptr(WASMLoaderContext *ctx, void *value)
|
|
{
|
|
if (ctx->p_code_compiled) {
|
|
#if WASM_CPU_SUPPORTS_UNALIGNED_ADDR_ACCESS == 0
|
|
bh_assert(((uintptr_t)ctx->p_code_compiled & 1) == 0);
|
|
#endif
|
|
STORE_PTR(ctx->p_code_compiled, value);
|
|
ctx->p_code_compiled += sizeof(void *);
|
|
}
|
|
else {
|
|
#if WASM_CPU_SUPPORTS_UNALIGNED_ADDR_ACCESS == 0
|
|
bh_assert((ctx->code_compiled_size & 1) == 0);
|
|
#endif
|
|
increase_compiled_code_space(ctx, sizeof(void *));
|
|
}
|
|
}
|
|
|
|
static void
|
|
wasm_loader_emit_backspace(WASMLoaderContext *ctx, uint32 size)
|
|
{
|
|
if (ctx->p_code_compiled) {
|
|
ctx->p_code_compiled -= size;
|
|
#if WASM_CPU_SUPPORTS_UNALIGNED_ADDR_ACCESS == 0
|
|
if (size == sizeof(uint8)) {
|
|
ctx->p_code_compiled--;
|
|
bh_assert(((uintptr_t)ctx->p_code_compiled & 1) == 0);
|
|
}
|
|
#endif
|
|
}
|
|
else {
|
|
ctx->code_compiled_size -= size;
|
|
#if WASM_CPU_SUPPORTS_UNALIGNED_ADDR_ACCESS == 0
|
|
if (size == sizeof(uint8)) {
|
|
ctx->code_compiled_size--;
|
|
bh_assert((ctx->code_compiled_size & 1) == 0);
|
|
}
|
|
#endif
|
|
}
|
|
}
|
|
|
|
static bool
|
|
preserve_referenced_local(WASMLoaderContext *loader_ctx, uint8 opcode,
|
|
uint32 local_index, uint32 local_type,
|
|
bool *preserved, char *error_buf,
|
|
uint32 error_buf_size)
|
|
{
|
|
uint32 i = 0;
|
|
int16 preserved_offset = (int16)local_index;
|
|
|
|
*preserved = false;
|
|
while (i < loader_ctx->stack_cell_num) {
|
|
uint8 cur_type = loader_ctx->frame_ref_bottom[i];
|
|
|
|
/* move previous local into dynamic space before a set/tee_local opcode
|
|
*/
|
|
if (loader_ctx->frame_offset_bottom[i] == (int16)local_index) {
|
|
if (!(*preserved)) {
|
|
*preserved = true;
|
|
skip_label();
|
|
preserved_offset = loader_ctx->preserved_local_offset;
|
|
if (loader_ctx->p_code_compiled) {
|
|
bh_assert(preserved_offset != (int16)local_index);
|
|
}
|
|
if (is_32bit_type(local_type)) {
|
|
/* Only increase preserve offset in the second traversal */
|
|
if (loader_ctx->p_code_compiled)
|
|
loader_ctx->preserved_local_offset++;
|
|
emit_label(EXT_OP_COPY_STACK_TOP);
|
|
}
|
|
else {
|
|
if (loader_ctx->p_code_compiled)
|
|
loader_ctx->preserved_local_offset += 2;
|
|
emit_label(EXT_OP_COPY_STACK_TOP_I64);
|
|
}
|
|
emit_operand(loader_ctx, local_index);
|
|
emit_operand(loader_ctx, preserved_offset);
|
|
emit_label(opcode);
|
|
}
|
|
loader_ctx->frame_offset_bottom[i] = preserved_offset;
|
|
}
|
|
|
|
if (is_32bit_type(cur_type))
|
|
i++;
|
|
else
|
|
i += 2;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
static bool
|
|
preserve_local_for_block(WASMLoaderContext *loader_ctx, uint8 opcode,
|
|
char *error_buf, uint32 error_buf_size)
|
|
{
|
|
uint32 i = 0;
|
|
bool preserve_local;
|
|
|
|
/* preserve locals before blocks to ensure that "tee/set_local" inside
|
|
blocks will not influence the value of these locals */
|
|
while (i < loader_ctx->stack_cell_num) {
|
|
int16 cur_offset = loader_ctx->frame_offset_bottom[i];
|
|
uint8 cur_type = loader_ctx->frame_ref_bottom[i];
|
|
|
|
if ((cur_offset < loader_ctx->start_dynamic_offset)
|
|
&& (cur_offset >= 0)) {
|
|
if (!(preserve_referenced_local(loader_ctx, opcode, cur_offset,
|
|
cur_type, &preserve_local,
|
|
error_buf, error_buf_size)))
|
|
return false;
|
|
}
|
|
|
|
if (is_32bit_type(cur_type == VALUE_TYPE_I32)) {
|
|
i++;
|
|
}
|
|
else {
|
|
i += 2;
|
|
}
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
static bool
|
|
add_label_patch_to_list(BranchBlock *frame_csp, uint8 patch_type,
|
|
uint8 *p_code_compiled, char *error_buf,
|
|
uint32 error_buf_size)
|
|
{
|
|
BranchBlockPatch *patch =
|
|
loader_malloc(sizeof(BranchBlockPatch), error_buf, error_buf_size);
|
|
if (!patch) {
|
|
return false;
|
|
}
|
|
patch->patch_type = patch_type;
|
|
patch->code_compiled = p_code_compiled;
|
|
if (!frame_csp->patch_list) {
|
|
frame_csp->patch_list = patch;
|
|
patch->next = NULL;
|
|
}
|
|
else {
|
|
patch->next = frame_csp->patch_list;
|
|
frame_csp->patch_list = patch;
|
|
}
|
|
return true;
|
|
}
|
|
|
|
static void
|
|
apply_label_patch(WASMLoaderContext *ctx, uint8 depth, uint8 patch_type)
|
|
{
|
|
BranchBlock *frame_csp = ctx->frame_csp - depth;
|
|
BranchBlockPatch *node = frame_csp->patch_list;
|
|
BranchBlockPatch *node_prev = NULL, *node_next;
|
|
|
|
if (!ctx->p_code_compiled)
|
|
return;
|
|
|
|
while (node) {
|
|
node_next = node->next;
|
|
if (node->patch_type == patch_type) {
|
|
STORE_PTR(node->code_compiled, ctx->p_code_compiled);
|
|
if (node_prev == NULL) {
|
|
frame_csp->patch_list = node_next;
|
|
}
|
|
else {
|
|
node_prev->next = node_next;
|
|
}
|
|
wasm_runtime_free(node);
|
|
}
|
|
else {
|
|
node_prev = node;
|
|
}
|
|
node = node_next;
|
|
}
|
|
}
|
|
|
|
static bool
|
|
wasm_loader_emit_br_info(WASMLoaderContext *ctx, BranchBlock *frame_csp,
|
|
bool is_br, char *error_buf, uint32 error_buf_size)
|
|
{
|
|
/* br info layout:
|
|
* a) arity of target block
|
|
* b) total cell num of arity values
|
|
* c) each arity value's cell num
|
|
* d) each arity value's src frame offset
|
|
* e) each arity values's dst dynamic offset
|
|
* f) branch target address
|
|
*
|
|
* Note: b-e are omitted when arity is 0 so that
|
|
* interpreter can recover the br info quickly.
|
|
*/
|
|
BlockType *block_type = &frame_csp->block_type;
|
|
uint8 *types = NULL, cell;
|
|
uint32 arity = 0;
|
|
int32 i;
|
|
int16 *frame_offset = ctx->frame_offset;
|
|
uint16 dynamic_offset;
|
|
|
|
/* Note: loop's arity is different from if and block. loop's arity is
|
|
* its parameter count while if and block arity is result count.
|
|
*/
|
|
if (frame_csp->label_type == LABEL_TYPE_LOOP)
|
|
arity = block_type_get_param_types(block_type, &types);
|
|
else
|
|
arity = block_type_get_result_types(block_type, &types);
|
|
|
|
/* Part a */
|
|
emit_uint32(ctx, arity);
|
|
|
|
if (arity) {
|
|
/* Part b */
|
|
emit_uint32(ctx, wasm_get_cell_num(types, arity));
|
|
|
|
/* Part c */
|
|
for (i = (int32)arity - 1; i >= 0; i--) {
|
|
cell = (uint8)wasm_value_type_cell_num(types[i]);
|
|
emit_byte(ctx, cell);
|
|
}
|
|
/* Part d */
|
|
for (i = (int32)arity - 1; i >= 0; i--) {
|
|
cell = (uint8)wasm_value_type_cell_num(types[i]);
|
|
frame_offset -= cell;
|
|
emit_operand(ctx, *(int16 *)(frame_offset));
|
|
}
|
|
/* Part e */
|
|
dynamic_offset =
|
|
frame_csp->dynamic_offset + wasm_get_cell_num(types, arity);
|
|
if (is_br)
|
|
ctx->dynamic_offset = dynamic_offset;
|
|
for (i = (int32)arity - 1; i >= 0; i--) {
|
|
cell = (uint8)wasm_value_type_cell_num(types[i]);
|
|
dynamic_offset -= cell;
|
|
emit_operand(ctx, dynamic_offset);
|
|
}
|
|
}
|
|
|
|
/* Part f */
|
|
if (frame_csp->label_type == LABEL_TYPE_LOOP) {
|
|
wasm_loader_emit_ptr(ctx, frame_csp->code_compiled);
|
|
}
|
|
else {
|
|
if (!add_label_patch_to_list(frame_csp, PATCH_END, ctx->p_code_compiled,
|
|
error_buf, error_buf_size))
|
|
return false;
|
|
/* label address, to be patched */
|
|
wasm_loader_emit_ptr(ctx, NULL);
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
static bool
|
|
wasm_loader_push_frame_offset(WASMLoaderContext *ctx, uint8 type,
|
|
bool disable_emit, int16 operand_offset,
|
|
char *error_buf, uint32 error_buf_size)
|
|
{
|
|
if (type == VALUE_TYPE_VOID)
|
|
return true;
|
|
|
|
/* only check memory overflow in first traverse */
|
|
if (ctx->p_code_compiled == NULL) {
|
|
if (!check_offset_push(ctx, error_buf, error_buf_size))
|
|
return false;
|
|
}
|
|
|
|
if (disable_emit)
|
|
*(ctx->frame_offset)++ = operand_offset;
|
|
else {
|
|
emit_operand(ctx, ctx->dynamic_offset);
|
|
*(ctx->frame_offset)++ = ctx->dynamic_offset;
|
|
ctx->dynamic_offset++;
|
|
if (ctx->dynamic_offset > ctx->max_dynamic_offset) {
|
|
ctx->max_dynamic_offset = ctx->dynamic_offset;
|
|
bh_assert(ctx->max_dynamic_offset < INT16_MAX);
|
|
}
|
|
}
|
|
|
|
if (is_32bit_type(type))
|
|
return true;
|
|
|
|
if (ctx->p_code_compiled == NULL) {
|
|
if (!check_offset_push(ctx, error_buf, error_buf_size))
|
|
return false;
|
|
}
|
|
|
|
ctx->frame_offset++;
|
|
if (!disable_emit) {
|
|
ctx->dynamic_offset++;
|
|
if (ctx->dynamic_offset > ctx->max_dynamic_offset) {
|
|
ctx->max_dynamic_offset = ctx->dynamic_offset;
|
|
bh_assert(ctx->max_dynamic_offset < INT16_MAX);
|
|
}
|
|
}
|
|
return true;
|
|
}
|
|
|
|
/* This function should be in front of wasm_loader_pop_frame_ref
|
|
as they both use ctx->stack_cell_num, and ctx->stack_cell_num
|
|
will be modified by wasm_loader_pop_frame_ref */
|
|
static bool
|
|
wasm_loader_pop_frame_offset(WASMLoaderContext *ctx, uint8 type,
|
|
char *error_buf, uint32 error_buf_size)
|
|
{
|
|
/* if ctx->frame_csp equals ctx->frame_csp_bottom,
|
|
then current block is the function block */
|
|
uint32 depth = ctx->frame_csp > ctx->frame_csp_bottom ? 1 : 0;
|
|
BranchBlock *cur_block = ctx->frame_csp - depth;
|
|
int32 available_stack_cell =
|
|
(int32)(ctx->stack_cell_num - cur_block->stack_cell_num);
|
|
|
|
/* Directly return success if current block is in stack
|
|
* polymorphic state while stack is empty. */
|
|
if (available_stack_cell <= 0 && cur_block->is_stack_polymorphic)
|
|
return true;
|
|
|
|
if (type == VALUE_TYPE_VOID)
|
|
return true;
|
|
|
|
if (is_32bit_type(type)) {
|
|
/* Check the offset stack bottom to ensure the frame offset
|
|
stack will not go underflow. But we don't thrown error
|
|
and return true here, because the error msg should be
|
|
given in wasm_loader_pop_frame_ref */
|
|
if (!check_offset_pop(ctx, 1))
|
|
return true;
|
|
|
|
ctx->frame_offset -= 1;
|
|
if ((*(ctx->frame_offset) > ctx->start_dynamic_offset)
|
|
&& (*(ctx->frame_offset) < ctx->max_dynamic_offset))
|
|
ctx->dynamic_offset -= 1;
|
|
}
|
|
else {
|
|
if (!check_offset_pop(ctx, 2))
|
|
return true;
|
|
|
|
ctx->frame_offset -= 2;
|
|
if ((*(ctx->frame_offset) > ctx->start_dynamic_offset)
|
|
&& (*(ctx->frame_offset) < ctx->max_dynamic_offset))
|
|
ctx->dynamic_offset -= 2;
|
|
}
|
|
emit_operand(ctx, *(ctx->frame_offset));
|
|
return true;
|
|
}
|
|
|
|
static bool
|
|
wasm_loader_push_frame_ref_offset(WASMLoaderContext *ctx, uint8 type,
|
|
bool disable_emit, int16 operand_offset,
|
|
char *error_buf, uint32 error_buf_size)
|
|
{
|
|
if (!(wasm_loader_push_frame_offset(ctx, type, disable_emit, operand_offset,
|
|
error_buf, error_buf_size)))
|
|
return false;
|
|
if (!(wasm_loader_push_frame_ref(ctx, type, error_buf, error_buf_size)))
|
|
return false;
|
|
|
|
return true;
|
|
}
|
|
|
|
static bool
|
|
wasm_loader_pop_frame_ref_offset(WASMLoaderContext *ctx, uint8 type,
|
|
char *error_buf, uint32 error_buf_size)
|
|
{
|
|
/* put wasm_loader_pop_frame_offset in front of wasm_loader_pop_frame_ref */
|
|
if (!wasm_loader_pop_frame_offset(ctx, type, error_buf, error_buf_size))
|
|
return false;
|
|
if (!wasm_loader_pop_frame_ref(ctx, type, error_buf, error_buf_size))
|
|
return false;
|
|
|
|
return true;
|
|
}
|
|
|
|
static bool
|
|
wasm_loader_push_pop_frame_ref_offset(WASMLoaderContext *ctx, uint8 pop_cnt,
|
|
uint8 type_push, uint8 type_pop,
|
|
bool disable_emit, int16 operand_offset,
|
|
char *error_buf, uint32 error_buf_size)
|
|
{
|
|
uint8 i;
|
|
|
|
for (i = 0; i < pop_cnt; i++) {
|
|
if (!wasm_loader_pop_frame_offset(ctx, type_pop, error_buf,
|
|
error_buf_size))
|
|
return false;
|
|
|
|
if (!wasm_loader_pop_frame_ref(ctx, type_pop, error_buf,
|
|
error_buf_size))
|
|
return false;
|
|
}
|
|
|
|
if (!wasm_loader_push_frame_offset(ctx, type_push, disable_emit,
|
|
operand_offset, error_buf,
|
|
error_buf_size))
|
|
return false;
|
|
|
|
if (!wasm_loader_push_frame_ref(ctx, type_push, error_buf, error_buf_size))
|
|
return false;
|
|
|
|
return true;
|
|
}
|
|
|
|
static bool
|
|
wasm_loader_get_const_offset(WASMLoaderContext *ctx, uint8 type, void *value,
|
|
int16 *offset, char *error_buf,
|
|
uint32 error_buf_size)
|
|
{
|
|
int8 bytes_to_increase;
|
|
int16 operand_offset = 0;
|
|
Const *c;
|
|
|
|
/* Search existing constant */
|
|
for (c = (Const *)ctx->const_buf;
|
|
(uint8 *)c < ctx->const_buf + ctx->num_const * sizeof(Const); c++) {
|
|
if ((type == c->value_type)
|
|
&& ((type == VALUE_TYPE_I64 && *(int64 *)value == c->value.i64)
|
|
|| (type == VALUE_TYPE_I32 && *(int32 *)value == c->value.i32)
|
|
#if WASM_ENABLE_REF_TYPES != 0
|
|
|| (type == VALUE_TYPE_FUNCREF
|
|
&& *(int32 *)value == c->value.i32)
|
|
|| (type == VALUE_TYPE_EXTERNREF
|
|
&& *(int32 *)value == c->value.i32)
|
|
#endif
|
|
|| (type == VALUE_TYPE_F64
|
|
&& (0 == memcmp(value, &(c->value.f64), sizeof(float64))))
|
|
|| (type == VALUE_TYPE_F32
|
|
&& (0
|
|
== memcmp(value, &(c->value.f32), sizeof(float32)))))) {
|
|
operand_offset = c->slot_index;
|
|
break;
|
|
}
|
|
if (c->value_type == VALUE_TYPE_I64 || c->value_type == VALUE_TYPE_F64)
|
|
operand_offset += 2;
|
|
else
|
|
operand_offset += 1;
|
|
}
|
|
|
|
if ((uint8 *)c == ctx->const_buf + ctx->num_const * sizeof(Const)) {
|
|
/* New constant, append to the const buffer */
|
|
if ((type == VALUE_TYPE_F64) || (type == VALUE_TYPE_I64)) {
|
|
bytes_to_increase = 2;
|
|
}
|
|
else {
|
|
bytes_to_increase = 1;
|
|
}
|
|
|
|
/* The max cell num of const buffer is 32768 since the valid index range
|
|
* is -32768 ~ -1. Return an invalid index 0 to indicate the buffer is
|
|
* full */
|
|
if (ctx->const_cell_num > INT16_MAX - bytes_to_increase + 1) {
|
|
*offset = 0;
|
|
return true;
|
|
}
|
|
|
|
if ((uint8 *)c == ctx->const_buf + ctx->const_buf_size) {
|
|
MEM_REALLOC(ctx->const_buf, ctx->const_buf_size,
|
|
ctx->const_buf_size + 4 * sizeof(Const));
|
|
ctx->const_buf_size += 4 * sizeof(Const);
|
|
c = (Const *)(ctx->const_buf + ctx->num_const * sizeof(Const));
|
|
}
|
|
c->value_type = type;
|
|
switch (type) {
|
|
case VALUE_TYPE_F64:
|
|
bh_memcpy_s(&(c->value.f64), sizeof(WASMValue), value,
|
|
sizeof(float64));
|
|
ctx->const_cell_num += 2;
|
|
/* The const buf will be reversed, we use the second cell */
|
|
/* of the i64/f64 const so the finnal offset is corrent */
|
|
operand_offset++;
|
|
break;
|
|
case VALUE_TYPE_I64:
|
|
c->value.i64 = *(int64 *)value;
|
|
ctx->const_cell_num += 2;
|
|
operand_offset++;
|
|
break;
|
|
case VALUE_TYPE_F32:
|
|
bh_memcpy_s(&(c->value.f32), sizeof(WASMValue), value,
|
|
sizeof(float32));
|
|
ctx->const_cell_num++;
|
|
break;
|
|
case VALUE_TYPE_I32:
|
|
c->value.i32 = *(int32 *)value;
|
|
ctx->const_cell_num++;
|
|
break;
|
|
#if WASM_ENABLE_REF_TYPES != 0
|
|
case VALUE_TYPE_EXTERNREF:
|
|
case VALUE_TYPE_FUNCREF:
|
|
c->value.i32 = *(int32 *)value;
|
|
ctx->const_cell_num++;
|
|
break;
|
|
#endif
|
|
default:
|
|
break;
|
|
}
|
|
c->slot_index = operand_offset;
|
|
ctx->num_const++;
|
|
LOG_OP("#### new const [%d]: %ld\n", ctx->num_const,
|
|
(int64)c->value.i64);
|
|
}
|
|
/* use negetive index for const */
|
|
operand_offset = -(operand_offset + 1);
|
|
*offset = operand_offset;
|
|
return true;
|
|
fail:
|
|
return false;
|
|
}
|
|
|
|
/*
|
|
PUSH(POP)_XXX = push(pop) frame_ref + push(pop) frame_offset
|
|
-- Mostly used for the binary / compare operation
|
|
PUSH(POP)_OFFSET_TYPE only push(pop) the frame_offset stack
|
|
-- Mostly used in block / control instructions
|
|
|
|
The POP will always emit the offset on the top of the frame_offset stack
|
|
PUSH can be used in two ways:
|
|
1. directly PUSH:
|
|
PUSH_XXX();
|
|
will allocate a dynamic space and emit
|
|
2. silent PUSH:
|
|
operand_offset = xxx; disable_emit = true;
|
|
PUSH_XXX();
|
|
only push the frame_offset stack, no emit
|
|
*/
|
|
#define PUSH_I32() \
|
|
do { \
|
|
if (!wasm_loader_push_frame_ref_offset(loader_ctx, VALUE_TYPE_I32, \
|
|
disable_emit, operand_offset, \
|
|
error_buf, error_buf_size)) \
|
|
goto fail; \
|
|
} while (0)
|
|
|
|
#define PUSH_F32() \
|
|
do { \
|
|
if (!wasm_loader_push_frame_ref_offset(loader_ctx, VALUE_TYPE_F32, \
|
|
disable_emit, operand_offset, \
|
|
error_buf, error_buf_size)) \
|
|
goto fail; \
|
|
} while (0)
|
|
|
|
#define PUSH_I64() \
|
|
do { \
|
|
if (!wasm_loader_push_frame_ref_offset(loader_ctx, VALUE_TYPE_I64, \
|
|
disable_emit, operand_offset, \
|
|
error_buf, error_buf_size)) \
|
|
goto fail; \
|
|
} while (0)
|
|
|
|
#define PUSH_F64() \
|
|
do { \
|
|
if (!wasm_loader_push_frame_ref_offset(loader_ctx, VALUE_TYPE_F64, \
|
|
disable_emit, operand_offset, \
|
|
error_buf, error_buf_size)) \
|
|
goto fail; \
|
|
} while (0)
|
|
|
|
#define PUSH_FUNCREF() \
|
|
do { \
|
|
if (!wasm_loader_push_frame_ref_offset(loader_ctx, VALUE_TYPE_FUNCREF, \
|
|
disable_emit, operand_offset, \
|
|
error_buf, error_buf_size)) \
|
|
goto fail; \
|
|
} while (0)
|
|
|
|
#define POP_I32() \
|
|
do { \
|
|
if (!wasm_loader_pop_frame_ref_offset(loader_ctx, VALUE_TYPE_I32, \
|
|
error_buf, error_buf_size)) \
|
|
goto fail; \
|
|
} while (0)
|
|
|
|
#define POP_F32() \
|
|
do { \
|
|
if (!wasm_loader_pop_frame_ref_offset(loader_ctx, VALUE_TYPE_F32, \
|
|
error_buf, error_buf_size)) \
|
|
goto fail; \
|
|
} while (0)
|
|
|
|
#define POP_I64() \
|
|
do { \
|
|
if (!wasm_loader_pop_frame_ref_offset(loader_ctx, VALUE_TYPE_I64, \
|
|
error_buf, error_buf_size)) \
|
|
goto fail; \
|
|
} while (0)
|
|
|
|
#define POP_F64() \
|
|
do { \
|
|
if (!wasm_loader_pop_frame_ref_offset(loader_ctx, VALUE_TYPE_F64, \
|
|
error_buf, error_buf_size)) \
|
|
goto fail; \
|
|
} while (0)
|
|
|
|
#define PUSH_OFFSET_TYPE(type) \
|
|
do { \
|
|
if (!(wasm_loader_push_frame_offset(loader_ctx, type, disable_emit, \
|
|
operand_offset, error_buf, \
|
|
error_buf_size))) \
|
|
goto fail; \
|
|
} while (0)
|
|
|
|
#define POP_OFFSET_TYPE(type) \
|
|
do { \
|
|
if (!(wasm_loader_pop_frame_offset(loader_ctx, type, error_buf, \
|
|
error_buf_size))) \
|
|
goto fail; \
|
|
} while (0)
|
|
|
|
#define PUSH_MEM_OFFSET() \
|
|
do { \
|
|
if (!wasm_loader_push_frame_ref_offset(loader_ctx, mem_offset_type, \
|
|
disable_emit, operand_offset, \
|
|
error_buf, error_buf_size)) \
|
|
goto fail; \
|
|
} while (0)
|
|
#define PUSH_PAGE_COUNT() PUSH_MEM_OFFSET()
|
|
|
|
#define POP_MEM_OFFSET() \
|
|
do { \
|
|
if (!wasm_loader_pop_frame_ref_offset(loader_ctx, mem_offset_type, \
|
|
error_buf, error_buf_size)) \
|
|
goto fail; \
|
|
} while (0)
|
|
|
|
#define POP_AND_PUSH(type_pop, type_push) \
|
|
do { \
|
|
if (!(wasm_loader_push_pop_frame_ref_offset( \
|
|
loader_ctx, 1, type_push, type_pop, disable_emit, \
|
|
operand_offset, error_buf, error_buf_size))) \
|
|
goto fail; \
|
|
} while (0)
|
|
|
|
/* type of POPs should be the same */
|
|
#define POP2_AND_PUSH(type_pop, type_push) \
|
|
do { \
|
|
if (!(wasm_loader_push_pop_frame_ref_offset( \
|
|
loader_ctx, 2, type_push, type_pop, disable_emit, \
|
|
operand_offset, error_buf, error_buf_size))) \
|
|
goto fail; \
|
|
} while (0)
|
|
|
|
#else /* WASM_ENABLE_FAST_INTERP */
|
|
|
|
#define PUSH_I32() \
|
|
do { \
|
|
if (!(wasm_loader_push_frame_ref(loader_ctx, VALUE_TYPE_I32, \
|
|
error_buf, error_buf_size))) \
|
|
goto fail; \
|
|
} while (0)
|
|
|
|
#define PUSH_F32() \
|
|
do { \
|
|
if (!(wasm_loader_push_frame_ref(loader_ctx, VALUE_TYPE_F32, \
|
|
error_buf, error_buf_size))) \
|
|
goto fail; \
|
|
} while (0)
|
|
|
|
#define PUSH_I64() \
|
|
do { \
|
|
if (!(wasm_loader_push_frame_ref(loader_ctx, VALUE_TYPE_I64, \
|
|
error_buf, error_buf_size))) \
|
|
goto fail; \
|
|
} while (0)
|
|
|
|
#define PUSH_F64() \
|
|
do { \
|
|
if (!(wasm_loader_push_frame_ref(loader_ctx, VALUE_TYPE_F64, \
|
|
error_buf, error_buf_size))) \
|
|
goto fail; \
|
|
} while (0)
|
|
|
|
#define PUSH_FUNCREF() \
|
|
do { \
|
|
if (!(wasm_loader_push_frame_ref(loader_ctx, VALUE_TYPE_FUNCREF, \
|
|
error_buf, error_buf_size))) \
|
|
goto fail; \
|
|
} while (0)
|
|
|
|
#define PUSH_MEM_OFFSET() \
|
|
do { \
|
|
if (!(wasm_loader_push_frame_ref(loader_ctx, mem_offset_type, \
|
|
error_buf, error_buf_size))) \
|
|
goto fail; \
|
|
} while (0)
|
|
|
|
#define PUSH_PAGE_COUNT() PUSH_MEM_OFFSET()
|
|
|
|
#define POP_I32() \
|
|
do { \
|
|
if (!(wasm_loader_pop_frame_ref(loader_ctx, VALUE_TYPE_I32, error_buf, \
|
|
error_buf_size))) \
|
|
goto fail; \
|
|
} while (0)
|
|
|
|
#define POP_F32() \
|
|
do { \
|
|
if (!(wasm_loader_pop_frame_ref(loader_ctx, VALUE_TYPE_F32, error_buf, \
|
|
error_buf_size))) \
|
|
goto fail; \
|
|
} while (0)
|
|
|
|
#define POP_I64() \
|
|
do { \
|
|
if (!(wasm_loader_pop_frame_ref(loader_ctx, VALUE_TYPE_I64, error_buf, \
|
|
error_buf_size))) \
|
|
goto fail; \
|
|
} while (0)
|
|
|
|
#define POP_F64() \
|
|
do { \
|
|
if (!(wasm_loader_pop_frame_ref(loader_ctx, VALUE_TYPE_F64, error_buf, \
|
|
error_buf_size))) \
|
|
goto fail; \
|
|
} while (0)
|
|
|
|
#define POP_FUNCREF() \
|
|
do { \
|
|
if (!(wasm_loader_pop_frame_ref(loader_ctx, VALUE_TYPE_FUNCREF, \
|
|
error_buf, error_buf_size))) \
|
|
goto fail; \
|
|
} while (0)
|
|
|
|
#define POP_MEM_OFFSET() \
|
|
do { \
|
|
if (!(wasm_loader_pop_frame_ref(loader_ctx, mem_offset_type, \
|
|
error_buf, error_buf_size))) \
|
|
goto fail; \
|
|
} while (0)
|
|
|
|
#define POP_AND_PUSH(type_pop, type_push) \
|
|
do { \
|
|
if (!(wasm_loader_push_pop_frame_ref(loader_ctx, 1, type_push, \
|
|
type_pop, error_buf, \
|
|
error_buf_size))) \
|
|
goto fail; \
|
|
} while (0)
|
|
|
|
/* type of POPs should be the same */
|
|
#define POP2_AND_PUSH(type_pop, type_push) \
|
|
do { \
|
|
if (!(wasm_loader_push_pop_frame_ref(loader_ctx, 2, type_push, \
|
|
type_pop, error_buf, \
|
|
error_buf_size))) \
|
|
goto fail; \
|
|
} while (0)
|
|
#endif /* WASM_ENABLE_FAST_INTERP */
|
|
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
|
|
static bool
|
|
reserve_block_ret(WASMLoaderContext *loader_ctx, uint8 opcode,
|
|
bool disable_emit, char *error_buf, uint32 error_buf_size)
|
|
{
|
|
int16 operand_offset = 0;
|
|
BranchBlock *block = (opcode == WASM_OP_ELSE) ? loader_ctx->frame_csp - 1
|
|
: loader_ctx->frame_csp;
|
|
BlockType *block_type = &block->block_type;
|
|
uint8 *return_types = NULL;
|
|
uint32 return_count = 0, value_count = 0, total_cel_num = 0;
|
|
int32 i = 0;
|
|
int16 dynamic_offset, dynamic_offset_org, *frame_offset = NULL,
|
|
*frame_offset_org = NULL;
|
|
|
|
return_count = block_type_get_result_types(block_type, &return_types);
|
|
|
|
/* If there is only one return value, use EXT_OP_COPY_STACK_TOP/_I64 instead
|
|
* of EXT_OP_COPY_STACK_VALUES for interpreter performance. */
|
|
if (return_count == 1) {
|
|
uint8 cell = (uint8)wasm_value_type_cell_num(return_types[0]);
|
|
if (cell <= 2 /* V128 isn't supported whose cell num is 4 */
|
|
&& block->dynamic_offset != *(loader_ctx->frame_offset - cell)) {
|
|
/* insert op_copy before else opcode */
|
|
if (opcode == WASM_OP_ELSE)
|
|
skip_label();
|
|
emit_label(cell == 1 ? EXT_OP_COPY_STACK_TOP
|
|
: EXT_OP_COPY_STACK_TOP_I64);
|
|
emit_operand(loader_ctx, *(loader_ctx->frame_offset - cell));
|
|
emit_operand(loader_ctx, block->dynamic_offset);
|
|
|
|
if (opcode == WASM_OP_ELSE) {
|
|
*(loader_ctx->frame_offset - cell) = block->dynamic_offset;
|
|
}
|
|
else {
|
|
loader_ctx->frame_offset -= cell;
|
|
loader_ctx->dynamic_offset = block->dynamic_offset;
|
|
PUSH_OFFSET_TYPE(return_types[0]);
|
|
wasm_loader_emit_backspace(loader_ctx, sizeof(int16));
|
|
}
|
|
if (opcode == WASM_OP_ELSE)
|
|
emit_label(opcode);
|
|
}
|
|
return true;
|
|
}
|
|
|
|
/* Copy stack top values to block's results which are in dynamic space.
|
|
* The instruction format:
|
|
* Part a: values count
|
|
* Part b: all values total cell num
|
|
* Part c: each value's cell_num, src offset and dst offset
|
|
* Part d: each value's src offset and dst offset
|
|
* Part e: each value's dst offset
|
|
*/
|
|
frame_offset = frame_offset_org = loader_ctx->frame_offset;
|
|
dynamic_offset = dynamic_offset_org =
|
|
block->dynamic_offset + wasm_get_cell_num(return_types, return_count);
|
|
|
|
/* First traversal to get the count of values needed to be copied. */
|
|
for (i = (int32)return_count - 1; i >= 0; i--) {
|
|
uint8 cells = (uint8)wasm_value_type_cell_num(return_types[i]);
|
|
|
|
frame_offset -= cells;
|
|
dynamic_offset -= cells;
|
|
if (dynamic_offset != *frame_offset) {
|
|
value_count++;
|
|
total_cel_num += cells;
|
|
}
|
|
}
|
|
|
|
if (value_count) {
|
|
uint32 j = 0;
|
|
uint8 *emit_data = NULL, *cells = NULL;
|
|
int16 *src_offsets = NULL;
|
|
uint16 *dst_offsets = NULL;
|
|
uint64 size =
|
|
(uint64)value_count
|
|
* (sizeof(*cells) + sizeof(*src_offsets) + sizeof(*dst_offsets));
|
|
|
|
/* Allocate memory for the emit data */
|
|
if (!(emit_data = loader_malloc(size, error_buf, error_buf_size)))
|
|
return false;
|
|
|
|
cells = emit_data;
|
|
src_offsets = (int16 *)(cells + value_count);
|
|
dst_offsets = (uint16 *)(src_offsets + value_count);
|
|
|
|
/* insert op_copy before else opcode */
|
|
if (opcode == WASM_OP_ELSE)
|
|
skip_label();
|
|
emit_label(EXT_OP_COPY_STACK_VALUES);
|
|
/* Part a) */
|
|
emit_uint32(loader_ctx, value_count);
|
|
/* Part b) */
|
|
emit_uint32(loader_ctx, total_cel_num);
|
|
|
|
/* Second traversal to get each value's cell num, src offset and dst
|
|
* offset. */
|
|
frame_offset = frame_offset_org;
|
|
dynamic_offset = dynamic_offset_org;
|
|
for (i = (int32)return_count - 1, j = 0; i >= 0; i--) {
|
|
uint8 cell = (uint8)wasm_value_type_cell_num(return_types[i]);
|
|
frame_offset -= cell;
|
|
dynamic_offset -= cell;
|
|
if (dynamic_offset != *frame_offset) {
|
|
/* cell num */
|
|
cells[j] = cell;
|
|
/* src offset */
|
|
src_offsets[j] = *frame_offset;
|
|
/* dst offset */
|
|
dst_offsets[j] = dynamic_offset;
|
|
j++;
|
|
}
|
|
if (opcode == WASM_OP_ELSE) {
|
|
*frame_offset = dynamic_offset;
|
|
}
|
|
else {
|
|
loader_ctx->frame_offset = frame_offset;
|
|
loader_ctx->dynamic_offset = dynamic_offset;
|
|
PUSH_OFFSET_TYPE(return_types[i]);
|
|
wasm_loader_emit_backspace(loader_ctx, sizeof(int16));
|
|
loader_ctx->frame_offset = frame_offset_org;
|
|
loader_ctx->dynamic_offset = dynamic_offset_org;
|
|
}
|
|
}
|
|
|
|
bh_assert(j == value_count);
|
|
|
|
/* Emit the cells, src_offsets and dst_offsets */
|
|
for (j = 0; j < value_count; j++)
|
|
emit_byte(loader_ctx, cells[j]);
|
|
for (j = 0; j < value_count; j++)
|
|
emit_operand(loader_ctx, src_offsets[j]);
|
|
for (j = 0; j < value_count; j++)
|
|
emit_operand(loader_ctx, dst_offsets[j]);
|
|
|
|
if (opcode == WASM_OP_ELSE)
|
|
emit_label(opcode);
|
|
|
|
wasm_runtime_free(emit_data);
|
|
}
|
|
|
|
return true;
|
|
|
|
fail:
|
|
return false;
|
|
}
|
|
|
|
#endif /* WASM_ENABLE_FAST_INTERP */
|
|
|
|
#define RESERVE_BLOCK_RET() \
|
|
do { \
|
|
if (!reserve_block_ret(loader_ctx, opcode, disable_emit, error_buf, \
|
|
error_buf_size)) \
|
|
goto fail; \
|
|
} while (0)
|
|
|
|
#define PUSH_TYPE(type) \
|
|
do { \
|
|
if (!(wasm_loader_push_frame_ref(loader_ctx, type, error_buf, \
|
|
error_buf_size))) \
|
|
goto fail; \
|
|
} while (0)
|
|
|
|
#define POP_TYPE(type) \
|
|
do { \
|
|
if (!(wasm_loader_pop_frame_ref(loader_ctx, type, error_buf, \
|
|
error_buf_size))) \
|
|
goto fail; \
|
|
} while (0)
|
|
|
|
#define PUSH_CSP(label_type, block_type, _start_addr) \
|
|
do { \
|
|
if (!wasm_loader_push_frame_csp(loader_ctx, label_type, block_type, \
|
|
_start_addr, error_buf, \
|
|
error_buf_size)) \
|
|
goto fail; \
|
|
} while (0)
|
|
|
|
#define POP_CSP() \
|
|
do { \
|
|
if (!wasm_loader_pop_frame_csp(loader_ctx, error_buf, error_buf_size)) \
|
|
goto fail; \
|
|
} while (0)
|
|
|
|
#define GET_LOCAL_INDEX_TYPE_AND_OFFSET() \
|
|
do { \
|
|
read_leb_uint32(p, p_end, local_idx); \
|
|
bh_assert(local_idx < param_count + local_count); \
|
|
local_type = local_idx < param_count \
|
|
? param_types[local_idx] \
|
|
: local_types[local_idx - param_count]; \
|
|
local_offset = local_offsets[local_idx]; \
|
|
} while (0)
|
|
|
|
#define CHECK_MEMORY() \
|
|
do { \
|
|
bh_assert(module->import_memory_count + module->memory_count > 0); \
|
|
} while (0)
|
|
|
|
static bool
|
|
wasm_loader_check_br(WASMLoaderContext *loader_ctx, uint32 depth, uint8 opcode,
|
|
char *error_buf, uint32 error_buf_size)
|
|
{
|
|
BranchBlock *target_block, *cur_block;
|
|
BlockType *target_block_type;
|
|
uint8 *types = NULL, *frame_ref;
|
|
uint32 arity = 0;
|
|
int32 i, available_stack_cell;
|
|
uint16 cell_num;
|
|
|
|
uint8 *frame_ref_old = loader_ctx->frame_ref;
|
|
uint8 *frame_ref_after_popped = NULL;
|
|
uint8 frame_ref_tmp[4] = { 0 };
|
|
uint8 *frame_ref_buf = frame_ref_tmp;
|
|
uint32 stack_cell_num_old = loader_ctx->stack_cell_num;
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
int16 *frame_offset_old = loader_ctx->frame_offset;
|
|
int16 *frame_offset_after_popped = NULL;
|
|
int16 frame_offset_tmp[4] = { 0 };
|
|
int16 *frame_offset_buf = frame_offset_tmp;
|
|
uint16 dynamic_offset_old = (loader_ctx->frame_csp - 1)->dynamic_offset;
|
|
#endif
|
|
bool ret = false;
|
|
|
|
bh_assert(loader_ctx->csp_num > 0);
|
|
if (loader_ctx->csp_num - 1 < depth) {
|
|
set_error_buf(error_buf, error_buf_size,
|
|
"unknown label, "
|
|
"unexpected end of section or function");
|
|
return false;
|
|
}
|
|
|
|
cur_block = loader_ctx->frame_csp - 1;
|
|
target_block = loader_ctx->frame_csp - (depth + 1);
|
|
target_block_type = &target_block->block_type;
|
|
frame_ref = loader_ctx->frame_ref;
|
|
|
|
/* Note: loop's arity is different from if and block. loop's arity is
|
|
* its parameter count while if and block arity is result count.
|
|
*/
|
|
if (target_block->label_type == LABEL_TYPE_LOOP)
|
|
arity = block_type_get_param_types(target_block_type, &types);
|
|
else
|
|
arity = block_type_get_result_types(target_block_type, &types);
|
|
|
|
/* If the stack is in polymorphic state, just clear the stack
|
|
* and then re-push the values to make the stack top values
|
|
* match block type. */
|
|
if (cur_block->is_stack_polymorphic) {
|
|
for (i = (int32)arity - 1; i >= 0; i--) {
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
POP_OFFSET_TYPE(types[i]);
|
|
#endif
|
|
POP_TYPE(types[i]);
|
|
}
|
|
|
|
/* Backup stack data since it may be changed in the below
|
|
push operations, and the stack data may be used when
|
|
checking other target blocks of opcode br_table */
|
|
if (opcode == WASM_OP_BR_TABLE) {
|
|
uint64 total_size;
|
|
|
|
frame_ref_after_popped = loader_ctx->frame_ref;
|
|
total_size = (uint64)sizeof(uint8)
|
|
* (frame_ref_old - frame_ref_after_popped);
|
|
if (total_size > sizeof(frame_ref_tmp)
|
|
&& !(frame_ref_buf = loader_malloc(total_size, error_buf,
|
|
error_buf_size))) {
|
|
goto fail;
|
|
}
|
|
bh_memcpy_s(frame_ref_buf, (uint32)total_size,
|
|
frame_ref_after_popped, (uint32)total_size);
|
|
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
frame_offset_after_popped = loader_ctx->frame_offset;
|
|
total_size = (uint64)sizeof(int16)
|
|
* (frame_offset_old - frame_offset_after_popped);
|
|
if (total_size > sizeof(frame_offset_tmp)
|
|
&& !(frame_offset_buf = loader_malloc(total_size, error_buf,
|
|
error_buf_size))) {
|
|
goto fail;
|
|
}
|
|
bh_memcpy_s(frame_offset_buf, (uint32)total_size,
|
|
frame_offset_after_popped, (uint32)total_size);
|
|
#endif
|
|
}
|
|
|
|
for (i = 0; i < (int32)arity; i++) {
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
bool disable_emit = true;
|
|
int16 operand_offset = 0;
|
|
PUSH_OFFSET_TYPE(types[i]);
|
|
#endif
|
|
PUSH_TYPE(types[i]);
|
|
}
|
|
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
emit_br_info(target_block, opcode == WASM_OP_BR);
|
|
#endif
|
|
|
|
/* Restore the stack data, note that frame_ref_bottom,
|
|
frame_reftype_map_bottom, frame_offset_bottom may be
|
|
re-allocated in the above push operations */
|
|
if (opcode == WASM_OP_BR_TABLE) {
|
|
uint32 total_size;
|
|
|
|
/* The stack operand num should not be smaller than before
|
|
after pop and push operations */
|
|
bh_assert(loader_ctx->stack_cell_num >= stack_cell_num_old);
|
|
loader_ctx->stack_cell_num = stack_cell_num_old;
|
|
loader_ctx->frame_ref =
|
|
loader_ctx->frame_ref_bottom + stack_cell_num_old;
|
|
total_size = (uint32)sizeof(uint8)
|
|
* (frame_ref_old - frame_ref_after_popped);
|
|
bh_memcpy_s((uint8 *)loader_ctx->frame_ref - total_size, total_size,
|
|
frame_ref_buf, total_size);
|
|
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
loader_ctx->frame_offset =
|
|
loader_ctx->frame_offset_bottom + stack_cell_num_old;
|
|
total_size = (uint32)sizeof(int16)
|
|
* (frame_offset_old - frame_offset_after_popped);
|
|
bh_memcpy_s((uint8 *)loader_ctx->frame_offset - total_size,
|
|
total_size, frame_offset_buf, total_size);
|
|
(loader_ctx->frame_csp - 1)->dynamic_offset = dynamic_offset_old;
|
|
#endif
|
|
}
|
|
|
|
ret = true;
|
|
goto cleanup_and_return;
|
|
}
|
|
|
|
available_stack_cell =
|
|
(int32)(loader_ctx->stack_cell_num - cur_block->stack_cell_num);
|
|
|
|
/* Check stack top values match target block type */
|
|
for (i = (int32)arity - 1; i >= 0; i--) {
|
|
if (!check_stack_top_values(frame_ref, available_stack_cell, types[i],
|
|
error_buf, error_buf_size)) {
|
|
goto fail;
|
|
}
|
|
cell_num = wasm_value_type_cell_num(types[i]);
|
|
frame_ref -= cell_num;
|
|
available_stack_cell -= cell_num;
|
|
}
|
|
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
emit_br_info(target_block, opcode == WASM_OP_BR);
|
|
#endif
|
|
|
|
ret = true;
|
|
|
|
cleanup_and_return:
|
|
fail:
|
|
if (frame_ref_buf && frame_ref_buf != frame_ref_tmp)
|
|
wasm_runtime_free(frame_ref_buf);
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
if (frame_offset_buf && frame_offset_buf != frame_offset_tmp)
|
|
wasm_runtime_free(frame_offset_buf);
|
|
#endif
|
|
|
|
return ret;
|
|
}
|
|
|
|
static BranchBlock *
|
|
check_branch_block(WASMLoaderContext *loader_ctx, uint8 **p_buf, uint8 *buf_end,
|
|
uint8 opcode, char *error_buf, uint32 error_buf_size)
|
|
{
|
|
uint8 *p = *p_buf, *p_end = buf_end;
|
|
BranchBlock *frame_csp_tmp;
|
|
uint32 depth;
|
|
|
|
read_leb_uint32(p, p_end, depth);
|
|
if (!wasm_loader_check_br(loader_ctx, depth, opcode, error_buf,
|
|
error_buf_size)) {
|
|
goto fail;
|
|
}
|
|
|
|
frame_csp_tmp = loader_ctx->frame_csp - depth - 1;
|
|
|
|
*p_buf = p;
|
|
return frame_csp_tmp;
|
|
fail:
|
|
return NULL;
|
|
}
|
|
|
|
static bool
|
|
check_block_stack(WASMLoaderContext *loader_ctx, BranchBlock *block,
|
|
char *error_buf, uint32 error_buf_size)
|
|
{
|
|
BlockType *block_type = &block->block_type;
|
|
uint8 *return_types = NULL;
|
|
uint32 return_count = 0;
|
|
int32 available_stack_cell, return_cell_num, i;
|
|
uint8 *frame_ref = NULL;
|
|
|
|
available_stack_cell =
|
|
(int32)(loader_ctx->stack_cell_num - block->stack_cell_num);
|
|
|
|
return_count = block_type_get_result_types(block_type, &return_types);
|
|
return_cell_num =
|
|
return_count > 0 ? wasm_get_cell_num(return_types, return_count) : 0;
|
|
|
|
/* If the stack is in polymorphic state, just clear the stack
|
|
* and then re-push the values to make the stack top values
|
|
* match block type. */
|
|
if (block->is_stack_polymorphic) {
|
|
for (i = (int32)return_count - 1; i >= 0; i--) {
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
POP_OFFSET_TYPE(return_types[i]);
|
|
#endif
|
|
POP_TYPE(return_types[i]);
|
|
}
|
|
|
|
/* Check stack is empty */
|
|
bh_assert(loader_ctx->stack_cell_num == block->stack_cell_num);
|
|
|
|
for (i = 0; i < (int32)return_count; i++) {
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
bool disable_emit = true;
|
|
int16 operand_offset = 0;
|
|
PUSH_OFFSET_TYPE(return_types[i]);
|
|
#endif
|
|
PUSH_TYPE(return_types[i]);
|
|
}
|
|
return true;
|
|
}
|
|
|
|
/* Check stack cell num equals return cell num */
|
|
bh_assert(available_stack_cell == return_cell_num);
|
|
|
|
/* Check stack values match return types */
|
|
frame_ref = loader_ctx->frame_ref;
|
|
for (i = (int32)return_count - 1; i >= 0; i--) {
|
|
if (!check_stack_top_values(frame_ref, available_stack_cell,
|
|
return_types[i], error_buf, error_buf_size))
|
|
return false;
|
|
frame_ref -= wasm_value_type_cell_num(return_types[i]);
|
|
available_stack_cell -= wasm_value_type_cell_num(return_types[i]);
|
|
}
|
|
|
|
(void)return_cell_num;
|
|
return true;
|
|
|
|
fail:
|
|
return false;
|
|
}
|
|
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
/* Copy parameters to dynamic space.
|
|
* 1) POP original parameter out;
|
|
* 2) Push and copy original values to dynamic space.
|
|
* The copy instruction format:
|
|
* Part a: param count
|
|
* Part b: all param total cell num
|
|
* Part c: each param's cell_num, src offset and dst offset
|
|
* Part d: each param's src offset
|
|
* Part e: each param's dst offset
|
|
*/
|
|
static bool
|
|
copy_params_to_dynamic_space(WASMLoaderContext *loader_ctx, bool is_if_block,
|
|
char *error_buf, uint32 error_buf_size)
|
|
{
|
|
bool ret = false;
|
|
int16 *frame_offset = NULL;
|
|
uint8 *cells = NULL, cell;
|
|
int16 *src_offsets = NULL;
|
|
uint8 *emit_data = NULL;
|
|
uint32 i;
|
|
BranchBlock *block = loader_ctx->frame_csp - 1;
|
|
BlockType *block_type = &block->block_type;
|
|
WASMFuncType *wasm_type = block_type->u.type;
|
|
uint32 param_count = block_type->u.type->param_count;
|
|
int16 condition_offset = 0;
|
|
bool disable_emit = false;
|
|
int16 operand_offset = 0;
|
|
|
|
uint64 size = (uint64)param_count * (sizeof(*cells) + sizeof(*src_offsets));
|
|
bh_assert(size > 0);
|
|
|
|
/* For if block, we also need copy the condition operand offset. */
|
|
if (is_if_block)
|
|
size += sizeof(*cells) + sizeof(*src_offsets);
|
|
|
|
/* Allocate memory for the emit data */
|
|
if (!(emit_data = loader_malloc(size, error_buf, error_buf_size)))
|
|
return false;
|
|
|
|
cells = emit_data;
|
|
src_offsets = (int16 *)(cells + param_count);
|
|
|
|
if (is_if_block)
|
|
condition_offset = *loader_ctx->frame_offset;
|
|
|
|
/* POP original parameter out */
|
|
for (i = 0; i < param_count; i++) {
|
|
POP_OFFSET_TYPE(wasm_type->types[param_count - i - 1]);
|
|
wasm_loader_emit_backspace(loader_ctx, sizeof(int16));
|
|
}
|
|
frame_offset = loader_ctx->frame_offset;
|
|
|
|
/* Get each param's cell num and src offset */
|
|
for (i = 0; i < param_count; i++) {
|
|
cell = (uint8)wasm_value_type_cell_num(wasm_type->types[i]);
|
|
cells[i] = cell;
|
|
src_offsets[i] = *frame_offset;
|
|
frame_offset += cell;
|
|
}
|
|
/* emit copy instruction */
|
|
emit_label(EXT_OP_COPY_STACK_VALUES);
|
|
/* Part a) */
|
|
emit_uint32(loader_ctx, is_if_block ? param_count + 1 : param_count);
|
|
/* Part b) */
|
|
emit_uint32(loader_ctx, is_if_block ? wasm_type->param_cell_num + 1
|
|
: wasm_type->param_cell_num);
|
|
/* Part c) */
|
|
for (i = 0; i < param_count; i++)
|
|
emit_byte(loader_ctx, cells[i]);
|
|
if (is_if_block)
|
|
emit_byte(loader_ctx, 1);
|
|
|
|
/* Part d) */
|
|
for (i = 0; i < param_count; i++)
|
|
emit_operand(loader_ctx, src_offsets[i]);
|
|
if (is_if_block)
|
|
emit_operand(loader_ctx, condition_offset);
|
|
|
|
/* Part e) */
|
|
/* Push to dynamic space. The push will emit the dst offset. */
|
|
for (i = 0; i < param_count; i++)
|
|
PUSH_OFFSET_TYPE(wasm_type->types[i]);
|
|
if (is_if_block)
|
|
PUSH_OFFSET_TYPE(VALUE_TYPE_I32);
|
|
|
|
ret = true;
|
|
|
|
fail:
|
|
/* Free the emit data */
|
|
wasm_runtime_free(emit_data);
|
|
|
|
return ret;
|
|
}
|
|
#endif
|
|
|
|
/* reset the stack to the state of before entering the last block */
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
#define RESET_STACK() \
|
|
do { \
|
|
loader_ctx->stack_cell_num = \
|
|
(loader_ctx->frame_csp - 1)->stack_cell_num; \
|
|
loader_ctx->frame_ref = \
|
|
loader_ctx->frame_ref_bottom + loader_ctx->stack_cell_num; \
|
|
loader_ctx->frame_offset = \
|
|
loader_ctx->frame_offset_bottom + loader_ctx->stack_cell_num; \
|
|
} while (0)
|
|
#else
|
|
#define RESET_STACK() \
|
|
do { \
|
|
loader_ctx->stack_cell_num = \
|
|
(loader_ctx->frame_csp - 1)->stack_cell_num; \
|
|
loader_ctx->frame_ref = \
|
|
loader_ctx->frame_ref_bottom + loader_ctx->stack_cell_num; \
|
|
} while (0)
|
|
#endif
|
|
|
|
/* set current block's stack polymorphic state */
|
|
#define SET_CUR_BLOCK_STACK_POLYMORPHIC_STATE(flag) \
|
|
do { \
|
|
BranchBlock *_cur_block = loader_ctx->frame_csp - 1; \
|
|
_cur_block->is_stack_polymorphic = flag; \
|
|
} while (0)
|
|
|
|
#define BLOCK_HAS_PARAM(block_type) \
|
|
(!block_type.is_value_type && block_type.u.type->param_count > 0)
|
|
|
|
#define PRESERVE_LOCAL_FOR_BLOCK() \
|
|
do { \
|
|
if (!(preserve_local_for_block(loader_ctx, opcode, error_buf, \
|
|
error_buf_size))) { \
|
|
goto fail; \
|
|
} \
|
|
} while (0)
|
|
|
|
static bool
|
|
wasm_loader_prepare_bytecode(WASMModule *module, WASMFunction *func,
|
|
uint32 cur_func_idx, char *error_buf,
|
|
uint32 error_buf_size)
|
|
{
|
|
uint8 *p = func->code, *p_end = func->code + func->code_size, *p_org;
|
|
uint32 param_count, local_count, global_count;
|
|
uint8 *param_types, *local_types, local_type, global_type, mem_offset_type;
|
|
BlockType func_block_type;
|
|
uint16 *local_offsets, local_offset;
|
|
uint32 count, local_idx, global_idx, u32, align, i;
|
|
mem_offset_t mem_offset;
|
|
int32 i32, i32_const = 0;
|
|
int64 i64_const;
|
|
uint8 opcode, u8;
|
|
bool return_value = false;
|
|
WASMLoaderContext *loader_ctx;
|
|
BranchBlock *frame_csp_tmp;
|
|
#if WASM_ENABLE_BULK_MEMORY != 0
|
|
uint32 segment_index;
|
|
#endif
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
uint8 *func_const_end, *func_const = NULL;
|
|
int16 operand_offset = 0;
|
|
uint8 last_op = 0;
|
|
bool disable_emit, preserve_local = false, if_condition_available = true;
|
|
;
|
|
float32 f32_const;
|
|
float64 f64_const;
|
|
|
|
LOG_OP("\nProcessing func | [%d] params | [%d] locals | [%d] return\n",
|
|
func->param_cell_num, func->local_cell_num, func->ret_cell_num);
|
|
#endif
|
|
#if WASM_ENABLE_MEMORY64 != 0
|
|
bool is_memory64 = has_module_memory64(module);
|
|
mem_offset_type = is_memory64 ? VALUE_TYPE_I64 : VALUE_TYPE_I32;
|
|
#else
|
|
mem_offset_type = VALUE_TYPE_I32;
|
|
#endif
|
|
|
|
global_count = module->import_global_count + module->global_count;
|
|
|
|
param_count = func->func_type->param_count;
|
|
param_types = func->func_type->types;
|
|
|
|
func_block_type.is_value_type = false;
|
|
func_block_type.u.type = func->func_type;
|
|
|
|
local_count = func->local_count;
|
|
local_types = func->local_types;
|
|
local_offsets = func->local_offsets;
|
|
|
|
if (!(loader_ctx = wasm_loader_ctx_init(func, error_buf, error_buf_size))) {
|
|
goto fail;
|
|
}
|
|
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
/* For the first traverse, the initial value of preserved_local_offset has
|
|
* not been determined, we use the INT16_MAX to represent that a slot has
|
|
* been copied to preserve space. For second traverse, this field will be
|
|
* set to the appropriate value in wasm_loader_ctx_reinit.
|
|
* This is for Issue #1230,
|
|
* https://github.com/bytecodealliance/wasm-micro-runtime/issues/1230, the
|
|
* drop opcodes need to know which slots are preserved, so those slots will
|
|
* not be treated as dynamically allocated slots */
|
|
loader_ctx->preserved_local_offset = INT16_MAX;
|
|
|
|
re_scan:
|
|
if (loader_ctx->code_compiled_size > 0) {
|
|
if (!wasm_loader_ctx_reinit(loader_ctx)) {
|
|
set_error_buf(error_buf, error_buf_size, "allocate memory failed");
|
|
goto fail;
|
|
}
|
|
p = func->code;
|
|
func->code_compiled = loader_ctx->p_code_compiled;
|
|
func->code_compiled_size = loader_ctx->code_compiled_size;
|
|
}
|
|
#endif
|
|
|
|
PUSH_CSP(LABEL_TYPE_FUNCTION, func_block_type, p);
|
|
|
|
while (p < p_end) {
|
|
opcode = *p++;
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
p_org = p;
|
|
disable_emit = false;
|
|
emit_label(opcode);
|
|
#endif
|
|
|
|
switch (opcode) {
|
|
case WASM_OP_UNREACHABLE:
|
|
RESET_STACK();
|
|
SET_CUR_BLOCK_STACK_POLYMORPHIC_STATE(true);
|
|
break;
|
|
|
|
case WASM_OP_NOP:
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
skip_label();
|
|
#endif
|
|
break;
|
|
|
|
case WASM_OP_IF:
|
|
{
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
BranchBlock *parent_block = loader_ctx->frame_csp - 1;
|
|
int32 available_stack_cell =
|
|
(int32)(loader_ctx->stack_cell_num
|
|
- parent_block->stack_cell_num);
|
|
|
|
if (available_stack_cell <= 0
|
|
&& parent_block->is_stack_polymorphic)
|
|
if_condition_available = false;
|
|
else
|
|
if_condition_available = true;
|
|
PRESERVE_LOCAL_FOR_BLOCK();
|
|
#endif
|
|
POP_I32();
|
|
goto handle_op_block_and_loop;
|
|
}
|
|
case WASM_OP_BLOCK:
|
|
case WASM_OP_LOOP:
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
PRESERVE_LOCAL_FOR_BLOCK();
|
|
#endif
|
|
handle_op_block_and_loop:
|
|
{
|
|
uint8 value_type;
|
|
BlockType block_type;
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
uint32 available_params = 0;
|
|
#endif
|
|
|
|
p_org = p - 1;
|
|
value_type = read_uint8(p);
|
|
if (is_byte_a_type(value_type)) {
|
|
/* If the first byte is one of these special values:
|
|
* 0x40/0x7F/0x7E/0x7D/0x7C, take it as the type of
|
|
* the single return value. */
|
|
block_type.is_value_type = true;
|
|
block_type.u.value_type.type = value_type;
|
|
}
|
|
else {
|
|
uint32 type_index;
|
|
/* Resolve the leb128 encoded type index as block type */
|
|
p--;
|
|
read_leb_uint32(p, p_end, type_index);
|
|
bh_assert(type_index < module->type_count);
|
|
block_type.is_value_type = false;
|
|
block_type.u.type = module->types[type_index];
|
|
#if WASM_ENABLE_FAST_INTERP == 0
|
|
/* If block use type index as block type, change the opcode
|
|
* to new extended opcode so that interpreter can resolve
|
|
* the block quickly.
|
|
*/
|
|
*p_org = EXT_OP_BLOCK + (opcode - WASM_OP_BLOCK);
|
|
#endif
|
|
}
|
|
|
|
/* Pop block parameters from stack */
|
|
if (BLOCK_HAS_PARAM(block_type)) {
|
|
WASMFuncType *wasm_type = block_type.u.type;
|
|
|
|
BranchBlock *cur_block = loader_ctx->frame_csp - 1;
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
uint32 cell_num;
|
|
available_params = block_type.u.type->param_count;
|
|
#endif
|
|
for (i = 0; i < block_type.u.type->param_count; i++) {
|
|
|
|
int32 available_stack_cell =
|
|
(int32)(loader_ctx->stack_cell_num
|
|
- cur_block->stack_cell_num);
|
|
if (available_stack_cell <= 0
|
|
&& cur_block->is_stack_polymorphic) {
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
available_params = i;
|
|
#endif
|
|
break;
|
|
}
|
|
|
|
POP_TYPE(
|
|
wasm_type->types[wasm_type->param_count - i - 1]);
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
/* decrease the frame_offset pointer accordingly to keep
|
|
* consistent with frame_ref stack */
|
|
cell_num = wasm_value_type_cell_num(
|
|
wasm_type->types[wasm_type->param_count - i - 1]);
|
|
loader_ctx->frame_offset -= cell_num;
|
|
#endif
|
|
}
|
|
}
|
|
|
|
PUSH_CSP(LABEL_TYPE_BLOCK + (opcode - WASM_OP_BLOCK),
|
|
block_type, p);
|
|
|
|
/* Pass parameters to block */
|
|
if (BLOCK_HAS_PARAM(block_type)) {
|
|
for (i = 0; i < block_type.u.type->param_count; i++) {
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
uint32 cell_num = wasm_value_type_cell_num(
|
|
block_type.u.type->types[i]);
|
|
if (i >= available_params) {
|
|
/* If there isn't enough data on stack, push a dummy
|
|
* offset to keep the stack consistent with
|
|
* frame_ref.
|
|
* Since the stack is already in polymorphic state,
|
|
* the opcode will not be executed, so the dummy
|
|
* offset won't cause any error */
|
|
uint32 n;
|
|
|
|
for (n = 0; n < cell_num; n++) {
|
|
if (loader_ctx->p_code_compiled == NULL) {
|
|
if (!check_offset_push(loader_ctx,
|
|
error_buf,
|
|
error_buf_size))
|
|
goto fail;
|
|
}
|
|
*loader_ctx->frame_offset++ = 0;
|
|
}
|
|
}
|
|
else {
|
|
loader_ctx->frame_offset += cell_num;
|
|
}
|
|
#endif
|
|
PUSH_TYPE(block_type.u.type->types[i]);
|
|
}
|
|
}
|
|
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
if (opcode == WASM_OP_BLOCK || opcode == WASM_OP_LOOP) {
|
|
skip_label();
|
|
if (BLOCK_HAS_PARAM(block_type)) {
|
|
/* Make sure params are in dynamic space */
|
|
if (!copy_params_to_dynamic_space(
|
|
loader_ctx, false, error_buf, error_buf_size))
|
|
goto fail;
|
|
}
|
|
if (opcode == WASM_OP_LOOP) {
|
|
(loader_ctx->frame_csp - 1)->code_compiled =
|
|
loader_ctx->p_code_compiled;
|
|
}
|
|
}
|
|
else if (opcode == WASM_OP_IF) {
|
|
BranchBlock *block = loader_ctx->frame_csp - 1;
|
|
/* If block has parameters, we should make sure they are in
|
|
* dynamic space. Otherwise, when else branch is missing,
|
|
* the later opcode may consume incorrect operand offset.
|
|
* Spec case:
|
|
* (func (export "params-id") (param i32) (result i32)
|
|
* (i32.const 1)
|
|
* (i32.const 2)
|
|
* (if (param i32 i32) (result i32 i32) (local.get 0)
|
|
* (then)) (i32.add)
|
|
* )
|
|
*
|
|
* So we should emit a copy instruction before the if.
|
|
*
|
|
* And we also need to save the parameter offsets and
|
|
* recover them before entering else branch.
|
|
*
|
|
*/
|
|
if (BLOCK_HAS_PARAM(block_type)) {
|
|
uint64 size;
|
|
|
|
/* In polymorphic state, there may be no if condition on
|
|
* the stack, so the offset may not emitted */
|
|
if (if_condition_available) {
|
|
/* skip the if condition operand offset */
|
|
wasm_loader_emit_backspace(loader_ctx,
|
|
sizeof(int16));
|
|
}
|
|
/* skip the if label */
|
|
skip_label();
|
|
/* Emit a copy instruction */
|
|
if (!copy_params_to_dynamic_space(
|
|
loader_ctx, true, error_buf, error_buf_size))
|
|
goto fail;
|
|
|
|
/* Emit the if instruction */
|
|
emit_label(opcode);
|
|
/* Emit the new condition operand offset */
|
|
POP_OFFSET_TYPE(VALUE_TYPE_I32);
|
|
|
|
/* Save top param_count values of frame_offset stack, so
|
|
* that we can recover it before executing else branch
|
|
*/
|
|
size = sizeof(int16)
|
|
* (uint64)block_type.u.type->param_cell_num;
|
|
if (!(block->param_frame_offsets = loader_malloc(
|
|
size, error_buf, error_buf_size)))
|
|
goto fail;
|
|
bh_memcpy_s(block->param_frame_offsets, (uint32)size,
|
|
loader_ctx->frame_offset
|
|
- size / sizeof(int16),
|
|
(uint32)size);
|
|
}
|
|
|
|
block->start_dynamic_offset = loader_ctx->dynamic_offset;
|
|
|
|
emit_empty_label_addr_and_frame_ip(PATCH_ELSE);
|
|
emit_empty_label_addr_and_frame_ip(PATCH_END);
|
|
}
|
|
#endif
|
|
break;
|
|
}
|
|
|
|
case WASM_OP_ELSE:
|
|
handle_op_else:
|
|
{
|
|
BranchBlock *block = NULL;
|
|
BlockType block_type = (loader_ctx->frame_csp - 1)->block_type;
|
|
bh_assert(loader_ctx->csp_num >= 2
|
|
/* the matched if is found */
|
|
&& (loader_ctx->frame_csp - 1)->label_type
|
|
== LABEL_TYPE_IF
|
|
/* duplicated else isn't found */
|
|
&& !(loader_ctx->frame_csp - 1)->else_addr);
|
|
block = loader_ctx->frame_csp - 1;
|
|
|
|
/* check whether if branch's stack matches its result type */
|
|
if (!check_block_stack(loader_ctx, block, error_buf,
|
|
error_buf_size))
|
|
goto fail;
|
|
|
|
block->else_addr = p - 1;
|
|
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
/* if the result of if branch is in local or const area, add a
|
|
* copy op */
|
|
RESERVE_BLOCK_RET();
|
|
|
|
emit_empty_label_addr_and_frame_ip(PATCH_END);
|
|
apply_label_patch(loader_ctx, 1, PATCH_ELSE);
|
|
#endif
|
|
RESET_STACK();
|
|
SET_CUR_BLOCK_STACK_POLYMORPHIC_STATE(false);
|
|
|
|
/* Pass parameters to if-false branch */
|
|
if (BLOCK_HAS_PARAM(block_type)) {
|
|
for (i = 0; i < block_type.u.type->param_count; i++)
|
|
PUSH_TYPE(block_type.u.type->types[i]);
|
|
}
|
|
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
/* Recover top param_count values of frame_offset stack */
|
|
if (BLOCK_HAS_PARAM((block_type))) {
|
|
uint32 size;
|
|
size = sizeof(int16) * block_type.u.type->param_cell_num;
|
|
bh_memcpy_s(loader_ctx->frame_offset, size,
|
|
block->param_frame_offsets, size);
|
|
loader_ctx->frame_offset += (size / sizeof(int16));
|
|
}
|
|
loader_ctx->dynamic_offset = block->start_dynamic_offset;
|
|
#endif
|
|
|
|
break;
|
|
}
|
|
|
|
case WASM_OP_END:
|
|
{
|
|
BranchBlock *cur_block = loader_ctx->frame_csp - 1;
|
|
|
|
/* check whether block stack matches its result type */
|
|
if (!check_block_stack(loader_ctx, cur_block, error_buf,
|
|
error_buf_size))
|
|
goto fail;
|
|
|
|
/* if there is no else branch, make a virtual else opcode for
|
|
easier integrity check and to copy the correct results to
|
|
the block return address for fast-interp mode:
|
|
change if block from `if ... end` to `if ... else end` */
|
|
if (cur_block->label_type == LABEL_TYPE_IF
|
|
&& !cur_block->else_addr) {
|
|
opcode = WASM_OP_ELSE;
|
|
p--;
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
p_org = p;
|
|
skip_label();
|
|
disable_emit = false;
|
|
emit_label(opcode);
|
|
#endif
|
|
goto handle_op_else;
|
|
}
|
|
|
|
POP_CSP();
|
|
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
skip_label();
|
|
/* copy the result to the block return address */
|
|
RESERVE_BLOCK_RET();
|
|
|
|
apply_label_patch(loader_ctx, 0, PATCH_END);
|
|
free_label_patch_list(loader_ctx->frame_csp);
|
|
if (loader_ctx->frame_csp->label_type == LABEL_TYPE_FUNCTION) {
|
|
int32 idx;
|
|
uint8 ret_type;
|
|
|
|
emit_label(WASM_OP_RETURN);
|
|
for (idx = (int32)func->func_type->result_count - 1;
|
|
idx >= 0; idx--) {
|
|
ret_type = *(func->func_type->types
|
|
+ func->func_type->param_count + idx);
|
|
POP_OFFSET_TYPE(ret_type);
|
|
}
|
|
}
|
|
#endif
|
|
if (loader_ctx->csp_num > 0) {
|
|
loader_ctx->frame_csp->end_addr = p - 1;
|
|
}
|
|
else {
|
|
/* end of function block, function will return */
|
|
bh_assert(p == p_end);
|
|
}
|
|
|
|
break;
|
|
}
|
|
|
|
case WASM_OP_BR:
|
|
{
|
|
if (!(frame_csp_tmp =
|
|
check_branch_block(loader_ctx, &p, p_end, opcode,
|
|
error_buf, error_buf_size)))
|
|
goto fail;
|
|
|
|
RESET_STACK();
|
|
SET_CUR_BLOCK_STACK_POLYMORPHIC_STATE(true);
|
|
break;
|
|
}
|
|
|
|
case WASM_OP_BR_IF:
|
|
{
|
|
POP_I32();
|
|
|
|
if (!(frame_csp_tmp =
|
|
check_branch_block(loader_ctx, &p, p_end, opcode,
|
|
error_buf, error_buf_size)))
|
|
goto fail;
|
|
|
|
break;
|
|
}
|
|
|
|
case WASM_OP_BR_TABLE:
|
|
{
|
|
uint8 *ret_types = NULL;
|
|
uint32 ret_count = 0, depth = 0;
|
|
#if WASM_ENABLE_FAST_INTERP == 0
|
|
BrTableCache *br_table_cache = NULL;
|
|
uint8 *p_depth_begin, *p_depth, *p_opcode = p - 1;
|
|
uint32 j;
|
|
#endif
|
|
|
|
read_leb_uint32(p, p_end, count);
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
emit_uint32(loader_ctx, count);
|
|
#endif
|
|
POP_I32();
|
|
|
|
/* Get each depth and check it */
|
|
p_org = p;
|
|
for (i = 0; i <= count; i++) {
|
|
read_leb_uint32(p, p_end, depth);
|
|
bh_assert(loader_ctx->csp_num > 0);
|
|
bh_assert(loader_ctx->csp_num - 1 >= depth);
|
|
(void)depth;
|
|
}
|
|
p = p_org;
|
|
|
|
#if WASM_ENABLE_FAST_INTERP == 0
|
|
p_depth_begin = p_depth = p;
|
|
#endif
|
|
for (i = 0; i <= count; i++) {
|
|
if (!(frame_csp_tmp =
|
|
check_branch_block(loader_ctx, &p, p_end, opcode,
|
|
error_buf, error_buf_size)))
|
|
goto fail;
|
|
|
|
#if WASM_ENABLE_FAST_INTERP == 0
|
|
depth = (uint32)(loader_ctx->frame_csp - 1 - frame_csp_tmp);
|
|
if (br_table_cache) {
|
|
br_table_cache->br_depths[i] = depth;
|
|
}
|
|
else {
|
|
if (depth > 255) {
|
|
/* The depth cannot be stored in one byte,
|
|
create br_table cache to store each depth */
|
|
if (!(br_table_cache = loader_malloc(
|
|
offsetof(BrTableCache, br_depths)
|
|
+ sizeof(uint32)
|
|
* (uint64)(count + 1),
|
|
error_buf, error_buf_size))) {
|
|
goto fail;
|
|
}
|
|
*p_opcode = EXT_OP_BR_TABLE_CACHE;
|
|
br_table_cache->br_table_op_addr = p_opcode;
|
|
br_table_cache->br_count = count;
|
|
/* Copy previous depths which are one byte */
|
|
for (j = 0; j < i; j++) {
|
|
br_table_cache->br_depths[j] = p_depth_begin[j];
|
|
}
|
|
br_table_cache->br_depths[i] = depth;
|
|
bh_list_insert(module->br_table_cache_list,
|
|
br_table_cache);
|
|
}
|
|
else {
|
|
/* The depth can be stored in one byte, use the
|
|
byte of the leb to store it */
|
|
*p_depth++ = (uint8)depth;
|
|
}
|
|
}
|
|
#endif
|
|
}
|
|
|
|
#if WASM_ENABLE_FAST_INTERP == 0
|
|
/* Set the tailing bytes to nop */
|
|
if (br_table_cache)
|
|
p_depth = p_depth_begin;
|
|
while (p_depth < p)
|
|
*p_depth++ = WASM_OP_NOP;
|
|
#endif
|
|
|
|
RESET_STACK();
|
|
SET_CUR_BLOCK_STACK_POLYMORPHIC_STATE(true);
|
|
|
|
(void)ret_count;
|
|
(void)ret_types;
|
|
break;
|
|
}
|
|
|
|
case WASM_OP_RETURN:
|
|
{
|
|
int32 idx;
|
|
uint8 ret_type;
|
|
for (idx = (int32)func->func_type->result_count - 1; idx >= 0;
|
|
idx--) {
|
|
ret_type = *(func->func_type->types
|
|
+ func->func_type->param_count + idx);
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
/* emit the offset after return opcode */
|
|
POP_OFFSET_TYPE(ret_type);
|
|
#endif
|
|
POP_TYPE(ret_type);
|
|
}
|
|
|
|
RESET_STACK();
|
|
SET_CUR_BLOCK_STACK_POLYMORPHIC_STATE(true);
|
|
|
|
break;
|
|
}
|
|
|
|
case WASM_OP_CALL:
|
|
#if WASM_ENABLE_TAIL_CALL != 0
|
|
case WASM_OP_RETURN_CALL:
|
|
#endif
|
|
{
|
|
WASMFuncType *func_type;
|
|
uint32 func_idx;
|
|
int32 idx;
|
|
|
|
read_leb_uint32(p, p_end, func_idx);
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
/* we need to emit func_idx before arguments */
|
|
emit_uint32(loader_ctx, func_idx);
|
|
#endif
|
|
|
|
bh_assert(func_idx < module->import_function_count
|
|
+ module->function_count);
|
|
|
|
if (func_idx < module->import_function_count)
|
|
func_type =
|
|
module->import_functions[func_idx].u.function.func_type;
|
|
else
|
|
func_type = module
|
|
->functions[func_idx
|
|
- module->import_function_count]
|
|
->func_type;
|
|
|
|
if (func_type->param_count > 0) {
|
|
for (idx = (int32)(func_type->param_count - 1); idx >= 0;
|
|
idx--) {
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
POP_OFFSET_TYPE(func_type->types[idx]);
|
|
#endif
|
|
POP_TYPE(func_type->types[idx]);
|
|
}
|
|
}
|
|
|
|
#if WASM_ENABLE_TAIL_CALL != 0
|
|
if (opcode == WASM_OP_CALL) {
|
|
#endif
|
|
for (i = 0; i < func_type->result_count; i++) {
|
|
PUSH_TYPE(func_type->types[func_type->param_count + i]);
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
/* Here we emit each return value's dynamic_offset. But
|
|
* in fact these offsets are continuous, so interpreter
|
|
* only need to get the first return value's offset.
|
|
*/
|
|
PUSH_OFFSET_TYPE(
|
|
func_type->types[func_type->param_count + i]);
|
|
#endif
|
|
}
|
|
#if WASM_ENABLE_TAIL_CALL != 0
|
|
}
|
|
else {
|
|
bh_assert(func_type->result_count
|
|
== func->func_type->result_count);
|
|
for (i = 0; i < func_type->result_count; i++) {
|
|
bh_assert(
|
|
func_type->types[func_type->param_count + i]
|
|
== func->func_type
|
|
->types[func->func_type->param_count + i]);
|
|
}
|
|
}
|
|
#endif
|
|
#if WASM_ENABLE_FAST_JIT != 0 || WASM_ENABLE_JIT != 0 \
|
|
|| WASM_ENABLE_WAMR_COMPILER != 0
|
|
func->has_op_func_call = true;
|
|
#endif
|
|
break;
|
|
}
|
|
|
|
case WASM_OP_CALL_INDIRECT:
|
|
#if WASM_ENABLE_TAIL_CALL != 0
|
|
case WASM_OP_RETURN_CALL_INDIRECT:
|
|
#endif
|
|
{
|
|
int32 idx;
|
|
WASMFuncType *func_type;
|
|
uint32 type_idx, table_idx;
|
|
|
|
bh_assert(module->import_table_count + module->table_count > 0);
|
|
|
|
read_leb_uint32(p, p_end, type_idx);
|
|
|
|
#if WASM_ENABLE_REF_TYPES != 0
|
|
read_leb_uint32(p, p_end, table_idx);
|
|
#else
|
|
CHECK_BUF(p, p_end, 1);
|
|
table_idx = read_uint8(p);
|
|
#endif
|
|
if (!check_table_index(module, table_idx, error_buf,
|
|
error_buf_size)) {
|
|
goto fail;
|
|
}
|
|
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
/* we need to emit before arguments */
|
|
emit_uint32(loader_ctx, type_idx);
|
|
emit_uint32(loader_ctx, table_idx);
|
|
#endif
|
|
|
|
/* skip elem idx */
|
|
POP_I32();
|
|
|
|
bh_assert(type_idx < module->type_count);
|
|
|
|
func_type = module->types[type_idx];
|
|
|
|
if (func_type->param_count > 0) {
|
|
for (idx = (int32)(func_type->param_count - 1); idx >= 0;
|
|
idx--) {
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
POP_OFFSET_TYPE(func_type->types[idx]);
|
|
#endif
|
|
POP_TYPE(func_type->types[idx]);
|
|
}
|
|
}
|
|
|
|
#if WASM_ENABLE_TAIL_CALL != 0
|
|
if (opcode == WASM_OP_CALL) {
|
|
#endif
|
|
for (i = 0; i < func_type->result_count; i++) {
|
|
PUSH_TYPE(func_type->types[func_type->param_count + i]);
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
PUSH_OFFSET_TYPE(
|
|
func_type->types[func_type->param_count + i]);
|
|
#endif
|
|
}
|
|
#if WASM_ENABLE_TAIL_CALL != 0
|
|
}
|
|
else {
|
|
bh_assert(func_type->result_count
|
|
== func->func_type->result_count);
|
|
for (i = 0; i < func_type->result_count; i++) {
|
|
bh_assert(
|
|
func_type->types[func_type->param_count + i]
|
|
== func->func_type
|
|
->types[func->func_type->param_count + i]);
|
|
}
|
|
}
|
|
#endif
|
|
|
|
#if WASM_ENABLE_FAST_JIT != 0 || WASM_ENABLE_JIT != 0 \
|
|
|| WASM_ENABLE_WAMR_COMPILER != 0
|
|
func->has_op_func_call = true;
|
|
#endif
|
|
#if WASM_ENABLE_JIT != 0 || WASM_ENABLE_WAMR_COMPILER != 0
|
|
func->has_op_call_indirect = true;
|
|
#endif
|
|
break;
|
|
}
|
|
|
|
#if WASM_ENABLE_EXCE_HANDLING != 0
|
|
case WASM_OP_TRY:
|
|
case WASM_OP_CATCH:
|
|
case WASM_OP_THROW:
|
|
case WASM_OP_RETHROW:
|
|
case WASM_OP_DELEGATE:
|
|
case WASM_OP_CATCH_ALL:
|
|
/* TODO */
|
|
set_error_buf(error_buf, error_buf_size, "unsupported opcode");
|
|
goto fail;
|
|
#endif
|
|
|
|
case WASM_OP_DROP:
|
|
{
|
|
BranchBlock *cur_block = loader_ctx->frame_csp - 1;
|
|
int32 available_stack_cell =
|
|
(int32)(loader_ctx->stack_cell_num
|
|
- cur_block->stack_cell_num);
|
|
|
|
bh_assert(!(available_stack_cell <= 0
|
|
&& !cur_block->is_stack_polymorphic));
|
|
|
|
if (available_stack_cell > 0) {
|
|
if (is_32bit_type(*(loader_ctx->frame_ref - 1))) {
|
|
loader_ctx->frame_ref--;
|
|
loader_ctx->stack_cell_num--;
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
skip_label();
|
|
loader_ctx->frame_offset--;
|
|
if ((*(loader_ctx->frame_offset)
|
|
> loader_ctx->start_dynamic_offset)
|
|
&& (*(loader_ctx->frame_offset)
|
|
< loader_ctx->max_dynamic_offset))
|
|
loader_ctx->dynamic_offset--;
|
|
#endif
|
|
}
|
|
else if (is_64bit_type(*(loader_ctx->frame_ref - 1))) {
|
|
loader_ctx->frame_ref -= 2;
|
|
loader_ctx->stack_cell_num -= 2;
|
|
#if WASM_ENABLE_FAST_INTERP == 0
|
|
*(p - 1) = WASM_OP_DROP_64;
|
|
#endif
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
skip_label();
|
|
loader_ctx->frame_offset -= 2;
|
|
if ((*(loader_ctx->frame_offset)
|
|
> loader_ctx->start_dynamic_offset)
|
|
&& (*(loader_ctx->frame_offset)
|
|
< loader_ctx->max_dynamic_offset))
|
|
loader_ctx->dynamic_offset -= 2;
|
|
#endif
|
|
}
|
|
else {
|
|
bh_assert(0);
|
|
}
|
|
}
|
|
else {
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
skip_label();
|
|
#endif
|
|
}
|
|
break;
|
|
}
|
|
|
|
case WASM_OP_SELECT:
|
|
{
|
|
uint8 ref_type;
|
|
BranchBlock *cur_block = loader_ctx->frame_csp - 1;
|
|
int32 available_stack_cell;
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
uint8 *p_code_compiled_tmp = loader_ctx->p_code_compiled;
|
|
#endif
|
|
|
|
POP_I32();
|
|
|
|
available_stack_cell = (int32)(loader_ctx->stack_cell_num
|
|
- cur_block->stack_cell_num);
|
|
|
|
bh_assert(!(available_stack_cell <= 0
|
|
&& !cur_block->is_stack_polymorphic));
|
|
|
|
if (available_stack_cell > 0) {
|
|
switch (*(loader_ctx->frame_ref - 1)) {
|
|
case REF_I32:
|
|
case REF_F32:
|
|
case REF_ANY:
|
|
break;
|
|
case REF_I64_2:
|
|
case REF_F64_2:
|
|
#if WASM_ENABLE_FAST_INTERP == 0
|
|
*(p - 1) = WASM_OP_SELECT_64;
|
|
#endif
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
if (loader_ctx->p_code_compiled) {
|
|
uint8 opcode_tmp = WASM_OP_SELECT_64;
|
|
#if WASM_ENABLE_LABELS_AS_VALUES != 0
|
|
#if WASM_CPU_SUPPORTS_UNALIGNED_ADDR_ACCESS != 0
|
|
*(void **)(p_code_compiled_tmp
|
|
- sizeof(void *)) =
|
|
handle_table[opcode_tmp];
|
|
#else
|
|
#if UINTPTR_MAX == UINT64_MAX
|
|
/* emit int32 relative offset in 64-bit target
|
|
*/
|
|
int32 offset =
|
|
(int32)((uint8 *)handle_table[opcode_tmp]
|
|
- (uint8 *)handle_table[0]);
|
|
*(int32 *)(p_code_compiled_tmp
|
|
- sizeof(int32)) = offset;
|
|
#else
|
|
/* emit uint32 label address in 32-bit target */
|
|
*(uint32 *)(p_code_compiled_tmp
|
|
- sizeof(uint32)) =
|
|
(uint32)(uintptr_t)handle_table[opcode_tmp];
|
|
#endif
|
|
#endif /* end of WASM_CPU_SUPPORTS_UNALIGNED_ADDR_ACCESS */
|
|
#else /* else of WASM_ENABLE_LABELS_AS_VALUES */
|
|
#if WASM_CPU_SUPPORTS_UNALIGNED_ADDR_ACCESS != 0
|
|
*(p_code_compiled_tmp - 1) = opcode_tmp;
|
|
#else
|
|
*(p_code_compiled_tmp - 2) = opcode_tmp;
|
|
#endif /* end of WASM_CPU_SUPPORTS_UNALIGNED_ADDR_ACCESS */
|
|
#endif /* end of WASM_ENABLE_LABELS_AS_VALUES */
|
|
}
|
|
#endif
|
|
break;
|
|
}
|
|
|
|
ref_type = *(loader_ctx->frame_ref - 1);
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
POP_OFFSET_TYPE(ref_type);
|
|
#endif
|
|
POP_TYPE(ref_type);
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
POP_OFFSET_TYPE(ref_type);
|
|
#endif
|
|
POP_TYPE(ref_type);
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
PUSH_OFFSET_TYPE(ref_type);
|
|
#endif
|
|
PUSH_TYPE(ref_type);
|
|
}
|
|
else {
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
PUSH_OFFSET_TYPE(VALUE_TYPE_ANY);
|
|
#endif
|
|
PUSH_TYPE(VALUE_TYPE_ANY);
|
|
}
|
|
break;
|
|
}
|
|
|
|
#if WASM_ENABLE_REF_TYPES != 0
|
|
case WASM_OP_SELECT_T:
|
|
{
|
|
uint8 vec_len, ref_type;
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
uint8 *p_code_compiled_tmp = loader_ctx->p_code_compiled;
|
|
#endif
|
|
|
|
read_leb_uint32(p, p_end, vec_len);
|
|
if (vec_len != 1) {
|
|
/* typed select must have exactly one result */
|
|
set_error_buf(error_buf, error_buf_size,
|
|
"invalid result arity");
|
|
goto fail;
|
|
}
|
|
|
|
CHECK_BUF(p, p_end, 1);
|
|
ref_type = read_uint8(p);
|
|
if (!is_value_type(ref_type)) {
|
|
set_error_buf(error_buf, error_buf_size,
|
|
"unknown value type");
|
|
goto fail;
|
|
}
|
|
|
|
POP_I32();
|
|
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
if (loader_ctx->p_code_compiled) {
|
|
uint8 opcode_tmp = WASM_OP_SELECT;
|
|
|
|
if (ref_type == VALUE_TYPE_F64
|
|
|| ref_type == VALUE_TYPE_I64)
|
|
opcode_tmp = WASM_OP_SELECT_64;
|
|
|
|
#if WASM_ENABLE_LABELS_AS_VALUES != 0
|
|
#if WASM_CPU_SUPPORTS_UNALIGNED_ADDR_ACCESS != 0
|
|
*(void **)(p_code_compiled_tmp - sizeof(void *)) =
|
|
handle_table[opcode_tmp];
|
|
#else
|
|
#if UINTPTR_MAX == UINT64_MAX
|
|
/* emit int32 relative offset in 64-bit target */
|
|
int32 offset = (int32)((uint8 *)handle_table[opcode_tmp]
|
|
- (uint8 *)handle_table[0]);
|
|
*(int32 *)(p_code_compiled_tmp - sizeof(int32)) = offset;
|
|
#else
|
|
/* emit uint32 label address in 32-bit target */
|
|
*(uint32 *)(p_code_compiled_tmp - sizeof(uint32)) =
|
|
(uint32)(uintptr_t)handle_table[opcode_tmp];
|
|
#endif
|
|
#endif /* end of WASM_CPU_SUPPORTS_UNALIGNED_ADDR_ACCESS */
|
|
#else /* else of WASM_ENABLE_LABELS_AS_VALUES */
|
|
#if WASM_CPU_SUPPORTS_UNALIGNED_ADDR_ACCESS != 0
|
|
*(p_code_compiled_tmp - 1) = opcode_tmp;
|
|
#else
|
|
*(p_code_compiled_tmp - 2) = opcode_tmp;
|
|
#endif /* end of WASM_CPU_SUPPORTS_UNALIGNED_ADDR_ACCESS */
|
|
#endif /* end of WASM_ENABLE_LABELS_AS_VALUES */
|
|
}
|
|
#endif /* WASM_ENABLE_FAST_INTERP != 0 */
|
|
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
POP_OFFSET_TYPE(ref_type);
|
|
POP_TYPE(ref_type);
|
|
POP_OFFSET_TYPE(ref_type);
|
|
POP_TYPE(ref_type);
|
|
PUSH_OFFSET_TYPE(ref_type);
|
|
PUSH_TYPE(ref_type);
|
|
#else
|
|
POP2_AND_PUSH(ref_type, ref_type);
|
|
#endif /* WASM_ENABLE_FAST_INTERP != 0 */
|
|
|
|
(void)vec_len;
|
|
break;
|
|
}
|
|
|
|
/* table.get x. tables[x]. [i32] -> [t] */
|
|
/* table.set x. tables[x]. [i32 t] -> [] */
|
|
case WASM_OP_TABLE_GET:
|
|
case WASM_OP_TABLE_SET:
|
|
{
|
|
uint8 decl_ref_type;
|
|
uint32 table_idx;
|
|
|
|
read_leb_uint32(p, p_end, table_idx);
|
|
if (!get_table_elem_type(module, table_idx, &decl_ref_type,
|
|
error_buf, error_buf_size))
|
|
goto fail;
|
|
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
emit_uint32(loader_ctx, table_idx);
|
|
#endif
|
|
|
|
if (opcode == WASM_OP_TABLE_GET) {
|
|
POP_I32();
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
PUSH_OFFSET_TYPE(decl_ref_type);
|
|
#endif
|
|
PUSH_TYPE(decl_ref_type);
|
|
}
|
|
else {
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
POP_OFFSET_TYPE(decl_ref_type);
|
|
#endif
|
|
POP_TYPE(decl_ref_type);
|
|
POP_I32();
|
|
}
|
|
break;
|
|
}
|
|
case WASM_OP_REF_NULL:
|
|
{
|
|
uint8 ref_type;
|
|
|
|
CHECK_BUF(p, p_end, 1);
|
|
ref_type = read_uint8(p);
|
|
if (ref_type != VALUE_TYPE_FUNCREF
|
|
&& ref_type != VALUE_TYPE_EXTERNREF) {
|
|
set_error_buf(error_buf, error_buf_size,
|
|
"unknown value type");
|
|
goto fail;
|
|
}
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
PUSH_OFFSET_TYPE(ref_type);
|
|
#endif
|
|
PUSH_TYPE(ref_type);
|
|
break;
|
|
}
|
|
case WASM_OP_REF_IS_NULL:
|
|
{
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
BranchBlock *cur_block = loader_ctx->frame_csp - 1;
|
|
int32 block_stack_cell_num =
|
|
(int32)(loader_ctx->stack_cell_num
|
|
- cur_block->stack_cell_num);
|
|
if (block_stack_cell_num <= 0) {
|
|
if (!cur_block->is_stack_polymorphic) {
|
|
set_error_buf(
|
|
error_buf, error_buf_size,
|
|
"type mismatch: expect data but stack was empty");
|
|
goto fail;
|
|
}
|
|
}
|
|
else {
|
|
if (*(loader_ctx->frame_ref - 1) == VALUE_TYPE_FUNCREF
|
|
|| *(loader_ctx->frame_ref - 1) == VALUE_TYPE_EXTERNREF
|
|
|| *(loader_ctx->frame_ref - 1) == VALUE_TYPE_ANY) {
|
|
if (!wasm_loader_pop_frame_ref_offset(
|
|
loader_ctx, *(loader_ctx->frame_ref - 1),
|
|
error_buf, error_buf_size)) {
|
|
goto fail;
|
|
}
|
|
}
|
|
else {
|
|
set_error_buf(error_buf, error_buf_size,
|
|
"type mismatch");
|
|
goto fail;
|
|
}
|
|
}
|
|
#else
|
|
if (!wasm_loader_pop_frame_ref(loader_ctx, VALUE_TYPE_FUNCREF,
|
|
error_buf, error_buf_size)
|
|
&& !wasm_loader_pop_frame_ref(loader_ctx,
|
|
VALUE_TYPE_EXTERNREF,
|
|
error_buf, error_buf_size)) {
|
|
goto fail;
|
|
}
|
|
#endif
|
|
PUSH_I32();
|
|
break;
|
|
}
|
|
case WASM_OP_REF_FUNC:
|
|
{
|
|
uint32 func_idx = 0;
|
|
read_leb_uint32(p, p_end, func_idx);
|
|
|
|
if (!check_function_index(module, func_idx, error_buf,
|
|
error_buf_size)) {
|
|
goto fail;
|
|
}
|
|
|
|
/* Refer to a forward-declared function:
|
|
the function must be an import, exported, or present in
|
|
a table elem segment or global initializer to be used as
|
|
the operand to ref.func */
|
|
if (func_idx >= module->import_function_count) {
|
|
WASMTableSeg *table_seg = module->table_segments;
|
|
bool func_declared = false;
|
|
uint32 j;
|
|
|
|
for (i = 0; i < module->global_count; i++) {
|
|
if (module->globals[i].type.val_type
|
|
== VALUE_TYPE_FUNCREF
|
|
&& module->globals[i].init_expr.init_expr_type
|
|
== INIT_EXPR_TYPE_FUNCREF_CONST
|
|
&& module->globals[i].init_expr.u.u32 == func_idx) {
|
|
func_declared = true;
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (!func_declared) {
|
|
/* Check whether the function is declared in table segs,
|
|
note that it doesn't matter whether the table seg's
|
|
mode is passive, active or declarative. */
|
|
for (i = 0; i < module->table_seg_count;
|
|
i++, table_seg++) {
|
|
if (table_seg->elem_type == VALUE_TYPE_FUNCREF) {
|
|
for (j = 0; j < table_seg->value_count; j++) {
|
|
if (table_seg->init_values[j].u.ref_index
|
|
== func_idx) {
|
|
func_declared = true;
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
if (!func_declared) {
|
|
/* Check whether the function is exported */
|
|
for (i = 0; i < module->export_count; i++) {
|
|
if (module->exports[i].kind == EXPORT_KIND_FUNC
|
|
&& module->exports[i].index == func_idx) {
|
|
func_declared = true;
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
bh_assert(func_declared);
|
|
(void)func_declared;
|
|
}
|
|
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
emit_uint32(loader_ctx, func_idx);
|
|
#endif
|
|
PUSH_FUNCREF();
|
|
break;
|
|
}
|
|
#endif /* WASM_ENABLE_REF_TYPES */
|
|
|
|
case WASM_OP_GET_LOCAL:
|
|
{
|
|
p_org = p - 1;
|
|
GET_LOCAL_INDEX_TYPE_AND_OFFSET();
|
|
PUSH_TYPE(local_type);
|
|
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
/* Get Local is optimized out */
|
|
skip_label();
|
|
disable_emit = true;
|
|
operand_offset = local_offset;
|
|
PUSH_OFFSET_TYPE(local_type);
|
|
#else
|
|
#if (WASM_ENABLE_WAMR_COMPILER == 0) && (WASM_ENABLE_JIT == 0) \
|
|
&& (WASM_ENABLE_FAST_JIT == 0)
|
|
if (local_offset < 0x80) {
|
|
*p_org++ = EXT_OP_GET_LOCAL_FAST;
|
|
if (is_32bit_type(local_type))
|
|
*p_org++ = (uint8)local_offset;
|
|
else
|
|
*p_org++ = (uint8)(local_offset | 0x80);
|
|
while (p_org < p)
|
|
*p_org++ = WASM_OP_NOP;
|
|
}
|
|
#endif
|
|
#endif
|
|
break;
|
|
}
|
|
|
|
case WASM_OP_SET_LOCAL:
|
|
{
|
|
p_org = p - 1;
|
|
GET_LOCAL_INDEX_TYPE_AND_OFFSET();
|
|
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
if (!(preserve_referenced_local(
|
|
loader_ctx, opcode, local_offset, local_type,
|
|
&preserve_local, error_buf, error_buf_size)))
|
|
goto fail;
|
|
|
|
if (local_offset < 256) {
|
|
skip_label();
|
|
if ((!preserve_local) && (LAST_OP_OUTPUT_I32())) {
|
|
if (loader_ctx->p_code_compiled)
|
|
STORE_U16(loader_ctx->p_code_compiled - 2,
|
|
local_offset);
|
|
loader_ctx->frame_offset--;
|
|
loader_ctx->dynamic_offset--;
|
|
}
|
|
else if ((!preserve_local) && (LAST_OP_OUTPUT_I64())) {
|
|
if (loader_ctx->p_code_compiled)
|
|
STORE_U16(loader_ctx->p_code_compiled - 2,
|
|
local_offset);
|
|
loader_ctx->frame_offset -= 2;
|
|
loader_ctx->dynamic_offset -= 2;
|
|
}
|
|
else {
|
|
if (is_32bit_type(local_type)) {
|
|
emit_label(EXT_OP_SET_LOCAL_FAST);
|
|
emit_byte(loader_ctx, (uint8)local_offset);
|
|
}
|
|
else {
|
|
emit_label(EXT_OP_SET_LOCAL_FAST_I64);
|
|
emit_byte(loader_ctx, (uint8)local_offset);
|
|
}
|
|
POP_OFFSET_TYPE(local_type);
|
|
}
|
|
}
|
|
else { /* local index larger than 255, reserve leb */
|
|
emit_uint32(loader_ctx, local_idx);
|
|
POP_OFFSET_TYPE(local_type);
|
|
}
|
|
#else
|
|
#if (WASM_ENABLE_WAMR_COMPILER == 0) && (WASM_ENABLE_JIT == 0) \
|
|
&& (WASM_ENABLE_FAST_JIT == 0)
|
|
if (local_offset < 0x80) {
|
|
*p_org++ = EXT_OP_SET_LOCAL_FAST;
|
|
if (is_32bit_type(local_type))
|
|
*p_org++ = (uint8)local_offset;
|
|
else
|
|
*p_org++ = (uint8)(local_offset | 0x80);
|
|
while (p_org < p)
|
|
*p_org++ = WASM_OP_NOP;
|
|
}
|
|
#endif
|
|
#endif
|
|
POP_TYPE(local_type);
|
|
break;
|
|
}
|
|
|
|
case WASM_OP_TEE_LOCAL:
|
|
{
|
|
p_org = p - 1;
|
|
GET_LOCAL_INDEX_TYPE_AND_OFFSET();
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
/* If the stack is in polymorphic state, do fake pop and push on
|
|
offset stack to keep the depth of offset stack to be the
|
|
same with ref stack */
|
|
BranchBlock *cur_block = loader_ctx->frame_csp - 1;
|
|
if (cur_block->is_stack_polymorphic) {
|
|
POP_OFFSET_TYPE(local_type);
|
|
PUSH_OFFSET_TYPE(local_type);
|
|
}
|
|
#endif
|
|
POP_TYPE(local_type);
|
|
PUSH_TYPE(local_type);
|
|
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
if (!(preserve_referenced_local(
|
|
loader_ctx, opcode, local_offset, local_type,
|
|
&preserve_local, error_buf, error_buf_size)))
|
|
goto fail;
|
|
|
|
if (local_offset < 256) {
|
|
skip_label();
|
|
if (is_32bit_type(local_type)) {
|
|
emit_label(EXT_OP_TEE_LOCAL_FAST);
|
|
emit_byte(loader_ctx, (uint8)local_offset);
|
|
}
|
|
else {
|
|
emit_label(EXT_OP_TEE_LOCAL_FAST_I64);
|
|
emit_byte(loader_ctx, (uint8)local_offset);
|
|
}
|
|
}
|
|
else { /* local index larger than 255, reserve leb */
|
|
emit_uint32(loader_ctx, local_idx);
|
|
}
|
|
emit_operand(loader_ctx,
|
|
*(loader_ctx->frame_offset
|
|
- wasm_value_type_cell_num(local_type)));
|
|
#else
|
|
#if (WASM_ENABLE_WAMR_COMPILER == 0) && (WASM_ENABLE_JIT == 0) \
|
|
&& (WASM_ENABLE_FAST_JIT == 0)
|
|
if (local_offset < 0x80) {
|
|
*p_org++ = EXT_OP_TEE_LOCAL_FAST;
|
|
if (is_32bit_type(local_type))
|
|
*p_org++ = (uint8)local_offset;
|
|
else
|
|
*p_org++ = (uint8)(local_offset | 0x80);
|
|
while (p_org < p)
|
|
*p_org++ = WASM_OP_NOP;
|
|
}
|
|
#endif
|
|
#endif
|
|
break;
|
|
}
|
|
|
|
case WASM_OP_GET_GLOBAL:
|
|
{
|
|
p_org = p - 1;
|
|
read_leb_uint32(p, p_end, global_idx);
|
|
bh_assert(global_idx < global_count);
|
|
|
|
global_type = global_idx < module->import_global_count
|
|
? module->import_globals[global_idx]
|
|
.u.global.type.val_type
|
|
: module
|
|
->globals[global_idx
|
|
- module->import_global_count]
|
|
.type.val_type;
|
|
|
|
PUSH_TYPE(global_type);
|
|
|
|
#if WASM_ENABLE_FAST_INTERP == 0
|
|
if (global_type == VALUE_TYPE_I64
|
|
|| global_type == VALUE_TYPE_F64) {
|
|
*p_org = WASM_OP_GET_GLOBAL_64;
|
|
}
|
|
#else /* else of WASM_ENABLE_FAST_INTERP */
|
|
if (is_64bit_type(global_type)) {
|
|
skip_label();
|
|
emit_label(WASM_OP_GET_GLOBAL_64);
|
|
}
|
|
emit_uint32(loader_ctx, global_idx);
|
|
PUSH_OFFSET_TYPE(global_type);
|
|
#endif /* end of WASM_ENABLE_FAST_INTERP */
|
|
break;
|
|
}
|
|
|
|
case WASM_OP_SET_GLOBAL:
|
|
{
|
|
bool is_mutable = false;
|
|
|
|
p_org = p - 1;
|
|
read_leb_uint32(p, p_end, global_idx);
|
|
bh_assert(global_idx < global_count);
|
|
|
|
is_mutable = global_idx < module->import_global_count
|
|
? module->import_globals[global_idx]
|
|
.u.global.type.is_mutable
|
|
: module
|
|
->globals[global_idx
|
|
- module->import_global_count]
|
|
.type.is_mutable;
|
|
bh_assert(is_mutable);
|
|
|
|
global_type = global_idx < module->import_global_count
|
|
? module->import_globals[global_idx]
|
|
.u.global.type.val_type
|
|
: module
|
|
->globals[global_idx
|
|
- module->import_global_count]
|
|
.type.val_type;
|
|
|
|
#if WASM_ENABLE_FAST_INTERP == 0
|
|
if (is_64bit_type(global_type)) {
|
|
*p_org = WASM_OP_SET_GLOBAL_64;
|
|
}
|
|
else if (module->aux_stack_size > 0
|
|
&& global_idx == module->aux_stack_top_global_index) {
|
|
*p_org = WASM_OP_SET_GLOBAL_AUX_STACK;
|
|
#if WASM_ENABLE_JIT != 0 || WASM_ENABLE_WAMR_COMPILER != 0
|
|
func->has_op_set_global_aux_stack = true;
|
|
#endif
|
|
}
|
|
#else /* else of WASM_ENABLE_FAST_INTERP */
|
|
if (is_64bit_type(global_type)) {
|
|
skip_label();
|
|
emit_label(WASM_OP_SET_GLOBAL_64);
|
|
}
|
|
else if (module->aux_stack_size > 0
|
|
&& global_idx == module->aux_stack_top_global_index) {
|
|
skip_label();
|
|
emit_label(WASM_OP_SET_GLOBAL_AUX_STACK);
|
|
}
|
|
emit_uint32(loader_ctx, global_idx);
|
|
POP_OFFSET_TYPE(global_type);
|
|
#endif /* end of WASM_ENABLE_FAST_INTERP */
|
|
|
|
POP_TYPE(global_type);
|
|
|
|
(void)is_mutable;
|
|
break;
|
|
}
|
|
|
|
/* load */
|
|
case WASM_OP_I32_LOAD:
|
|
case WASM_OP_I32_LOAD8_S:
|
|
case WASM_OP_I32_LOAD8_U:
|
|
case WASM_OP_I32_LOAD16_S:
|
|
case WASM_OP_I32_LOAD16_U:
|
|
case WASM_OP_I64_LOAD:
|
|
case WASM_OP_I64_LOAD8_S:
|
|
case WASM_OP_I64_LOAD8_U:
|
|
case WASM_OP_I64_LOAD16_S:
|
|
case WASM_OP_I64_LOAD16_U:
|
|
case WASM_OP_I64_LOAD32_S:
|
|
case WASM_OP_I64_LOAD32_U:
|
|
case WASM_OP_F32_LOAD:
|
|
case WASM_OP_F64_LOAD:
|
|
/* store */
|
|
case WASM_OP_I32_STORE:
|
|
case WASM_OP_I32_STORE8:
|
|
case WASM_OP_I32_STORE16:
|
|
case WASM_OP_I64_STORE:
|
|
case WASM_OP_I64_STORE8:
|
|
case WASM_OP_I64_STORE16:
|
|
case WASM_OP_I64_STORE32:
|
|
case WASM_OP_F32_STORE:
|
|
case WASM_OP_F64_STORE:
|
|
{
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
/* change F32/F64 into I32/I64 */
|
|
if (opcode == WASM_OP_F32_LOAD) {
|
|
skip_label();
|
|
emit_label(WASM_OP_I32_LOAD);
|
|
}
|
|
else if (opcode == WASM_OP_F64_LOAD) {
|
|
skip_label();
|
|
emit_label(WASM_OP_I64_LOAD);
|
|
}
|
|
else if (opcode == WASM_OP_F32_STORE) {
|
|
skip_label();
|
|
emit_label(WASM_OP_I32_STORE);
|
|
}
|
|
else if (opcode == WASM_OP_F64_STORE) {
|
|
skip_label();
|
|
emit_label(WASM_OP_I64_STORE);
|
|
}
|
|
#endif
|
|
CHECK_MEMORY();
|
|
read_leb_uint32(p, p_end, align); /* align */
|
|
read_leb_mem_offset(p, p_end, mem_offset); /* offset */
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
emit_uint32(loader_ctx, mem_offset);
|
|
#endif
|
|
#if WASM_ENABLE_JIT != 0 || WASM_ENABLE_WAMR_COMPILER != 0
|
|
func->has_memory_operations = true;
|
|
#endif
|
|
switch (opcode) {
|
|
/* load */
|
|
case WASM_OP_I32_LOAD:
|
|
case WASM_OP_I32_LOAD8_S:
|
|
case WASM_OP_I32_LOAD8_U:
|
|
case WASM_OP_I32_LOAD16_S:
|
|
case WASM_OP_I32_LOAD16_U:
|
|
POP_AND_PUSH(mem_offset_type, VALUE_TYPE_I32);
|
|
break;
|
|
case WASM_OP_I64_LOAD:
|
|
case WASM_OP_I64_LOAD8_S:
|
|
case WASM_OP_I64_LOAD8_U:
|
|
case WASM_OP_I64_LOAD16_S:
|
|
case WASM_OP_I64_LOAD16_U:
|
|
case WASM_OP_I64_LOAD32_S:
|
|
case WASM_OP_I64_LOAD32_U:
|
|
POP_AND_PUSH(mem_offset_type, VALUE_TYPE_I64);
|
|
break;
|
|
case WASM_OP_F32_LOAD:
|
|
POP_AND_PUSH(mem_offset_type, VALUE_TYPE_F32);
|
|
break;
|
|
case WASM_OP_F64_LOAD:
|
|
POP_AND_PUSH(mem_offset_type, VALUE_TYPE_F64);
|
|
break;
|
|
/* store */
|
|
case WASM_OP_I32_STORE:
|
|
case WASM_OP_I32_STORE8:
|
|
case WASM_OP_I32_STORE16:
|
|
POP_I32();
|
|
POP_MEM_OFFSET();
|
|
break;
|
|
case WASM_OP_I64_STORE:
|
|
case WASM_OP_I64_STORE8:
|
|
case WASM_OP_I64_STORE16:
|
|
case WASM_OP_I64_STORE32:
|
|
POP_I64();
|
|
POP_MEM_OFFSET();
|
|
break;
|
|
case WASM_OP_F32_STORE:
|
|
POP_F32();
|
|
POP_MEM_OFFSET();
|
|
break;
|
|
case WASM_OP_F64_STORE:
|
|
POP_F64();
|
|
POP_MEM_OFFSET();
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
break;
|
|
}
|
|
|
|
case WASM_OP_MEMORY_SIZE:
|
|
CHECK_MEMORY();
|
|
/* reserved byte 0x00 */
|
|
bh_assert(*p == 0x00);
|
|
p++;
|
|
PUSH_PAGE_COUNT();
|
|
|
|
module->possible_memory_grow = true;
|
|
#if WASM_ENABLE_JIT != 0 || WASM_ENABLE_WAMR_COMPILER != 0
|
|
func->has_memory_operations = true;
|
|
#endif
|
|
break;
|
|
|
|
case WASM_OP_MEMORY_GROW:
|
|
CHECK_MEMORY();
|
|
/* reserved byte 0x00 */
|
|
bh_assert(*p == 0x00);
|
|
p++;
|
|
POP_AND_PUSH(mem_offset_type, mem_offset_type);
|
|
|
|
module->possible_memory_grow = true;
|
|
#if WASM_ENABLE_FAST_JIT != 0 || WASM_ENABLE_JIT != 0 \
|
|
|| WASM_ENABLE_WAMR_COMPILER != 0
|
|
func->has_op_memory_grow = true;
|
|
#endif
|
|
#if WASM_ENABLE_JIT != 0 || WASM_ENABLE_WAMR_COMPILER != 0
|
|
func->has_memory_operations = true;
|
|
#endif
|
|
break;
|
|
|
|
case WASM_OP_I32_CONST:
|
|
read_leb_int32(p, p_end, i32_const);
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
skip_label();
|
|
disable_emit = true;
|
|
GET_CONST_OFFSET(VALUE_TYPE_I32, i32_const);
|
|
|
|
if (operand_offset == 0) {
|
|
disable_emit = false;
|
|
emit_label(WASM_OP_I32_CONST);
|
|
emit_uint32(loader_ctx, i32_const);
|
|
}
|
|
#else
|
|
(void)i32_const;
|
|
#endif
|
|
PUSH_I32();
|
|
break;
|
|
|
|
case WASM_OP_I64_CONST:
|
|
read_leb_int64(p, p_end, i64_const);
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
skip_label();
|
|
disable_emit = true;
|
|
GET_CONST_OFFSET(VALUE_TYPE_I64, i64_const);
|
|
|
|
if (operand_offset == 0) {
|
|
disable_emit = false;
|
|
emit_label(WASM_OP_I64_CONST);
|
|
emit_uint64(loader_ctx, i64_const);
|
|
}
|
|
#endif
|
|
PUSH_I64();
|
|
break;
|
|
|
|
case WASM_OP_F32_CONST:
|
|
CHECK_BUF(p, p_end, sizeof(float32));
|
|
p += sizeof(float32);
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
skip_label();
|
|
disable_emit = true;
|
|
bh_memcpy_s((uint8 *)&f32_const, sizeof(float32), p_org,
|
|
sizeof(float32));
|
|
GET_CONST_F32_OFFSET(VALUE_TYPE_F32, f32_const);
|
|
|
|
if (operand_offset == 0) {
|
|
disable_emit = false;
|
|
emit_label(WASM_OP_F32_CONST);
|
|
emit_float32(loader_ctx, f32_const);
|
|
}
|
|
#endif
|
|
PUSH_F32();
|
|
break;
|
|
|
|
case WASM_OP_F64_CONST:
|
|
CHECK_BUF(p, p_end, sizeof(float64));
|
|
p += sizeof(float64);
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
skip_label();
|
|
disable_emit = true;
|
|
/* Some MCU may require 8-byte align */
|
|
bh_memcpy_s((uint8 *)&f64_const, sizeof(float64), p_org,
|
|
sizeof(float64));
|
|
GET_CONST_F64_OFFSET(VALUE_TYPE_F64, f64_const);
|
|
|
|
if (operand_offset == 0) {
|
|
disable_emit = false;
|
|
emit_label(WASM_OP_F64_CONST);
|
|
emit_float64(loader_ctx, f64_const);
|
|
}
|
|
#endif
|
|
PUSH_F64();
|
|
break;
|
|
|
|
case WASM_OP_I32_EQZ:
|
|
POP_AND_PUSH(VALUE_TYPE_I32, VALUE_TYPE_I32);
|
|
break;
|
|
|
|
case WASM_OP_I32_EQ:
|
|
case WASM_OP_I32_NE:
|
|
case WASM_OP_I32_LT_S:
|
|
case WASM_OP_I32_LT_U:
|
|
case WASM_OP_I32_GT_S:
|
|
case WASM_OP_I32_GT_U:
|
|
case WASM_OP_I32_LE_S:
|
|
case WASM_OP_I32_LE_U:
|
|
case WASM_OP_I32_GE_S:
|
|
case WASM_OP_I32_GE_U:
|
|
POP2_AND_PUSH(VALUE_TYPE_I32, VALUE_TYPE_I32);
|
|
break;
|
|
|
|
case WASM_OP_I64_EQZ:
|
|
POP_AND_PUSH(VALUE_TYPE_I64, VALUE_TYPE_I32);
|
|
break;
|
|
|
|
case WASM_OP_I64_EQ:
|
|
case WASM_OP_I64_NE:
|
|
case WASM_OP_I64_LT_S:
|
|
case WASM_OP_I64_LT_U:
|
|
case WASM_OP_I64_GT_S:
|
|
case WASM_OP_I64_GT_U:
|
|
case WASM_OP_I64_LE_S:
|
|
case WASM_OP_I64_LE_U:
|
|
case WASM_OP_I64_GE_S:
|
|
case WASM_OP_I64_GE_U:
|
|
POP2_AND_PUSH(VALUE_TYPE_I64, VALUE_TYPE_I32);
|
|
break;
|
|
|
|
case WASM_OP_F32_EQ:
|
|
case WASM_OP_F32_NE:
|
|
case WASM_OP_F32_LT:
|
|
case WASM_OP_F32_GT:
|
|
case WASM_OP_F32_LE:
|
|
case WASM_OP_F32_GE:
|
|
POP2_AND_PUSH(VALUE_TYPE_F32, VALUE_TYPE_I32);
|
|
break;
|
|
|
|
case WASM_OP_F64_EQ:
|
|
case WASM_OP_F64_NE:
|
|
case WASM_OP_F64_LT:
|
|
case WASM_OP_F64_GT:
|
|
case WASM_OP_F64_LE:
|
|
case WASM_OP_F64_GE:
|
|
POP2_AND_PUSH(VALUE_TYPE_F64, VALUE_TYPE_I32);
|
|
break;
|
|
|
|
case WASM_OP_I32_CLZ:
|
|
case WASM_OP_I32_CTZ:
|
|
case WASM_OP_I32_POPCNT:
|
|
POP_AND_PUSH(VALUE_TYPE_I32, VALUE_TYPE_I32);
|
|
break;
|
|
|
|
case WASM_OP_I32_ADD:
|
|
case WASM_OP_I32_SUB:
|
|
case WASM_OP_I32_MUL:
|
|
case WASM_OP_I32_DIV_S:
|
|
case WASM_OP_I32_DIV_U:
|
|
case WASM_OP_I32_REM_S:
|
|
case WASM_OP_I32_REM_U:
|
|
case WASM_OP_I32_AND:
|
|
case WASM_OP_I32_OR:
|
|
case WASM_OP_I32_XOR:
|
|
case WASM_OP_I32_SHL:
|
|
case WASM_OP_I32_SHR_S:
|
|
case WASM_OP_I32_SHR_U:
|
|
case WASM_OP_I32_ROTL:
|
|
case WASM_OP_I32_ROTR:
|
|
POP2_AND_PUSH(VALUE_TYPE_I32, VALUE_TYPE_I32);
|
|
break;
|
|
|
|
case WASM_OP_I64_CLZ:
|
|
case WASM_OP_I64_CTZ:
|
|
case WASM_OP_I64_POPCNT:
|
|
POP_AND_PUSH(VALUE_TYPE_I64, VALUE_TYPE_I64);
|
|
break;
|
|
|
|
case WASM_OP_I64_ADD:
|
|
case WASM_OP_I64_SUB:
|
|
case WASM_OP_I64_MUL:
|
|
case WASM_OP_I64_DIV_S:
|
|
case WASM_OP_I64_DIV_U:
|
|
case WASM_OP_I64_REM_S:
|
|
case WASM_OP_I64_REM_U:
|
|
case WASM_OP_I64_AND:
|
|
case WASM_OP_I64_OR:
|
|
case WASM_OP_I64_XOR:
|
|
case WASM_OP_I64_SHL:
|
|
case WASM_OP_I64_SHR_S:
|
|
case WASM_OP_I64_SHR_U:
|
|
case WASM_OP_I64_ROTL:
|
|
case WASM_OP_I64_ROTR:
|
|
POP2_AND_PUSH(VALUE_TYPE_I64, VALUE_TYPE_I64);
|
|
break;
|
|
|
|
case WASM_OP_F32_ABS:
|
|
case WASM_OP_F32_NEG:
|
|
case WASM_OP_F32_CEIL:
|
|
case WASM_OP_F32_FLOOR:
|
|
case WASM_OP_F32_TRUNC:
|
|
case WASM_OP_F32_NEAREST:
|
|
case WASM_OP_F32_SQRT:
|
|
POP_AND_PUSH(VALUE_TYPE_F32, VALUE_TYPE_F32);
|
|
break;
|
|
|
|
case WASM_OP_F32_ADD:
|
|
case WASM_OP_F32_SUB:
|
|
case WASM_OP_F32_MUL:
|
|
case WASM_OP_F32_DIV:
|
|
case WASM_OP_F32_MIN:
|
|
case WASM_OP_F32_MAX:
|
|
case WASM_OP_F32_COPYSIGN:
|
|
POP2_AND_PUSH(VALUE_TYPE_F32, VALUE_TYPE_F32);
|
|
break;
|
|
|
|
case WASM_OP_F64_ABS:
|
|
case WASM_OP_F64_NEG:
|
|
case WASM_OP_F64_CEIL:
|
|
case WASM_OP_F64_FLOOR:
|
|
case WASM_OP_F64_TRUNC:
|
|
case WASM_OP_F64_NEAREST:
|
|
case WASM_OP_F64_SQRT:
|
|
POP_AND_PUSH(VALUE_TYPE_F64, VALUE_TYPE_F64);
|
|
break;
|
|
|
|
case WASM_OP_F64_ADD:
|
|
case WASM_OP_F64_SUB:
|
|
case WASM_OP_F64_MUL:
|
|
case WASM_OP_F64_DIV:
|
|
case WASM_OP_F64_MIN:
|
|
case WASM_OP_F64_MAX:
|
|
case WASM_OP_F64_COPYSIGN:
|
|
POP2_AND_PUSH(VALUE_TYPE_F64, VALUE_TYPE_F64);
|
|
break;
|
|
|
|
case WASM_OP_I32_WRAP_I64:
|
|
POP_AND_PUSH(VALUE_TYPE_I64, VALUE_TYPE_I32);
|
|
break;
|
|
|
|
case WASM_OP_I32_TRUNC_S_F32:
|
|
case WASM_OP_I32_TRUNC_U_F32:
|
|
POP_AND_PUSH(VALUE_TYPE_F32, VALUE_TYPE_I32);
|
|
break;
|
|
|
|
case WASM_OP_I32_TRUNC_S_F64:
|
|
case WASM_OP_I32_TRUNC_U_F64:
|
|
POP_AND_PUSH(VALUE_TYPE_F64, VALUE_TYPE_I32);
|
|
break;
|
|
|
|
case WASM_OP_I64_EXTEND_S_I32:
|
|
case WASM_OP_I64_EXTEND_U_I32:
|
|
POP_AND_PUSH(VALUE_TYPE_I32, VALUE_TYPE_I64);
|
|
break;
|
|
|
|
case WASM_OP_I64_TRUNC_S_F32:
|
|
case WASM_OP_I64_TRUNC_U_F32:
|
|
POP_AND_PUSH(VALUE_TYPE_F32, VALUE_TYPE_I64);
|
|
break;
|
|
|
|
case WASM_OP_I64_TRUNC_S_F64:
|
|
case WASM_OP_I64_TRUNC_U_F64:
|
|
POP_AND_PUSH(VALUE_TYPE_F64, VALUE_TYPE_I64);
|
|
break;
|
|
|
|
case WASM_OP_F32_CONVERT_S_I32:
|
|
case WASM_OP_F32_CONVERT_U_I32:
|
|
POP_AND_PUSH(VALUE_TYPE_I32, VALUE_TYPE_F32);
|
|
break;
|
|
|
|
case WASM_OP_F32_CONVERT_S_I64:
|
|
case WASM_OP_F32_CONVERT_U_I64:
|
|
POP_AND_PUSH(VALUE_TYPE_I64, VALUE_TYPE_F32);
|
|
break;
|
|
|
|
case WASM_OP_F32_DEMOTE_F64:
|
|
POP_AND_PUSH(VALUE_TYPE_F64, VALUE_TYPE_F32);
|
|
break;
|
|
|
|
case WASM_OP_F64_CONVERT_S_I32:
|
|
case WASM_OP_F64_CONVERT_U_I32:
|
|
POP_AND_PUSH(VALUE_TYPE_I32, VALUE_TYPE_F64);
|
|
break;
|
|
|
|
case WASM_OP_F64_CONVERT_S_I64:
|
|
case WASM_OP_F64_CONVERT_U_I64:
|
|
POP_AND_PUSH(VALUE_TYPE_I64, VALUE_TYPE_F64);
|
|
break;
|
|
|
|
case WASM_OP_F64_PROMOTE_F32:
|
|
POP_AND_PUSH(VALUE_TYPE_F32, VALUE_TYPE_F64);
|
|
break;
|
|
|
|
case WASM_OP_I32_REINTERPRET_F32:
|
|
POP_AND_PUSH(VALUE_TYPE_F32, VALUE_TYPE_I32);
|
|
break;
|
|
|
|
case WASM_OP_I64_REINTERPRET_F64:
|
|
POP_AND_PUSH(VALUE_TYPE_F64, VALUE_TYPE_I64);
|
|
break;
|
|
|
|
case WASM_OP_F32_REINTERPRET_I32:
|
|
POP_AND_PUSH(VALUE_TYPE_I32, VALUE_TYPE_F32);
|
|
break;
|
|
|
|
case WASM_OP_F64_REINTERPRET_I64:
|
|
POP_AND_PUSH(VALUE_TYPE_I64, VALUE_TYPE_F64);
|
|
break;
|
|
|
|
case WASM_OP_I32_EXTEND8_S:
|
|
case WASM_OP_I32_EXTEND16_S:
|
|
POP_AND_PUSH(VALUE_TYPE_I32, VALUE_TYPE_I32);
|
|
break;
|
|
|
|
case WASM_OP_I64_EXTEND8_S:
|
|
case WASM_OP_I64_EXTEND16_S:
|
|
case WASM_OP_I64_EXTEND32_S:
|
|
POP_AND_PUSH(VALUE_TYPE_I64, VALUE_TYPE_I64);
|
|
break;
|
|
|
|
case WASM_OP_MISC_PREFIX:
|
|
{
|
|
uint32 opcode1;
|
|
|
|
read_leb_uint32(p, p_end, opcode1);
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
emit_byte(loader_ctx, ((uint8)opcode1));
|
|
#endif
|
|
switch (opcode1) {
|
|
case WASM_OP_I32_TRUNC_SAT_S_F32:
|
|
case WASM_OP_I32_TRUNC_SAT_U_F32:
|
|
POP_AND_PUSH(VALUE_TYPE_F32, VALUE_TYPE_I32);
|
|
break;
|
|
case WASM_OP_I32_TRUNC_SAT_S_F64:
|
|
case WASM_OP_I32_TRUNC_SAT_U_F64:
|
|
POP_AND_PUSH(VALUE_TYPE_F64, VALUE_TYPE_I32);
|
|
break;
|
|
case WASM_OP_I64_TRUNC_SAT_S_F32:
|
|
case WASM_OP_I64_TRUNC_SAT_U_F32:
|
|
POP_AND_PUSH(VALUE_TYPE_F32, VALUE_TYPE_I64);
|
|
break;
|
|
case WASM_OP_I64_TRUNC_SAT_S_F64:
|
|
case WASM_OP_I64_TRUNC_SAT_U_F64:
|
|
POP_AND_PUSH(VALUE_TYPE_F64, VALUE_TYPE_I64);
|
|
break;
|
|
#if WASM_ENABLE_BULK_MEMORY != 0
|
|
case WASM_OP_MEMORY_INIT:
|
|
{
|
|
read_leb_uint32(p, p_end, segment_index);
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
emit_uint32(loader_ctx, segment_index);
|
|
#endif
|
|
bh_assert(module->import_memory_count
|
|
+ module->memory_count
|
|
> 0);
|
|
|
|
bh_assert(*p == 0x00);
|
|
p++;
|
|
|
|
bh_assert(segment_index < module->data_seg_count);
|
|
bh_assert(module->data_seg_count1 > 0);
|
|
|
|
POP_I32();
|
|
POP_I32();
|
|
POP_MEM_OFFSET();
|
|
#if WASM_ENABLE_JIT != 0 || WASM_ENABLE_WAMR_COMPILER != 0
|
|
func->has_memory_operations = true;
|
|
#endif
|
|
break;
|
|
}
|
|
case WASM_OP_DATA_DROP:
|
|
{
|
|
read_leb_uint32(p, p_end, segment_index);
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
emit_uint32(loader_ctx, segment_index);
|
|
#endif
|
|
bh_assert(segment_index < module->data_seg_count);
|
|
bh_assert(module->data_seg_count1 > 0);
|
|
#if WASM_ENABLE_JIT != 0 || WASM_ENABLE_WAMR_COMPILER != 0
|
|
func->has_memory_operations = true;
|
|
#endif
|
|
break;
|
|
}
|
|
case WASM_OP_MEMORY_COPY:
|
|
{
|
|
CHECK_BUF(p, p_end, sizeof(int16));
|
|
/* both src and dst memory index should be 0 */
|
|
bh_assert(*(int16 *)p == 0x0000);
|
|
p += 2;
|
|
|
|
bh_assert(module->import_memory_count
|
|
+ module->memory_count
|
|
> 0);
|
|
|
|
POP_MEM_OFFSET();
|
|
POP_MEM_OFFSET();
|
|
POP_MEM_OFFSET();
|
|
#if WASM_ENABLE_JIT != 0 || WASM_ENABLE_WAMR_COMPILER != 0
|
|
func->has_memory_operations = true;
|
|
#endif
|
|
break;
|
|
}
|
|
case WASM_OP_MEMORY_FILL:
|
|
{
|
|
bh_assert(*p == 0);
|
|
p++;
|
|
|
|
bh_assert(module->import_memory_count
|
|
+ module->memory_count
|
|
> 0);
|
|
|
|
POP_MEM_OFFSET();
|
|
POP_I32();
|
|
POP_MEM_OFFSET();
|
|
#if WASM_ENABLE_JIT != 0 || WASM_ENABLE_WAMR_COMPILER != 0
|
|
func->has_memory_operations = true;
|
|
#endif
|
|
break;
|
|
}
|
|
#endif /* WASM_ENABLE_BULK_MEMORY */
|
|
#if WASM_ENABLE_REF_TYPES != 0
|
|
case WASM_OP_TABLE_INIT:
|
|
{
|
|
uint8 seg_ref_type, tbl_ref_type;
|
|
uint32 table_seg_idx, table_idx;
|
|
|
|
read_leb_uint32(p, p_end, table_seg_idx);
|
|
read_leb_uint32(p, p_end, table_idx);
|
|
|
|
if (!get_table_elem_type(module, table_idx,
|
|
&tbl_ref_type, error_buf,
|
|
error_buf_size))
|
|
goto fail;
|
|
|
|
if (!get_table_seg_elem_type(module, table_seg_idx,
|
|
&seg_ref_type, error_buf,
|
|
error_buf_size))
|
|
goto fail;
|
|
|
|
if (seg_ref_type != tbl_ref_type) {
|
|
set_error_buf(error_buf, error_buf_size,
|
|
"type mismatch");
|
|
goto fail;
|
|
}
|
|
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
emit_uint32(loader_ctx, table_seg_idx);
|
|
emit_uint32(loader_ctx, table_idx);
|
|
#endif
|
|
POP_I32();
|
|
POP_I32();
|
|
POP_I32();
|
|
break;
|
|
}
|
|
case WASM_OP_ELEM_DROP:
|
|
{
|
|
uint32 table_seg_idx;
|
|
read_leb_uint32(p, p_end, table_seg_idx);
|
|
if (!get_table_seg_elem_type(module, table_seg_idx,
|
|
NULL, error_buf,
|
|
error_buf_size))
|
|
goto fail;
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
emit_uint32(loader_ctx, table_seg_idx);
|
|
#endif
|
|
break;
|
|
}
|
|
case WASM_OP_TABLE_COPY:
|
|
{
|
|
uint8 src_ref_type, dst_ref_type;
|
|
uint32 src_tbl_idx, dst_tbl_idx;
|
|
|
|
read_leb_uint32(p, p_end, src_tbl_idx);
|
|
if (!get_table_elem_type(module, src_tbl_idx,
|
|
&src_ref_type, error_buf,
|
|
error_buf_size))
|
|
goto fail;
|
|
|
|
read_leb_uint32(p, p_end, dst_tbl_idx);
|
|
if (!get_table_elem_type(module, dst_tbl_idx,
|
|
&dst_ref_type, error_buf,
|
|
error_buf_size))
|
|
goto fail;
|
|
|
|
if (src_ref_type != dst_ref_type) {
|
|
set_error_buf(error_buf, error_buf_size,
|
|
"type mismatch");
|
|
goto fail;
|
|
}
|
|
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
emit_uint32(loader_ctx, src_tbl_idx);
|
|
emit_uint32(loader_ctx, dst_tbl_idx);
|
|
#endif
|
|
POP_I32();
|
|
POP_I32();
|
|
POP_I32();
|
|
break;
|
|
}
|
|
case WASM_OP_TABLE_SIZE:
|
|
{
|
|
uint32 table_idx;
|
|
|
|
read_leb_uint32(p, p_end, table_idx);
|
|
/* TODO: shall we create a new function to check
|
|
table idx instead of using below function? */
|
|
if (!get_table_elem_type(module, table_idx, NULL,
|
|
error_buf, error_buf_size))
|
|
goto fail;
|
|
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
emit_uint32(loader_ctx, table_idx);
|
|
#endif
|
|
|
|
PUSH_I32();
|
|
break;
|
|
}
|
|
case WASM_OP_TABLE_GROW:
|
|
case WASM_OP_TABLE_FILL:
|
|
{
|
|
uint8 decl_ref_type;
|
|
uint32 table_idx;
|
|
|
|
read_leb_uint32(p, p_end, table_idx);
|
|
if (!get_table_elem_type(module, table_idx,
|
|
&decl_ref_type, error_buf,
|
|
error_buf_size))
|
|
goto fail;
|
|
|
|
if (opcode1 == WASM_OP_TABLE_GROW) {
|
|
if (table_idx < module->import_table_count) {
|
|
module->import_tables[table_idx]
|
|
.u.table.possible_grow = true;
|
|
}
|
|
else {
|
|
module
|
|
->tables[table_idx
|
|
- module->import_table_count]
|
|
.possible_grow = true;
|
|
}
|
|
}
|
|
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
emit_uint32(loader_ctx, table_idx);
|
|
#endif
|
|
|
|
POP_I32();
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
POP_OFFSET_TYPE(decl_ref_type);
|
|
#endif
|
|
POP_TYPE(decl_ref_type);
|
|
if (opcode1 == WASM_OP_TABLE_GROW)
|
|
PUSH_I32();
|
|
else
|
|
POP_I32();
|
|
break;
|
|
}
|
|
#endif /* WASM_ENABLE_REF_TYPES */
|
|
default:
|
|
bh_assert(0);
|
|
break;
|
|
}
|
|
break;
|
|
}
|
|
|
|
#if WASM_ENABLE_SHARED_MEMORY != 0
|
|
case WASM_OP_ATOMIC_PREFIX:
|
|
{
|
|
uint32 opcode1;
|
|
|
|
read_leb_uint32(p, p_end, opcode1);
|
|
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
emit_byte(loader_ctx, opcode1);
|
|
#endif
|
|
if (opcode1 != WASM_OP_ATOMIC_FENCE) {
|
|
CHECK_MEMORY();
|
|
read_leb_uint32(p, p_end, align); /* align */
|
|
read_leb_mem_offset(p, p_end, mem_offset); /* offset */
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
emit_uint32(loader_ctx, mem_offset);
|
|
#endif
|
|
}
|
|
#if WASM_ENABLE_JIT != 0 || WASM_ENABLE_WAMR_COMPILER != 0
|
|
func->has_memory_operations = true;
|
|
#endif
|
|
switch (opcode1) {
|
|
case WASM_OP_ATOMIC_NOTIFY:
|
|
POP_I32();
|
|
POP_MEM_OFFSET();
|
|
PUSH_I32();
|
|
break;
|
|
case WASM_OP_ATOMIC_WAIT32:
|
|
POP_I64();
|
|
POP_I32();
|
|
POP_MEM_OFFSET();
|
|
PUSH_I32();
|
|
break;
|
|
case WASM_OP_ATOMIC_WAIT64:
|
|
POP_I64();
|
|
POP_I64();
|
|
POP_MEM_OFFSET();
|
|
PUSH_I32();
|
|
break;
|
|
case WASM_OP_ATOMIC_FENCE:
|
|
/* reserved byte 0x00 */
|
|
bh_assert(*p == 0x00);
|
|
p++;
|
|
break;
|
|
case WASM_OP_ATOMIC_I32_LOAD:
|
|
case WASM_OP_ATOMIC_I32_LOAD8_U:
|
|
case WASM_OP_ATOMIC_I32_LOAD16_U:
|
|
POP_AND_PUSH(mem_offset_type, VALUE_TYPE_I32);
|
|
break;
|
|
case WASM_OP_ATOMIC_I32_STORE:
|
|
case WASM_OP_ATOMIC_I32_STORE8:
|
|
case WASM_OP_ATOMIC_I32_STORE16:
|
|
POP_I32();
|
|
POP_MEM_OFFSET();
|
|
break;
|
|
case WASM_OP_ATOMIC_I64_LOAD:
|
|
case WASM_OP_ATOMIC_I64_LOAD8_U:
|
|
case WASM_OP_ATOMIC_I64_LOAD16_U:
|
|
case WASM_OP_ATOMIC_I64_LOAD32_U:
|
|
POP_AND_PUSH(mem_offset_type, VALUE_TYPE_I64);
|
|
break;
|
|
case WASM_OP_ATOMIC_I64_STORE:
|
|
case WASM_OP_ATOMIC_I64_STORE8:
|
|
case WASM_OP_ATOMIC_I64_STORE16:
|
|
case WASM_OP_ATOMIC_I64_STORE32:
|
|
POP_I64();
|
|
POP_MEM_OFFSET();
|
|
break;
|
|
case WASM_OP_ATOMIC_RMW_I32_ADD:
|
|
case WASM_OP_ATOMIC_RMW_I32_ADD8_U:
|
|
case WASM_OP_ATOMIC_RMW_I32_ADD16_U:
|
|
case WASM_OP_ATOMIC_RMW_I32_SUB:
|
|
case WASM_OP_ATOMIC_RMW_I32_SUB8_U:
|
|
case WASM_OP_ATOMIC_RMW_I32_SUB16_U:
|
|
case WASM_OP_ATOMIC_RMW_I32_AND:
|
|
case WASM_OP_ATOMIC_RMW_I32_AND8_U:
|
|
case WASM_OP_ATOMIC_RMW_I32_AND16_U:
|
|
case WASM_OP_ATOMIC_RMW_I32_OR:
|
|
case WASM_OP_ATOMIC_RMW_I32_OR8_U:
|
|
case WASM_OP_ATOMIC_RMW_I32_OR16_U:
|
|
case WASM_OP_ATOMIC_RMW_I32_XOR:
|
|
case WASM_OP_ATOMIC_RMW_I32_XOR8_U:
|
|
case WASM_OP_ATOMIC_RMW_I32_XOR16_U:
|
|
case WASM_OP_ATOMIC_RMW_I32_XCHG:
|
|
case WASM_OP_ATOMIC_RMW_I32_XCHG8_U:
|
|
case WASM_OP_ATOMIC_RMW_I32_XCHG16_U:
|
|
POP_I32();
|
|
POP_MEM_OFFSET();
|
|
PUSH_I32();
|
|
break;
|
|
case WASM_OP_ATOMIC_RMW_I64_ADD:
|
|
case WASM_OP_ATOMIC_RMW_I64_ADD8_U:
|
|
case WASM_OP_ATOMIC_RMW_I64_ADD16_U:
|
|
case WASM_OP_ATOMIC_RMW_I64_ADD32_U:
|
|
case WASM_OP_ATOMIC_RMW_I64_SUB:
|
|
case WASM_OP_ATOMIC_RMW_I64_SUB8_U:
|
|
case WASM_OP_ATOMIC_RMW_I64_SUB16_U:
|
|
case WASM_OP_ATOMIC_RMW_I64_SUB32_U:
|
|
case WASM_OP_ATOMIC_RMW_I64_AND:
|
|
case WASM_OP_ATOMIC_RMW_I64_AND8_U:
|
|
case WASM_OP_ATOMIC_RMW_I64_AND16_U:
|
|
case WASM_OP_ATOMIC_RMW_I64_AND32_U:
|
|
case WASM_OP_ATOMIC_RMW_I64_OR:
|
|
case WASM_OP_ATOMIC_RMW_I64_OR8_U:
|
|
case WASM_OP_ATOMIC_RMW_I64_OR16_U:
|
|
case WASM_OP_ATOMIC_RMW_I64_OR32_U:
|
|
case WASM_OP_ATOMIC_RMW_I64_XOR:
|
|
case WASM_OP_ATOMIC_RMW_I64_XOR8_U:
|
|
case WASM_OP_ATOMIC_RMW_I64_XOR16_U:
|
|
case WASM_OP_ATOMIC_RMW_I64_XOR32_U:
|
|
case WASM_OP_ATOMIC_RMW_I64_XCHG:
|
|
case WASM_OP_ATOMIC_RMW_I64_XCHG8_U:
|
|
case WASM_OP_ATOMIC_RMW_I64_XCHG16_U:
|
|
case WASM_OP_ATOMIC_RMW_I64_XCHG32_U:
|
|
POP_I64();
|
|
POP_MEM_OFFSET();
|
|
PUSH_I64();
|
|
break;
|
|
case WASM_OP_ATOMIC_RMW_I32_CMPXCHG:
|
|
case WASM_OP_ATOMIC_RMW_I32_CMPXCHG8_U:
|
|
case WASM_OP_ATOMIC_RMW_I32_CMPXCHG16_U:
|
|
POP_I32();
|
|
POP_I32();
|
|
POP_MEM_OFFSET();
|
|
PUSH_I32();
|
|
break;
|
|
case WASM_OP_ATOMIC_RMW_I64_CMPXCHG:
|
|
case WASM_OP_ATOMIC_RMW_I64_CMPXCHG8_U:
|
|
case WASM_OP_ATOMIC_RMW_I64_CMPXCHG16_U:
|
|
case WASM_OP_ATOMIC_RMW_I64_CMPXCHG32_U:
|
|
POP_I64();
|
|
POP_I64();
|
|
POP_MEM_OFFSET();
|
|
PUSH_I64();
|
|
break;
|
|
default:
|
|
bh_assert(0);
|
|
break;
|
|
}
|
|
break;
|
|
}
|
|
#endif /* end of WASM_ENABLE_SHARED_MEMORY */
|
|
|
|
default:
|
|
bh_assert(0);
|
|
break;
|
|
}
|
|
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
last_op = opcode;
|
|
#endif
|
|
}
|
|
|
|
if (loader_ctx->csp_num > 0) {
|
|
set_error_buf(error_buf, error_buf_size,
|
|
"function body must end with END opcode");
|
|
goto fail;
|
|
}
|
|
|
|
#if WASM_ENABLE_FAST_INTERP != 0
|
|
if (loader_ctx->p_code_compiled == NULL)
|
|
goto re_scan;
|
|
|
|
func->const_cell_num = loader_ctx->const_cell_num;
|
|
if (func->const_cell_num > 0) {
|
|
int32 j;
|
|
|
|
if (!(func->consts = func_const = loader_malloc(
|
|
func->const_cell_num * 4, error_buf, error_buf_size)))
|
|
goto fail;
|
|
|
|
func_const_end = func->consts + func->const_cell_num * 4;
|
|
/* reverse the const buf */
|
|
for (j = loader_ctx->num_const - 1; j >= 0; j--) {
|
|
Const *c = (Const *)(loader_ctx->const_buf + j * sizeof(Const));
|
|
if (c->value_type == VALUE_TYPE_F64
|
|
|| c->value_type == VALUE_TYPE_I64) {
|
|
bh_memcpy_s(func_const, (uint32)(func_const_end - func_const),
|
|
&(c->value.f64), (uint32)sizeof(int64));
|
|
func_const += sizeof(int64);
|
|
}
|
|
else {
|
|
bh_memcpy_s(func_const, (uint32)(func_const_end - func_const),
|
|
&(c->value.f32), (uint32)sizeof(int32));
|
|
func_const += sizeof(int32);
|
|
}
|
|
}
|
|
}
|
|
|
|
func->max_stack_cell_num = loader_ctx->preserved_local_offset
|
|
- loader_ctx->start_dynamic_offset + 1;
|
|
#else
|
|
func->max_stack_cell_num = loader_ctx->max_stack_cell_num;
|
|
#endif
|
|
func->max_block_num = loader_ctx->max_csp_num;
|
|
return_value = true;
|
|
|
|
fail:
|
|
wasm_loader_ctx_destroy(loader_ctx);
|
|
|
|
(void)u8;
|
|
(void)u32;
|
|
(void)i32;
|
|
(void)i64_const;
|
|
(void)global_count;
|
|
(void)local_count;
|
|
(void)local_offset;
|
|
(void)p_org;
|
|
(void)mem_offset;
|
|
(void)align;
|
|
#if WASM_ENABLE_BULK_MEMORY != 0
|
|
(void)segment_index;
|
|
#endif
|
|
return return_value;
|
|
}
|