mirror of
https://github.com/bytecodealliance/wasm-micro-runtime.git
synced 2025-05-14 21:51:27 +00:00
Implement xtensa XIP (#1202)
Lookup table for i32.const and i64.const for xtensa XIP Lookup const offset from table for load/store opcodes for xtensa XIP Fill capability flags for xtensa XIP Enable lower switch pass for xtensa XIP
This commit is contained in:
parent
b75ae3363d
commit
6922f3ac68
|
@ -61,8 +61,10 @@ static const aot_intrinsic g_intrinsic_mapping[] = {
|
||||||
{ "f64_promote_f32", "aot_intrinsic_f32_to_f64", AOT_INTRINSIC_FLAG_F32_TO_F64 },
|
{ "f64_promote_f32", "aot_intrinsic_f32_to_f64", AOT_INTRINSIC_FLAG_F32_TO_F64 },
|
||||||
{ "f32_cmp", "aot_intrinsic_f32_cmp", AOT_INTRINSIC_FLAG_F32_CMP },
|
{ "f32_cmp", "aot_intrinsic_f32_cmp", AOT_INTRINSIC_FLAG_F32_CMP },
|
||||||
{ "f64_cmp", "aot_intrinsic_f64_cmp", AOT_INTRINSIC_FLAG_F64_CMP },
|
{ "f64_cmp", "aot_intrinsic_f64_cmp", AOT_INTRINSIC_FLAG_F64_CMP },
|
||||||
{ "f32.const", NULL, AOT_INTRINSIC_FLAG_F32_CONST},
|
{ "i32.const", NULL, AOT_INTRINSIC_FLAG_I32_CONST },
|
||||||
{ "f64.const", NULL, AOT_INTRINSIC_FLAG_F64_CONST},
|
{ "i64.const", NULL, AOT_INTRINSIC_FLAG_I64_CONST },
|
||||||
|
{ "f32.const", NULL, AOT_INTRINSIC_FLAG_F32_CONST },
|
||||||
|
{ "f64.const", NULL, AOT_INTRINSIC_FLAG_F64_CONST },
|
||||||
};
|
};
|
||||||
/* clang-format on */
|
/* clang-format on */
|
||||||
|
|
||||||
|
@ -619,6 +621,19 @@ aot_intrinsic_fill_capability_flags(AOTCompContext *comp_ctx)
|
||||||
add_f64_common_intrinsics(comp_ctx);
|
add_f64_common_intrinsics(comp_ctx);
|
||||||
add_common_float_integer_convertion(comp_ctx);
|
add_common_float_integer_convertion(comp_ctx);
|
||||||
}
|
}
|
||||||
|
else if (!strncmp(comp_ctx->target_arch, "xtensa", 6)) {
|
||||||
|
/*
|
||||||
|
* Note: Use builtin intrinsics since hardware float operation
|
||||||
|
* will cause rodata relocation
|
||||||
|
*/
|
||||||
|
add_f32_common_intrinsics(comp_ctx);
|
||||||
|
add_f64_common_intrinsics(comp_ctx);
|
||||||
|
add_common_float_integer_convertion(comp_ctx);
|
||||||
|
add_intrinsic_capability(comp_ctx, AOT_INTRINSIC_FLAG_F32_CONST);
|
||||||
|
add_intrinsic_capability(comp_ctx, AOT_INTRINSIC_FLAG_F64_CONST);
|
||||||
|
add_intrinsic_capability(comp_ctx, AOT_INTRINSIC_FLAG_I32_CONST);
|
||||||
|
add_intrinsic_capability(comp_ctx, AOT_INTRINSIC_FLAG_I64_CONST);
|
||||||
|
}
|
||||||
else {
|
else {
|
||||||
/*
|
/*
|
||||||
* Use constant value table by default
|
* Use constant value table by default
|
||||||
|
|
|
@ -58,6 +58,7 @@ extern "C" {
|
||||||
#define AOT_INTRINSIC_FLAG_F32_TO_F64 AOT_INTRINSIC_FLAG(0, 24)
|
#define AOT_INTRINSIC_FLAG_F32_TO_F64 AOT_INTRINSIC_FLAG(0, 24)
|
||||||
#define AOT_INTRINSIC_FLAG_F32_CMP AOT_INTRINSIC_FLAG(0, 25)
|
#define AOT_INTRINSIC_FLAG_F32_CMP AOT_INTRINSIC_FLAG(0, 25)
|
||||||
#define AOT_INTRINSIC_FLAG_F32_CONST AOT_INTRINSIC_FLAG(0, 26)
|
#define AOT_INTRINSIC_FLAG_F32_CONST AOT_INTRINSIC_FLAG(0, 26)
|
||||||
|
#define AOT_INTRINSIC_FLAG_I32_CONST AOT_INTRINSIC_FLAG(0, 27)
|
||||||
|
|
||||||
#define AOT_INTRINSIC_FLAG_F64_FADD AOT_INTRINSIC_FLAG(1, 0)
|
#define AOT_INTRINSIC_FLAG_F64_FADD AOT_INTRINSIC_FLAG(1, 0)
|
||||||
#define AOT_INTRINSIC_FLAG_F64_FSUB AOT_INTRINSIC_FLAG(1, 1)
|
#define AOT_INTRINSIC_FLAG_F64_FSUB AOT_INTRINSIC_FLAG(1, 1)
|
||||||
|
@ -86,6 +87,7 @@ extern "C" {
|
||||||
#define AOT_INTRINSIC_FLAG_F64_TO_F32 AOT_INTRINSIC_FLAG(1, 24)
|
#define AOT_INTRINSIC_FLAG_F64_TO_F32 AOT_INTRINSIC_FLAG(1, 24)
|
||||||
#define AOT_INTRINSIC_FLAG_F64_CMP AOT_INTRINSIC_FLAG(1, 25)
|
#define AOT_INTRINSIC_FLAG_F64_CMP AOT_INTRINSIC_FLAG(1, 25)
|
||||||
#define AOT_INTRINSIC_FLAG_F64_CONST AOT_INTRINSIC_FLAG(1, 26)
|
#define AOT_INTRINSIC_FLAG_F64_CONST AOT_INTRINSIC_FLAG(1, 26)
|
||||||
|
#define AOT_INTRINSIC_FLAG_I64_CONST AOT_INTRINSIC_FLAG(1, 27)
|
||||||
/* clang-format on */
|
/* clang-format on */
|
||||||
|
|
||||||
float32
|
float32
|
||||||
|
|
|
@ -482,7 +482,7 @@ load_native_symbol_section(const uint8 *buf, const uint8 *buf_end,
|
||||||
|
|
||||||
for (i = cnt - 1; i >= 0; i--) {
|
for (i = cnt - 1; i >= 0; i--) {
|
||||||
read_string(p, p_end, symbol);
|
read_string(p, p_end, symbol);
|
||||||
if (!strncmp(symbol, "f32#", 4)) {
|
if (!strncmp(symbol, "f32#", 4) || !strncmp(symbol, "i32#", 4)) {
|
||||||
uint32 u32;
|
uint32 u32;
|
||||||
/* Resolve the raw int bits of f32 const */
|
/* Resolve the raw int bits of f32 const */
|
||||||
if (!str2uint32(symbol + 4, &u32)) {
|
if (!str2uint32(symbol + 4, &u32)) {
|
||||||
|
@ -492,7 +492,8 @@ load_native_symbol_section(const uint8 *buf, const uint8 *buf_end,
|
||||||
}
|
}
|
||||||
*(uint32 *)(&module->native_symbol_list[i]) = u32;
|
*(uint32 *)(&module->native_symbol_list[i]) = u32;
|
||||||
}
|
}
|
||||||
else if (!strncmp(symbol, "f64#", 4)) {
|
else if (!strncmp(symbol, "f64#", 4)
|
||||||
|
|| !strncmp(symbol, "i64#", 4)) {
|
||||||
uint64 u64;
|
uint64 u64;
|
||||||
/* Resolve the raw int bits of f64 const */
|
/* Resolve the raw int bits of f64 const */
|
||||||
if (!str2uint64(symbol + 4, &u64)) {
|
if (!str2uint64(symbol + 4, &u64)) {
|
||||||
|
|
|
@ -2742,8 +2742,9 @@ aot_require_lower_switch_pass(AOTCompContext *comp_ctx)
|
||||||
{
|
{
|
||||||
bool ret = false;
|
bool ret = false;
|
||||||
|
|
||||||
/* IR switch/case will cause .rodata relocation on riscv */
|
/* IR switch/case will cause .rodata relocation on riscv/xtensa */
|
||||||
if (!strncmp(comp_ctx->target_arch, "riscv", 5)) {
|
if (!strncmp(comp_ctx->target_arch, "riscv", 5)
|
||||||
|
|| !strncmp(comp_ctx->target_arch, "xtensa", 6)) {
|
||||||
ret = true;
|
ret = true;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -10,8 +10,23 @@ bool
|
||||||
aot_compile_op_i32_const(AOTCompContext *comp_ctx, AOTFuncContext *func_ctx,
|
aot_compile_op_i32_const(AOTCompContext *comp_ctx, AOTFuncContext *func_ctx,
|
||||||
int32 i32_const)
|
int32 i32_const)
|
||||||
{
|
{
|
||||||
LLVMValueRef value = I32_CONST((uint32)i32_const);
|
LLVMValueRef value;
|
||||||
CHECK_LLVM_CONST(value);
|
|
||||||
|
if (comp_ctx->is_indirect_mode
|
||||||
|
&& aot_intrinsic_check_capability(comp_ctx, "i32.const")) {
|
||||||
|
WASMValue wasm_value;
|
||||||
|
wasm_value.i32 = i32_const;
|
||||||
|
value = aot_load_const_from_table(comp_ctx, func_ctx->native_symbol,
|
||||||
|
&wasm_value, VALUE_TYPE_I32);
|
||||||
|
if (!value) {
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
else {
|
||||||
|
value = I32_CONST((uint32)i32_const);
|
||||||
|
CHECK_LLVM_CONST(value);
|
||||||
|
}
|
||||||
|
|
||||||
PUSH_I32(value);
|
PUSH_I32(value);
|
||||||
return true;
|
return true;
|
||||||
fail:
|
fail:
|
||||||
|
@ -22,8 +37,23 @@ bool
|
||||||
aot_compile_op_i64_const(AOTCompContext *comp_ctx, AOTFuncContext *func_ctx,
|
aot_compile_op_i64_const(AOTCompContext *comp_ctx, AOTFuncContext *func_ctx,
|
||||||
int64 i64_const)
|
int64 i64_const)
|
||||||
{
|
{
|
||||||
LLVMValueRef value = I64_CONST((uint64)i64_const);
|
LLVMValueRef value;
|
||||||
CHECK_LLVM_CONST(value);
|
|
||||||
|
if (comp_ctx->is_indirect_mode
|
||||||
|
&& aot_intrinsic_check_capability(comp_ctx, "i64.const")) {
|
||||||
|
WASMValue wasm_value;
|
||||||
|
wasm_value.i64 = i64_const;
|
||||||
|
value = aot_load_const_from_table(comp_ctx, func_ctx->native_symbol,
|
||||||
|
&wasm_value, VALUE_TYPE_I64);
|
||||||
|
if (!value) {
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
else {
|
||||||
|
value = I64_CONST((uint64)i64_const);
|
||||||
|
CHECK_LLVM_CONST(value);
|
||||||
|
}
|
||||||
|
|
||||||
PUSH_I64(value);
|
PUSH_I64(value);
|
||||||
return true;
|
return true;
|
||||||
fail:
|
fail:
|
||||||
|
|
|
@ -97,7 +97,19 @@ aot_check_memory_overflow(AOTCompContext *comp_ctx, AOTFuncContext *func_ctx,
|
||||||
|
|
||||||
is_target_64bit = (comp_ctx->pointer_size == sizeof(uint64)) ? true : false;
|
is_target_64bit = (comp_ctx->pointer_size == sizeof(uint64)) ? true : false;
|
||||||
|
|
||||||
CHECK_LLVM_CONST(offset_const);
|
if (comp_ctx->is_indirect_mode
|
||||||
|
&& aot_intrinsic_check_capability(comp_ctx, "i32.const")) {
|
||||||
|
WASMValue wasm_value;
|
||||||
|
wasm_value.i32 = offset;
|
||||||
|
offset_const = aot_load_const_from_table(
|
||||||
|
comp_ctx, func_ctx->native_symbol, &wasm_value, VALUE_TYPE_I32);
|
||||||
|
if (!offset_const) {
|
||||||
|
return NULL;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
else {
|
||||||
|
CHECK_LLVM_CONST(offset_const);
|
||||||
|
}
|
||||||
|
|
||||||
/* Get memory base address and memory data size */
|
/* Get memory base address and memory data size */
|
||||||
if (func_ctx->mem_space_unchanged
|
if (func_ctx->mem_space_unchanged
|
||||||
|
|
|
@ -800,9 +800,10 @@ is_targeting_soft_float(AOTCompContext *comp_ctx, bool is_f32)
|
||||||
* so user must specify '--cpu-features=+soft-float' to wamrc if the target
|
* so user must specify '--cpu-features=+soft-float' to wamrc if the target
|
||||||
* doesn't have or enable FPU on arm, x86 or mips. */
|
* doesn't have or enable FPU on arm, x86 or mips. */
|
||||||
if (is_target_arm(comp_ctx) || is_target_x86(comp_ctx)
|
if (is_target_arm(comp_ctx) || is_target_x86(comp_ctx)
|
||||||
|| is_target_mips(comp_ctx))
|
|| is_target_mips(comp_ctx)) {
|
||||||
ret = strstr(feature_string, "+soft-float") ? true : false;
|
ret = strstr(feature_string, "+soft-float") ? true : false;
|
||||||
else if (is_target_xtensa(comp_ctx))
|
}
|
||||||
|
else if (is_target_xtensa(comp_ctx)) {
|
||||||
/* Note:
|
/* Note:
|
||||||
* 1. The Floating-Point Coprocessor Option of xtensa only support
|
* 1. The Floating-Point Coprocessor Option of xtensa only support
|
||||||
* single-precision floating-point operations, so must use soft-float
|
* single-precision floating-point operations, so must use soft-float
|
||||||
|
@ -811,7 +812,11 @@ is_targeting_soft_float(AOTCompContext *comp_ctx, bool is_f32)
|
||||||
* so user must specify '--cpu-features=-fp' to wamrc if the target
|
* so user must specify '--cpu-features=-fp' to wamrc if the target
|
||||||
* doesn't have or enable Floating-Point Coprocessor Option on xtensa.
|
* doesn't have or enable Floating-Point Coprocessor Option on xtensa.
|
||||||
*/
|
*/
|
||||||
ret = (!is_f32 || strstr(feature_string, "-fp")) ? true : false;
|
if (comp_ctx->disable_llvm_intrinsics)
|
||||||
|
ret = false;
|
||||||
|
else
|
||||||
|
ret = (!is_f32 || strstr(feature_string, "-fp")) ? true : false;
|
||||||
|
}
|
||||||
else if (is_target_riscv(comp_ctx)) {
|
else if (is_target_riscv(comp_ctx)) {
|
||||||
/*
|
/*
|
||||||
* Note: Use builtin intrinsics since hardware float operation
|
* Note: Use builtin intrinsics since hardware float operation
|
||||||
|
@ -823,8 +828,9 @@ is_targeting_soft_float(AOTCompContext *comp_ctx, bool is_f32)
|
||||||
else
|
else
|
||||||
ret = !strstr(feature_string, "+d") ? true : false;
|
ret = !strstr(feature_string, "+d") ? true : false;
|
||||||
}
|
}
|
||||||
else
|
else {
|
||||||
ret = true;
|
ret = true;
|
||||||
|
}
|
||||||
|
|
||||||
LLVMDisposeMessage(feature_string);
|
LLVMDisposeMessage(feature_string);
|
||||||
return ret;
|
return ret;
|
||||||
|
|
|
@ -2731,6 +2731,16 @@ aot_load_const_from_table(AOTCompContext *comp_ctx, LLVMValueRef base,
|
||||||
int32 index;
|
int32 index;
|
||||||
|
|
||||||
switch (value_type) {
|
switch (value_type) {
|
||||||
|
case VALUE_TYPE_I32:
|
||||||
|
/* Store the raw int bits of i32 const as a hex string */
|
||||||
|
snprintf(buf, sizeof(buf), "i32#%08" PRIX32, value->i32);
|
||||||
|
const_ptr_type = INT32_PTR_TYPE;
|
||||||
|
break;
|
||||||
|
case VALUE_TYPE_I64:
|
||||||
|
/* Store the raw int bits of i64 const as a hex string */
|
||||||
|
snprintf(buf, sizeof(buf), "i64#%016" PRIX64, value->i64);
|
||||||
|
const_ptr_type = INT64_PTR_TYPE;
|
||||||
|
break;
|
||||||
case VALUE_TYPE_F32:
|
case VALUE_TYPE_F32:
|
||||||
/* Store the raw int bits of f32 const as a hex string */
|
/* Store the raw int bits of f32 const as a hex string */
|
||||||
snprintf(buf, sizeof(buf), "f32#%08" PRIX32, value->i32);
|
snprintf(buf, sizeof(buf), "f32#%08" PRIX32, value->i32);
|
||||||
|
|
|
@ -581,7 +581,7 @@ If you don't have it installed, follow the instructions [here](https://docs.espr
|
||||||
ESP-IDF also installs the toolchains needed for compiling WAMR and ESP-IDF.
|
ESP-IDF also installs the toolchains needed for compiling WAMR and ESP-IDF.
|
||||||
A small demonstration of how to use WAMR and ESP-IDF can be found under [product_mini](/product-mini/platforms/esp-idf).
|
A small demonstration of how to use WAMR and ESP-IDF can be found under [product_mini](/product-mini/platforms/esp-idf).
|
||||||
The demo builds WAMR for ESP-IDF and runs a small wasm program.
|
The demo builds WAMR for ESP-IDF and runs a small wasm program.
|
||||||
In order to run it for your specific Espressif chip, edit the ['build_and_run.sh'](/product-mini/platforms/esp-idf/build_and_run.sh) file and put the correct toolchain file (see #Cross-compilation) and `IDF_TARGET`.
|
In order to run it for your specific Espressif chip, edit the [build_and_run.sh](/product-mini/platforms/esp-idf/build_and_run.sh) file and put the correct toolchain file (see #Cross-compilation) and `IDF_TARGET`.
|
||||||
Before compiling it is also necessary to call ESP-IDF's `export.sh` script to bring all compile time relevant information in scope.
|
Before compiling it is also necessary to call ESP-IDF's `export.sh` script to bring all compile time relevant information in scope.
|
||||||
|
|
||||||
Docker
|
Docker
|
||||||
|
|
Loading…
Reference in New Issue
Block a user