mirror of
				https://github.com/bytecodealliance/wasm-micro-runtime.git
				synced 2025-10-31 05:11:19 +00:00 
			
		
		
		
	 16a4d71b34
			
		
	
	
		16a4d71b34
		
			
		
	
	
	
	
		
			
			Implement the GC (Garbage Collection) feature for interpreter mode, AOT mode and LLVM-JIT mode, and support most features of the latest spec proposal, and also enable the stringref feature. Use `cmake -DWAMR_BUILD_GC=1/0` to enable/disable the feature, and `wamrc --enable-gc` to generate the AOT file with GC supported. And update the AOT file version from 2 to 3 since there are many AOT ABI breaks, including the changes of AOT file format, the changes of AOT module/memory instance layouts, the AOT runtime APIs for the AOT code to invoke and so on.
		
			
				
	
	
		
			528 lines
		
	
	
		
			15 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			528 lines
		
	
	
		
			15 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * Copyright (C) 2019 Intel Corporation.  All rights reserved.
 | |
|  * SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
 | |
|  */
 | |
| 
 | |
| #include "ems_gc_internal.h"
 | |
| 
 | |
| static gc_handle_t
 | |
| gc_init_internal(gc_heap_t *heap, char *base_addr, gc_size_t heap_max_size)
 | |
| {
 | |
|     hmu_tree_node_t *root = NULL, *q = NULL;
 | |
|     int ret;
 | |
| 
 | |
|     memset(heap, 0, sizeof *heap);
 | |
|     memset(base_addr, 0, heap_max_size);
 | |
| 
 | |
|     ret = os_mutex_init(&heap->lock);
 | |
|     if (ret != BHT_OK) {
 | |
|         LOG_ERROR("[GC_ERROR]failed to init lock\n");
 | |
|         return NULL;
 | |
|     }
 | |
| 
 | |
|     /* init all data structures*/
 | |
|     heap->current_size = heap_max_size;
 | |
|     heap->base_addr = (gc_uint8 *)base_addr;
 | |
|     heap->heap_id = (gc_handle_t)heap;
 | |
| 
 | |
|     heap->total_free_size = heap->current_size;
 | |
|     heap->highmark_size = 0;
 | |
| #if WASM_ENABLE_GC != 0
 | |
|     heap->gc_threshold_factor = GC_DEFAULT_THRESHOLD_FACTOR;
 | |
|     gc_update_threshold(heap);
 | |
| #endif
 | |
| 
 | |
|     root = heap->kfc_tree_root = (hmu_tree_node_t *)heap->kfc_tree_root_buf;
 | |
|     memset(root, 0, sizeof *root);
 | |
|     root->size = sizeof *root;
 | |
|     hmu_set_ut(&root->hmu_header, HMU_FC);
 | |
|     hmu_set_size(&root->hmu_header, sizeof *root);
 | |
| 
 | |
|     q = (hmu_tree_node_t *)heap->base_addr;
 | |
|     memset(q, 0, sizeof *q);
 | |
|     hmu_set_ut(&q->hmu_header, HMU_FC);
 | |
|     hmu_set_size(&q->hmu_header, heap->current_size);
 | |
| 
 | |
|     ASSERT_TREE_NODE_ALIGNED_ACCESS(q);
 | |
|     ASSERT_TREE_NODE_ALIGNED_ACCESS(root);
 | |
| 
 | |
|     hmu_mark_pinuse(&q->hmu_header);
 | |
|     root->right = q;
 | |
|     q->parent = root;
 | |
|     q->size = heap->current_size;
 | |
| 
 | |
|     bh_assert(root->size <= HMU_FC_NORMAL_MAX_SIZE);
 | |
| 
 | |
|     return heap;
 | |
| }
 | |
| 
 | |
| gc_handle_t
 | |
| gc_init_with_pool(char *buf, gc_size_t buf_size)
 | |
| {
 | |
|     char *buf_end = buf + buf_size;
 | |
|     char *buf_aligned = (char *)(((uintptr_t)buf + 7) & (uintptr_t)~7);
 | |
|     char *base_addr = buf_aligned + sizeof(gc_heap_t);
 | |
|     gc_heap_t *heap = (gc_heap_t *)buf_aligned;
 | |
|     gc_size_t heap_max_size;
 | |
| 
 | |
|     if (buf_size < APP_HEAP_SIZE_MIN) {
 | |
|         LOG_ERROR("[GC_ERROR]heap init buf size (%" PRIu32 ") < %" PRIu32 "\n",
 | |
|                   buf_size, (uint32)APP_HEAP_SIZE_MIN);
 | |
|         return NULL;
 | |
|     }
 | |
| 
 | |
|     base_addr =
 | |
|         (char *)(((uintptr_t)base_addr + 7) & (uintptr_t)~7) + GC_HEAD_PADDING;
 | |
|     heap_max_size = (uint32)(buf_end - base_addr) & (uint32)~7;
 | |
| 
 | |
| #if WASM_ENABLE_MEMORY_TRACING != 0
 | |
|     os_printf("Heap created, total size: %u\n", buf_size);
 | |
|     os_printf("   heap struct size: %u\n", sizeof(gc_heap_t));
 | |
|     os_printf("   actual heap size: %u\n", heap_max_size);
 | |
|     os_printf("   padding bytes: %u\n",
 | |
|               buf_size - sizeof(gc_heap_t) - heap_max_size);
 | |
| #endif
 | |
|     return gc_init_internal(heap, base_addr, heap_max_size);
 | |
| }
 | |
| 
 | |
| gc_handle_t
 | |
| gc_init_with_struct_and_pool(char *struct_buf, gc_size_t struct_buf_size,
 | |
|                              char *pool_buf, gc_size_t pool_buf_size)
 | |
| {
 | |
|     gc_heap_t *heap = (gc_heap_t *)struct_buf;
 | |
|     char *base_addr = pool_buf + GC_HEAD_PADDING;
 | |
|     char *pool_buf_end = pool_buf + pool_buf_size;
 | |
|     gc_size_t heap_max_size;
 | |
| 
 | |
|     if ((((uintptr_t)struct_buf) & 7) != 0) {
 | |
|         LOG_ERROR("[GC_ERROR]heap init struct buf not 8-byte aligned\n");
 | |
|         return NULL;
 | |
|     }
 | |
| 
 | |
|     if (struct_buf_size < sizeof(gc_handle_t)) {
 | |
|         LOG_ERROR("[GC_ERROR]heap init struct buf size (%" PRIu32 ") < %zu\n",
 | |
|                   struct_buf_size, sizeof(gc_handle_t));
 | |
|         return NULL;
 | |
|     }
 | |
| 
 | |
|     if ((((uintptr_t)pool_buf) & 7) != 0) {
 | |
|         LOG_ERROR("[GC_ERROR]heap init pool buf not 8-byte aligned\n");
 | |
|         return NULL;
 | |
|     }
 | |
| 
 | |
|     if (pool_buf_size < APP_HEAP_SIZE_MIN) {
 | |
|         LOG_ERROR("[GC_ERROR]heap init buf size (%" PRIu32 ") < %u\n",
 | |
|                   pool_buf_size, APP_HEAP_SIZE_MIN);
 | |
|         return NULL;
 | |
|     }
 | |
| 
 | |
|     heap_max_size = (uint32)(pool_buf_end - base_addr) & (uint32)~7;
 | |
| 
 | |
| #if WASM_ENABLE_MEMORY_TRACING != 0
 | |
|     os_printf("Heap created, total size: %u\n",
 | |
|               struct_buf_size + pool_buf_size);
 | |
|     os_printf("   heap struct size: %u\n", sizeof(gc_heap_t));
 | |
|     os_printf("   actual heap size: %u\n", heap_max_size);
 | |
|     os_printf("   padding bytes: %u\n", pool_buf_size - heap_max_size);
 | |
| #endif
 | |
|     return gc_init_internal(heap, base_addr, heap_max_size);
 | |
| }
 | |
| 
 | |
| int
 | |
| gc_destroy_with_pool(gc_handle_t handle)
 | |
| {
 | |
|     gc_heap_t *heap = (gc_heap_t *)handle;
 | |
|     int ret = GC_SUCCESS;
 | |
| 
 | |
| #if WASM_ENABLE_GC != 0
 | |
|     gc_size_t i = 0;
 | |
| 
 | |
|     if (heap->extra_info_node_cnt > 0) {
 | |
|         for (i = 0; i < heap->extra_info_node_cnt; i++) {
 | |
|             extra_info_node_t *node = heap->extra_info_nodes[i];
 | |
| #if BH_ENABLE_GC_VERIFY != 0
 | |
|             os_printf("Memory leak detected: gc object [%p] not claimed\n",
 | |
|                       node->obj);
 | |
| #endif
 | |
|             bh_assert(heap->is_reclaim_enabled);
 | |
|             node->finalizer(node->obj, node->data);
 | |
| 
 | |
|             BH_FREE(heap->extra_info_nodes[i]);
 | |
|         }
 | |
| 
 | |
|         if (heap->extra_info_nodes != heap->extra_info_normal_nodes) {
 | |
|             BH_FREE(heap->extra_info_nodes);
 | |
|         }
 | |
|     }
 | |
| #endif
 | |
| 
 | |
| #if BH_ENABLE_GC_VERIFY != 0
 | |
|     hmu_t *cur = (hmu_t *)heap->base_addr;
 | |
|     hmu_t *end = (hmu_t *)((char *)heap->base_addr + heap->current_size);
 | |
| 
 | |
|     if (
 | |
| #if BH_ENABLE_GC_CORRUPTION_CHECK != 0
 | |
|         !heap->is_heap_corrupted &&
 | |
| #endif
 | |
|         (hmu_t *)((char *)cur + hmu_get_size(cur)) != end) {
 | |
|         LOG_WARNING("Memory leak detected:\n");
 | |
|         gci_dump(heap);
 | |
|         ret = GC_ERROR;
 | |
|     }
 | |
| #endif
 | |
| 
 | |
|     os_mutex_destroy(&heap->lock);
 | |
|     memset(heap->base_addr, 0, heap->current_size);
 | |
|     memset(heap, 0, sizeof(gc_heap_t));
 | |
|     return ret;
 | |
| }
 | |
| 
 | |
| #if WASM_ENABLE_GC != 0
 | |
| #if WASM_ENABLE_THREAD_MGR == 0
 | |
| void
 | |
| gc_enable_gc_reclaim(gc_handle_t handle, void *exec_env)
 | |
| {
 | |
|     gc_heap_t *heap = (gc_heap_t *)handle;
 | |
| 
 | |
|     heap->is_reclaim_enabled = 1;
 | |
|     heap->exec_env = exec_env;
 | |
| }
 | |
| #else
 | |
| void
 | |
| gc_enable_gc_reclaim(gc_handle_t handle, void *cluster)
 | |
| {
 | |
|     gc_heap_t *heap = (gc_heap_t *)handle;
 | |
| 
 | |
|     heap->is_reclaim_enabled = 1;
 | |
|     heap->cluster = cluster;
 | |
| }
 | |
| #endif
 | |
| #endif
 | |
| 
 | |
| uint32
 | |
| gc_get_heap_struct_size()
 | |
| {
 | |
|     return sizeof(gc_heap_t);
 | |
| }
 | |
| 
 | |
| static void
 | |
| adjust_ptr(uint8 **p_ptr, intptr_t offset)
 | |
| {
 | |
|     if (*p_ptr)
 | |
|         *p_ptr = (uint8 *)((intptr_t)(*p_ptr) + offset);
 | |
| }
 | |
| 
 | |
| int
 | |
| gc_migrate(gc_handle_t handle, char *pool_buf_new, gc_size_t pool_buf_size)
 | |
| {
 | |
|     gc_heap_t *heap = (gc_heap_t *)handle;
 | |
|     char *base_addr_new = pool_buf_new + GC_HEAD_PADDING;
 | |
|     char *pool_buf_end = pool_buf_new + pool_buf_size;
 | |
|     intptr_t offset = (uint8 *)base_addr_new - (uint8 *)heap->base_addr;
 | |
|     hmu_t *cur = NULL, *end = NULL;
 | |
|     hmu_tree_node_t *tree_node;
 | |
|     uint8 **p_left, **p_right, **p_parent;
 | |
|     gc_size_t heap_max_size, size;
 | |
| 
 | |
|     if ((((uintptr_t)pool_buf_new) & 7) != 0) {
 | |
|         LOG_ERROR("[GC_ERROR]heap migrate pool buf not 8-byte aligned\n");
 | |
|         return GC_ERROR;
 | |
|     }
 | |
| 
 | |
|     heap_max_size = (uint32)(pool_buf_end - base_addr_new) & (uint32)~7;
 | |
| 
 | |
|     if (pool_buf_end < base_addr_new || heap_max_size < heap->current_size) {
 | |
|         LOG_ERROR("[GC_ERROR]heap migrate invlaid pool buf size\n");
 | |
|         return GC_ERROR;
 | |
|     }
 | |
| 
 | |
|     if (offset == 0)
 | |
|         return 0;
 | |
| 
 | |
| #if BH_ENABLE_GC_CORRUPTION_CHECK != 0
 | |
|     if (heap->is_heap_corrupted) {
 | |
|         LOG_ERROR("[GC_ERROR]Heap is corrupted, heap migrate failed.\n");
 | |
|         return GC_ERROR;
 | |
|     }
 | |
| #endif
 | |
| 
 | |
|     heap->base_addr = (uint8 *)base_addr_new;
 | |
| 
 | |
|     ASSERT_TREE_NODE_ALIGNED_ACCESS(heap->kfc_tree_root);
 | |
| 
 | |
|     p_left = (uint8 **)((uint8 *)heap->kfc_tree_root
 | |
|                         + offsetof(hmu_tree_node_t, left));
 | |
|     p_right = (uint8 **)((uint8 *)heap->kfc_tree_root
 | |
|                          + offsetof(hmu_tree_node_t, right));
 | |
|     p_parent = (uint8 **)((uint8 *)heap->kfc_tree_root
 | |
|                           + offsetof(hmu_tree_node_t, parent));
 | |
|     adjust_ptr(p_left, offset);
 | |
|     adjust_ptr(p_right, offset);
 | |
|     adjust_ptr(p_parent, offset);
 | |
| 
 | |
|     cur = (hmu_t *)heap->base_addr;
 | |
|     end = (hmu_t *)((char *)heap->base_addr + heap->current_size);
 | |
| 
 | |
|     while (cur < end) {
 | |
|         size = hmu_get_size(cur);
 | |
| 
 | |
| #if BH_ENABLE_GC_CORRUPTION_CHECK != 0
 | |
|         if (size <= 0 || size > (uint32)((uint8 *)end - (uint8 *)cur)) {
 | |
|             LOG_ERROR("[GC_ERROR]Heap is corrupted, heap migrate failed.\n");
 | |
|             heap->is_heap_corrupted = true;
 | |
|             return GC_ERROR;
 | |
|         }
 | |
| #endif
 | |
| 
 | |
|         if (hmu_get_ut(cur) == HMU_FC && !HMU_IS_FC_NORMAL(size)) {
 | |
|             tree_node = (hmu_tree_node_t *)cur;
 | |
| 
 | |
|             ASSERT_TREE_NODE_ALIGNED_ACCESS(tree_node);
 | |
| 
 | |
|             p_left = (uint8 **)((uint8 *)tree_node
 | |
|                                 + offsetof(hmu_tree_node_t, left));
 | |
|             p_right = (uint8 **)((uint8 *)tree_node
 | |
|                                  + offsetof(hmu_tree_node_t, right));
 | |
|             p_parent = (uint8 **)((uint8 *)tree_node
 | |
|                                   + offsetof(hmu_tree_node_t, parent));
 | |
|             adjust_ptr(p_left, offset);
 | |
|             adjust_ptr(p_right, offset);
 | |
|             if (tree_node->parent != heap->kfc_tree_root)
 | |
|                 /* The root node belongs to heap structure,
 | |
|                    it is fixed part and isn't changed. */
 | |
|                 adjust_ptr(p_parent, offset);
 | |
|         }
 | |
|         cur = (hmu_t *)((char *)cur + size);
 | |
|     }
 | |
| 
 | |
| #if BH_ENABLE_GC_CORRUPTION_CHECK != 0
 | |
|     if (cur != end) {
 | |
|         LOG_ERROR("[GC_ERROR]Heap is corrupted, heap migrate failed.\n");
 | |
|         heap->is_heap_corrupted = true;
 | |
|         return GC_ERROR;
 | |
|     }
 | |
| #else
 | |
|     bh_assert(cur == end);
 | |
| #endif
 | |
| 
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| bool
 | |
| gc_is_heap_corrupted(gc_handle_t handle)
 | |
| {
 | |
| #if BH_ENABLE_GC_CORRUPTION_CHECK != 0
 | |
|     gc_heap_t *heap = (gc_heap_t *)handle;
 | |
| 
 | |
|     return heap->is_heap_corrupted ? true : false;
 | |
| #else
 | |
|     return false;
 | |
| #endif
 | |
| }
 | |
| 
 | |
| #if BH_ENABLE_GC_VERIFY != 0
 | |
| void
 | |
| gci_verify_heap(gc_heap_t *heap)
 | |
| {
 | |
|     hmu_t *cur = NULL, *end = NULL;
 | |
| 
 | |
|     bh_assert(heap && gci_is_heap_valid(heap));
 | |
|     cur = (hmu_t *)heap->base_addr;
 | |
|     end = (hmu_t *)(heap->base_addr + heap->current_size);
 | |
|     while (cur < end) {
 | |
|         hmu_verify(heap, cur);
 | |
|         cur = (hmu_t *)((gc_uint8 *)cur + hmu_get_size(cur));
 | |
|     }
 | |
|     bh_assert(cur == end);
 | |
| }
 | |
| #endif
 | |
| 
 | |
| void
 | |
| gc_heap_stat(void *heap_ptr, gc_stat_t *stat)
 | |
| {
 | |
|     hmu_t *cur = NULL, *end = NULL;
 | |
|     hmu_type_t ut;
 | |
|     gc_size_t size;
 | |
|     gc_heap_t *heap = (gc_heap_t *)heap_ptr;
 | |
| 
 | |
|     memset(stat, 0, sizeof(gc_stat_t));
 | |
|     cur = (hmu_t *)heap->base_addr;
 | |
|     end = (hmu_t *)((char *)heap->base_addr + heap->current_size);
 | |
| 
 | |
|     while (cur < end) {
 | |
|         ut = hmu_get_ut(cur);
 | |
|         size = hmu_get_size(cur);
 | |
|         bh_assert(size > 0);
 | |
| 
 | |
|         if (ut == HMU_FC || ut == HMU_FM
 | |
|             || (ut == HMU_VO && hmu_is_vo_freed(cur))
 | |
|             || (ut == HMU_WO && !hmu_is_wo_marked(cur))) {
 | |
|             if (ut == HMU_VO)
 | |
|                 stat->vo_free += size;
 | |
|             if (ut == HMU_WO)
 | |
|                 stat->wo_free += size;
 | |
|             stat->free += size;
 | |
|             stat->free_block++;
 | |
|             if (size / sizeof(int) < GC_HEAP_STAT_SIZE - 1)
 | |
|                 stat->free_sizes[size / sizeof(int)] += 1;
 | |
|             else
 | |
|                 stat->free_sizes[GC_HEAP_STAT_SIZE - 1] += 1;
 | |
|         }
 | |
|         else {
 | |
|             if (ut == HMU_VO)
 | |
|                 stat->vo_usage += size;
 | |
|             if (ut == HMU_WO)
 | |
|                 stat->wo_usage += size;
 | |
|             stat->usage += size;
 | |
|             stat->usage_block++;
 | |
|             if (size / sizeof(int) < GC_HEAP_STAT_SIZE - 1)
 | |
|                 stat->usage_sizes[size / sizeof(int)] += 1;
 | |
|             else
 | |
|                 stat->usage_sizes[GC_HEAP_STAT_SIZE - 1] += 1;
 | |
|         }
 | |
| 
 | |
|         cur = (hmu_t *)((char *)cur + size);
 | |
|     }
 | |
| }
 | |
| 
 | |
| void
 | |
| gc_print_stat(void *heap_ptr, int verbose)
 | |
| {
 | |
|     gc_stat_t stat;
 | |
|     int i;
 | |
| 
 | |
|     bh_assert(heap_ptr != NULL);
 | |
|     gc_heap_t *heap = (gc_heap_t *)(heap_ptr);
 | |
| 
 | |
|     gc_heap_stat(heap, &stat);
 | |
| 
 | |
|     os_printf("# stat %s %p use %d free %d \n", "instance", heap, stat.usage,
 | |
|               stat.free);
 | |
|     os_printf("# stat %s %p wo_usage %d vo_usage %d \n", "instance", heap,
 | |
|               stat.wo_usage, stat.vo_usage);
 | |
|     os_printf("# stat %s %p wo_free %d vo_free %d \n", "instance", heap,
 | |
|               stat.wo_free, stat.vo_free);
 | |
| #if WASM_ENABLE_GC == 0
 | |
|     os_printf("# stat free size %" PRIu32 " high %" PRIu32 "\n",
 | |
|               heap->total_free_size, heap->highmark_size);
 | |
| #else
 | |
|     os_printf("# stat gc %" PRIu32 " free size %" PRIu32 " high %" PRIu32 "\n",
 | |
|               heap->total_gc_count, heap->total_free_size, heap->highmark_size);
 | |
| #endif
 | |
|     if (verbose) {
 | |
|         os_printf("usage sizes: \n");
 | |
|         for (i = 0; i < GC_HEAP_STAT_SIZE; i++)
 | |
|             if (stat.usage_sizes[i])
 | |
|                 os_printf(" %d: %d; ", i * 4, stat.usage_sizes[i]);
 | |
|         os_printf(" \n");
 | |
|         os_printf("free sizes: \n");
 | |
|         for (i = 0; i < GC_HEAP_STAT_SIZE; i++)
 | |
|             if (stat.free_sizes[i])
 | |
|                 os_printf(" %d: %d; ", i * 4, stat.free_sizes[i]);
 | |
|     }
 | |
| }
 | |
| 
 | |
| void *
 | |
| gc_heap_stats(void *heap_arg, uint32 *stats, int size)
 | |
| {
 | |
|     int i;
 | |
|     gc_heap_t *heap = (gc_heap_t *)heap_arg;
 | |
| 
 | |
|     if (!gci_is_heap_valid(heap)) {
 | |
|         for (i = 0; i < size; i++)
 | |
|             stats[i] = 0;
 | |
|         return NULL;
 | |
|     }
 | |
| 
 | |
|     for (i = 0; i < size; i++) {
 | |
|         switch (i) {
 | |
|             case GC_STAT_TOTAL:
 | |
|                 stats[i] = heap->current_size;
 | |
|                 break;
 | |
|             case GC_STAT_FREE:
 | |
|                 stats[i] = heap->total_free_size;
 | |
|                 break;
 | |
|             case GC_STAT_HIGHMARK:
 | |
|                 stats[i] = heap->highmark_size;
 | |
|                 break;
 | |
| #if WASM_ENABLE_GC != 0
 | |
|             case GC_STAT_COUNT:
 | |
|                 stats[i] = heap->total_gc_count;
 | |
|                 break;
 | |
|             case GC_STAT_TIME:
 | |
|                 stats[i] = heap->total_gc_time;
 | |
|                 break;
 | |
| #endif
 | |
|             default:
 | |
|                 break;
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     return heap;
 | |
| }
 | |
| 
 | |
| void
 | |
| gc_traverse_tree(hmu_tree_node_t *node, gc_size_t *stats, int *n)
 | |
| {
 | |
|     if (!node)
 | |
|         return;
 | |
| 
 | |
|     if (*n > 0)
 | |
|         gc_traverse_tree(node->right, stats, n);
 | |
| 
 | |
|     if (*n > 0) {
 | |
|         (*n)--;
 | |
|         stats[*n] = node->size;
 | |
|     }
 | |
| 
 | |
|     if (*n > 0)
 | |
|         gc_traverse_tree(node->left, stats, n);
 | |
| }
 | |
| 
 | |
| void
 | |
| gc_show_stat(void *heap)
 | |
| {
 | |
| 
 | |
|     uint32 stats[GC_STAT_MAX];
 | |
| 
 | |
|     heap = gc_heap_stats(heap, stats, GC_STAT_MAX);
 | |
| 
 | |
|     os_printf("\n[GC stats %p] %" PRIu32 " %" PRIu32 " %" PRIu32 " %" PRIu32
 | |
|               " %" PRIu32 "\n",
 | |
|               heap, stats[0], stats[1], stats[2], stats[3], stats[4]);
 | |
| }
 | |
| 
 | |
| #if WASM_ENABLE_GC != 0
 | |
| void
 | |
| gc_show_fragment(void *heap_arg)
 | |
| {
 | |
|     uint32 stats[3];
 | |
|     int n = 3;
 | |
|     gc_heap_t *heap = (gc_heap_t *)heap_arg;
 | |
| 
 | |
|     memset(stats, 0, n * sizeof(int));
 | |
|     gct_vm_mutex_lock(&heap->lock);
 | |
|     gc_traverse_tree(heap->kfc_tree_root, (gc_size_t *)stats, &n);
 | |
|     gct_vm_mutex_unlock(&heap->lock);
 | |
|     os_printf("\n[GC %p top sizes] %" PRIu32 " %" PRIu32 " %" PRIu32 "\n", heap,
 | |
|               stats[0], stats[1], stats[2]);
 | |
| }
 | |
| 
 | |
| #if WASM_ENABLE_GC_PERF_PROFILING != 0
 | |
| void
 | |
| gc_dump_perf_profiling(gc_handle_t *handle)
 | |
| {
 | |
|     gc_heap_t *gc_heap_handle = (void *)handle;
 | |
|     if (gc_heap_handle) {
 | |
|         os_printf("\nGC performance summary\n");
 | |
|         os_printf("    Total GC time (ms): %u\n",
 | |
|                   gc_heap_handle->total_gc_time);
 | |
|         os_printf("    Max GC time (ms): %u\n", gc_heap_handle->max_gc_time);
 | |
|     }
 | |
|     else {
 | |
|         os_printf("Failed to dump GC performance\n");
 | |
|     }
 | |
| }
 | |
| #endif
 | |
| #endif
 |