wasm-micro-runtime/core
liang.he 0599351262
wasi-nn: Add a new target for llama.cpp as a wasi-nn backend (#3709)
Minimum support:
- [x] accept (WasmEdge) customized model parameters. metadata.
- [x] Target [wasmedge-ggml examples](https://github.com/second-state/WasmEdge-WASINN-examples/tree/master/wasmedge-ggml)
  - [x] basic
  - [x] chatml
  - [x] gemma
  - [x] llama
  - [x] qwen

---

In the future, to support if required:
- [ ] Target [wasmedge-ggml examples](https://github.com/second-state/WasmEdge-WASINN-examples/tree/master/wasmedge-ggml)
  - [ ] command-r. (>70G memory requirement)
  - [ ] embedding. (embedding mode)
  - [ ] grammar. (use the grammar option to constrain the model to generate the JSON output)
  - [ ] llama-stream. (new APIS `compute_single`, `get_output_single`, `fini_single`)
  - [ ] llava. (image representation)
  - [ ] llava-base64-stream. (image representation)
  - [ ] multimodel. (image representation)
- [ ] Target [llamaedge](https://github.com/LlamaEdge/LlamaEdge)
2024-09-10 08:45:18 +08:00
..
deps Separate app-manager and app-framework from WAMR (#3129) 2024-02-20 18:12:36 +08:00
iwasm wasi-nn: Add a new target for llama.cpp as a wasi-nn backend (#3709) 2024-09-10 08:45:18 +08:00
shared Improve posix mmap retry logic (#3714) 2024-09-03 11:03:24 +08:00
config.h Add support for multi-memory proposal in classic interpreter (#3742) 2024-08-21 12:22:23 +08:00
version.h Update version to 2.1.2 and update release notes (#3718) 2024-08-20 12:00:39 +08:00