forked from pool/wasmedge
OBS-URL: https://build.opensuse.org/package/show/devel:languages:javascript/wasmedge?expand=0&rev=6
131 lines
5.5 KiB
Plaintext
131 lines
5.5 KiB
Plaintext
-------------------------------------------------------------------
|
|
Tue Jul 16 08:47:26 UTC 2024 - Jan Engelhardt <jengelh@inai.de>
|
|
|
|
- Add fmt11.patch to resolve FTBFS
|
|
|
|
-------------------------------------------------------------------
|
|
Sun Nov 26 11:32:45 UTC 2023 - Dirk Müller <dmueller@suse.com>
|
|
|
|
- update to 0.13.5:
|
|
* [Component] share loading entry for component and module
|
|
(#2945)
|
|
* Initial support for the component model proposal.
|
|
* This PR allows WasmEdge to recognize the component and module
|
|
format.
|
|
* Provide options for enabling OpenBLAS, Metal, and cuBLAS.
|
|
* Bump llama.cpp to b1383
|
|
* Build thirdparty/ggml only when the ggml backend is enabled.
|
|
* Enable the ggml plugin on the macOS platform.
|
|
* Introduce `AUTO` detection. Wasm application will no longer
|
|
need to specify the hardware spec (e.g., CPU or GPU). It will
|
|
auto-detect by the runtime.
|
|
* Unified the preload options with case-insensitive matching
|
|
* Introduce `metadata` for setting the ggml options.
|
|
* The following options are supported:
|
|
* `enable-log`: `true` to enable logging. (default: `false`)
|
|
* `stream-stdout`: `true` to print the inferred tokens in the
|
|
streaming mode to standard output. (default: `false`)
|
|
* `ctx-size`: Set the context size the same as the `--ctx-size`
|
|
parameter in llama.cpp. (default: `512`)
|
|
* `n-predict`: Set the number of tokens to predict, the same as
|
|
the `--n-predict` parameter in llama.cpp. (default: `512`)
|
|
* `n-gpu-layers`: Set the number of layers to store in VRAM,
|
|
the same as the `--n-gpu-layers` parameter in llama.cpp.
|
|
(default: `0`)
|
|
* `reverse-prompt`: Set the token pattern at which you want to
|
|
halt the generation. Similar to the `--reverse-prompt`
|
|
parameter in llama.cpp. (default: `""`)
|
|
* `batch-size`: Set the number of batch sizes for prompt
|
|
processing, the same as the `--batch-size` parameter in
|
|
llama.cpp. (default: `512`)
|
|
* Notice: Because of the limitation of the WASI-NN proposal,
|
|
there is no way to set the metadata during the loading
|
|
process. The current workaround will re-load the model when
|
|
`n_gpu_layers` is set to a non-zero value.
|
|
* Installer: Support WASI-NN ggml plugin on both macOS Intel
|
|
model (CPU only) and macOS Apple Silicon model. (#2882)
|
|
* [Java Bindings] provide platform-specific jni and jar for
|
|
Java bindings (#2980)
|
|
* [C API]:
|
|
* Provide getData API for FunctionInstance (#2937)
|
|
* Add the API to set WASI-NN preloads. (#2827)
|
|
* [Plugin]:
|
|
* [zlib]:
|
|
* initial support of the zlib plugin (#2562)
|
|
* With a simple building guide and basic working examples
|
|
* [MSVC] Support MSVC for building WasmEdge
|
|
* [AOT] Support LLVM 17
|
|
* Fixed issues:
|
|
* [Installer]: Double quote the strings to prevent splitting in
|
|
env file (#2994)
|
|
* [AOT]:
|
|
* Validate AOT section header fields
|
|
* Add invariant attribute for memory and global pointer
|
|
* [C API]:
|
|
* Fix the wrong logic of getting types from exports.
|
|
* [Example] Fix get-string with the latest C++ internal getSpan
|
|
API. Fixes #2887 (#2929)
|
|
* [CI] install llvm@16 to fix macOS build (#2878)
|
|
* Misc:
|
|
* [Example] Update wit-bindgen version from 0.7.0 to 0.11.0
|
|
(#2770)
|
|
* Thank all the contributors who made this release possible!
|
|
* dm4, hydai, Lîm Tsú-thuàn, Meenu Yadav, michael1017, proohit,
|
|
Saikat Dey, Shen-Ta Hsieh, Shreyas Atre, Wang Jikai, Wck-
|
|
iipi, YiYing He
|
|
* If you want to build from source, please use
|
|
WasmEdge-0.13.5-src.tar.gz instead of the zip or tarball
|
|
provided by GitHub directly.
|
|
* ### 0.13.4 (2023-09-05)
|
|
* Features:
|
|
* [C API] Provide API for registering the Pre- and Post- host
|
|
functions
|
|
* Pre host function will be triggered before calling every host
|
|
function
|
|
* Post host function will be triggered after calling every host
|
|
function
|
|
* [CI] Update llvm-windows from 13.0.3 to 16.0.6
|
|
* WasmEdge supports multiple LLVM version, users can choose
|
|
whatever they want.
|
|
* This change is for CI.
|
|
* [CI] build alpine static libraries (#2699)
|
|
* This provides pre-built static libraries using musl-libc on
|
|
alpine.
|
|
* [Plugin] add wasmedge\_rustls\_plugin (#2762)
|
|
* [Plugin] implement opencvmini `rectangle` and `cvtColor`
|
|
(#2705)
|
|
* [Test] Migrating spec test from RapidJSON to SIMDJSON (#2659)
|
|
* [WASI Socket] AF\_UNIX Support (#2216)
|
|
* This is disable by default.
|
|
* How to enable this feature:
|
|
* CLI: Use `--allow-af-unix`.
|
|
* C API: Use `WasmEdge\_ConfigureSetAllowAFUNIX`.
|
|
* [WASI-NN] Add ggml backend for llama (#2763)
|
|
* Integrate llama.cpp as a new WASI-NN backend.
|
|
* [WASI-NN] Add load\_by\_name implementation into wasi-nn
|
|
plugin (#2742)
|
|
* Support named\_model feature.
|
|
* [WASI-NN] Added support for Tuple Type Output Tensors in
|
|
Pytorch Backend (#2564)
|
|
* Fixed issues:
|
|
* [AOT] Fix fallback case of `compileVectorExtAddPairwise`.
|
|
(#2736)
|
|
* [AOT] Fix the neontbl1 codegen error on macOS (#2738)
|
|
* [Runtime] fix memory.init oob. issue #2743 (#2758)
|
|
* [Runtime] fix table.init oob. issue #2744 (#2756)
|
|
* [System] Remove "inline" from Fault::emitFault (#2695)
|
|
(#2720)
|
|
* [Test] Use std::filesystem::u8path instead of a `const char*`
|
|
Path (#2706)
|
|
* [Utils] Installer: Fix checking of shell paths (#2752)
|
|
* [Utils] Installer: Formatting and Better source message
|
|
(#2721)
|
|
* [WASI] Avoid undefined function `FindHolderBase::reset`
|
|
* [WASI] itimerspec with 0 timeout will disarm timer, +1 to
|
|
workaround (#2730)
|
|
|
|
-------------------------------------------------------------------
|
|
Mon Aug 21 12:40:52 UTC 2023 - Alexandre Vicenzi <alexandre.vicenzi@suse.com>
|
|
|
|
- Initial revision
|