Run AI models locally on your machine with node.js bindings for llama.cpp. Enforce a JSON schema on the model output on the generation level
Search, hash, sort, fingerprint, and fuzzy-match strings faster via SWAR, SIMD, and GPGPU
Native module for An another Node binding of llama.cpp (win32-x64-cuda)
Native module for An another Node binding of whisper.cpp (linux-x64-cuda)
Native module for An another Node binding of whisper.cpp (linux-x64-vulkan)