@realtimex/node-llama-cpp
Run AI models locally on your machine with node.js bindings for llama.cpp. Enforce a JSON schema on the model output on the generation level
Ecosystem
npmjs.org
npmjs.org
Latest Release
about 23 hours ago
0.132.0
about 23 hours ago
Versions
132
132
Downloads
13,554 last month
13,554 last month
Loading...
Readme
Loading...
Links
| Registry | npmjs.org |
| Source | Repository |
| Homepage | Homepage |
| JSON API | View JSON |
| CodeMeta | codemeta.json |
Package Details
| PURL |
pkg:npm/%40realtimex/node-llama-cpp
spec |
| License | MIT |
| Namespace | realtimex |
| First Release | about 1 month ago |
| Last Synced | about 22 hours ago |
Keywords
llama llama-cpp llama.cpp bindings ai cmake cmake-js prebuilt-binaries llm gguf metal cuda vulkan grammar embedding rerank reranking json-grammar json-schema-grammar functions function-calling token-prediction speculative-decoding temperature minP topK topP seed xtc json-schema raspberry-pi self-hosted local catai mistral deepseek qwen qwq gpt gpt-oss typescript lora batching gpu
llama llama-cpp llama.cpp bindings ai cmake cmake-js prebuilt-binaries llm gguf metal cuda vulkan grammar embedding rerank reranking json-grammar json-schema-grammar functions function-calling token-prediction speculative-decoding temperature minP topK topP seed xtc json-schema raspberry-pi self-hosted local catai mistral deepseek qwen qwq gpt gpt-oss typescript lora batching gpu
Repository
| Stars | 1 on GitHub |
| Forks | 0 on GitHub |