@realtimex/node-llama-cpp
Run AI models locally on your machine with node.js bindings for llama.cpp. Enforce a JSON schema on the model output on the generation level
Ecosystem
npmjs.org
npmjs.org
Latest Release
3 days ago
0.3.1
3 days ago
Versions
5
5
Downloads
173 last month
173 last month
Loading...
Readme
Loading...
Links
| Registry | npmjs.org |
| Source | Repository |
| Homepage | Homepage |
| JSON API | View JSON |
| CodeMeta | codemeta.json |
Package Details
| PURL |
pkg:npm/%40realtimex/node-llama-cpp
spec |
| License | MIT |
| Namespace | realtimex |
| First Release | 5 days ago |
| Last Synced | 3 days ago |
Keywords
llama llama-cpp llama.cpp bindings ai cmake cmake-js prebuilt-binaries llm gguf metal cuda vulkan grammar embedding rerank reranking json-grammar json-schema-grammar functions function-calling token-prediction speculative-decoding temperature minP topK topP seed xtc json-schema raspberry-pi self-hosted local catai mistral deepseek qwen qwq gpt gpt-oss typescript lora batching gpu
llama llama-cpp llama.cpp bindings ai cmake cmake-js prebuilt-binaries llm gguf metal cuda vulkan grammar embedding rerank reranking json-grammar json-schema-grammar functions function-calling token-prediction speculative-decoding temperature minP topK topP seed xtc json-schema raspberry-pi self-hosted local catai mistral deepseek qwen qwq gpt gpt-oss typescript lora batching gpu
Repository
| Stars | 0 on GitHub |
| Forks | 0 on GitHub |