Files
wifi-densepose/vendor/ruvector/examples/ruvLLM/esp32-flash/npm/package.json

66 lines
1.3 KiB
JSON

{
"name": "ruvllm-esp32",
"version": "0.3.1",
"description": "RuvLLM ESP32 - Tiny LLM inference for ESP32 microcontrollers with INT8 quantization, RAG, HNSW vector search, and multi-chip federation. Run AI on $4 hardware.",
"keywords": [
"esp32",
"llm",
"ai",
"inference",
"embedded",
"microcontroller",
"rag",
"vector-search",
"hnsw",
"quantization",
"edge-ai",
"iot",
"machine-learning",
"neural-network",
"esp32-s3",
"xtensa",
"riscv",
"offline-ai",
"tiny-ml",
"semantic-memory"
],
"author": "RuVector Team",
"license": "MIT",
"repository": {
"type": "git",
"url": "https://github.com/ruvnet/ruvector.git",
"directory": "examples/ruvLLM/esp32-flash"
},
"homepage": "https://github.com/ruvnet/ruvector/tree/main/examples/ruvLLM/esp32-flash",
"bugs": {
"url": "https://github.com/ruvnet/ruvector/issues"
},
"bin": {
"ruvllm-esp32": "./bin/cli.js"
},
"files": [
"bin/",
"binaries/",
"scripts/",
"templates/",
"web-flasher/",
"README.md"
],
"scripts": {
"postinstall": "node bin/postinstall.js"
},
"engines": {
"node": ">=16.0.0"
},
"os": [
"darwin",
"linux",
"win32"
],
"cpu": [
"x64",
"arm64"
],
"preferGlobal": true
}