28 lines
744 B
JSON
28 lines
744 B
JSON
{
|
|
"name": "ruvector-attention-wasm",
|
|
"collaborators": [
|
|
"Ruvector Team"
|
|
],
|
|
"description": "High-performance WebAssembly attention mechanisms: Multi-Head, Flash, Hyperbolic, MoE, CGT Sheaf Attention with GPU acceleration for transformers and LLMs",
|
|
"version": "2.0.5",
|
|
"license": "MIT",
|
|
"repository": {
|
|
"type": "git",
|
|
"url": "https://github.com/ruvnet/ruvector"
|
|
},
|
|
"files": [
|
|
"ruvector_attention_wasm_bg.wasm",
|
|
"ruvector_attention_wasm.js",
|
|
"ruvector_attention_wasm.d.ts"
|
|
],
|
|
"main": "ruvector_attention_wasm.js",
|
|
"homepage": "https://ruv.io/ruvector",
|
|
"types": "ruvector_attention_wasm.d.ts",
|
|
"keywords": [
|
|
"wasm",
|
|
"attention",
|
|
"transformer",
|
|
"flash-attention",
|
|
"llm"
|
|
]
|
|
} |