Performance Whitepaper
Performance Whitepaper
Complete performance comparison of VXI SRAM-CIM processor vs mainstream Edge AI solutions, plus interactive performance simulator.
Benchmark Comparison
Performance Comparison
VXI SRAM-CIM vs RK3588 NPU vs Jetson Orin NX — LLM on-device inference benchmark data
| Metric | VXI | RK3588 | Orin NX |
|---|---|---|---|
| 1.5B Decode (tok/s) | 35-45 | 5-8 | 25-30 |
| 7B Decode (tok/s) | 15-20 | N/A | 8-12 |
| System Efficiency (TOPS/W) | 27.8 | 3.2 | 5.6 |
| Power (W) | 5 | 8 | 15 |
| KV Cache on-chip | 16 MB | 0 | 0 |
| 3-Year TCO (USD) | ~$15 | ~$45 | ~$120 |
* Data above are pre-silicon design targets. Actual performance is subject to production silicon measurements.
Performance Simulator
Performance Simulator
Select a model, adjust parameters, and instantly estimate VXI processor inference performance and resource requirements.
Model
Parameters
Simulation Results
Inference Speed
~37.4
tok/s
Throughput
~37.4
tok/s (batch=1)
Weight Size
715 MB
KV Cache
14.0 MB
Weights require batch loading(715 / 64 MB)
KV Cache fully resides on-chip(14.0 / 32 MB)
Recommended SKU
VXI-XC8
~8.4W estimated power
* This is a simplified estimation. Actual performance varies with model architecture, quantization method, batch size, etc.
Need More Detailed Performance Analysis?
Our technical team can provide customized performance reports for your specific models and scenarios.
Contact Technical Team