INT4/INT5/INT8 and FP16 inference on CPU for RWKV language model
-
Updated
Mar 23, 2025 - C++
INT4/INT5/INT8 and FP16 inference on CPU for RWKV language model
🧠Build and optimize machine learning models with this lightweight tensor library featuring cross-platform support and no runtime memory allocations.
Add a description, image, and links to the rwkv topic page so that developers can more easily learn about it.
To associate your repository with the rwkv topic, visit your repo's landing page and select "manage topics."