nvfp4
Here are 6 public repositories matching this topic...
LLM fine-tuning with LoRA + NVFP4/MXFP8 on NVIDIA DGX Spark (Blackwell GB10)
-
Updated
Dec 22, 2025 - Python
Code for the paper "ARCQuant: Boosting NVFP4 Quantization with Augmented Residual Channels for LLMs"
-
Updated
Jan 13, 2026 - Cuda
Production LLM deployment specs for NVIDIA Blackwell GPUs (RTX Pro 6000, DGX Spark). Includes vLLM configurations, benchmarks, load balancer, and throughput calculators for NVFP4/FP8/MoE models.
-
Updated
Jan 16, 2026 - Python
🔧 Fine-tune large language models efficiently on NVIDIA DGX Spark with LoRA adapters and optimized quantization for high performance.
-
Updated
Jan 19, 2026 - Python
Improve this page
Add a description, image, and links to the nvfp4 topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the nvfp4 topic, visit your repo's landing page and select "manage topics."