blackwell
Here are 22 public repositories matching this topic...
đź”§ Fine-tune large language models efficiently on NVIDIA DGX Spark with LoRA adapters and optimized quantization for high performance.
-
Updated
Jan 11, 2026 - Python
GEN3C: Generative Novel 3D Captions - Adapted for NVIDIA Blackwell GPU architecture (sm_120). Includes automatic GPU detection, CPU-based T5 text encoding for Blackwell compatibility, and full backward compatibility with older GPUs.
-
Updated
Oct 23, 2025 - Jupyter Notebook
Repository for Campbells-Luggs-Blackwells family history web site
-
Updated
Jul 23, 2022 - HTML
📊 Summarize merged PRs daily with vLLM, ensuring you stay updated on key changes and enhancements in your projects.
-
Updated
Jan 11, 2026
🚀 Build and explore OpenAI's GPT-OSS model from scratch in Python, unlocking the mechanics of large language models.
-
Updated
Jan 11, 2026 - Python
📦 A fully automated method for installing Nvidia drivers on Arch Linux
-
Updated
Jan 11, 2026 - Shell
An empirical study of benchmarking LLM inference with KV cache offloading using vLLM and LMCache on NVIDIA GB200 with high-bandwidth NVLink-C2C .
-
Updated
Dec 20, 2025 - Python
LLM fine-tuning with LoRA + NVFP4/MXFP8 on NVIDIA DGX Spark (Blackwell GB10)
-
Updated
Dec 22, 2025 - Python
Cross-platform FlashAttention-2 Triton implementation for Turing+ GPUs with custom configuration mode
-
Updated
Dec 16, 2025 - Python
Pre-built wheels for llama-cpp-python across platforms and CUDA versions
-
Updated
Nov 9, 2025
RTX 5090 & RTX 5060 Docker container with PyTorch + TensorFlow. First fully-tested Blackwell GPU support for ML/AI. CUDA 12.8, Python 3.11, Ubuntu 24.04. Works with RTX 50-series (5090/5080/5070/5060) and RTX 40-series.
-
Updated
Jul 8, 2025 - Shell
One-command vLLM installation for NVIDIA DGX Spark with Blackwell GB10 GPUs (sm_121 architecture)
-
Updated
Oct 28, 2025 - Shell
Prebuilt DeepSpeed wheels for Windows with NVIDIA GPU support. Supports GTX 10 - RTX 50 series. Compiled with pytorch 2.7, 2.8 and cuda 12.8
-
Updated
Aug 18, 2025
QuTLASS: CUTLASS-Powered Quantized BLAS for Deep Learning
-
Updated
Nov 11, 2025 - C++
Parallax is a distributed model serving framework that lets you build your own AI cluster anywhere
-
Updated
Jan 11, 2026 - Python
TensorRT LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and supports state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT LLM also contains components to create Python and C++ runtimes that orchestrate the inference execution in a performant way.
-
Updated
Jan 11, 2026 - Python
Improve this page
Add a description, image, and links to the blackwell topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the blackwell topic, visit your repo's landing page and select "manage topics."