Thanks to visit codestin.com
Credit goes to github.com

Skip to content
Change the repository type filter

All

    Repositories list

    • nv-ingest

      Public
      NeMo Retriever extraction is a scalable, performance-oriented document content and metadata extraction microservice. NeMo Retriever extraction uses specialized NVIDIA NIM microservices to find, contextualize, and extract text, tables, charts and images that you can use in downstream generative applications.
      Python
      2812.8k10132Updated Dec 19, 2025Dec 19, 2025
    • cccl

      Public
      CUDA Core Compute Libraries
      C++
      3092.1k1.1k200Updated Dec 19, 2025Dec 19, 2025
    • Ongoing research training transformer models at scale
      Python
      3.4k15k344251Updated Dec 19, 2025Dec 19, 2025
    • CUDA Python: Performance meets Productivity
      Cython
      2333.1k20215Updated Dec 19, 2025Dec 19, 2025
    • Fuser

      Public
      A Fusion Code Generator for NVIDIA GPUs (commonly known as "nvFuser")
      C++
      74366209218Updated Dec 19, 2025Dec 19, 2025
    • The NVIDIA NeMo Agent toolkit is an open-source library for efficiently connecting and optimizing teams of AI agents.
      Python
      4621.6k6236Updated Dec 19, 2025Dec 19, 2025
    • physicsnemo

      Public
      Open-source deep-learning framework for building, training, and fine-tuning deep learning models using state-of-the-art Physics-ML methods
      Python
      5192.2k3943Updated Dec 19, 2025Dec 19, 2025
    • cloudai

      Public
      CloudAI Benchmark Framework
      Python
      407716Updated Dec 19, 2025Dec 19, 2025
    • NVSentinel is a cross-platform fault remediation service designed to rapidly remediate runtime node-level issues in GPU-accelerated computing environments
      Go
      29127338Updated Dec 19, 2025Dec 19, 2025
    • TensorRT LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and supports state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT LLM also contains components to create Python and C++ runtimes that orchestrate the inference execution in a performant way.
      Python
      2k12k537481Updated Dec 19, 2025Dec 19, 2025
    • Model-Optimizer

      Public
      A unified library of SOTA model optimization techniques like quantization, pruning, distillation, speculative decoding, etc. It compresses deep learning models for downstream deployment frameworks like TensorRT-LLM, TensorRT, vLLM, etc. to optimize inference speed.
      Python
      2181.7k5454Updated Dec 19, 2025Dec 19, 2025
    • JAX-Toolbox
      Python
      683688039Updated Dec 19, 2025Dec 19, 2025
    • kokoro

      Public
      JavaScript
      0100Updated Dec 19, 2025Dec 19, 2025
    • The CUDA target for Numba
      Python
      512339924Updated Dec 19, 2025Dec 19, 2025
    • Ubuntu kernels which are optimized for NVIDIA server systems
      C
      477307Updated Dec 19, 2025Dec 19, 2025
    • Nsight Python is a Python kernel profiling interface based on NVIDIA Nsight Tools
      Python
      67551Updated Dec 19, 2025Dec 19, 2025
    • DALI

      Public
      A GPU-accelerated library containing highly optimized building blocks and an execution engine for data processing to accelerate deep learning training and inference applications.
      C++
      6555.6k22229Updated Dec 19, 2025Dec 19, 2025
    • NVTX

      Public
      The NVIDIA® Tools Extension SDK (NVTX) is a C-based Application Programming Interface (API) for annotating events, code ranges, and resources in your applications.
      C++
      6648734Updated Dec 19, 2025Dec 19, 2025
    • C++ and Python support for the CUDA Quantum programming model for heterogeneous quantum-classical workflows
      C++
      31387540580Updated Dec 19, 2025Dec 19, 2025
    • edk2

      Public
      NVIDIA fork of tianocore/edk2
      C
      1626015Updated Dec 19, 2025Dec 19, 2025
    • C++
      11100Updated Dec 19, 2025Dec 19, 2025
    • cuEquivariance is a math library that is a collective of low-level primitives and tensor ops to accelerate widely-used models, like DiffDock, MACE, Allegro and NEQUIP, based on equivariant neural networks. Also includes kernels for accelerated structure prediction.
      Python
      23336125Updated Dec 19, 2025Dec 19, 2025
    • NVIDIA GPU Operator creates, configures, and manages GPUs in Kubernetes
      Go
      4312.5k9482Updated Dec 19, 2025Dec 19, 2025
    • doca-platform

      Public
      DOCA Platform manages provisioning and service orchestration for Bluefield DPUs
      Go
      166400Updated Dec 19, 2025Dec 19, 2025
    • AMGX

      Public
      Distributed multigrid linear solver library on GPU
      Cuda
      1666281112Updated Dec 19, 2025Dec 19, 2025
    • Experimental projects related to TensorRT
      MLIR
      221173712Updated Dec 19, 2025Dec 19, 2025
    • earth2studio

      Public
      Open-source deep-learning framework for exploring, building and deploying AI weather/climate workflows.
      Python
      85316108Updated Dec 19, 2025Dec 19, 2025
    • TileGym

      Public
      Helpful kernel tutorials and examples for tile-based GPU programming
      Python
      2245500Updated Dec 19, 2025Dec 19, 2025
    • stdexec

      Public
      `std::execution`, the proposed C++ framework for asynchronous and parallel programming.
      C++
      2222.1k11412Updated Dec 19, 2025Dec 19, 2025
    • OSMO

      Public
      The developer-first platform for scaling complex Physical AI workloads across heterogeneous compute—unifying training GPUs, simulation clusters, and edge devices in a simple YAML
      Python
      6611412Updated Dec 19, 2025Dec 19, 2025