Skip to content
@NVIDIA

NVIDIA Corporation

Pinned Loading

  1. cuopt cuopt Public

    GPU accelerated decision optimization

    Cuda 650 111

  2. cuopt-examples cuopt-examples Public

    NVIDIA cuOpt examples for decision optimization

    Jupyter Notebook 395 62

  3. open-gpu-kernel-modules open-gpu-kernel-modules Public

    NVIDIA Linux open GPU kernel module source

    C 16.6k 1.6k

  4. aistore aistore Public

    AIStore: scalable storage for AI applications

    Go 1.7k 231

  5. nvidia-container-toolkit nvidia-container-toolkit Public

    Build and run containers leveraging NVIDIA GPUs

    Go 4k 459

  6. GenerativeAIExamples GenerativeAIExamples Public

    Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.

    Jupyter Notebook 3.7k 956

Repositories

Showing 10 of 648 repositories
  • gpu-driver-container Public

    The NVIDIA GPU driver container allows the provisioning of the NVIDIA driver through the use of containers.

    NVIDIA/gpu-driver-container’s past year of commit activity
    Shell 152 Apache-2.0 69 26 26 Updated Jan 13, 2026
  • Model-Optimizer Public

    A unified library of SOTA model optimization techniques like quantization, pruning, distillation, speculative decoding, etc. It compresses deep learning models for downstream deployment frameworks like TensorRT-LLM, TensorRT, vLLM, etc. to optimize inference speed.

    NVIDIA/Model-Optimizer’s past year of commit activity
    Python 1,804 Apache-2.0 235 60 68 Updated Jan 13, 2026
  • NVSentinel Public

    NVSentinel is a cross-platform fault remediation service designed to rapidly remediate runtime node-level issues in GPU-accelerated computing environments

    NVIDIA/NVSentinel’s past year of commit activity
    Go 155 Apache-2.0 34 32 23 Updated Jan 13, 2026
  • TensorRT-LLM Public

    TensorRT LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and supports state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT LLM also contains components to create Python and C++ runtimes that orchestrate the inference execution in a performant way.

    NVIDIA/TensorRT-LLM’s past year of commit activity
    Python 12,615 2,005 517 475 Updated Jan 13, 2026
  • Megatron-LM Public

    Ongoing research training transformer models at scale

    NVIDIA/Megatron-LM’s past year of commit activity
    Python 14,890 3,481 311 (1 issue needs help) 250 Updated Jan 13, 2026
  • cccl Public

    CUDA Core Compute Libraries

    NVIDIA/cccl’s past year of commit activity
    C++ 2,123 320 1,141 (5 issues need help) 200 Updated Jan 13, 2026
  • KAI-Scheduler Public

    KAI Scheduler is an open source Kubernetes Native scheduler for AI workloads at large scale

    NVIDIA/KAI-Scheduler’s past year of commit activity
    Go 1,067 Apache-2.0 135 26 67 Updated Jan 13, 2026
  • spark-rapids-jni Public

    RAPIDS Accelerator JNI For Apache Spark

    NVIDIA/spark-rapids-jni’s past year of commit activity
    Cuda 52 Apache-2.0 78 85 7 Updated Jan 13, 2026
  • doca-sosreport Public Forked from sosreport/sos

    A unified tool for collecting system logs and other debug information

    NVIDIA/doca-sosreport’s past year of commit activity
    Python 5 GPL-2.0 606 0 4 Updated Jan 13, 2026
  • cuda-quantum Public

    C++ and Python support for the CUDA Quantum programming model for heterogeneous quantum-classical workflows

    NVIDIA/cuda-quantum’s past year of commit activity
    C++ 886 319 413 (16 issues need help) 92 Updated Jan 13, 2026