Skip to content
@NVIDIA

NVIDIA Corporation

Pinned Loading

  1. cuopt cuopt Public

    GPU accelerated decision optimization

    Cuda 788 151

  2. cuopt-examples cuopt-examples Public

    NVIDIA cuOpt examples for decision optimization

    Jupyter Notebook 426 74

  3. open-gpu-kernel-modules open-gpu-kernel-modules Public

    NVIDIA Linux open GPU kernel module source

    C 16.8k 1.6k

  4. aistore aistore Public

    AIStore: scalable storage for AI applications

    Go 1.8k 243

  5. nvidia-container-toolkit nvidia-container-toolkit Public

    Build and run containers leveraging NVIDIA GPUs

    Go 4.2k 499

  6. GenerativeAIExamples GenerativeAIExamples Public

    Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.

    Jupyter Notebook 3.9k 1k

Repositories

Showing 10 of 706 repositories
  • NeMo-speech-data-processor Public

    A toolkit for processing speech data and creating speech datasets

    NVIDIA/NeMo-speech-data-processor’s past year of commit activity
    Python 202 Apache-2.0 43 6 18 Updated Mar 29, 2026
  • Megatron-LM Public

    Ongoing research training transformer models at scale

    NVIDIA/Megatron-LM’s past year of commit activity
    Python 15,833 3,768 333 (1 issue needs help) 328 Updated Mar 29, 2026
  • Megatron-Energon Public

    Megatron's multi-modal data loader

    NVIDIA/Megatron-Energon’s past year of commit activity
    Python 328 40 15 11 Updated Mar 29, 2026
  • cccl Public

    CUDA Core Compute Libraries

    NVIDIA/cccl’s past year of commit activity
    C++ 2,242 371 1,285 (6 issues need help) 236 Updated Mar 29, 2026
  • stdexec Public

    `std::execution`, the proposed C++ framework for asynchronous and parallel programming.

    NVIDIA/stdexec’s past year of commit activity
    C++ 2,282 Apache-2.0 234 127 12 Updated Mar 29, 2026
  • Model-Optimizer Public

    A unified library of SOTA model optimization techniques like quantization, pruning, distillation, speculative decoding, etc. It compresses deep learning models for downstream deployment frameworks like TensorRT-LLM, TensorRT, vLLM, etc. to optimize inference speed.

    NVIDIA/Model-Optimizer’s past year of commit activity
    Python 2,258 Apache-2.0 319 67 123 Updated Mar 29, 2026
  • cuopt-examples Public

    NVIDIA cuOpt examples for decision optimization

    NVIDIA/cuopt-examples’s past year of commit activity
    Jupyter Notebook 426 Apache-2.0 74 1 6 Updated Mar 29, 2026
  • nova Public Forked from torvalds/linux

    Linux kernel source tree

    NVIDIA/nova’s past year of commit activity
    C 6 64,161 0 5 Updated Mar 29, 2026
  • NVFlare Public

    NVIDIA Federated Learning Application Runtime Environment

    NVIDIA/NVFlare’s past year of commit activity
    Python 915 Apache-2.0 246 14 21 Updated Mar 29, 2026
  • TensorRT-LLM Public

    TensorRT LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and supports state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT LLM also contains components to create Python and C++ runtimes that orchestrate the inference execution in a performant way.

    NVIDIA/TensorRT-LLM’s past year of commit activity
    Python 13,210 2,222 562 639 Updated Mar 29, 2026