Skip to content
@NVIDIA

NVIDIA Corporation

Pinned Loading

  1. cuopt cuopt Public

    GPU accelerated decision optimization

    Cuda 777 148

  2. cuopt-examples cuopt-examples Public

    NVIDIA cuOpt examples for decision optimization

    Jupyter Notebook 424 72

  3. open-gpu-kernel-modules open-gpu-kernel-modules Public

    NVIDIA Linux open GPU kernel module source

    C 16.8k 1.6k

  4. aistore aistore Public

    AIStore: scalable storage for AI applications

    Go 1.8k 243

  5. nvidia-container-toolkit nvidia-container-toolkit Public

    Build and run containers leveraging NVIDIA GPUs

    Go 4.2k 496

  6. GenerativeAIExamples GenerativeAIExamples Public

    Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.

    Jupyter Notebook 3.9k 1k

Repositories

Showing 10 of 704 repositories
  • cccl Public

    CUDA Core Compute Libraries

    NVIDIA/cccl’s past year of commit activity
    C++ 2,234 365 1,279 (6 issues need help) 227 Updated Mar 23, 2026
  • gpu-operator Public

    NVIDIA GPU Operator creates, configures, and manages GPUs in Kubernetes

    NVIDIA/gpu-operator’s past year of commit activity
    Go 2,600 Apache-2.0 466 58 44 Updated Mar 23, 2026
  • nv-sflow Public

    A Python CLI workflow orchestrator with pluggable backends (e.g. local, Slurm) for running declarative YAML DAGs, collecting logs, and organizing outputs consistently.

    NVIDIA/nv-sflow’s past year of commit activity
    Python 11 Apache-2.0 1 0 0 Updated Mar 23, 2026
  • Model-Optimizer Public

    A unified library of SOTA model optimization techniques like quantization, pruning, distillation, speculative decoding, etc. It compresses deep learning models for downstream deployment frameworks like TensorRT-LLM, TensorRT, vLLM, etc. to optimize inference speed.

    NVIDIA/Model-Optimizer’s past year of commit activity
    Python 2,223 Apache-2.0 312 70 116 Updated Mar 23, 2026
  • TensorRT-LLM Public

    TensorRT LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and supports state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT LLM also contains components to create Python and C++ runtimes that orchestrate the inference execution in a performant way.

    NVIDIA/TensorRT-LLM’s past year of commit activity
    Python 13,169 2,209 555 600 Updated Mar 23, 2026
  • cuCascade Public

    GPU Memory Reservation Library

    NVIDIA/cuCascade’s past year of commit activity
    C++ 38 Apache-2.0 17 6 0 Updated Mar 23, 2026
  • Megatron-LM Public

    Ongoing research training transformer models at scale

    NVIDIA/Megatron-LM’s past year of commit activity
    Python 15,769 3,736 327 (1 issue needs help) 338 Updated Mar 23, 2026
  • makani Public

    Massively parallel training of machine-learning based weather and climate models

    NVIDIA/makani’s past year of commit activity
    Python 363 68 6 5 Updated Mar 23, 2026
  • doca-platform Public

    DOCA Platform manages provisioning and service orchestration for Bluefield DPUs

    NVIDIA/doca-platform’s past year of commit activity
    Go 81 Apache-2.0 21 0 1 Updated Mar 23, 2026
  • pldm Public Forked from openbmc/pldm
    NVIDIA/pldm’s past year of commit activity
    C++ 6 Apache-2.0 66 0 2 Updated Mar 23, 2026