Skip to content
@NVIDIA

NVIDIA Corporation

Pinned Loading

  1. cuopt cuopt Public

    GPU accelerated decision optimization

    Cuda 663 112

  2. cuopt-examples cuopt-examples Public

    NVIDIA cuOpt examples for decision optimization

    Jupyter Notebook 395 62

  3. open-gpu-kernel-modules open-gpu-kernel-modules Public

    NVIDIA Linux open GPU kernel module source

    C 16.6k 1.6k

  4. aistore aistore Public

    AIStore: scalable storage for AI applications

    Go 1.7k 232

  5. nvidia-container-toolkit nvidia-container-toolkit Public

    Build and run containers leveraging NVIDIA GPUs

    Go 4k 461

  6. GenerativeAIExamples GenerativeAIExamples Public

    Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.

    Jupyter Notebook 3.7k 957

Repositories

Showing 10 of 654 repositories
  • cccl Public

    CUDA Core Compute Libraries

    NVIDIA/cccl’s past year of commit activity
    C++ 2,128 321 1,165 (4 issues need help) 211 Updated Jan 18, 2026
  • Megatron-LM Public

    Ongoing research training transformer models at scale

    NVIDIA/Megatron-LM’s past year of commit activity
    Python 14,939 3,498 306 (1 issue needs help) 252 Updated Jan 18, 2026
  • JAX-Toolbox Public

    JAX-Toolbox

    NVIDIA/JAX-Toolbox’s past year of commit activity
    Python 377 Apache-2.0 68 80 41 Updated Jan 18, 2026
  • personaplex Public

    PersonaPlex code.

    NVIDIA/personaplex’s past year of commit activity
    Python 75 MIT 9 1 0 Updated Jan 18, 2026
  • KAI-Scheduler Public

    KAI Scheduler is an open source Kubernetes Native scheduler for AI workloads at large scale

    NVIDIA/KAI-Scheduler’s past year of commit activity
    Go 1,073 Apache-2.0 135 29 64 Updated Jan 18, 2026
  • TensorRT-LLM Public

    TensorRT LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and supports state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT LLM also contains components to create Python and C++ runtimes that orchestrate the inference execution in a performant way.

    NVIDIA/TensorRT-LLM’s past year of commit activity
    Python 12,659 2,017 515 477 Updated Jan 18, 2026
  • k8s-test-infra Public

    K8s-test-infra

    NVIDIA/k8s-test-infra’s past year of commit activity
    Go 10 Apache-2.0 8 0 6 Updated Jan 18, 2026
  • Model-Optimizer Public

    A unified library of SOTA model optimization techniques like quantization, pruning, distillation, speculative decoding, etc. It compresses deep learning models for downstream deployment frameworks like TensorRT-LLM, TensorRT, vLLM, etc. to optimize inference speed.

    NVIDIA/Model-Optimizer’s past year of commit activity
    Python 1,835 Apache-2.0 239 62 68 Updated Jan 18, 2026
  • nv-ingest Public

    NeMo Retriever extraction is a scalable, performance-oriented document content and metadata extraction microservice. NeMo Retriever extraction uses specialized NVIDIA NIM microservices to find, contextualize, and extract text, tables, charts and images that you can use in downstream generative applications.

    NVIDIA/nv-ingest’s past year of commit activity
    Python 2,803 Apache-2.0 285 101 (1 issue needs help) 40 Updated Jan 18, 2026
  • mig-parted Public

    MIG Partition Editor for NVIDIA GPUs

    NVIDIA/mig-parted’s past year of commit activity
    Go 235 Apache-2.0 54 22 22 Updated Jan 18, 2026