From 7c2e8af0523685667be72589c53e3c4741093fc4 Mon Sep 17 00:00:00 2001 From: Lawrence Lane Date: Mon, 26 Jan 2026 11:22:02 -0500 Subject: [PATCH 1/6] staging Signed-off-by: Lawrence Lane --- GH-TOPICS.MD | 52 ++++++++ profile/README.md | 317 ++++++++++++++++++++++++++++++++++------------ 2 files changed, 287 insertions(+), 82 deletions(-) create mode 100644 GH-TOPICS.MD diff --git a/GH-TOPICS.MD b/GH-TOPICS.MD new file mode 100644 index 0000000..973ab13 --- /dev/null +++ b/GH-TOPICS.MD @@ -0,0 +1,52 @@ +## GitHub Topics Strategy + +### Topic Categories + +| Category | Topics | Purpose | +|----------|--------|---------| +| **Pipeline Stage** | `stage-data`, `stage-training`, `stage-alignment`, `stage-evaluation`, `stage-deployment`, `stage-safety` | Filter by workflow stage | +| **Model Type** | `model-llm`, `model-vlm`, `model-speech`, `model-diffusion`, `model-omni` | Filter by model modality | +| **Training Method** | `method-pretraining`, `method-sft`, `method-lora`, `method-rl`, `method-dpo`, `method-grpo` | Filter by training technique | +| **Backend** | `backend-megatron`, `backend-pytorch`, `backend-vllm`, `backend-tensorrt` | Filter by infrastructure | +| **Meta** | `nvidia-nemo` | All repos in the framework | + + + +### Example Filter Links + +Once topics are applied, we you add convenience links to your project's README and docs: + +```markdown +**Browse by pipeline stage:** +[Data](https://github.com/orgs/NVIDIA-NeMo/repositories?q=topic:stage-data) · +[Training](https://github.com/orgs/NVIDIA-NeMo/repositories?q=topic:stage-training) · +[Alignment](https://github.com/orgs/NVIDIA-NeMo/repositories?q=topic:stage-alignment) · +[Evaluation](https://github.com/orgs/NVIDIA-NeMo/repositories?q=topic:stage-evaluation) · +[Deployment](https://github.com/orgs/NVIDIA-NeMo/repositories?q=topic:stage-deployment) + +**Browse by model type:** +[LLM](https://github.com/orgs/NVIDIA-NeMo/repositories?q=topic:model-llm) · +[VLM](https://github.com/orgs/NVIDIA-NeMo/repositories?q=topic:model-vlm) · +[Speech](https://github.com/orgs/NVIDIA-NeMo/repositories?q=topic:model-speech) · +[Diffusion](https://github.com/orgs/NVIDIA-NeMo/repositories?q=topic:model-diffusion) +``` \ No newline at end of file diff --git a/profile/README.md b/profile/README.md index 47f739c..e29b3ca 100644 --- a/profile/README.md +++ b/profile/README.md @@ -3,90 +3,243 @@ SPDX-FileCopyrightText: Copyright (c) 2024-2025 NVIDIA CORPORATION & AFFILIATES. SPDX-License-Identifier: Apache-2.0 --> -## NVIDIA NeMo Framework Overview - -NeMo Framework is NVIDIA's GPU accelerated, fully open-source, end-to-end training framework for large language models (LLMs), multi-modal models, diffusion and speech models. It enables seamless scaling of pretraining, post-training, and reinforcement learning workloads from single GPU to thousand-node clusters for both 🤗Hugging Face/PyTorch and Megatron models. This GitHub organization includes a suite of libraries and recipe collections to help users train models from end to end. - -NeMo Framework is also a part of the NVIDIA NeMo software suite for managing the AI agent lifecycle. - -## Latest 📣 announcements and 🗣️ discussions -### 🐳 NeMo AutoModel -- [10/6/2025][Enabling PyTorch Native Pipeline Parallelism for 🤗 Hugging Face Transformer Models](https://github.com/NVIDIA-NeMo/Automodel/discussions/589) -- [9/22/2025][Fine-tune Hugging Face Models Instantly with Day-0 Support with NVIDIA NeMo AutoModel](https://github.com/NVIDIA-NeMo/Automodel/discussions/477) -- [9/18/2025][🚀 NeMo Framework Now Supports Google Gemma 3n: Efficient Multimodal Fine-tuning Made Simple](https://github.com/NVIDIA-NeMo/Automodel/discussions/494) - -### 🔬 NeMo RL -- [10/1/2025][On-policy Distillation](https://github.com/NVIDIA-NeMo/RL/discussions/1445) -- [9/27/2025][FP8 Quantization in NeMo RL](https://github.com/NVIDIA-NeMo/RL/discussions/1216) -- [8/15/2025][NeMo-RL: Journey of Optimizing Weight Transfer in Large MoE Models by 10x](https://github.com/NVIDIA-NeMo/RL/discussions/1189) - -### 💬 NeMo Speech -- [8/1/2025][Guide to Fine-tune Nvidia NeMo models with Granary Data](https://github.com/NVIDIA-NeMo/NeMo/discussions/14758) - -More to come and stay tuned! - -## Getting Started - -||Installation|Checkpoint Conversion HF<>Megatron|LLM example recipes and scripts|VLM example recipes and scripts| -|-|-|-|-|-| -|1 ~ 1,000 GPUs|[NeMo Automodel](https://github.com/NVIDIA-NeMo/Automodel?tab=readme-ov-file#getting-started), [NeMo RL](https://github.com/NVIDIA-NeMo/RL?tab=readme-ov-file#prerequisites)|No Need|[Pre-training](https://github.com/NVIDIA-NeMo/Automodel?tab=readme-ov-file#llm-pre-training), [SFT](https://github.com/NVIDIA-NeMo/Automodel?tab=readme-ov-file#llm-supervised-fine-tuning-sft), [LoRA](https://github.com/NVIDIA-NeMo/Automodel?tab=readme-ov-file#llm-parameter-efficient-fine-tuning-peft), [DPO](https://github.com/NVIDIA-NeMo/RL/blob/main/examples/run_dpo.py), [GRPO](https://github.com/NVIDIA-NeMo/RL/blob/main/examples/run_grpo_math.py)|[SFT](https://github.com/NVIDIA-NeMo/Automodel?tab=readme-ov-file#vlm-supervised-fine-tuning-sft), [LoRA](https://github.com/NVIDIA-NeMo/Automodel?tab=readme-ov-file#vlm-parameter-efficient-fine-tuning-peft), [GRPO](https://github.com/NVIDIA-NeMo/RL/blob/main/examples/run_vlm_grpo.py) -|Over 1,000 GPUs|[NeMo Megatron-Bridge](https://github.com/NVIDIA-NeMo/Megatron-Bridge?tab=readme-ov-file#-installation), [NeMo RL](https://github.com/NVIDIA-NeMo/RL?tab=readme-ov-file#prerequisites)|[Conversion](https://github.com/NVIDIA-NeMo/Megatron-Bridge/blob/main/examples/conversion/README.md)|[Pretrain, SFT, and LoRA](https://github.com/NVIDIA-NeMo/Megatron-Bridge/blob/main/src/megatron/bridge/recipes/llama/llama3.py), [DPO](https://github.com/NVIDIA-NeMo/RL/blob/main/examples/run_dpo.py) with [megatron_cfg](https://github.com/NVIDIA-NeMo/RL/blob/fa379fffbc9c5580301fa748dbba269c7d90f883/examples/configs/dpo.yaml#L99), [GRPO](https://github.com/NVIDIA-NeMo/RL/blob/main/examples/run_grpo_math.py) with [megatron_cfg](https://github.com/NVIDIA-NeMo/RL/blob/fa379fffbc9c5580301fa748dbba269c7d90f883/examples/configs/grpo_math_1B_megatron.yaml#L79)|[SFT, LoRA](https://github.com/NVIDIA-NeMo/Megatron-Bridge/blob/main/src/megatron/bridge/recipes/qwen_vl/qwen25_vl.py), [GRPO megatron config](https://github.com/NVIDIA-NeMo/RL/blob/main/examples/configs/vlm_grpo_3B_megatron.yaml)| - -## Repo organization under NeMo Framework - -### Summary of key functionalities and container strategy of each repo - -Visit the individual repos to find out more 🔍, raise :bug:, contribute ✍️ and participate in discussion forums 🗣️! - -Note: The NeMo Framework is currently in the process of restructuring. The original NeMo 2.0 repository will now focus specifically on speech-related components, while other parts of the framework are being modularized into separate libraries such as NeMo Automodel, NeMo Gym, NeMo RL, and more. This transition aims to make NeMo more modular and developer-friendly. -

- -|Repo|Key Functionality & Documentation Link|Training Loop|Training Backends|Infernece Backends|Model Coverage|Container| -|-|-|-|-|-|-|-| -|[NeMo Megatron-Bridge](https://github.com/NVIDIA-NeMo/Megatron-Bridge)|[Pretraining, LoRA, SFT](https://docs.nvidia.com/nemo/megatron-bridge/latest/)|PyT native loop|Megatron-core|NA|LLM & VLM|NeMo Framework Container -|[NeMo AutoModel](https://github.com/NVIDIA-NeMo/Automodel)|[Pretraining, LoRA, SFT](https://docs.nvidia.com/nemo/automodel/latest/index.html)|PyT native loop|PyTorch|NA|LLM, VLM, Omni, VFM|NeMo AutoModel Container| -|[Previous NeMo 2.0 Repo -> will be repurposed to focus on Speech](https://github.com/NVIDIA-NeMo/NeMo)|[Pretraining,SFT](https://docs.nvidia.com/nemo-framework/user-guide/latest/speech_ai/index.html)|PyTorch Lightning Loop|Megatron-core & PyTorch|RIVA|Speech|NA| -|[NeMo RL](https://github.com/NVIDIA-NeMo/RL)|[SFT, RL](https://docs.nvidia.com/nemo/rl/latest/index.html)|PyT native loop|Megatron-core & PyTorch|vLLM|LLM, VLM|NeMo RL container| -|[NeMo Gym](https://github.com/NVIDIA-NeMo/Gym)|[RL Environment, integrate with RL Framework](https://docs.nvidia.com/nemo/gym/latest/index.html)|NA|NA|NA|NA|NeMo RL Container (WIP)| -|[NeMo Aligner (deprecated)](https://github.com/NVIDIA/NeMo-Aligner)|SFT, RL|PyT Lightning Loop|Megatron-core|TRTLLM|LLM|NA -|[NeMo Curator](https://github.com/NVIDIA-NeMo/Curator)|[Data curation](https://docs.nvidia.com/nemo/curator/latest/)|NA|NA|NA|Agnostic|NeMo Curator Container| -|[NeMo Evaluator](https://github.com/NVIDIA-NeMo/Evaluator)|[Model evaluation](https://docs.nvidia.com/nemo/evaluator/latest/)|NA|NA||Agnostic|NeMo Framework Container| -|[NeMo Export-Deploy](https://github.com/NVIDIA-NeMo/Export-Deploy)|[Export to Production](https://docs.nvidia.com/nemo/export-deploy/latest/index.html)|NA|NA|vLLM, TRT, TRTLLM, ONNX|Agnostic|NeMo Framework Container| -|[NeMo Run](https://github.com/NVIDIA-NeMo/Run)|[Experiment launcher](https://docs.nvidia.com/nemo/run/latest/)|NA|NA|NA|Agnostic|NeMo Framework Container| -|[NeMo Guardrails](https://github.com/NVIDIA-NeMo/Guardrails)|[Guardrail model response](https://docs.nvidia.com/nemo/guardrails/latest/)|NA|NA|NA||NA| -|[NeMo Skills](https://github.com/NVIDIA-NeMo/Skills)|[Reference pipeline for SDG & Eval](https://nvidia.github.io/NeMo-Skills/)|NA|NA|NA|Agnostic|NA| -|[NeMo Emerging Optimizers](https://github.com/NVIDIA-NeMo/Emerging-Optimizers)|[Collection of Optimizers](https://docs.nvidia.com/nemo/emerging-optimizers/0.1.0/index.html)|NA|Agnostic|NA|NA|NA| -|[NeMo DFM](https://github.com/NVIDIA-NeMo/DFM/tree/main)|[Diffusion foundation model training](https://github.com/NVIDIA-NeMo/DFM/tree/main/docs)|PyT native loop|Megatron-core and PyTorch|NA|Diffusion models|NA| -|[Nemotron](https://github.com/NVIDIA-NeMo/Nemotron)|Developer asset hub for Nemotron models|NA|NA|NA|Nemotron models|NA| -|[NeMo Data Designer](https://github.com/NVIDIA-NeMo/DataDesigner)|Synthetic data generation library|NA|NA|NA|NA|NA| - -
- Table 1. NeMo Framework Repos -
-

- -### Diagram Ilustration of Repos under NeMo Framework (WIP) - - ![image](/RepoDiagram.png) - -
- Figure 1. NeMo Framework Repo Overview -
-

- -### Some background motivations and historical contexts -The NeMo GitHub Org and its repo collections are created to address the following problems -* **Need for composability**: The [Previous NeMo 2.0 version](https://github.com/NVIDIA/NeMo) is monolithic and encompasses too many things, making it hard for users to find what they need. Container size is also an issue. Breaking down the Monolithic repo into a series of functional-focused repos to facilitate code discovery. -* **Need for customizability**: The [Previous NeMo 2.0 version](https://github.com/NVIDIA/NeMo) uses PyTorch Lighting as the default trainer loop, which provides some out of the box functionality but making it hard to customize. [NeMo Megatron-Bridge](https://github.com/NVIDIA-NeMo/Megatron-Bridge), [NeMo AutoModel](https://github.com/NVIDIA-NeMo/Automodel), and [NeMo RL](https://github.com/NVIDIA-NeMo/RL) have adopted pytorch native custom loop to improve flexibility and ease of use for developers. +# NVIDIA NeMo Framework - +This GitHub org contains libraries for training, data curation, evaluation, alignment, and deployment. Scale from a single GPU to 10,000+ nodes with day-0 Hugging Face support or Megatron backends for maximum throughput. + +--- + +## Choose Your Path + + + + + + + +
+ +### Get Started + +**Start with [NeMo AutoModel](https://github.com/NVIDIA-NeMo/Automodel)** – the simplest path to fine-tuning Hugging Face models on NVIDIA GPUs. + +```bash +pip install nemo-automodel +``` + +```python +from nemo_automodel import AutoModelForCausalLM, Trainer + +model = AutoModelForCausalLM.from_pretrained( + "meta-llama/Llama-3.3-70B-Instruct" +) +trainer = Trainer(model=model, train_dataset=dataset) +trainer.train() +``` + +[→ AutoModel Quick Start](https://docs.nvidia.com/nemo/automodel/latest/launcher/local-workstation.html#quick-start-choose-your-job-launch-option) + + + +### Scale Training + +Choose your training approach: + +- **< 1,000 GPUs**: [AutoModel](https://github.com/NVIDIA-NeMo/Automodel) +- **1,000+ GPUs**: [Megatron-Bridge](https://github.com/NVIDIA-NeMo/Megatron-Bridge) +- **RLHF/DPO**: [NeMo RL](https://github.com/NVIDIA-NeMo/RL) + +[→ Training Recipes](#training-recipes) + +### Manage Experiments + +[NeMo Run](https://github.com/NVIDIA-NeMo/Run) for launching and tracking experiments across: + +- Local machines +- SLURM clusters +- Kubernetes + +[→ Run Documentation](https://docs.nvidia.com/nemo/run/latest/) + + + +### Explore Libraries + +- [Curator](https://github.com/NVIDIA-NeMo/Curator) – Data curation at scale +- [Evaluator](https://github.com/NVIDIA-NeMo/Evaluator) – Model benchmarking +- [Export-Deploy](https://github.com/NVIDIA-NeMo/Export-Deploy) – Production deployment +- [Guardrails](https://github.com/NVIDIA-NeMo/Guardrails) – Safety rails + +[→ All Libraries](#all-libraries) + +### Use Containers + +Pull optimized containers to get started fast. + +- [NeMo Framework](https://catalog.ngc.nvidia.com/orgs/nvidia/containers/nemo) +- [NeMo AutoModel](https://catalog.ngc.nvidia.com/orgs/nvidia/containers/nemo-automodel) +- [NeMo RL](https://catalog.ngc.nvidia.com/orgs/nvidia/containers/nemo-rl) +- [NeMo Curator](https://catalog.ngc.nvidia.com/orgs/nvidia/containers/nemo-curator) + +[→ Explore NGC Catalog](https://catalog.ngc.nvidia.com/orgs/nvidia/teams/nemo/containers) +
+ +
+📋 Decision Guide — Which library should I use? + +| I want to... | Models | Scale | Library | Docs | +|--------------|--------|-------|---------|------| +| **Train/fine-tune** | LLM, VLM | ≤1K GPUs | [AutoModel](https://github.com/NVIDIA-NeMo/Automodel) | [docs](https://docs.nvidia.com/nemo/automodel/latest/) | +| **Train at scale** | LLM, VLM | 1K+ GPUs | [Megatron-Bridge](https://github.com/NVIDIA-NeMo/Megatron-Bridge) | [docs](https://docs.nvidia.com/nemo/megatron-bridge/latest/) | +| **Align** (DPO/GRPO) | LLM, VLM | Any | [NeMo RL](https://github.com/NVIDIA-NeMo/RL) | [docs](https://docs.nvidia.com/nemo/rl/latest/) | +| **Curate data** | — | Any | [Curator](https://github.com/NVIDIA-NeMo/Curator) | [docs](https://docs.nvidia.com/nemo/curator/latest/) | +| **Evaluate** | Any | — | [Evaluator](https://github.com/NVIDIA-NeMo/Evaluator) | [docs](https://docs.nvidia.com/nemo/evaluator/latest/) | +| **Deploy** | Any | — | [Export-Deploy](https://github.com/NVIDIA-NeMo/Export-Deploy) | [docs](https://docs.nvidia.com/nemo/export-deploy/latest/) | +| **Speech AI** | ASR, TTS | Any | [NeMo Speech](https://github.com/NVIDIA-NeMo/NeMo) | [docs](https://docs.nvidia.com/nemo/speech/latest/) | + +
+ +--- + +## Training Recipes + +| Library | LLM Recipes | VLM Recipes | +|---------|-------------|-------------| +| [AutoModel](https://github.com/NVIDIA-NeMo/Automodel) | [Llama](https://github.com/NVIDIA-NeMo/Automodel/tree/main/examples/llm_finetune/llama3_2), [Qwen](https://github.com/NVIDIA-NeMo/Automodel/tree/main/examples/llm_finetune/qwen), [Gemma](https://github.com/NVIDIA-NeMo/Automodel/tree/main/examples/llm_finetune/gemma), [DeepSeek V3](https://github.com/NVIDIA-NeMo/Automodel/blob/main/examples/llm_pretrain/deepseekv3_pretrain.yaml), [Mistral](https://github.com/NVIDIA-NeMo/Automodel/tree/main/examples/llm_finetune/mistral), [Phi](https://github.com/NVIDIA-NeMo/Automodel/tree/main/examples/llm_finetune/phi) | [Gemma 3 VL](https://github.com/NVIDIA-NeMo/Automodel/tree/main/examples/vlm_finetune/gemma3), [Qwen2.5 VL](https://github.com/NVIDIA-NeMo/Automodel/tree/main/examples/vlm_finetune/qwen2_5), [Gemma 3n VL](https://github.com/NVIDIA-NeMo/Automodel/tree/main/examples/vlm_finetune/gemma3n) | +| [Megatron-Bridge](https://github.com/NVIDIA-NeMo/Megatron-Bridge) | [Llama](https://github.com/NVIDIA-NeMo/Megatron-Bridge/blob/main/src/megatron/bridge/recipes/llama/llama3.py), [Qwen](https://github.com/NVIDIA-NeMo/Megatron-Bridge/blob/main/src/megatron/bridge/recipes/qwen/qwen2.py), [DeepSeek V3](https://github.com/NVIDIA-NeMo/Megatron-Bridge/blob/main/src/megatron/bridge/recipes/deepseek/deepseek_v3.py), [Gemma 3](https://github.com/NVIDIA-NeMo/Megatron-Bridge/blob/main/src/megatron/bridge/recipes/gemma/gemma3.py), [Nemotron](https://github.com/NVIDIA-NeMo/Megatron-Bridge/blob/main/src/megatron/bridge/recipes/nemotronh/nemotronh.py) | [Gemma 3 VL](https://github.com/NVIDIA-NeMo/Megatron-Bridge/blob/main/src/megatron/bridge/recipes/gemma3_vl/gemma3_vl.py), [Qwen2.5 VL](https://github.com/NVIDIA-NeMo/Megatron-Bridge/blob/main/src/megatron/bridge/recipes/qwen_vl/qwen25_vl.py), [Qwen3 VL](https://github.com/NVIDIA-NeMo/Megatron-Bridge/blob/main/src/megatron/bridge/recipes/qwen_vl/qwen3vl.py) | +| [NeMo RL](https://github.com/NVIDIA-NeMo/RL) | [GRPO](https://github.com/NVIDIA-NeMo/RL/blob/main/examples/run_grpo_math.py), [DPO](https://github.com/NVIDIA-NeMo/RL/blob/main/examples/run_dpo.py), [SFT](https://github.com/NVIDIA-NeMo/RL/blob/main/examples/run_sft.py) | [GRPO](https://github.com/NVIDIA-NeMo/RL/blob/main/docs/guides/grpo.md), [SFT](https://github.com/NVIDIA-NeMo/RL/blob/main/docs/guides/sft.md) | + + +--- + +## All Libraries + +### Pipeline Overview + +```mermaid +flowchart LR + subgraph Data + Curator + DataDesigner[Data Designer] + Skills + end + + subgraph Training + AutoModel + MBridge[Megatron-Bridge] + end + + subgraph Alignment + RL[NeMo RL] + end + + subgraph Evaluation + Evaluator + end + + subgraph Deployment + Export[Export-Deploy] + Guardrails + end + + Gym[NeMo Gym] + + Data --> Training + Training --> Alignment + Training --> Evaluation + Alignment --> Evaluation + Evaluation --> Deployment + + Gym -.-> RL + Skills -.-> Evaluator +``` + +### Data + +| Repo | Description | Docs | Container | +|------|-------------|------|-----------| +| [Curator](https://github.com/NVIDIA-NeMo/Curator) | Data curation at scale | [docs](https://docs.nvidia.com/nemo/curator/latest/) | [NeMo Curator](https://catalog.ngc.nvidia.com/orgs/nvidia/containers/nemo-curator) | +| [Data Designer](https://github.com/NVIDIA-NeMo/DataDesigner) | Synthetic data generation | [docs](https://nvidia-nemo.github.io/DataDesigner/latest/) | — | +| [Skills](https://github.com/NVIDIA-NeMo/Skills) | SDG pipelines (math, code, science datasets) | [docs](https://nvidia-nemo.github.io/Skills/) | — | + +### Training + +| Repo | Description | Backend | Models | Docs | Container | +|------|-------------|---------|--------|------|-----------| +| [AutoModel](https://github.com/NVIDIA-NeMo/Automodel) | Pretraining, SFT, LoRA | PyTorch | LLM, VLM, Omni | [docs](https://docs.nvidia.com/nemo/automodel/latest/) | [NeMo AutoModel](https://catalog.ngc.nvidia.com/orgs/nvidia/containers/nemo-automodel) | +| [Megatron-Bridge](https://github.com/NVIDIA-NeMo/Megatron-Bridge) | Pretraining, SFT, LoRA | Megatron-core | LLM, VLM | [docs](https://docs.nvidia.com/nemo/megatron-bridge/latest/) | [NeMo Framework](https://catalog.ngc.nvidia.com/orgs/nvidia/containers/nemo) | +| [NeMo Speech](https://github.com/NVIDIA-NeMo/NeMo) | Pretraining, SFT | Megatron-core | Speech | [docs](https://docs.nvidia.com/nemo-framework/user-guide/latest/speech_ai/index.html) | — | +| [DFM](https://github.com/NVIDIA-NeMo/DFM) | Diffusion training | Megatron-core | Diffusion | [docs](https://github.com/NVIDIA-NeMo/DFM/tree/main/docs) | — | + +### Alignment + +| Repo | Description | Backend | Models | Docs | Container | +|------|-------------|---------|--------|------|-----------| +| [NeMo RL](https://github.com/NVIDIA-NeMo/RL) | SFT, DPO, GRPO | Megatron-core, vLLM | LLM, VLM | [docs](https://docs.nvidia.com/nemo/rl/latest/) | [NeMo RL](https://catalog.ngc.nvidia.com/orgs/nvidia/containers/nemo-rl) | +| [Gym](https://github.com/NVIDIA-NeMo/Gym) | RL environments | — | LLM, VLM | [docs](https://docs.nvidia.com/nemo/gym/latest/index.html) | — | + +### Evaluation + +| Repo | Description | Docs | Container | +|------|-------------|------|-----------| +| [Evaluator](https://github.com/NVIDIA-NeMo/Evaluator) | Model benchmarking | [docs](https://docs.nvidia.com/nemo/evaluator/latest/) | [NeMo Framework](https://catalog.ngc.nvidia.com/orgs/nvidia/containers/nemo) | +| [Skills](https://github.com/NVIDIA-NeMo/Skills) | Evaluation pipelines (math, code, science, etc.) | [docs](https://nvidia-nemo.github.io/Skills/) | — | + +### Deployment + +| Repo | Description | Backends | Docs | Container | +|------|-------------|----------|------|-----------| +| [Export-Deploy](https://github.com/NVIDIA-NeMo/Export-Deploy) | Export to production | vLLM, TRT-LLM, ONNX | [docs](https://docs.nvidia.com/nemo/export-deploy/latest/) | [NeMo Framework](https://catalog.ngc.nvidia.com/orgs/nvidia/containers/nemo) | +| [Guardrails](https://github.com/NVIDIA-NeMo/Guardrails) | Safety rails | — | [docs](https://docs.nvidia.com/nemo/guardrails/latest/) | — | + +### Infrastructure + +| Repo | Description | Docs | Container | +|------|-------------|------|-----------| +| [Run](https://github.com/NVIDIA-NeMo/Run) | Experiment launcher | [docs](https://docs.nvidia.com/nemo/run/latest/) | [NeMo Framework](https://catalog.ngc.nvidia.com/orgs/nvidia/containers/nemo) | +| [Emerging-Optimizers](https://github.com/NVIDIA-NeMo/Emerging-Optimizers) | Collection of optimizers | [docs](https://docs.nvidia.com/nemo/emerging-optimizers/latest/index.html) | — | +| [Nemotron](https://github.com/NVIDIA-NeMo/Nemotron) | Recipes for Nemotron models | [docs](https://github.com/NVIDIA-NeMo/Nemotron#readme) | — | + +### Architecture Reference + +![Framework Architecture](RepoDiagram.png) + +*Architectural layers and dependencies across the NeMo Framework.* + +--- + +## Community + + + + + + +
+ +### 💬 Get Involved + +**[GitHub Discussions](https://github.com/orgs/NVIDIA-NeMo/discussions)** — Questions, ideas, and announcements + +- [All Repositories](https://github.com/orgs/NVIDIA-NeMo/repositories) +- [Contributing Guide](https://github.com/NVIDIA/NeMo/blob/stable/CONTRIBUTING.md) +- [Release Notes](https://docs.nvidia.com/nemo/releases/) + + + +### 📣 Latest + +**🐳 AutoModel** +- [PyTorch Native Pipeline Parallelism for HF Models](https://github.com/orgs/NVIDIA-NeMo/discussions) *(Oct 2025)* +- [Day-0 Hugging Face Support](https://github.com/orgs/NVIDIA-NeMo/discussions) *(Sep 2025)* +- [Gemma 3n Multimodal Fine-tuning](https://github.com/orgs/NVIDIA-NeMo/discussions) *(Sep 2025)* + +**🔬 NeMo RL** — [On-policy Distillation](https://github.com/orgs/NVIDIA-NeMo/discussions), [FP8 Quantization](https://github.com/orgs/NVIDIA-NeMo/discussions), [10× MoE Weight Transfer](https://github.com/orgs/NVIDIA-NeMo/discussions) + +
## License -Apache 2.0 licensed with third-party attributions documented in each repository. +Apache 2.0. Third-party attributions in each repository. From ac8672747786b56236d9c8f0b61a767647be2616 Mon Sep 17 00:00:00 2001 From: Lawrence Lane Date: Mon, 26 Jan 2026 11:29:20 -0500 Subject: [PATCH 2/6] update Signed-off-by: Lawrence Lane --- profile/README.md | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/profile/README.md b/profile/README.md index e29b3ca..1563c32 100644 --- a/profile/README.md +++ b/profile/README.md @@ -50,7 +50,7 @@ Choose your training approach: [→ Training Recipes](#training-recipes) -### Manage Experiments +### Experiment [NeMo Run](https://github.com/NVIDIA-NeMo/Run) for launching and tracking experiments across: @@ -109,7 +109,7 @@ Pull optimized containers to get started fast. |---------|-------------|-------------| | [AutoModel](https://github.com/NVIDIA-NeMo/Automodel) | [Llama](https://github.com/NVIDIA-NeMo/Automodel/tree/main/examples/llm_finetune/llama3_2), [Qwen](https://github.com/NVIDIA-NeMo/Automodel/tree/main/examples/llm_finetune/qwen), [Gemma](https://github.com/NVIDIA-NeMo/Automodel/tree/main/examples/llm_finetune/gemma), [DeepSeek V3](https://github.com/NVIDIA-NeMo/Automodel/blob/main/examples/llm_pretrain/deepseekv3_pretrain.yaml), [Mistral](https://github.com/NVIDIA-NeMo/Automodel/tree/main/examples/llm_finetune/mistral), [Phi](https://github.com/NVIDIA-NeMo/Automodel/tree/main/examples/llm_finetune/phi) | [Gemma 3 VL](https://github.com/NVIDIA-NeMo/Automodel/tree/main/examples/vlm_finetune/gemma3), [Qwen2.5 VL](https://github.com/NVIDIA-NeMo/Automodel/tree/main/examples/vlm_finetune/qwen2_5), [Gemma 3n VL](https://github.com/NVIDIA-NeMo/Automodel/tree/main/examples/vlm_finetune/gemma3n) | | [Megatron-Bridge](https://github.com/NVIDIA-NeMo/Megatron-Bridge) | [Llama](https://github.com/NVIDIA-NeMo/Megatron-Bridge/blob/main/src/megatron/bridge/recipes/llama/llama3.py), [Qwen](https://github.com/NVIDIA-NeMo/Megatron-Bridge/blob/main/src/megatron/bridge/recipes/qwen/qwen2.py), [DeepSeek V3](https://github.com/NVIDIA-NeMo/Megatron-Bridge/blob/main/src/megatron/bridge/recipes/deepseek/deepseek_v3.py), [Gemma 3](https://github.com/NVIDIA-NeMo/Megatron-Bridge/blob/main/src/megatron/bridge/recipes/gemma/gemma3.py), [Nemotron](https://github.com/NVIDIA-NeMo/Megatron-Bridge/blob/main/src/megatron/bridge/recipes/nemotronh/nemotronh.py) | [Gemma 3 VL](https://github.com/NVIDIA-NeMo/Megatron-Bridge/blob/main/src/megatron/bridge/recipes/gemma3_vl/gemma3_vl.py), [Qwen2.5 VL](https://github.com/NVIDIA-NeMo/Megatron-Bridge/blob/main/src/megatron/bridge/recipes/qwen_vl/qwen25_vl.py), [Qwen3 VL](https://github.com/NVIDIA-NeMo/Megatron-Bridge/blob/main/src/megatron/bridge/recipes/qwen_vl/qwen3vl.py) | -| [NeMo RL](https://github.com/NVIDIA-NeMo/RL) | [GRPO](https://github.com/NVIDIA-NeMo/RL/blob/main/examples/run_grpo_math.py), [DPO](https://github.com/NVIDIA-NeMo/RL/blob/main/examples/run_dpo.py), [SFT](https://github.com/NVIDIA-NeMo/RL/blob/main/examples/run_sft.py) | [GRPO](https://github.com/NVIDIA-NeMo/RL/blob/main/docs/guides/grpo.md), [SFT](https://github.com/NVIDIA-NeMo/RL/blob/main/docs/guides/sft.md) | +| [NeMo RL](https://github.com/NVIDIA-NeMo/RL) | [GRPO](https://github.com/NVIDIA-NeMo/RL/blob/main/examples/run_grpo_math.py), [DPO](https://github.com/NVIDIA-NeMo/RL/blob/main/examples/run_dpo.py), [SFT](https://github.com/NVIDIA-NeMo/RL/blob/main/examples/run_sft.py) | [GRPO](https://github.com/NVIDIA-NeMo/RL/blob/main/examples/run_vlm_grpo.py), [SFT](https://github.com/NVIDIA-NeMo/RL/blob/main/examples/run_vlm_sft.py) | --- @@ -221,8 +221,8 @@ flowchart LR **[GitHub Discussions](https://github.com/orgs/NVIDIA-NeMo/discussions)** — Questions, ideas, and announcements - [All Repositories](https://github.com/orgs/NVIDIA-NeMo/repositories) -- [Contributing Guide](https://github.com/NVIDIA/NeMo/blob/stable/CONTRIBUTING.md) -- [Release Notes](https://docs.nvidia.com/nemo/releases/) +- Follow each repos CONTRIBUTING guide to get started +- [Release Notes](https://docs.nvidia.com/nemo-framework/user-guide/latest/changelog.html) From d759a438f7adcb4a3f8bb390594243b6497c466b Mon Sep 17 00:00:00 2001 From: Lawrence Lane Date: Mon, 26 Jan 2026 11:33:41 -0500 Subject: [PATCH 3/6] alignment Signed-off-by: Lawrence Lane --- profile/README.md | 3 --- 1 file changed, 3 deletions(-) diff --git a/profile/README.md b/profile/README.md index 1563c32..14d46de 100644 --- a/profile/README.md +++ b/profile/README.md @@ -42,8 +42,6 @@ trainer.train() ### Scale Training -Choose your training approach: - - **< 1,000 GPUs**: [AutoModel](https://github.com/NVIDIA-NeMo/Automodel) - **1,000+ GPUs**: [Megatron-Bridge](https://github.com/NVIDIA-NeMo/Megatron-Bridge) - **RLHF/DPO**: [NeMo RL](https://github.com/NVIDIA-NeMo/RL) @@ -68,7 +66,6 @@ Choose your training approach: - [Curator](https://github.com/NVIDIA-NeMo/Curator) – Data curation at scale - [Evaluator](https://github.com/NVIDIA-NeMo/Evaluator) – Model benchmarking - [Export-Deploy](https://github.com/NVIDIA-NeMo/Export-Deploy) – Production deployment -- [Guardrails](https://github.com/NVIDIA-NeMo/Guardrails) – Safety rails [→ All Libraries](#all-libraries) From 6971797f2e6fdf3c9be19054922b1323657d574e Mon Sep 17 00:00:00 2001 From: Lawrence Lane Date: Mon, 26 Jan 2026 11:34:26 -0500 Subject: [PATCH 4/6] word Signed-off-by: Lawrence Lane --- profile/README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/profile/README.md b/profile/README.md index 14d46de..5a5d830 100644 --- a/profile/README.md +++ b/profile/README.md @@ -63,7 +63,7 @@ trainer.train() ### Explore Libraries -- [Curator](https://github.com/NVIDIA-NeMo/Curator) – Data curation at scale +- [Curator](https://github.com/NVIDIA-NeMo/Curator) – Data curation - [Evaluator](https://github.com/NVIDIA-NeMo/Evaluator) – Model benchmarking - [Export-Deploy](https://github.com/NVIDIA-NeMo/Export-Deploy) – Production deployment From 1682c52ba73ee8950ec40f908cfcd8df37f17589 Mon Sep 17 00:00:00 2001 From: Lawrence Lane Date: Wed, 28 Jan 2026 10:01:01 -0500 Subject: [PATCH 5/6] update Signed-off-by: Lawrence Lane --- profile/README.md | 9 +++++++-- 1 file changed, 7 insertions(+), 2 deletions(-) diff --git a/profile/README.md b/profile/README.md index 5a5d830..55d77a4 100644 --- a/profile/README.md +++ b/profile/README.md @@ -169,6 +169,7 @@ flowchart LR | [Megatron-Bridge](https://github.com/NVIDIA-NeMo/Megatron-Bridge) | Pretraining, SFT, LoRA | Megatron-core | LLM, VLM | [docs](https://docs.nvidia.com/nemo/megatron-bridge/latest/) | [NeMo Framework](https://catalog.ngc.nvidia.com/orgs/nvidia/containers/nemo) | | [NeMo Speech](https://github.com/NVIDIA-NeMo/NeMo) | Pretraining, SFT | Megatron-core | Speech | [docs](https://docs.nvidia.com/nemo-framework/user-guide/latest/speech_ai/index.html) | — | | [DFM](https://github.com/NVIDIA-NeMo/DFM) | Diffusion training | Megatron-core | Diffusion | [docs](https://github.com/NVIDIA-NeMo/DFM/tree/main/docs) | — | +| [Emerging-Optimizers](https://github.com/NVIDIA-NeMo/Emerging-Optimizers) | Collection of optimizers | — | — | [docs](https://docs.nvidia.com/nemo/emerging-optimizers/latest/index.html) | — | ### Alignment @@ -191,13 +192,17 @@ flowchart LR | [Export-Deploy](https://github.com/NVIDIA-NeMo/Export-Deploy) | Export to production | vLLM, TRT-LLM, ONNX | [docs](https://docs.nvidia.com/nemo/export-deploy/latest/) | [NeMo Framework](https://catalog.ngc.nvidia.com/orgs/nvidia/containers/nemo) | | [Guardrails](https://github.com/NVIDIA-NeMo/Guardrails) | Safety rails | — | [docs](https://docs.nvidia.com/nemo/guardrails/latest/) | — | +### Models and Recipes + +| Repo | Description | Docs | Container | +|------|-------------|------|-----------| +| [Nemotron](https://github.com/NVIDIA-NeMo/Nemotron) | Recipes for Nemotron models | [docs](https://github.com/NVIDIA-NeMo/Nemotron#readme) | — | + ### Infrastructure | Repo | Description | Docs | Container | |------|-------------|------|-----------| | [Run](https://github.com/NVIDIA-NeMo/Run) | Experiment launcher | [docs](https://docs.nvidia.com/nemo/run/latest/) | [NeMo Framework](https://catalog.ngc.nvidia.com/orgs/nvidia/containers/nemo) | -| [Emerging-Optimizers](https://github.com/NVIDIA-NeMo/Emerging-Optimizers) | Collection of optimizers | [docs](https://docs.nvidia.com/nemo/emerging-optimizers/latest/index.html) | — | -| [Nemotron](https://github.com/NVIDIA-NeMo/Nemotron) | Recipes for Nemotron models | [docs](https://github.com/NVIDIA-NeMo/Nemotron#readme) | — | ### Architecture Reference From 38e37c3cd4dae45ef448d1e082bdb175fd6c8cc8 Mon Sep 17 00:00:00 2001 From: Lawrence Lane Date: Mon, 2 Feb 2026 14:01:02 -0500 Subject: [PATCH 6/6] linkfixes Signed-off-by: Lawrence Lane --- profile/README.md | 10 ++++++---- 1 file changed, 6 insertions(+), 4 deletions(-) diff --git a/profile/README.md b/profile/README.md index 55d77a4..074bc68 100644 --- a/profile/README.md +++ b/profile/README.md @@ -232,11 +232,13 @@ flowchart LR ### 📣 Latest **🐳 AutoModel** -- [PyTorch Native Pipeline Parallelism for HF Models](https://github.com/orgs/NVIDIA-NeMo/discussions) *(Oct 2025)* -- [Day-0 Hugging Face Support](https://github.com/orgs/NVIDIA-NeMo/discussions) *(Sep 2025)* -- [Gemma 3n Multimodal Fine-tuning](https://github.com/orgs/NVIDIA-NeMo/discussions) *(Sep 2025)* +- [Enabling PyTorch Native Pipeline Parallelism for HF Models](https://github.com/NVIDIA-NeMo/Automodel/discussions/589) *(Oct 2025)* +- [Day-0 Hugging Face Support](https://github.com/NVIDIA-NeMo/Automodel/discussions/477) *(Sep 2025)* +- [Gemma 3n Multimodal Fine-tuning](https://github.com/NVIDIA-NeMo/Automodel/discussions/494) *(Sep 2025)* -**🔬 NeMo RL** — [On-policy Distillation](https://github.com/orgs/NVIDIA-NeMo/discussions), [FP8 Quantization](https://github.com/orgs/NVIDIA-NeMo/discussions), [10× MoE Weight Transfer](https://github.com/orgs/NVIDIA-NeMo/discussions) +**🔬 NeMo RL** — [On-policy Distillation](https://github.com/NVIDIA-NeMo/RL/discussions/1445), [FP8 Quantization](https://github.com/NVIDIA-NeMo/RL/discussions/1216), [10× MoE Weight Transfer](https://github.com/NVIDIA-NeMo/RL/discussions/1189) + +**💬 NeMo Speech** — [Fine-tune NeMo models with Granary Data](https://github.com/NVIDIA-NeMo/NeMo/discussions/14758)