Skip to content
Open
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
65 changes: 5 additions & 60 deletions profile/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -3,72 +3,17 @@ SPDX-FileCopyrightText: Copyright (c) 2024-2025 NVIDIA CORPORATION & AFFILIATES.
SPDX-License-Identifier: Apache-2.0
-->

## NVIDIA NeMo Framework Overview

NeMo Framework is NVIDIA's GPU accelerated, fully open-source, end-to-end training framework for large language models (LLMs), multi-modal models, diffusion and speech models. It enables seamless scaling of pretraining, post-training, and reinforcement learning workloads from single GPU to thousand-node clusters for both 🤗Hugging Face/PyTorch and Megatron models. This GitHub organization includes a suite of libraries and recipe collections to help users train models from end to end.

NeMo Framework is also a part of the NVIDIA NeMo software suite for managing the AI agent lifecycle.

## Latest 📣 announcements and 🗣️ discussions

### 💬 NeMo Assist
- [4/13/2026][Try NeMo Assist to chat with the NeMo Framework docs and code.](https://nemo-assist.nvidia.com/)

## Getting Started

||Installation|Checkpoint Conversion HF<>Megatron|LLM example recipes and scripts|VLM example recipes and scripts|
|Data|Pretraining|RL|Inference|E2E|
|-|-|-|-|-|
|1 ~ 1,000 GPUs|[NeMo Automodel](https://github.com/NVIDIA-NeMo/Automodel?tab=readme-ov-file#getting-started), [NeMo RL](https://github.com/NVIDIA-NeMo/RL?tab=readme-ov-file#prerequisites)|No Need|[Pre-training](https://github.com/NVIDIA-NeMo/Automodel?tab=readme-ov-file#llm-pre-training), [SFT](https://github.com/NVIDIA-NeMo/Automodel?tab=readme-ov-file#llm-supervised-fine-tuning-sft), [LoRA](https://github.com/NVIDIA-NeMo/Automodel?tab=readme-ov-file#llm-parameter-efficient-fine-tuning-peft), [DPO](https://github.com/NVIDIA-NeMo/RL/blob/main/examples/run_dpo.py), [GRPO](https://github.com/NVIDIA-NeMo/RL/blob/main/examples/run_grpo_math.py)|[SFT](https://github.com/NVIDIA-NeMo/Automodel?tab=readme-ov-file#vlm-supervised-fine-tuning-sft), [LoRA](https://github.com/NVIDIA-NeMo/Automodel?tab=readme-ov-file#vlm-parameter-efficient-fine-tuning-peft), [GRPO](https://github.com/NVIDIA-NeMo/RL/blob/main/examples/run_vlm_grpo.py)
|Over 1,000 GPUs|[NeMo Megatron-Bridge](https://github.com/NVIDIA-NeMo/Megatron-Bridge?tab=readme-ov-file#-installation), [NeMo RL](https://github.com/NVIDIA-NeMo/RL?tab=readme-ov-file#prerequisites)|[Conversion](https://github.com/NVIDIA-NeMo/Megatron-Bridge/blob/main/examples/conversion/README.md)|[Pretrain, SFT, and LoRA](https://github.com/NVIDIA-NeMo/Megatron-Bridge/blob/main/src/megatron/bridge/recipes/llama/llama3.py), [DPO](https://github.com/NVIDIA-NeMo/RL/blob/main/examples/run_dpo.py) with [megatron_cfg](https://github.com/NVIDIA-NeMo/RL/blob/fa379fffbc9c5580301fa748dbba269c7d90f883/examples/configs/dpo.yaml#L99), [GRPO](https://github.com/NVIDIA-NeMo/RL/blob/main/examples/run_grpo_math.py) with [megatron_cfg](https://github.com/NVIDIA-NeMo/RL/blob/fa379fffbc9c5580301fa748dbba269c7d90f883/examples/configs/grpo_math_1B_megatron.yaml#L79)|[SFT, LoRA](https://github.com/NVIDIA-NeMo/Megatron-Bridge/blob/main/src/megatron/bridge/recipes/qwen_vl/qwen25_vl.py), [GRPO megatron config](https://github.com/NVIDIA-NeMo/RL/blob/main/examples/configs/vlm_grpo_3B_megatron.yaml)|

## Repo organization under NeMo Framework

### Summary of key functionalities and container strategy of each repo
|[Curator](https://github.com/NVIDIA-NeMo/Curator), [Anonymizer](https://github.com/NVIDIA-NeMo/Anonymizer), [Data Designer](https://github.com/NVIDIA-NeMo/DataDesigner), [Safe Synthesizer](https://github.com/NVIDIA-NeMo/Safe-Synthesizer), [SDG-PGM](https://github.com/NVIDIA-NeMo/SDG-PGMs) |[Megatron-Bridge](https://github.com/NVIDIA-NeMo/Megatron-Bridge), [Automodel](https://github.com/NVIDIA-NeMo/Automodel), [Speech](https://github.com/NVIDIA-NeMo/NeMo), [Emerging Optimizers](https://github.com/NVIDIA-NeMo/Emerging-Optimizers)|[RL](https://github.com/NVIDIA-NeMo/RL), [Gym](https://github.com/NVIDIA-NeMo/Gym), [Pro-Agent-Server](https://github.com/NVIDIA-NeMo/ProRL-Agent-Server) |[Guardrails](https://github.com/NVIDIA-NeMo/Guardrails), [Evaluator](https://github.com/NVIDIA-NeMo/Evaluator), [Export-Deploy](https://github.com/NVIDIA-NeMo/Export-Deploy)| [Skills](https://github.com/NVIDIA-NeMo/Skills), [Nemotron](https://github.com/NVIDIA-NeMo/Nemotron), [Run](https://github.com/NVIDIA-NeMo/Run)|

Visit the individual repos to find out more 🔍, raise :bug:, contribute ✍️ and participate in discussion forums 🗣️!
## Latest 📣 announcements

Note: The NeMo Framework is currently in the process of restructuring. The original NeMo 2.0 repository will now focus specifically on speech-related components, while other parts of the framework are being modularized into separate libraries such as NeMo Automodel, NeMo Gym, NeMo RL, and more. This transition aims to make NeMo more modular and developer-friendly.
<p></p>

|Repo|Key Functionality & Documentation Link|Training Loop|Training Backends|Inference Backends|Model Coverage|Container|
|-|-|-|-|-|-|-|
|[NeMo Megatron-Bridge](https://github.com/NVIDIA-NeMo/Megatron-Bridge)|[Pretraining, LoRA, SFT](https://docs.nvidia.com/nemo/megatron-bridge/latest/)|PyT native loop|Megatron-core|NA|LLM & VLM|NeMo Framework Container
|[NeMo AutoModel](https://github.com/NVIDIA-NeMo/Automodel)|[Pretraining, LoRA, SFT](https://docs.nvidia.com/nemo/automodel/latest/index.html)|PyT native loop|PyTorch|NA|LLM, VLM, Omni, VFM|NeMo AutoModel Container|
|[Previous NeMo 2.0 Repo -> will be repurposed to focus on Speech](https://github.com/NVIDIA-NeMo/NeMo)|[Pretraining,SFT](https://docs.nvidia.com/nemo-framework/user-guide/latest/speech_ai/index.html)|PyTorch Lightning Loop|Megatron-core & PyTorch|RIVA|Speech|NA|
|[NeMo RL](https://github.com/NVIDIA-NeMo/RL)|[SFT, RL](https://docs.nvidia.com/nemo/rl/latest/index.html)|PyT native loop|Megatron-core & PyTorch|vLLM|LLM, VLM|NeMo RL container|
|[NeMo Gym](https://github.com/NVIDIA-NeMo/Gym)|[RL Environment, integrate with RL Framework](https://docs.nvidia.com/nemo/gym/latest/index.html)|NA|NA|NA|NA|NeMo RL Container (WIP)|
|[NeMo Aligner (deprecated)](https://github.com/NVIDIA/NeMo-Aligner)|SFT, RL|PyT Lightning Loop|Megatron-core|TRTLLM|LLM|NA
|[NeMo Curator](https://github.com/NVIDIA-NeMo/Curator)|[Data curation](https://docs.nvidia.com/nemo/curator/latest/)|NA|NA|NA|Agnostic|NeMo Curator Container|
|[NeMo Evaluator](https://github.com/NVIDIA-NeMo/Evaluator)|[Model evaluation](https://docs.nvidia.com/nemo/evaluator/latest/)|NA|NA||Agnostic|NeMo Framework Container|
|[NeMo Export-Deploy](https://github.com/NVIDIA-NeMo/Export-Deploy)|[Export to Production](https://docs.nvidia.com/nemo/export-deploy/latest/index.html)|NA|NA|vLLM, TRT, TRTLLM, ONNX|Agnostic|NeMo Framework Container|
|[NeMo Run](https://github.com/NVIDIA-NeMo/Run)|[Experiment launcher](https://docs.nvidia.com/nemo/run/latest/)|NA|NA|NA|Agnostic|NeMo Framework Container|
|[NeMo Guardrails](https://github.com/NVIDIA-NeMo/Guardrails)|[Guardrail model response](https://docs.nvidia.com/nemo/guardrails/latest/)|NA|NA|NA||NA|
|[NeMo Skills](https://github.com/NVIDIA-NeMo/Skills)|[Reference pipeline for SDG & Eval](https://nvidia.github.io/NeMo-Skills/)|NA|NA|NA|Agnostic|NA|
|[NeMo Emerging Optimizers](https://github.com/NVIDIA-NeMo/Emerging-Optimizers)|[Collection of Optimizers](https://docs.nvidia.com/nemo/emerging-optimizers/0.1.0/index.html)|NA|Agnostic|NA|NA|NA|
|[NeMo DFM](https://github.com/NVIDIA-NeMo/DFM/tree/main)|[Diffusion foundation model training](https://github.com/NVIDIA-NeMo/DFM/tree/main/docs)|PyT native loop|Megatron-core and PyTorch|NA|Diffusion models|NA|
|[Nemotron](https://github.com/NVIDIA-NeMo/Nemotron)|Developer asset hub for Nemotron models|NA|NA|NA|Nemotron models|NA|
|[NeMo Data Designer](https://github.com/NVIDIA-NeMo/DataDesigner)|[Synthetic data generation library](https://docs.nvidia.com/nemo/microservices/latest/design-synthetic-data-from-scratch-or-seeds/)|NA|NA|NA|NA|NA|
|[NeMo Safe Synthesizer](https://github.com/NVIDIA-NeMo/Safe-Synthesizer)|[Privacy-preserving synthetic tabular data](https://docs.nvidia.com/nemo/microservices/latest/generate-private-synthetic-data/)|HF Trainer|PyTorch, Opacus|vLLM|LLM|NA|
|[NeMo Anonymizer](https://github.com/NVIDIA-NeMo/Anonymizer)|PII detection and anonymization|NA|NA|NA|Agnostic|NA|

<div align="center">
Table 1. NeMo Framework Repos
</div>
<p></p>

### Diagram Ilustration of Repos under NeMo Framework (WIP)

![image](/RepoDiagram.png)

<div align="center">
Figure 1. NeMo Framework Repo Overview
</div>
<p></p>

### Some background motivations and historical contexts
The NeMo GitHub Org and its repo collections are created to address the following problems
* **Need for composability**: The [Previous NeMo 2.0 version](https://github.com/NVIDIA/NeMo) is monolithic and encompasses too many things, making it hard for users to find what they need. Container size is also an issue. Breaking down the Monolithic repo into a series of functional-focused repos to facilitate code discovery.
* **Need for customizability**: The [Previous NeMo 2.0 version](https://github.com/NVIDIA/NeMo) uses PyTorch Lighting as the default trainer loop, which provides some out of the box functionality but making it hard to customize. [NeMo Megatron-Bridge](https://github.com/NVIDIA-NeMo/Megatron-Bridge), [NeMo AutoModel](https://github.com/NVIDIA-NeMo/Automodel), and [NeMo RL](https://github.com/NVIDIA-NeMo/RL) have adopted pytorch native custom loop to improve flexibility and ease of use for developers.
### 💬 NeMo Assist
- [4/13/2026][Try NeMo Assist to chat with the NeMo Framework docs and code.](https://nemo-assist.nvidia.com/)

<!--
## Contribution & Support
Expand Down