Skip to content

Conversation

@sbhavani
Copy link
Collaborator

@sbhavani sbhavani commented Feb 1, 2026

Description

Updates README add missing format support documentation, update news section, and fix broken/outdated links.

Type of change

  • Documentation change (change only to the documentation, either a fix or a new content)
  • Bug fix (non-breaking change which fixes an issue)
  • New feature (non-breaking change which adds functionality)
  • Breaking change (fix or feature that would cause existing functionality to not work as expected)
  • Infra/Build change
  • Code refactoring

Changes

  • Add MXFP8 and NVFP4 format support to highlights and description
  • Update FP8 convergence table with MXFP8 results from arxiv paper
  • Remove outdated JAX Toolbox links and "available on request" entries
  • Update Docker container versions to 26.01
  • Fix DeepSpeed and Lightning integration links
  • Add Nemotron 3 paper to Latest News
  • Add quickstart notebook link after PyTorch example

Checklist:

  • I have read and followed the contributing guidelines
  • The functionality is complete
  • I have commented my code, particularly in hard-to-understand areas
  • I have made corresponding changes to the documentation
  • My changes generate no new warnings
  • I have added tests that prove my fix is effective or that my feature works
  • New and existing unit tests pass locally with my changes

@greptile-apps
Copy link
Contributor

greptile-apps bot commented Feb 1, 2026

Greptile Overview

Greptile Summary

Updated README documentation to reflect current Transformer Engine capabilities and fix outdated information. Added MXFP8 and NVFP4 format support documentation for Blackwell GPUs, updated the FP8 convergence table with new MXFP8 results from arxiv paper, updated Docker container versions from 25.08 to 26.01, fixed broken/outdated integration links (DeepSpeed and Lightning), added Nemotron 3 paper to Latest News, and corrected notebook references to use the existing fp8_primer.ipynb instead of the non-existent quickstart.ipynb.

Key changes:

  • Added MXFP8 and NVFP4 format support mentions in description and highlights
  • Updated convergence table with LLM-8B and MoE-16B Megatron Core results
  • Removed outdated "available on request" entries and broken JAX Toolbox links
  • Updated Docker container versions to current release (26.01)
  • Fixed integration documentation links to point to valid resources

Minor issue:

  • Inconsistent spacing in convergence table for "Megatron Core" entries (3 spaces instead of 2)

Confidence Score: 4/5

  • Safe to merge with minor formatting inconsistency
  • Documentation-only changes that improve accuracy and fix broken links. The referenced notebook file exists, convergence table updates appear legitimate with valid arxiv reference, Docker versions are reasonable progression, and integration links point to valid resources. Only minor style issue is inconsistent table spacing.
  • No files require special attention - minor style fixes optional

Important Files Changed

Filename Overview
README.rst Documentation updates to add MXFP8/NVFP4 support info, update convergence table, fix links, and update container versions. Minor formatting inconsistency in convergence table.

Sequence Diagram

sequenceDiagram
    participant Dev as Developer
    participant README as README.rst
    participant Users as Documentation Users
    
    Dev->>README: Add MXFP8/NVFP4 format support info
    Dev->>README: Update FP8 convergence table with MXFP8 results
    Dev->>README: Remove outdated JAX Toolbox links
    Dev->>README: Update Docker container versions (25.08 → 26.01)
    Dev->>README: Fix integration links (DeepSpeed, Lightning)
    Dev->>README: Add Nemotron 3 paper to Latest News
    Dev->>README: Fix notebook references (quickstart → fp8_primer)
    README->>Users: Provide updated documentation
    Users->>README: Access current format support info
    Users->>README: View latest convergence results
    Users->>README: Use correct Docker versions
    Users->>README: Follow working integration links
Loading

Copy link
Contributor

@greptile-apps greptile-apps bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

1 file reviewed, 1 comment

Edit Code Review Agent Settings | Greptile

README.rst Outdated
loss = out.sum()
loss.backward()

For a tutorial with more details, see the `Quickstart Notebook <https://github.com/NVIDIA/TransformerEngine/blob/main/docs/examples/quickstart.ipynb>`_.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

The referenced quickstart.ipynb file does not exist in docs/examples/. The actual notebooks in that directory are fp8_primer.ipynb, advanced_optimizations.ipynb, and te_jax_integration.ipynb. Consider using one of these existing notebooks or creating the quickstart notebook before merging.

…t info

- Add MXFP8 and NVFP4 format support to highlights and description
- Update FP8 convergence table with MXFP8 results from arxiv paper
- Remove outdated JAX-Toolbox links and "available on request" entries
- Update Docker container versions to 26.01
- Fix DeepSpeed and Lightning integration links
- Add Nemotron 3 paper to Latest News
- Add quickstart notebook link after PyTorch example

Signed-off-by: Santosh Bhavani <santosh.bhavani@live.com>
Copy link
Contributor

@greptile-apps greptile-apps bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

1 file reviewed, 1 comment

Edit Code Review Agent Settings | Greptile

README.rst Outdated
Comment on lines 354 to 358
| LLM-8B | Megatron Core | https://arxiv.org/abs/2506.08027 |
+------------+------------------+---------------------------------------------------------------------------------------------------------+
| MPT-13B | Mosaic Composer | https://www.databricks.com/blog/turbocharged-training-optimizing-databricks-mosaic-ai-stack-fp8 |
+------------+------------------+---------------------------------------------------------------------------------------------------------+
| GPT-22B | NeMo Framework | Available on request |
| MoE-16B | Megatron Core | https://arxiv.org/abs/2506.08027 |
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

extra whitespace in "Megatron Core" framework names - inconsistent with other rows

Suggested change
| LLM-8B | Megatron Core | https://arxiv.org/abs/2506.08027 |
+------------+------------------+---------------------------------------------------------------------------------------------------------+
| MPT-13B | Mosaic Composer | https://www.databricks.com/blog/turbocharged-training-optimizing-databricks-mosaic-ai-stack-fp8 |
+------------+------------------+---------------------------------------------------------------------------------------------------------+
| GPT-22B | NeMo Framework | Available on request |
| MoE-16B | Megatron Core | https://arxiv.org/abs/2506.08027 |
| LLM-8B | Megatron Core | https://arxiv.org/abs/2506.08027 |
+------------+------------------+---------------------------------------------------------------------------------------------------------+
| MPT-13B | Mosaic Composer | https://www.databricks.com/blog/turbocharged-training-optimizing-databricks-mosaic-ai-stack-fp8 |
+------------+------------------+---------------------------------------------------------------------------------------------------------+
| MoE-16B | Megatron Core | https://arxiv.org/abs/2506.08027 |

Note: If this suggestion doesn't match your team's coding style, reply to this and let me know. I'll remember it for next time!

- Replace quickstart.ipynb link with fp8_primer.ipynb (file exists)
- Fix extra whitespace in Megatron Core table rows

Signed-off-by: Santosh Bhavani <santosh.bhavani@live.com>
Copy link
Contributor

@greptile-apps greptile-apps bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

1 file reviewed, 2 comments

Edit Code Review Agent Settings | Greptile

| LLama2-7B | Alibaba Pai | https://mp.weixin.qq.com/s/NQT0uKXLbXyh5031zBdeBQ |
+------------+------------------+---------------------------------------------------------------------------------------------------------+
| T5-11B | JAX/T5x | Available on request |
| LLM-8B | Megatron Core | https://arxiv.org/abs/2506.08027 |
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

extra space before "Megatron Core" - should have 2 spaces like other rows, not 3

Suggested change
| LLM-8B | Megatron Core | https://arxiv.org/abs/2506.08027 |
| LLM-8B | Megatron Core | https://arxiv.org/abs/2506.08027 |

| MPT-13B | Mosaic Composer | https://www.databricks.com/blog/turbocharged-training-optimizing-databricks-mosaic-ai-stack-fp8 |
+------------+------------------+---------------------------------------------------------------------------------------------------------+
| GPT-22B | NeMo Framework | Available on request |
| MoE-16B | Megatron Core | https://arxiv.org/abs/2506.08027 |
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

extra space before "Megatron Core" - should have 2 spaces like other rows, not 3

Suggested change
| MoE-16B | Megatron Core | https://arxiv.org/abs/2506.08027 |
| MoE-16B | Megatron Core | https://arxiv.org/abs/2506.08027 |

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant