This module is used to export megatron core models to different inference frameworks. Currently we support TRTLLM export . In the future we will be adding support for VLLM etc.
Follow the examples of TensorRT Model Optimizer to perform post training quantization, followed by an export to a HF-like checkpoint for TensorRT-LLM, vLLM, and SGLang deployment.
Follow the instructions in trtllm_export to do export to TRTLLM checkpoint format alone.