Skip to content

Commit 54a56e8

Browse files
committed
fix spelling errors
Signed-off-by: noeyy-mino <174223378+noeyy-mino@users.noreply.github.com>
1 parent 2bad66c commit 54a56e8

6 files changed

Lines changed: 11 additions & 11 deletions

File tree

CHANGELOG.rst

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -255,7 +255,7 @@ NVIDIA Model Optimizer Changelog
255255
- NeMo and Megatron-LM distributed checkpoint (``torch-dist``) stored with legacy version can no longer be loaded. The remedy is to load the legacy distributed checkpoint with 0.29 and store a ``torch`` checkpoint and resume with 0.31 to convert to a new format. The following changes only apply to storing and resuming distributed checkpoint.
256256
- ``quantizer_state`` of :class:`TensorQuantizer <modelopt.torch.quantization.nn.modules.TensorQuantizer>` is now stored in ``extra_state`` of :class:`QuantModule <modelopt.torch.quantization.nn.module.QuantModule>` where it used to be stored in the sharded ``modelopt_state``.
257257
- The dtype and shape of ``amax`` and ``pre_quant_scale`` stored in the distributed checkpoint are now restored. Some dtype and shape are previously changed to make all decoder layers to have homogeneous structure in the checkpoint.
258-
- Together with megatron.core-0.13, quantized model will store and resume distributed checkpoint in a heterogenous format.
258+
- Together with megatron.core-0.13, quantized model will store and resume distributed checkpoint in a heterogeneous format.
259259
- auto_quantize API now accepts a list of quantization config dicts as the list of quantization choices.
260260
- This API previously accepts a list of strings of quantization format names. It was therefore limited to only pre-defined quantization formats unless through some hacks.
261261
- With this change, now user can easily use their own custom quantization formats for auto_quantize.

docs/source/guides/7_nas.rst

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -570,7 +570,7 @@ NAS-based training
570570
During training of an search space, we simultaneously train both the model's weights and
571571
architecture:
572572

573-
* Using :mod:`modelopt.torch.nas<modelopt.torch.nas>` you can re-use your existing
573+
* Using :mod:`modelopt.torch.nas<modelopt.torch.nas>` you can reuse your existing
574574
training loop to train the search space.
575575

576576
* During search space training the entire collection of subnets is automatically trained together

examples/diffusers/README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ Cache Diffusion is a technique that reuses cached outputs from previous diffusio
1212
| :------------: | :------------: | :------------: | :------------: |
1313
| Pre-Requisites | Required & optional packages to use this technique | \[[Link](#pre-requisites)\] | |
1414
| Getting Started | Learn how to optimize your models using quantization/cache diffusion to reduce precision and improve inference efficiency | \[[Link](#getting-started)\] | \[[docs](https://nvidia.github.io/Model-Optimizer/guides/1_quantization.html)\] |
15-
| Support Matrix | View the support matrix to see quantization/cahce diffusion compatibility and feature availability across different models | \[[Link](#support-matrix)\] | \[[docs](https://nvidia.github.io/Model-Optimizer/guides/1_quantization.html)\] |
15+
| Support Matrix | View the support matrix to see quantization/cache diffusion compatibility and feature availability across different models | \[[Link](#support-matrix)\] | \[[docs](https://nvidia.github.io/Model-Optimizer/guides/1_quantization.html)\] |
1616
| Cache Diffusion | Caching technique to accelerate inference without compromising quality | \[[Link](#cache-diffusion)\] | |
1717
| Post Training Quantization (PTQ) | Example scripts on how to run PTQ on diffusion models | \[[Link](#post-training-quantization-ptq)\] | \[[docs](https://nvidia.github.io/Model-Optimizer/guides/1_quantization.html)\] |
1818
| Quantization Aware Training (QAT) | Example scripts on how to run QAT on diffusion models | \[[Link](#quantization-aware-training-qat)\] | \[[docs](https://nvidia.github.io/Model-Optimizer/guides/1_quantization.html)\] |
@@ -389,7 +389,7 @@ This simple code demonstrates how to evaluate images generated by diffusion (or
389389
### Install Requirements
390390

391391
```bash
392-
pip install -r eval/requirments.txt
392+
pip install -r eval/requirements.txt
393393
```
394394

395395
### Data Format

examples/llm_qat/notebooks/QAT_QAD_Walkthrough.ipynb

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -23,22 +23,22 @@
2323
"id": "c3f7f931-ac38-494e-aea8-ca2cd6d05794",
2424
"metadata": {},
2525
"source": [
26-
"## Installing Prerequisites and Dependancies"
26+
"## Installing Prerequisites and Dependencies"
2727
]
2828
},
2929
{
3030
"cell_type": "markdown",
3131
"id": "d7d4f25f-e569-42cf-8022-bb7cc6f9ea6e",
3232
"metadata": {},
3333
"source": [
34-
"If you haven't already, install the required dependencies for this notebook. Key dependancies include:\n",
34+
"If you haven't already, install the required dependencies for this notebook. Key dependencies include:\n",
3535
"\n",
3636
"- nvidia-modelopt\n",
3737
"- torch\n",
3838
"- transformers\n",
3939
"- jupyterlab\n",
4040
"\n",
41-
"This repo contains a `examples/llm_qat/notebooks/requirements.txt` file that can be used to install all required dependancies."
41+
"This repo contains a `examples/llm_qat/notebooks/requirements.txt` file that can be used to install all required dependencies."
4242
]
4343
},
4444
{
@@ -414,7 +414,7 @@
414414
"id": "e471ef6c-1346-4e5e-8782-5e9f2bc38f8a",
415415
"metadata": {},
416416
"source": [
417-
"Once you have quantized the model you can now start the post-training process for QAT. The training process will calculate validation loss at 50 step intervals and save the model. These can be controled by adjusting the `eval_steps` and `output_dir` above along with other `training_args`."
417+
"Once you have quantized the model you can now start the post-training process for QAT. The training process will calculate validation loss at 50 step intervals and save the model. These can be controlled by adjusting the `eval_steps` and `output_dir` above along with other `training_args`."
418418
]
419419
},
420420
{
@@ -799,7 +799,7 @@
799799
"id": "eca645fb-d8d2-4c98-9cb2-afbae7d30d6c",
800800
"metadata": {},
801801
"source": [
802-
"## Stop the TensorRT-LLM Docker contrainer"
802+
"## Stop the TensorRT-LLM Docker container"
803803
]
804804
},
805805
{

examples/specdec_bench/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22

33
## Installation
44

5-
This benchmark is meant to be a lightweight layer ontop of an existing vLLM/SGLang/TRTLLM installation. For example, no install
5+
This benchmark is meant to be a lightweight layer on top of an existing vLLM/SGLang/TRTLLM installation. For example, no install
66
is required if one is running in the following dockers: `vllm/vllm-openai:v0.11.0` (vLLM), `lmsysorg/sglang:v0.5.4.post2` (SGLang), or
77
`nvcr.io/nvidia/tensorrt-llm/release:1.2.0` (TRT-LLM).
88

examples/speculative_decoding/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -89,7 +89,7 @@ For large models, you can export intermediate hidden states to disk and train on
8989

9090
### Dumpping Hidden States to Disk
9191

92-
We support two backends for generating base model hidden states. For better effciency, it is recommended to use TRT-LLM:
92+
We support two backends for generating base model hidden states. For better efficiency, it is recommended to use TRT-LLM:
9393

9494
```bash
9595
python collect_hidden_states/compute_hidden_states_trtllm.py \

0 commit comments

Comments
 (0)