Skip to content

Commit

Permalink
Addresses documentation bugs (#245)
Browse files Browse the repository at this point in the history
Signed-off-by: Terry Kong <terryk@nvidia.com>
  • Loading branch information
terrykong committed Jul 22, 2024
1 parent 479e839 commit c098da8
Show file tree
Hide file tree
Showing 2 changed files with 5 additions and 3 deletions.
6 changes: 4 additions & 2 deletions docs/user-guide/rlhf.rst
Original file line number Diff line number Diff line change
Expand Up @@ -23,7 +23,7 @@ After running this script you should have the files ``{train,test}_comparisons.j
Reward Model Training
#####################

The reward model is used to score how good a response is. It is trained using a pairwise comparison loss and therefore requires a dataset of response pairs, where one response in the pair is ranked higher than the other. A good reward model is cruical for the success of the PPO training.
The reward model is used to score how good a response is. It is trained using a pairwise comparison loss and therefore requires a dataset of response pairs, where one response in the pair is ranked higher than the other. A good reward model is crucial for the success of the PPO training.

Data Preprocessing
%%%%%%%%%%%%%%%%%%
Expand Down Expand Up @@ -170,6 +170,8 @@ To launch the server:
.. code-block:: bash
#!/bin/bash
# Example: If using the reward model trained from the above configuration, you can find
the trained reward model checkpoint here: "/results/checkpoints/megatron_gpt.nemo"
CHECKPOINT_NEMO_FILE="/path/to/trained_rm.nemo"
GPFS="/path/to/nemo-aligner-repo"
Expand Down Expand Up @@ -368,4 +370,4 @@ Once you've completed RLHF training, you can serve your model using the `megatro
Scaling the tutorial to bigger models
#####################################
The above tutorial is a way to get started with RLHF but is not the most optimal performant or convergence configuration. When running RLHF fully, we expect around +0.4 to +0.5 on the MT-bench score. It is cruical to start with a good SFT model and monitor the response length.
The above tutorial is a way to get started with RLHF but is not the most optimal performant or convergence configuration. When running RLHF fully, we expect around +0.4 to +0.5 on the MT-bench score. It is crucial to start with a good SFT model and monitor the response length.
2 changes: 1 addition & 1 deletion nemo_aligner/package_info.py
Original file line number Diff line number Diff line change
Expand Up @@ -27,7 +27,7 @@
__package_name__ = "nemo_aligner"
__contact_names__ = "NVIDIA"
__contact_emails__ = "nemo-toolkit@nvidia.com"
__homepage__ = "https://docs.nvidia.com/nemo-framework/user-guide/latest/ModelAlignment/index.html"
__homepage__ = "https://docs.nvidia.com/nemo-framework/user-guide/latest/modelalignment/index.html"
__repository_url__ = "https://github.com/NVIDIA/NeMo-Aligner"
__download_url__ = "https://github.com/NVIDIA/NeMo-Aligner/releases"
__description__ = "NeMo-Aligner - a toolkit for model alignment"
Expand Down

0 comments on commit c098da8

Please sign in to comment.