Add training doc signposting to TRL (#14439)
Signed-off-by: Harry Mellor <19981378+hmellor@users.noreply.github.com>
This commit is contained in:
parent
b8b0ccbd2d
commit
be0b399d74
@ -100,6 +100,14 @@ features/compatibility_matrix
|
|||||||
|
|
||||||
% Details about running vLLM
|
% Details about running vLLM
|
||||||
|
|
||||||
|
:::{toctree}
|
||||||
|
:caption: Training
|
||||||
|
:maxdepth: 1
|
||||||
|
|
||||||
|
training/trl.md
|
||||||
|
|
||||||
|
:::
|
||||||
|
|
||||||
:::{toctree}
|
:::{toctree}
|
||||||
:caption: Inference and Serving
|
:caption: Inference and Serving
|
||||||
:maxdepth: 1
|
:maxdepth: 1
|
||||||
|
13
docs/source/training/trl.md
Normal file
13
docs/source/training/trl.md
Normal file
@ -0,0 +1,13 @@
|
|||||||
|
# Transformers Reinforcement Learning
|
||||||
|
|
||||||
|
Transformers Reinforcement Learning (TRL) is a full stack library that provides a set of tools to train transformer language models with methods like Supervised Fine-Tuning (SFT), Group Relative Policy Optimization (GRPO), Direct Preference Optimization (DPO), Reward Modeling, and more. The library is integrated with 🤗 transformers.
|
||||||
|
|
||||||
|
Online methods such as GRPO or Online DPO require the model to generate completions. vLLM can be used to generate these completions!
|
||||||
|
|
||||||
|
See the guide [vLLM for fast generation in online methods](https://huggingface.co/docs/trl/main/en/speeding_up_training#vllm-for-fast-generation-in-online-methods) in the TRL documentation for more information.
|
||||||
|
|
||||||
|
:::{seealso}
|
||||||
|
For more information on the `use_vllm` flag you can provide to the configs of these online methods, see:
|
||||||
|
- [`trl.GRPOConfig.use_vllm`](https://huggingface.co/docs/trl/main/en/grpo_trainer#trl.GRPOConfig.use_vllm)
|
||||||
|
- [`trl.OnlineDPOConfig.use_vllm`](https://huggingface.co/docs/trl/main/en/online_dpo_trainer#trl.OnlineDPOConfig.use_vllm)
|
||||||
|
:::
|
Loading…
x
Reference in New Issue
Block a user