You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Why are the dataset requirements for ppo not the same as for Preference Dataset: stage: dpo? Both should be RLHF techniques, PPO compute rewards which are typically needed not for SFT-style.
reacted with thumbs up emoji reacted with thumbs down emoji reacted with laugh emoji reacted with hooray emoji reacted with confused emoji reacted with heart emoji reacted with rocket emoji reacted with eyes emoji
-
Why are the dataset requirements for
ppo
not the same as forPreference Dataset
:stage: dpo
? Both should be RLHF techniques, PPO compute rewards which are typically needed not for SFT-style.LLaMA-Factory/examples/train_lora/llama3_lora_ppo.yaml
Line 12 in c93d55b
LLaMA-Factory/examples/train_lora/llama3_lora_dpo.yaml
Line 13 in c93d55b
Beta Was this translation helpful? Give feedback.
All reactions