Replies: 1 comment
-
Now verl can't load and continue training from a pre-trained LoRA adapter, after the pr (#3523) it can work. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I’m implementing a multi-stage reinforcement learning (RL) pipeline for reasoning tasks using GRPO, and I’d like to:
Setup:
Qwen3
In standard Hugging Face + PEFT workflows, I can load a pre-trained LoRA adapter like this:
However, in my current GRPO trainer config, LoRA is initialized from scratch via these parameters:
There doesn’t appear to be a config option (e.g.,
lora_path
orpretrained_adapter_name_or_path
) to load an existing adapter instead of initializing new LoRA weights.Question:
How can I configure the GRPO trainer to load and continue training from a pre-trained LoRA adapter? Is this supported, and if so, what’s the correct way to specify the adapter path in the config or code?
Beta Was this translation helpful? Give feedback.
All reactions