Datasets:

ArXiv:
troub1emaker commited on
Commit
af44d85
·
verified ·
1 Parent(s): 15f3658

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. examples/ppo_trainer/run_qwen2-7b_rm_seq_balance_nsys.sh +81 -0
  2. examples/sglang_multiturn/run_qwen2.5-3b_gsm8k_multiturn_4xgpu_server.sh +60 -0
  3. examples/sglang_multiturn/run_qwen2.5-3b_gsm8k_multiturn_server.sh +63 -0
  4. examples/skypilot/README.md +107 -0
  5. examples/skypilot/verl-grpo.yaml +99 -0
  6. examples/skypilot/verl-multiturn-tools.yaml +91 -0
  7. examples/skypilot/verl-ppo.yaml +109 -0
  8. examples/slurm/ray_on_slurm.slurm +98 -0
  9. examples/split_placement/README.md +61 -0
  10. examples/split_placement/main_ppo_split.py +216 -0
  11. examples/split_placement/run_deepseek7b_llm.sh +37 -0
  12. outputs/2026-01-13/23-34-32/.hydra/overrides.yaml +47 -0
  13. outputs/2026-01-13/23-57-30/.hydra/config.yaml +617 -0
  14. outputs/2026-01-14/00-16-12/.hydra/overrides.yaml +47 -0
  15. outputs/2026-01-14/00-17-48/.hydra/config.yaml +617 -0
  16. outputs/2026-01-14/09-28-49/.hydra/overrides.yaml +47 -0
  17. outputs/2026-01-14/09-30-16/.hydra/overrides.yaml +47 -0
  18. outputs/2026-01-14/09-35-00/.hydra/overrides.yaml +47 -0
  19. outputs/2026-01-14/14-26-40/.hydra/hydra.yaml +218 -0
  20. outputs/2026-01-14/14-26-40/.hydra/overrides.yaml +47 -0
  21. outputs/2026-01-14/15-08-55/.hydra/config.yaml +616 -0
  22. outputs/2026-01-14/15-08-55/.hydra/overrides.yaml +47 -0
  23. outputs/2026-01-14/15-16-37/.hydra/config.yaml +616 -0
  24. outputs/2026-01-14/15-16-37/.hydra/hydra.yaml +218 -0
  25. outputs/2026-01-14/15-31-37/.hydra/config.yaml +617 -0
  26. outputs/2026-01-14/15-31-37/.hydra/overrides.yaml +47 -0
  27. outputs/2026-01-14/15-31-37/main_sgrpo.log +0 -0
  28. outputs/2026-01-14/15-45-38/.hydra/hydra.yaml +218 -0
  29. outputs/2026-01-14/15-45-38/main_sgrpo.log +0 -0
  30. outputs/2026-01-14/16-00-30/.hydra/config.yaml +616 -0
  31. outputs/2026-01-14/16-00-30/.hydra/overrides.yaml +47 -0
  32. outputs/2026-01-14/16-00-30/main_sgrpo.log +0 -0
  33. outputs/2026-01-14/16-50-33/.hydra/config.yaml +616 -0
  34. outputs/2026-01-14/16-54-53/.hydra/overrides.yaml +47 -0
  35. outputs/2026-01-14/16-54-53/main_sgrpo.log +0 -0
  36. outputs/2026-01-14/17-46-32/.hydra/hydra.yaml +218 -0
  37. outputs/2026-01-14/17-46-32/.hydra/overrides.yaml +47 -0
  38. outputs/2026-01-14/17-46-32/main_sgrpo.log +0 -0
  39. outputs/2026-01-14/19-57-10/.hydra/overrides.yaml +47 -0
  40. outputs/2026-01-14/19-57-10/main_sgrpo.log +0 -0
  41. outputs/2026-01-14/19-57-55/.hydra/overrides.yaml +47 -0
  42. outputs/2026-01-14/19-57-55/main_sgrpo.log +0 -0
  43. outputs/2026-01-14/20-02-28/.hydra/config.yaml +616 -0
  44. outputs/2026-01-14/20-02-28/main_sgrpo.log +0 -0
  45. outputs/2026-01-14/20-19-23/.hydra/config.yaml +616 -0
  46. outputs/2026-01-14/20-19-23/main_sgrpo.log +0 -0
  47. outputs/2026-01-14/20-27-06/.hydra/hydra.yaml +218 -0
  48. outputs/2026-01-14/20-54-11/.hydra/hydra.yaml +218 -0
  49. outputs/2026-01-14/21-01-00/.hydra/config.yaml +616 -0
  50. outputs/2026-01-14/21-41-16/.hydra/overrides.yaml +47 -0
examples/ppo_trainer/run_qwen2-7b_rm_seq_balance_nsys.sh ADDED
@@ -0,0 +1,81 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ set -x
2
+
3
+ gsm8k_train_path=$HOME/data/gsm8k/train.parquet
4
+ gsm8k_test_path=$HOME/data/gsm8k/test.parquet
5
+ math_train_path=$HOME/data/math/train.parquet
6
+ math_test_path=$HOME/data/math/test.parquet
7
+
8
+ train_files=${train_files:-"$gsm8k_train_path"}
9
+ test_files=${test_files:-"$gsm8k_test_path"}
10
+
11
+ PROFILE_STEPS="[1,2,5]" # or [] or null
12
+ PROFILE_RANKS_ALL=False # or True
13
+ PROFILE_RANKS=[0,4]
14
+ DISCRETE=True # or True
15
+
16
+ python3 -m verl.trainer.main_ppo \
17
+ algorithm.adv_estimator=gae \
18
+ data.train_files="$train_files" \
19
+ data.val_files="$test_files" \
20
+ data.train_batch_size=4096 \
21
+ data.max_prompt_length=4096 \
22
+ data.max_response_length=4096 \
23
+ data.filter_overlong_prompts=True \
24
+ data.truncation='error' \
25
+ data.return_raw_chat=True \
26
+ actor_rollout_ref.model.path=Qwen/Qwen2-7B-Instruct \
27
+ actor_rollout_ref.actor.optim.lr=1e-6 \
28
+ actor_rollout_ref.model.use_remove_padding=True \
29
+ actor_rollout_ref.model.enable_gradient_checkpointing=True \
30
+ actor_rollout_ref.actor.ppo_mini_batch_size=512 \
31
+ actor_rollout_ref.actor.ppo_micro_batch_size_per_gpu=2 \
32
+ actor_rollout_ref.actor.use_dynamic_bsz=True \
33
+ actor_rollout_ref.actor.ppo_max_token_len_per_gpu=12000 \
34
+ actor_rollout_ref.actor.fsdp_config.param_offload=False \
35
+ actor_rollout_ref.actor.fsdp_config.optimizer_offload=False \
36
+ actor_rollout_ref.actor.use_kl_loss=False \
37
+ actor_rollout_ref.actor.profiler.enable=True \
38
+ actor_rollout_ref.actor.profiler.ranks=$PROFILE_RANKS \
39
+ actor_rollout_ref.actor.profiler.all_ranks=$PROFILE_RANKS_ALL \
40
+ actor_rollout_ref.rollout.tensor_model_parallel_size=2 \
41
+ actor_rollout_ref.rollout.name=vllm \
42
+ actor_rollout_ref.rollout.gpu_memory_utilization=0.5 \
43
+ actor_rollout_ref.rollout.log_prob_max_token_len_per_gpu=24000 \
44
+ critic.optim.lr=1e-5 \
45
+ critic.model.use_remove_padding=True \
46
+ critic.model.path=Qwen/Qwen2-7B-Instruct \
47
+ critic.model.enable_gradient_checkpointing=True \
48
+ critic.ppo_micro_batch_size_per_gpu=2 \
49
+ critic.use_dynamic_bsz=True \
50
+ critic.ppo_max_token_len_per_gpu=98304 \
51
+ critic.model.fsdp_config.param_offload=False \
52
+ critic.model.fsdp_config.optimizer_offload=False \
53
+ critic.profiler.enable=True \
54
+ critic.profiler.ranks=$PROFILE_RANKS \
55
+ critic.profiler.all_ranks=$PROFILE_RANKS_ALL \
56
+ reward_model.enable=True \
57
+ reward_model.model.path=sfairXC/FsfairX-LLaMA3-RM-v0.1\
58
+ reward_model.model.use_remove_padding=True \
59
+ reward_model.model.fsdp_config.param_offload=True \
60
+ reward_model.micro_batch_size_per_gpu=32 \
61
+ reward_model.use_dynamic_bsz=True \
62
+ reward_model.forward_max_token_len_per_gpu=98304 \
63
+ reward_model.profiler.enable=True \
64
+ reward_model.profiler.ranks=$PROFILE_RANKS \
65
+ reward_model.profiler.all_ranks=$PROFILE_RANKS_ALL \
66
+ algorithm.use_kl_in_reward=False \
67
+ trainer.critic_warmup=0 \
68
+ trainer.logger='["console","wandb"]' \
69
+ trainer.project_name='verl_example_gsm8k' \
70
+ trainer.experiment_name='qwen2-7b_hybrid_rm_bsz8k_p4k_r4k_seq_packing' \
71
+ trainer.n_gpus_per_node=8 \
72
+ trainer.val_before_train=False \
73
+ trainer.nnodes=1 \
74
+ trainer.save_freq=-1 \
75
+ trainer.test_freq=-1 \
76
+ trainer.total_epochs=15 \
77
+ trainer.total_training_steps=6 \
78
+ global_profiler.profile_continuous_steps=True \
79
+ global_profiler.tool=nsys \
80
+ global_profiler.steps=$PROFILE_STEPS \
81
+ global_profiler.global_tool_config.nsys.discrete=$DISCRETE $@
examples/sglang_multiturn/run_qwen2.5-3b_gsm8k_multiturn_4xgpu_server.sh ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # run on 4xH100
2
+ # make sure your current working directory is the root of the project
3
+
4
+ set -x
5
+ export HYDRA_FULL_ERROR=1
6
+ ulimit -n 65535
7
+
8
+ PROJECT_DIR="$(pwd)"
9
+ CONFIG_PATH="$PROJECT_DIR/examples/sglang_multiturn/config"
10
+
11
+ python3 -m verl.trainer.main_ppo \
12
+ --config-path="$CONFIG_PATH" \
13
+ --config-name='gsm8k_multiturn_grpo_server' \
14
+ algorithm.adv_estimator=grpo \
15
+ data.train_batch_size=256 \
16
+ data.max_prompt_length=1024 \
17
+ data.max_response_length=1024 \
18
+ data.filter_overlong_prompts=True \
19
+ data.truncation='error' \
20
+ data.return_raw_chat=True \
21
+ actor_rollout_ref.model.path=Qwen/Qwen2.5-3B-Instruct \
22
+ actor_rollout_ref.actor.optim.lr=1e-6 \
23
+ actor_rollout_ref.model.use_remove_padding=True \
24
+ actor_rollout_ref.actor.ppo_mini_batch_size=256 \
25
+ actor_rollout_ref.actor.ppo_micro_batch_size_per_gpu=32 \
26
+ actor_rollout_ref.actor.use_kl_loss=True \
27
+ actor_rollout_ref.actor.kl_loss_coef=0.001 \
28
+ actor_rollout_ref.actor.kl_loss_type=low_var_kl \
29
+ actor_rollout_ref.actor.entropy_coeff=0 \
30
+ actor_rollout_ref.model.enable_gradient_checkpointing=True \
31
+ actor_rollout_ref.actor.fsdp_config.param_offload=False \
32
+ actor_rollout_ref.actor.fsdp_config.optimizer_offload=False \
33
+ actor_rollout_ref.rollout.log_prob_micro_batch_size_per_gpu=32 \
34
+ actor_rollout_ref.rollout.tensor_model_parallel_size=4 \
35
+ actor_rollout_ref.rollout.name=sglang \
36
+ actor_rollout_ref.rollout.gpu_memory_utilization=0.6 \
37
+ actor_rollout_ref.rollout.n=16 \
38
+ actor_rollout_ref.ref.log_prob_micro_batch_size_per_gpu=32 \
39
+ actor_rollout_ref.ref.fsdp_config.param_offload=True \
40
+ algorithm.use_kl_in_reward=False \
41
+ trainer.critic_warmup=0 \
42
+ trainer.logger='["console", "wandb"]' \
43
+ trainer.project_name='gsm8k_async_rl_server' \
44
+ trainer.experiment_name='qwen2.5-3b_function_rm-gsm8k-async-sgl-multi-w-tool-verify-n16-4cards' \
45
+ trainer.n_gpus_per_node=4 \
46
+ trainer.nnodes=1 \
47
+ trainer.save_freq=-1 \
48
+ trainer.test_freq=20 \
49
+ trainer.total_epochs=15 \
50
+ actor_rollout_ref.actor.ppo_max_token_len_per_gpu=8192 \
51
+ actor_rollout_ref.rollout.log_prob_max_token_len_per_gpu=8192 \
52
+ actor_rollout_ref.ref.log_prob_max_token_len_per_gpu=8192 \
53
+ critic.ppo_max_token_len_per_gpu=8192 \
54
+ critic.forward_max_token_len_per_gpu=8192 \
55
+ data.train_files=$HOME/data/gsm8k/train.parquet \
56
+ data.val_files=$HOME/data/gsm8k/test.parquet \
57
+ actor_rollout_ref.rollout.multi_turn.tool_config_path="$PROJECT_DIR/examples/sglang_multiturn/config/tool_config/gsm8k_tool_config.yaml" \
58
+ actor_rollout_ref.rollout.multi_turn.interaction_config_path="$PROJECT_DIR/examples/sglang_multiturn/config/interaction_config/gsm8k_interaction_config.yaml" \
59
+ actor_rollout_ref.rollout.multi_turn.max_user_turns=1 \
60
+ $@
examples/sglang_multiturn/run_qwen2.5-3b_gsm8k_multiturn_server.sh ADDED
@@ -0,0 +1,63 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # run on 8xH100
2
+ # make sure your current working directory is the root of the project
3
+
4
+ set -x
5
+
6
+ ulimit -n 65535
7
+
8
+ PROJECT_DIR="$(pwd)"
9
+ CONFIG_PATH="$PROJECT_DIR/examples/sglang_multiturn/config"
10
+
11
+ function now() {
12
+ date '+%d-%H-%M'
13
+ }
14
+
15
+ EXPERIMENT_NAME="qwen2.5-3b_baseline_$(now)"
16
+
17
+ python3 -m verl.trainer.main_ppo \
18
+ --config-path="$CONFIG_PATH" \
19
+ --config-name='gsm8k_multiturn_grpo_server' \
20
+ algorithm.adv_estimator=grpo \
21
+ data.train_batch_size=256 \
22
+ data.max_prompt_length=1024 \
23
+ data.max_response_length=1024 \
24
+ data.filter_overlong_prompts=True \
25
+ data.truncation='error' \
26
+ data.return_raw_chat=True \
27
+ actor_rollout_ref.model.path=Qwen/Qwen2.5-3B-Instruct \
28
+ actor_rollout_ref.actor.optim.lr=1e-6 \
29
+ actor_rollout_ref.model.use_remove_padding=True \
30
+ actor_rollout_ref.actor.ppo_mini_batch_size=256 \
31
+ actor_rollout_ref.actor.ppo_micro_batch_size_per_gpu=32 \
32
+ actor_rollout_ref.actor.use_kl_loss=True \
33
+ actor_rollout_ref.actor.kl_loss_coef=0.001 \
34
+ actor_rollout_ref.actor.kl_loss_type=low_var_kl \
35
+ actor_rollout_ref.actor.entropy_coeff=0 \
36
+ actor_rollout_ref.model.enable_gradient_checkpointing=True \
37
+ actor_rollout_ref.actor.fsdp_config.param_offload=False \
38
+ actor_rollout_ref.actor.fsdp_config.optimizer_offload=False \
39
+ actor_rollout_ref.rollout.log_prob_micro_batch_size_per_gpu=32 \
40
+ actor_rollout_ref.rollout.tensor_model_parallel_size=1 \
41
+ actor_rollout_ref.rollout.name=sglang \
42
+ actor_rollout_ref.rollout.gpu_memory_utilization=0.85 \
43
+ actor_rollout_ref.rollout.multi_stage_wake_up=True \
44
+ actor_rollout_ref.rollout.n=16 \
45
+ actor_rollout_ref.ref.log_prob_micro_batch_size_per_gpu=32 \
46
+ actor_rollout_ref.ref.fsdp_config.param_offload=True \
47
+ actor_rollout_ref.rollout.over_sample_rate=0 \
48
+ algorithm.use_kl_in_reward=False \
49
+ trainer.critic_warmup=0 \
50
+ trainer.logger='["console","wandb"]' \
51
+ trainer.project_name='multi-turn-grpo-qwen2.5-3b-sglang' \
52
+ trainer.experiment_name=$EXPERIMENT_NAME \
53
+ trainer.n_gpus_per_node=8 \
54
+ trainer.nnodes=1 \
55
+ trainer.save_freq=-1 \
56
+ trainer.test_freq=20 \
57
+ trainer.val_before_train=True \
58
+ data.train_files=$HOME/data/gsm8k/train.parquet \
59
+ data.val_files=$HOME/data/gsm8k/test.parquet \
60
+ actor_rollout_ref.rollout.multi_turn.tool_config_path="$PROJECT_DIR/examples/sglang_multiturn/config/tool_config/gsm8k_tool_config.yaml" \
61
+ trainer.total_epochs=15 \
62
+ actor_rollout_ref.rollout.update_weights_bucket_megabytes=512 $@
63
+
examples/skypilot/README.md ADDED
@@ -0,0 +1,107 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # verl with SkyPilot
2
+
3
+ Run verl reinforcement learning training jobs on Kubernetes clusters or cloud platforms with GPU nodes using [SkyPilot](https://github.com/skypilot-org/skypilot).
4
+
5
+ ## Installation and Configuration
6
+
7
+ ### Step 1: Install SkyPilot
8
+
9
+ Choose the installation based on your target platform:
10
+
11
+ ```bash
12
+ # For Kubernetes only
13
+ pip install "skypilot[kubernetes]"
14
+
15
+ # For AWS
16
+ pip install "skypilot[aws]"
17
+
18
+ # For Google Cloud Platform
19
+ pip install "skypilot[gcp]"
20
+
21
+ # For Azure
22
+ pip install "skypilot[azure]"
23
+
24
+ # For multiple platforms
25
+ pip install "skypilot[kubernetes,aws,gcp,azure]"
26
+ ```
27
+
28
+ ### Step 2: Configure Your Platform
29
+
30
+ See https://docs.skypilot.co/en/latest/getting-started/installation.html
31
+
32
+ ### Step 3: Set Up Environment Variables
33
+
34
+ Export necessary API keys for experiment tracking:
35
+
36
+ ```bash
37
+ # For Weights & Biases tracking
38
+ export WANDB_API_KEY="your-wandb-api-key"
39
+
40
+ # For HuggingFace gated models (if needed)
41
+ export HF_TOKEN="your-huggingface-token"
42
+ ```
43
+
44
+ ## Examples
45
+
46
+ ### PPO Training
47
+ ```bash
48
+ sky launch -c verl-ppo verl-ppo.yaml --secret WANDB_API_KEY -y
49
+ ```
50
+ Runs PPO training on GSM8K dataset using Qwen2.5-0.5B-Instruct model across 2 nodes with H100 GPUs. Based on examples in [`../ppo_trainer/`](../ppo_trainer/).
51
+
52
+ ### GRPO Training
53
+ ```bash
54
+ sky launch -c verl-grpo verl-grpo.yaml --secret WANDB_API_KEY -y
55
+ ```
56
+ Runs GRPO (Group Relative Policy Optimization) training on MATH dataset using Qwen2.5-7B-Instruct model. Memory-optimized configuration for 2 nodes. Based on examples in [`../grpo_trainer/`](../grpo_trainer/).
57
+
58
+ ### Multi-turn Tool Usage Training
59
+ ```bash
60
+ sky launch -c verl-multiturn verl-multiturn-tools.yaml --secret WANDB_API_KEY --secret HF_TOKEN -y
61
+ ```
62
+ Single-node training with 8xH100 GPUs for multi-turn tool usage with Qwen2.5-3B-Instruct. Includes tool and interaction configurations for GSM8K. Based on examples in [`../sglang_multiturn/`](../sglang_multiturn/) but uses vLLM instead of sglang.
63
+
64
+ ## Configuration
65
+
66
+ The example YAML files are pre-configured with:
67
+
68
+ - **Infrastructure**: Kubernetes clusters (`infra: k8s`) - can be changed to `infra: aws` or `infra: gcp`, etc.
69
+ - **Docker Image**: verl's official Docker image with CUDA 12.6 support
70
+ - **Setup**: Automatically clones and installs verl from source
71
+ - **Datasets**: Downloads required datasets during setup phase
72
+ - **Ray Cluster**: Configures distributed training across nodes
73
+ - **Logging**: Supports Weights & Biases via `--secret WANDB_API_KEY`
74
+ - **Models**: Supports gated HuggingFace models via `--secret HF_TOKEN`
75
+
76
+ ## Launch Command Options
77
+
78
+ - `-c <name>`: Cluster name for managing the job
79
+ - `--secret KEY`: Pass secrets for API keys (can be used multiple times)
80
+ - `-y`: Skip confirmation prompt
81
+
82
+ ## Monitoring Your Jobs
83
+
84
+ ### Check cluster status
85
+ ```bash
86
+ sky status
87
+ ```
88
+
89
+ ### View logs
90
+ ```bash
91
+ sky logs verl-ppo # View logs for the PPO job
92
+ ```
93
+
94
+ ### SSH into head node
95
+ ```bash
96
+ ssh verl-ppo
97
+ ```
98
+
99
+ ### Access Ray dashboard
100
+ ```bash
101
+ sky status --endpoint 8265 verl-ppo # Get dashboard URL
102
+ ```
103
+
104
+ ### Stop a cluster
105
+ ```bash
106
+ sky down verl-ppo
107
+ ```
examples/skypilot/verl-grpo.yaml ADDED
@@ -0,0 +1,99 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ resources:
2
+ infra: k8s
3
+ accelerators: H100:1
4
+ memory: 128+
5
+ image_id: docker:verlai/verl:base-verl0.5-cu126-cudnn9.8-torch2.7.0-fa2.7.4
6
+ ports: 8265
7
+
8
+ num_nodes: 2
9
+
10
+ secrets:
11
+ WANDB_API_KEY:
12
+
13
+ setup: |
14
+ rm -rf verl
15
+ git clone https://github.com/volcengine/verl.git
16
+ cd verl
17
+ pip3 install -v -e .[vllm]
18
+ pip3 install flashinfer-python
19
+ echo "Downloading Math dataset..."
20
+ mkdir -p ~/data/math
21
+ python3 "$(pwd)/examples/data_preprocess/math_dataset.py" --local_dir ~/data/math
22
+ echo "Math dataset download completed"
23
+
24
+ run: |
25
+ HEAD_IP=$(echo "$SKYPILOT_NODE_IPS" | head -n1)
26
+ NUM_NODES=$SKYPILOT_NUM_NODES
27
+ NUM_GPUS_PER_NODE=$SKYPILOT_NUM_GPUS_PER_NODE
28
+
29
+ if [ "$SKYPILOT_NODE_RANK" == "0" ]; then
30
+ echo "Starting Ray head node..."
31
+ ps aux | grep ray | grep 6379 &> /dev/null || ray start --head --disable-usage-stats \
32
+ --port=6379 \
33
+ --dashboard-host=0.0.0.0 \
34
+ --dashboard-port=8265
35
+
36
+ # Wait for all worker nodes to join
37
+ retry_count=0
38
+ max_retries=30
39
+ while [ $retry_count -lt $max_retries ]; do
40
+ connected_nodes=$(ray status 2>/dev/null | grep -c "node_" || echo "0")
41
+ echo "Connected nodes: $connected_nodes/$NUM_NODES (attempt $((retry_count+1))/$max_retries)"
42
+
43
+ if [ "$connected_nodes" -ge "$NUM_NODES" ]; then
44
+ echo "All nodes connected to Ray cluster"
45
+ break
46
+ fi
47
+
48
+ retry_count=$((retry_count+1))
49
+ sleep 10
50
+ done
51
+
52
+ python3 -m verl.trainer.main_ppo \
53
+ algorithm.adv_estimator=grpo \
54
+ data.train_files=$HOME/data/math/train.parquet \
55
+ data.val_files=$HOME/data/math/test.parquet \
56
+ data.train_batch_size=32 \
57
+ data.max_prompt_length=256 \
58
+ data.max_response_length=256 \
59
+ data.filter_overlong_prompts=True \
60
+ data.truncation='error' \
61
+ actor_rollout_ref.model.path=Qwen/Qwen2.5-7B-Instruct \
62
+ actor_rollout_ref.actor.optim.lr=1e-6 \
63
+ actor_rollout_ref.model.use_remove_padding=True \
64
+ actor_rollout_ref.actor.ppo_mini_batch_size=16 \
65
+ actor_rollout_ref.actor.ppo_micro_batch_size_per_gpu=4 \
66
+ actor_rollout_ref.actor.ppo_epochs=1 \
67
+ actor_rollout_ref.actor.use_kl_loss=False \
68
+ actor_rollout_ref.actor.entropy_coeff=0 \
69
+ actor_rollout_ref.model.enable_gradient_checkpointing=True \
70
+ actor_rollout_ref.actor.fsdp_config.param_offload=True \
71
+ actor_rollout_ref.actor.fsdp_config.optimizer_offload=True \
72
+ actor_rollout_ref.rollout.log_prob_micro_batch_size_per_gpu=16 \
73
+ actor_rollout_ref.rollout.tensor_model_parallel_size=1 \
74
+ actor_rollout_ref.rollout.name=vllm \
75
+ actor_rollout_ref.rollout.gpu_memory_utilization=0.4 \
76
+ actor_rollout_ref.rollout.n=1 \
77
+ actor_rollout_ref.rollout.enable_chunked_prefill=True \
78
+ actor_rollout_ref.rollout.max_num_batched_tokens=2048 \
79
+ actor_rollout_ref.ref.log_prob_micro_batch_size_per_gpu=16 \
80
+ actor_rollout_ref.ref.fsdp_config.param_offload=True \
81
+ algorithm.use_kl_in_reward=False \
82
+ trainer.critic_warmup=0 \
83
+ trainer.logger=[console,wandb] \
84
+ trainer.project_name=verl_math_grpo_demo \
85
+ trainer.experiment_name=qwen25_7b_grpo \
86
+ trainer.n_gpus_per_node=$NUM_GPUS_PER_NODE \
87
+ trainer.nnodes=$NUM_NODES \
88
+ trainer.save_freq=-1 \
89
+ trainer.test_freq=-1 \
90
+ trainer.total_epochs=1
91
+
92
+ else
93
+ sleep 15
94
+ echo "Starting Ray worker node..."
95
+ ps aux | grep ray | grep $HEAD_IP:6379 &> /dev/null || ray start --address $HEAD_IP:6379 --disable-usage-stats
96
+ sleep 10
97
+ fi
98
+
99
+ echo "Node setup and Ray start script finished for rank $SKYPILOT_NODE_RANK."
examples/skypilot/verl-multiturn-tools.yaml ADDED
@@ -0,0 +1,91 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ resources:
2
+ infra: k8s
3
+ accelerators: H100:8
4
+ memory: 128+
5
+ image_id: docker:verlai/verl:base-verl0.5-cu126-cudnn9.8-torch2.7.0-fa2.7.4
6
+ ports: 8265
7
+
8
+ num_nodes: 1
9
+
10
+ secrets:
11
+ WANDB_API_KEY:
12
+ HF_TOKEN: # in case you're using gated models from the HF hub
13
+
14
+ setup: |
15
+ rm -rf verl
16
+ git clone https://github.com/volcengine/verl.git
17
+ cd verl
18
+ pip3 install -v -e .[vllm]
19
+ pip3 install flashinfer-python
20
+ pip install "transformers<4.54.0" # https://github.com/vllm-project/vllm-ascend/issues/2046
21
+ # Download GSM8K dataset for multiturn tool training
22
+ echo "Downloading GSM8K dataset..."
23
+ mkdir -p ~/data/gsm8k
24
+ python3 "$(pwd)/examples/data_preprocess/gsm8k.py" --local_dir ~/data/gsm8k
25
+ echo "GSM8K dataset download completed"
26
+
27
+ run: |
28
+ NUM_GPUS_PER_NODE=$SKYPILOT_NUM_GPUS_PER_NODE
29
+ PROJECT_DIR="$(pwd)/verl"
30
+ CONFIG_PATH="$PROJECT_DIR/examples/sglang_multiturn/config"
31
+
32
+ # Single node setup - no worker coordination needed
33
+ echo "Starting Ray head node..."
34
+ ps aux | grep ray | grep 6379 &> /dev/null || ray start --head --disable-usage-stats \
35
+ --port=6379 \
36
+ --dashboard-host=0.0.0.0 \
37
+ --dashboard-port=8265
38
+
39
+ cd verl
40
+
41
+ python3 -m verl.trainer.main_ppo \
42
+ --config-path="$CONFIG_PATH" \
43
+ --config-name='gsm8k_multiturn_grpo' \
44
+ algorithm.adv_estimator=grpo \
45
+ data.train_batch_size=512 \
46
+ data.max_prompt_length=1024 \
47
+ data.max_response_length=1024 \
48
+ data.filter_overlong_prompts=True \
49
+ data.truncation='error' \
50
+ data.return_raw_chat=True \
51
+ data.train_files=$HOME/data/gsm8k/train.parquet \
52
+ data.val_files=$HOME/data/gsm8k/test.parquet \
53
+ actor_rollout_ref.model.path=Qwen/Qwen2.5-3B-Instruct \
54
+ actor_rollout_ref.actor.optim.lr=1e-6 \
55
+ actor_rollout_ref.model.use_remove_padding=True \
56
+ actor_rollout_ref.actor.ppo_mini_batch_size=512 \
57
+ actor_rollout_ref.actor.ppo_micro_batch_size_per_gpu=32 \
58
+ actor_rollout_ref.actor.use_kl_loss=True \
59
+ actor_rollout_ref.actor.kl_loss_coef=0.001 \
60
+ actor_rollout_ref.actor.kl_loss_type=low_var_kl \
61
+ actor_rollout_ref.actor.entropy_coeff=0 \
62
+ actor_rollout_ref.model.enable_gradient_checkpointing=True \
63
+ actor_rollout_ref.actor.fsdp_config.param_offload=False \
64
+ actor_rollout_ref.actor.fsdp_config.optimizer_offload=False \
65
+ actor_rollout_ref.rollout.log_prob_micro_batch_size_per_gpu=64 \
66
+ actor_rollout_ref.rollout.tensor_model_parallel_size=4 \
67
+ actor_rollout_ref.rollout.name=vllm \
68
+ actor_rollout_ref.rollout.gpu_memory_utilization=0.5 \
69
+ actor_rollout_ref.rollout.n=16 \
70
+ actor_rollout_ref.ref.log_prob_micro_batch_size_per_gpu=64 \
71
+ actor_rollout_ref.ref.fsdp_config.param_offload=True \
72
+ algorithm.use_kl_in_reward=False \
73
+ trainer.critic_warmup=0 \
74
+ trainer.logger=[console,wandb] \
75
+ trainer.project_name=verl_multiturn_tools \
76
+ trainer.experiment_name=qwen25_7b_gsm8k_multiturn_tools \
77
+ trainer.n_gpus_per_node=$NUM_GPUS_PER_NODE \
78
+ trainer.nnodes=1 \
79
+ trainer.save_freq=10 \
80
+ trainer.test_freq=5 \
81
+ trainer.total_epochs=10 \
82
+ actor_rollout_ref.actor.ppo_max_token_len_per_gpu=8192 \
83
+ actor_rollout_ref.rollout.log_prob_max_token_len_per_gpu=8192 \
84
+ actor_rollout_ref.ref.log_prob_max_token_len_per_gpu=8192 \
85
+ critic.ppo_max_token_len_per_gpu=8192 \
86
+ critic.forward_max_token_len_per_gpu=8192 \
87
+ actor_rollout_ref.rollout.multi_turn.tool_config_path="$PROJECT_DIR/examples/sglang_multiturn/config/tool_config/gsm8k_tool_config.yaml" \
88
+ actor_rollout_ref.rollout.multi_turn.interaction_config_path="$PROJECT_DIR/examples/sglang_multiturn/config/interaction_config/gsm8k_interaction_config.yaml" \
89
+ actor_rollout_ref.rollout.multi_turn.max_user_turns=1
90
+
91
+ echo "Node setup and Ray start script finished for rank $SKYPILOT_NODE_RANK."
examples/skypilot/verl-ppo.yaml ADDED
@@ -0,0 +1,109 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ resources:
2
+ infra: k8s
3
+ accelerators: H100:1
4
+ memory: 128+
5
+ image_id: docker:verlai/verl:base-verl0.5-cu126-cudnn9.8-torch2.7.0-fa2.7.4
6
+ ports: 8265
7
+
8
+ num_nodes: 2
9
+
10
+ secrets:
11
+ WANDB_API_KEY:
12
+
13
+ setup: |
14
+ rm -rf verl
15
+ git clone https://github.com/volcengine/verl.git
16
+ cd verl
17
+ pip3 install -v -e .[vllm]
18
+ pip3 install flashinfer-python
19
+ # Download GSM8K dataset - alternative approach
20
+ echo "Downloading GSM8K dataset..."
21
+ mkdir -p ~/data/gsm8k
22
+ # Check if the script exists and use absolute path
23
+ if [ -f "$(pwd)/examples/data_preprocess/gsm8k.py" ]; then
24
+ python3 "$(pwd)/examples/data_preprocess/gsm8k.py" --local_dir ~/data/gsm8k
25
+ else
26
+ echo "Warning: gsm8k.py script not found, skipping dataset download"
27
+ # You might want to download the dataset manually or use a different approach
28
+ fi
29
+ echo "GSM8K dataset download completed"
30
+
31
+ run: |
32
+ # Get the Head node's IP and total number of nodes
33
+ HEAD_IP=$(echo "$SKYPILOT_NODE_IPS" | head -n1)
34
+ NUM_NODES=$SKYPILOT_NUM_NODES
35
+
36
+ # login wandb
37
+ # python3 -c "import wandb; wandb.login(relogin=True, key='$WANDB_API_KEY')"
38
+
39
+ if [ "$SKYPILOT_NODE_RANK" == "0" ]; then
40
+ # Head node starts Ray Head
41
+ echo "Starting Ray head node..."
42
+ ps aux | grep ray | grep 6379 &> /dev/null || ray start --head --disable-usage-stats \
43
+ --port=6379 \
44
+ --dashboard-host=0.0.0.0 \
45
+ --dashboard-port=8265
46
+
47
+ # Wait for all worker nodes to join the cluster with better checking
48
+ echo "Waiting for all nodes to join Ray cluster..."
49
+ retry_count=0
50
+ max_retries=30
51
+ while [ $retry_count -lt $max_retries ]; do
52
+ connected_nodes=$(ray status 2>/dev/null | grep -c "node_" || echo "0")
53
+ echo "Connected nodes: $connected_nodes/$NUM_NODES (attempt $((retry_count+1))/$max_retries)"
54
+
55
+ if [ "$connected_nodes" -ge "$NUM_NODES" ]; then
56
+ echo "All nodes connected to Ray cluster"
57
+ break
58
+ fi
59
+
60
+ retry_count=$((retry_count+1))
61
+ sleep 10
62
+ done
63
+
64
+ if [ $retry_count -eq $max_retries ]; then
65
+ echo "WARNING: Not all nodes connected to Ray cluster after $max_retries attempts"
66
+ echo "Current Ray status:"
67
+ ray status
68
+ fi
69
+
70
+ python3 -m verl.trainer.main_ppo \
71
+ data.train_files=$HOME/data/gsm8k/train.parquet \
72
+ data.val_files=$HOME/data/gsm8k/test.parquet \
73
+ data.train_batch_size=256 \
74
+ data.max_prompt_length=512 \
75
+ data.max_response_length=256 \
76
+ actor_rollout_ref.model.path=Qwen/Qwen2.5-0.5B-Instruct \
77
+ actor_rollout_ref.actor.optim.lr=1e-6 \
78
+ actor_rollout_ref.actor.ppo_mini_batch_size=64 \
79
+ actor_rollout_ref.actor.ppo_micro_batch_size_per_gpu=4 \
80
+ actor_rollout_ref.rollout.log_prob_micro_batch_size_per_gpu=8 \
81
+ actor_rollout_ref.rollout.tensor_model_parallel_size=1 \
82
+ actor_rollout_ref.rollout.name=vllm \
83
+ actor_rollout_ref.rollout.gpu_memory_utilization=0.4 \
84
+ actor_rollout_ref.ref.log_prob_micro_batch_size_per_gpu=4 \
85
+ critic.optim.lr=1e-5 \
86
+ critic.model.path=Qwen/Qwen2.5-0.5B-Instruct \
87
+ critic.ppo_micro_batch_size_per_gpu=4 \
88
+ algorithm.kl_ctrl.kl_coef=0.001 \
89
+ trainer.logger=[console,wandb] \
90
+ trainer.val_before_train=False \
91
+ trainer.default_hdfs_dir=null \
92
+ trainer.n_gpus_per_node=1 \
93
+ trainer.nnodes=2 \
94
+ trainer.save_freq=20 \
95
+ trainer.test_freq=20 \
96
+ trainer.total_epochs=2 \
97
+ trainer.project_name=verl_examples \
98
+ trainer.experiment_name=experiment_name_gsm8k
99
+
100
+ else
101
+ # Wait for Ray Head to start
102
+ sleep 15
103
+ # Worker node starts Ray Worker
104
+ echo "Starting Ray worker node..."
105
+ ps aux | grep ray | grep $HEAD_IP:6379 &> /dev/null || ray start --address $HEAD_IP:6379 --disable-usage-stats
106
+ sleep 10
107
+ fi
108
+
109
+ echo "Node setup and Ray start script finished for rank $SKYPILOT_NODE_RANK."
examples/slurm/ray_on_slurm.slurm ADDED
@@ -0,0 +1,98 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/bin/bash
2
+ #SBATCH --job-name=verl-ray-on-slurm
3
+ #SBATCH --nodes=2
4
+ #SBATCH --ntasks-per-node=1
5
+ #SBATCH --mem=200G
6
+ #SBATCH --partition=your-partition
7
+ #SBATCH --time=01:00:00
8
+ #SBATCH --account=your-account
9
+ #SBATCH --gpus-per-node=4
10
+ #SBATCH --cpus-per-task=64
11
+ #SBATCH --output=slurm-%j.out
12
+ #SBATCH --error=slurm-%j.err
13
+
14
+ # load necessary modules
15
+
16
+ # replace these information with your own
17
+ verl_workdir=/path/to/verl
18
+ train_files=/path/to/gsm8k/train.parquet
19
+ val_files=/path/to/gsm8k/test.parquet
20
+ apptainer_image_path=/path/to/verl-ngc.sif
21
+ # replace these information with your own
22
+
23
+ # Getting the node names
24
+ nodes=$(scontrol show hostnames "$SLURM_JOB_NODELIST")
25
+ nodes_array=("$nodes")
26
+
27
+ head_node=${nodes_array[0]}
28
+ head_node_ip=$(srun --nodes=1 --ntasks=1 -w "$head_node" hostname --ip-address)
29
+
30
+ # if we detect a space character in the head node IP, we'll
31
+ # convert it to an ipv4 address. This step is optional.
32
+ if [[ "$head_node_ip" == *" "* ]]; then
33
+ IFS=' ' read -ra ADDR <<<"$head_node_ip"
34
+ if [[ ${#ADDR[0]} -gt 16 ]]; then
35
+ head_node_ip=${ADDR[1]}
36
+ else
37
+ head_node_ip=${ADDR[0]}
38
+ fi
39
+ echo "IPV6 address detected. We split the IPV4 address as $head_node_ip"
40
+ fi
41
+
42
+ port=6379
43
+ ip_head=$head_node_ip:$port
44
+ export ip_head
45
+ echo "IP Head: $ip_head"
46
+
47
+ # make sure we set environment variables before Ray initialization
48
+
49
+ printenv
50
+
51
+ echo "Starting HEAD at $head_node"
52
+ srun --nodes=1 --ntasks=1 -w "$head_node" \
53
+ apptainer run --nv --bind $verl_workdir $apptainer_image_path \
54
+ ray start --head --node-ip-address="$head_node_ip" --port=$port \
55
+ --num-cpus "${SLURM_CPUS_PER_TASK}" --num-gpus "${SLURM_GPUS_PER_NODE}" --block &
56
+ # optional, though may be useful in certain versions of Ray < 1.0.
57
+ sleep 10
58
+
59
+ # number of nodes other than the head node
60
+ worker_num=$((SLURM_JOB_NUM_NODES - 1))
61
+
62
+ for ((i = 1; i <= worker_num; i++)); do
63
+ node_i=${nodes_array[$i]}
64
+ echo "Starting WORKER $i at $node_i"
65
+ srun --nodes=1 --ntasks=1 -w "$node_i" \
66
+ apptainer run --nv --bind $verl_workdir $apptainer_image_path \
67
+ ray start --address "$ip_head" --num-cpus "${SLURM_CPUS_PER_TASK}" --num-gpus "${SLURM_GPUS_PER_NODE}" --block &
68
+ sleep 5
69
+ done
70
+
71
+ PYTHONUNBUFFERED=1 srun --overlap --nodes=1 --ntasks=1 -w "$head_node" \
72
+ apptainer run --nv --bind $verl_workdir $apptainer_image_path \
73
+ python3 -m verl.trainer.main_ppo \
74
+ algorithm.adv_estimator=gae \
75
+ data.train_files=$train_files \
76
+ data.val_files=$val_files \
77
+ data.train_batch_size=256 \
78
+ data.max_prompt_length=512 \
79
+ data.max_response_length=256 \
80
+ actor_rollout_ref.model.path=Qwen/Qwen2.5-0.5B-Instruct \
81
+ actor_rollout_ref.actor.optim.lr=1e-6 \
82
+ actor_rollout_ref.actor.ppo_mini_batch_size=64 \
83
+ actor_rollout_ref.actor.ppo_micro_batch_size_per_gpu=4 \
84
+ actor_rollout_ref.actor.use_kl_loss=False \
85
+ actor_rollout_ref.rollout.log_prob_micro_batch_size_per_gpu=8 \
86
+ actor_rollout_ref.rollout.tensor_model_parallel_size=1 \
87
+ actor_rollout_ref.rollout.gpu_memory_utilization=0.4 \
88
+ critic.optim.lr=1e-5 \
89
+ critic.model.path=Qwen/Qwen2.5-0.5B-Instruct \
90
+ critic.ppo_micro_batch_size_per_gpu=4 \
91
+ algorithm.use_kl_in_reward=False \
92
+ trainer.logger=console \
93
+ trainer.val_before_train=False \
94
+ trainer.n_gpus_per_node="${SLURM_GPUS_PER_NODE}" \
95
+ trainer.nnodes="${SLURM_NNODES}" \
96
+ trainer.save_freq=10 \
97
+ trainer.test_freq=10 \
98
+ trainer.total_epochs=15 2>&1 | tee verl_demo_slurm.log
examples/split_placement/README.md ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Split Placement Example
2
+ Here we introduce how to run the naive implementation of the split placement of PPO algorithm.
3
+ We will release the complete version of flexible placement in the near future.
4
+
5
+ For quickstart, you can only follow Step 2 to modify the code and then follow Step 4 to execute the split placement example.
6
+
7
+ ### Step 1: Placing the models to different GPUs
8
+ Specify the placement and resource allocation. In the example, we place the actor and reference in the first half of the GPUs while map the critic and reward model (if any) to the second half of the GPUs.
9
+ ```python
10
+ actor_rollout_ref_pool_id = 'actor_rollout_ref_pool'
11
+ critic_pool_id = 'critic_pool'
12
+ if config.trainer.nnodes // 2 == 0 and config.trainer.n_gpus_per_node // 2 > 0:
13
+ resource_pool_spec = {
14
+ actor_rollout_ref_pool_id: [config.trainer.n_gpus_per_node // 2] * config.trainer.nnodes,
15
+ critic_pool_id: [config.trainer.n_gpus_per_node // 2] * config.trainer.nnodes,
16
+ }
17
+ else:
18
+ resource_pool_spec = {
19
+ actor_rollout_ref_pool_id: [config.trainer.n_gpus_per_node] * (config.trainer.nnodes // 2),
20
+ critic_pool_id: [config.trainer.n_gpus_per_node] * (config.trainer.nnodes // 2),
21
+ }
22
+ print(f'resource_pool_spec: {resource_pool_spec}')
23
+ mapping = {
24
+ Role.ActorRollout: actor_rollout_ref_pool_id,
25
+ Role.Critic: critic_pool_id,
26
+ Role.RefPolicy: actor_rollout_ref_pool_id,
27
+ }
28
+ mapping[Role.RewardModel] = critic_pool_id
29
+ ```
30
+
31
+ ### Step 2: Make the models executed asynchronously
32
+ Based on the model placement, we need to make the models executed asynchronously.
33
+
34
+ To do so, you need to turn off the `blocking` flag (i.e., `blocking=False`) in our decorator of some model operations.
35
+ For example, we hope the actor update and critic update can be executed in parallel, then we need to make the following modification in `fsdp_workers.py`
36
+
37
+ ```
38
+ @register(dispatch_mode=Dispatch.DP_COMPUTE_PROTO, blocking=False)
39
+ def update_actor(self, data: DataProto):
40
+ ...
41
+
42
+ @register(dispatch_mode=Dispatch.DP_COMPUTE_PROTO, blocking=False)
43
+ def update_critic(self, data: DataProto):
44
+ ...
45
+ ```
46
+
47
+ We can also parallelize the computation of `ref_log_prob` and `values` and `rewards` in the split placement. For simplicity of the tutorial, we don't do this in this example.
48
+
49
+ ### Step 3: Execute these operation in parallel in the single controller process
50
+ To implement the parallel execution of the actor and critic update, the only thing we need to modify in the `ray_trainer.py` is to `get` the concurrent `futures` on the single controller process.
51
+
52
+ ```python
53
+ critic_output = critic_output.get()
54
+ actor_output = actor_output.get()
55
+ ```
56
+
57
+ ### Step 4: Run the split placement example
58
+
59
+ ```
60
+ bash run_deepseek7b_llm.sh
61
+ ```
examples/split_placement/main_ppo_split.py ADDED
@@ -0,0 +1,216 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright 2024 Bytedance Ltd. and/or its affiliates
2
+ #
3
+ # Licensed under the Apache License, Version 2.0 (the "License");
4
+ # you may not use this file except in compliance with the License.
5
+ # You may obtain a copy of the License at
6
+ #
7
+ # http://www.apache.org/licenses/LICENSE-2.0
8
+ #
9
+ # Unless required by applicable law or agreed to in writing, software
10
+ # distributed under the License is distributed on an "AS IS" BASIS,
11
+ # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12
+ # See the License for the specific language governing permissions and
13
+ # limitations under the License.
14
+ """
15
+ Note that we don't combine the main with ray_trainer as ray_trainer is used by other main.
16
+ """
17
+
18
+ import hydra
19
+ import ray
20
+ import torch
21
+ from omegaconf import OmegaConf
22
+ from split_monkey_patch import fit
23
+
24
+ from verl import DataProto
25
+ from verl.trainer.ppo.ray_trainer import RayPPOTrainer
26
+ from verl.utils.reward_score import gsm8k, math_reward
27
+
28
+
29
+ def _select_rm_score_fn(data_source):
30
+ if data_source == "openai/gsm8k":
31
+ return gsm8k.compute_score
32
+ elif data_source == "lighteval/MATH":
33
+ return math_reward.compute_score
34
+ else:
35
+ raise NotImplementedError
36
+
37
+
38
+ class RewardManager:
39
+ def __init__(self, tokenizer, num_examine) -> None:
40
+ self.tokenizer = tokenizer
41
+ self.num_examine = num_examine # the number of batches of decoded responses to print to the console
42
+
43
+ def __call__(self, data: DataProto, return_dict: bool = False):
44
+ """We will expand this function gradually based on the available datasets"""
45
+
46
+ # If there is rm score, we directly return rm score. Otherwise, we compute via rm_score_fn
47
+ if "rm_scores" in data.batch.keys():
48
+ return data.batch["rm_scores"]
49
+
50
+ reward_tensor = torch.zeros_like(data.batch["responses"], dtype=torch.float32)
51
+
52
+ already_print_data_sources = {}
53
+
54
+ for i in range(len(data)):
55
+ data_item = data[i] # DataProtoItem
56
+
57
+ prompt_ids = data_item.batch["prompts"]
58
+
59
+ prompt_length = prompt_ids.shape[-1]
60
+
61
+ valid_prompt_length = data_item.batch["attention_mask"][:prompt_length].sum()
62
+ valid_prompt_ids = prompt_ids[-valid_prompt_length:]
63
+
64
+ response_ids = data_item.batch["responses"]
65
+ valid_response_length = data_item.batch["attention_mask"][prompt_length:].sum()
66
+ valid_response_ids = response_ids[:valid_response_length]
67
+
68
+ # decode
69
+ sequences = torch.cat((valid_prompt_ids, valid_response_ids))
70
+ sequences_str = self.tokenizer.decode(sequences)
71
+
72
+ ground_truth = data_item.non_tensor_batch["reward_model"]["ground_truth"]
73
+
74
+ # select rm_score
75
+ data_source = data_item.non_tensor_batch["data_source"]
76
+ compute_score_fn = _select_rm_score_fn(data_source)
77
+
78
+ score = compute_score_fn(solution_str=sequences_str, ground_truth=ground_truth)
79
+ reward_tensor[i, valid_response_length - 1] = score
80
+
81
+ if data_source not in already_print_data_sources:
82
+ already_print_data_sources[data_source] = 0
83
+
84
+ if already_print_data_sources[data_source] < self.num_examine:
85
+ already_print_data_sources[data_source] += 1
86
+ print(sequences_str)
87
+
88
+ if return_dict:
89
+ return {"reward_tensor": reward_tensor}
90
+ else:
91
+ return reward_tensor
92
+
93
+
94
+ @hydra.main(config_path="config", config_name="ppo_trainer_split", version_base=None)
95
+ def main(config):
96
+ if not ray.is_initialized():
97
+ # this is for local ray cluster
98
+ default_runtime_env = {"env_vars": {"TOKENIZERS_PARALLELISM": "true", "NCCL_DEBUG": "WARN"}}
99
+ ray_init_kwargs = config.ray_kwargs.get("ray_init", {})
100
+ runtime_env_kwargs = ray_init_kwargs.get("runtime_env", {})
101
+ runtime_env = OmegaConf.merge(default_runtime_env, runtime_env_kwargs)
102
+ ray_init_kwargs = OmegaConf.create({**ray_init_kwargs, "runtime_env": runtime_env})
103
+ print(f"ray init kwargs: {ray_init_kwargs}")
104
+ ray.init(**OmegaConf.to_container(ray_init_kwargs))
105
+
106
+ ray.get(main_task.remote(config))
107
+
108
+
109
+ @ray.remote
110
+ def main_task(config):
111
+ # print initial config
112
+ from pprint import pprint
113
+
114
+ from omegaconf import OmegaConf
115
+
116
+ from verl.utils.fs import copy_to_local
117
+
118
+ pprint(OmegaConf.to_container(config, resolve=True)) # resolve=True will eval symbol values
119
+ OmegaConf.resolve(config)
120
+
121
+ # download the checkpoint from hdfs
122
+ local_path = copy_to_local(config.actor_rollout_ref.model.path)
123
+
124
+ # instantiate tokenizer
125
+ from verl.utils import hf_tokenizer
126
+
127
+ tokenizer = hf_tokenizer(local_path)
128
+
129
+ # define worker classes
130
+ if config.actor_rollout_ref.actor.strategy in {"fsdp", "fsdp2"}:
131
+ assert config.critic.strategy in {"fsdp", "fsdp2"}
132
+ from verl.single_controller.ray import RayWorkerGroup
133
+ from verl.workers.fsdp_workers import ActorRolloutRefWorker, CriticWorker
134
+
135
+ ray_worker_group_cls = RayWorkerGroup
136
+
137
+ elif config.actor_rollout_ref.actor.strategy == "megatron":
138
+ assert config.actor_rollout_ref.actor.strategy == config.critic.strategy
139
+ from verl.single_controller.ray import RayWorkerGroup
140
+ from verl.workers.megatron_workers import ActorRolloutRefWorker, CriticWorker
141
+
142
+ ray_worker_group_cls = RayWorkerGroup
143
+
144
+ else:
145
+ raise NotImplementedError
146
+
147
+ from verl.trainer.ppo.ray_trainer import ResourcePoolManager, Role
148
+
149
+ role_worker_mapping = {
150
+ Role.ActorRollout: ray.remote(ActorRolloutRefWorker),
151
+ Role.Critic: ray.remote(CriticWorker),
152
+ }
153
+
154
+ # NOTE: initialze two resource pool
155
+ actor_rollout_ref_pool_id = "actor_rollout_ref_pool"
156
+ critic_pool_id = "critic_pool"
157
+ if config.trainer.nnodes // 2 == 0 and config.trainer.n_gpus_per_node // 2 > 0:
158
+ resource_pool_spec = {
159
+ actor_rollout_ref_pool_id: [config.trainer.n_gpus_per_node // 2] * config.trainer.nnodes,
160
+ critic_pool_id: [config.trainer.n_gpus_per_node // 2] * config.trainer.nnodes,
161
+ }
162
+ else:
163
+ resource_pool_spec = {
164
+ actor_rollout_ref_pool_id: [config.trainer.n_gpus_per_node] * (config.trainer.nnodes // 2),
165
+ critic_pool_id: [config.trainer.n_gpus_per_node] * (config.trainer.nnodes // 2),
166
+ }
167
+ print(f"resource_pool_spec: {resource_pool_spec}")
168
+ mapping = {
169
+ Role.ActorRollout: actor_rollout_ref_pool_id,
170
+ Role.Critic: critic_pool_id,
171
+ }
172
+
173
+ # use reference model
174
+ if config.algorithm.use_kl_in_reward or config.actor_rollout_ref.actor.use_kl_loss:
175
+ role_worker_mapping[Role.RefPolicy] = ray.remote(ActorRolloutRefWorker)
176
+ mapping[Role.RefPolicy] = actor_rollout_ref_pool_id
177
+
178
+ # we should adopt a multi-source reward function here
179
+ # - for rule-based rm, we directly call a reward score
180
+ # - for model-based rm, we call a model
181
+ # - for code related prompt, we send to a sandbox if there are test cases
182
+ # - finally, we combine all the rewards together
183
+ # - The reward type depends on the tag of the data
184
+ if config.reward_model.enable:
185
+ if config.reward_model.strategy in {"fsdp", "fsdp2"}:
186
+ from verl.workers.fsdp_workers import RewardModelWorker
187
+ elif config.reward_model.strategy == "megatron":
188
+ from verl.workers.megatron_workers import RewardModelWorker
189
+ else:
190
+ raise NotImplementedError
191
+ role_worker_mapping[Role.RewardModel] = ray.remote(RewardModelWorker)
192
+ mapping[Role.RewardModel] = critic_pool_id
193
+
194
+ reward_fn = RewardManager(tokenizer=tokenizer, num_examine=0)
195
+
196
+ # Note that we always use function-based RM for validation
197
+ val_reward_fn = RewardManager(tokenizer=tokenizer, num_examine=1)
198
+
199
+ resource_pool_manager = ResourcePoolManager(resource_pool_spec=resource_pool_spec, mapping=mapping)
200
+
201
+ RayPPOTrainer.fit = fit
202
+ trainer = RayPPOTrainer(
203
+ config=config,
204
+ tokenizer=tokenizer,
205
+ role_worker_mapping=role_worker_mapping,
206
+ resource_pool_manager=resource_pool_manager,
207
+ ray_worker_group_cls=ray_worker_group_cls,
208
+ reward_fn=reward_fn,
209
+ val_reward_fn=val_reward_fn,
210
+ )
211
+ trainer.init_workers()
212
+ trainer.fit()
213
+
214
+
215
+ if __name__ == "__main__":
216
+ main()
examples/split_placement/run_deepseek7b_llm.sh ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ set -x
2
+
3
+ python3 main_ppo_split.py \
4
+ algorithm.adv_estimator=gae \
5
+ data.train_files=$HOME/data/gsm8k/train.parquet \
6
+ data.val_files=$HOME/data/gsm8k/test.parquet \
7
+ data.train_batch_size=1024 \
8
+ data.max_prompt_length=512 \
9
+ data.max_response_length=512 \
10
+ data.filter_overlong_prompts=True \
11
+ data.truncation='error' \
12
+ actor_rollout_ref.model.path=deepseek-ai/deepseek-llm-7b-chat \
13
+ actor_rollout_ref.actor.optim.lr=1e-6 \
14
+ actor_rollout_ref.actor.ppo_mini_batch_size=256 \
15
+ actor_rollout_ref.actor.ppo_micro_batch_size_per_gpu=8 \
16
+ actor_rollout_ref.actor.fsdp_config.param_offload=False \
17
+ actor_rollout_ref.actor.fsdp_config.optimizer_offload=False \
18
+ actor_rollout_ref.actor.use_kl_loss=False \
19
+ actor_rollout_ref.rollout.log_prob_micro_batch_size_per_gpu=8 \
20
+ actor_rollout_ref.rollout.tensor_model_parallel_size=4 \
21
+ actor_rollout_ref.rollout.name=vllm \
22
+ actor_rollout_ref.rollout.gpu_memory_utilization=0.4 \
23
+ critic.optim.lr=1e-5 \
24
+ critic.model.path=deepseek-ai/deepseek-llm-7b-chat \
25
+ critic.model.enable_gradient_checkpointing=False \
26
+ critic.ppo_micro_batch_size_per_gpu=8 \
27
+ critic.model.fsdp_config.param_offload=False \
28
+ critic.model.fsdp_config.optimizer_offload=False \
29
+ algorithm.use_kl_in_reward=False \
30
+ trainer.critic_warmup=0 \
31
+ trainer.logger='["console","wandb"]' \
32
+ trainer.project_name='verl_example_gsm8k' \
33
+ trainer.experiment_name='deepseek_llm_7b_function_rm' \
34
+ trainer.n_gpus_per_node=8 \
35
+ trainer.nnodes=1 \
36
+ trainer.save_freq=-1 \
37
+ trainer.total_epochs=15 $@
outputs/2026-01-13/23-34-32/.hydra/overrides.yaml ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ - algorithm.adv_estimator=sgrpo
2
+ - algorithm.norm_adv_by_std_in_grpo=False
3
+ - algorithm.use_kl_in_reward=False
4
+ - data.train_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/train.parquet
5
+ - data.val_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/test.parquet
6
+ - data.train_batch_size=1024
7
+ - data.max_prompt_length=512
8
+ - data.max_response_length=1024
9
+ - data.filter_overlong_prompts=True
10
+ - data.truncation=error
11
+ - data.return_raw_chat=True
12
+ - actor_rollout_ref.model.path=/mnt/tidal-alsh01/dataset/redtrans/zhangruiqi/models/Qwen3-8B
13
+ - actor_rollout_ref.actor.optim.lr=1e-6
14
+ - actor_rollout_ref.model.use_remove_padding=True
15
+ - actor_rollout_ref.actor.ppo_mini_batch_size=256
16
+ - actor_rollout_ref.actor.ppo_micro_batch_size_per_gpu=32
17
+ - actor_rollout_ref.actor.use_kl_loss=True
18
+ - actor_rollout_ref.actor.kl_loss_coef=0.001
19
+ - actor_rollout_ref.actor.kl_loss_type=low_var_kl
20
+ - actor_rollout_ref.actor.entropy_coeff=0
21
+ - actor_rollout_ref.model.enable_gradient_checkpointing=True
22
+ - actor_rollout_ref.actor.fsdp_config.param_offload=False
23
+ - actor_rollout_ref.actor.fsdp_config.optimizer_offload=False
24
+ - actor_rollout_ref.rollout.log_prob_micro_batch_size_per_gpu=32
25
+ - actor_rollout_ref.rollout.tensor_model_parallel_size=2
26
+ - actor_rollout_ref.rollout.name=vllm
27
+ - actor_rollout_ref.rollout.gpu_memory_utilization=0.6
28
+ - actor_rollout_ref.rollout.n=1
29
+ - actor_rollout_ref.rollout.prompt_length=1536
30
+ - actor_rollout_ref.rollout.response_length=512
31
+ - actor_rollout_ref.ref.log_prob_micro_batch_size_per_gpu=32
32
+ - actor_rollout_ref.ref.fsdp_config.param_offload=False
33
+ - sgrpo.num_truncations=4
34
+ - sgrpo.force_answer_prompt="Time is limited, stop thinking and start answering.\n</think>\n\n"
35
+ - sgrpo.answer_max_tokens=256
36
+ - trainer.critic_warmup=0
37
+ - trainer.logger=["console","wandb"]
38
+ - trainer.project_name=verl_sgrpo_example_gsm8k
39
+ - trainer.experiment_name=qwen3_8b_sgrpo
40
+ - trainer.resume_mode=disable
41
+ - trainer.n_gpus_per_node=8
42
+ - trainer.nnodes=1
43
+ - trainer.save_freq=20
44
+ - trainer.test_freq=5
45
+ - trainer.total_epochs=15
46
+ - trainer.rollout_data_dir=./outputs/sgrpo_samples
47
+ - trainer.rollout_data_dump_freq=20
outputs/2026-01-13/23-57-30/.hydra/config.yaml ADDED
@@ -0,0 +1,617 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ actor_rollout_ref:
2
+ actor:
3
+ optim:
4
+ _target_: verl.workers.config.FSDPOptimizerConfig
5
+ optimizer: AdamW
6
+ optimizer_impl: torch.optim
7
+ lr: 1.0e-06
8
+ lr_warmup_steps_ratio: 0.0
9
+ total_training_steps: -1
10
+ weight_decay: 0.01
11
+ lr_warmup_steps: -1
12
+ betas:
13
+ - 0.9
14
+ - 0.999
15
+ clip_grad: 1.0
16
+ min_lr_ratio: 0.0
17
+ num_cycles: 0.5
18
+ lr_scheduler_type: constant
19
+ warmup_style: null
20
+ override_optimizer_config: null
21
+ fsdp_config:
22
+ _target_: verl.workers.config.FSDPEngineConfig
23
+ wrap_policy:
24
+ min_num_params: 0
25
+ param_offload: false
26
+ optimizer_offload: false
27
+ offload_policy: false
28
+ reshard_after_forward: true
29
+ fsdp_size: -1
30
+ forward_prefetch: false
31
+ model_dtype: fp32
32
+ use_orig_params: false
33
+ seed: 42
34
+ full_determinism: false
35
+ ulysses_sequence_parallel_size: 1
36
+ entropy_from_logits_with_chunking: false
37
+ use_torch_compile: true
38
+ entropy_checkpointing: false
39
+ forward_only: false
40
+ strategy: fsdp
41
+ dtype: bfloat16
42
+ _target_: verl.workers.config.FSDPActorConfig
43
+ rollout_n: ${oc.select:actor_rollout_ref.rollout.n,1}
44
+ strategy: fsdp
45
+ ppo_mini_batch_size: 256
46
+ ppo_micro_batch_size: null
47
+ ppo_micro_batch_size_per_gpu: 32
48
+ use_dynamic_bsz: false
49
+ ppo_max_token_len_per_gpu: 16384
50
+ clip_ratio: 0.2
51
+ clip_ratio_low: 0.2
52
+ clip_ratio_high: 0.2
53
+ freeze_vision_tower: false
54
+ policy_loss:
55
+ _target_: verl.workers.config.PolicyLossConfig
56
+ loss_mode: vanilla
57
+ clip_cov_ratio: 0.0002
58
+ clip_cov_lb: 1.0
59
+ clip_cov_ub: 5.0
60
+ kl_cov_ratio: 0.0002
61
+ ppo_kl_coef: 0.1
62
+ clip_ratio_c: 3.0
63
+ loss_agg_mode: token-mean
64
+ loss_scale_factor: null
65
+ entropy_coeff: 0
66
+ calculate_entropy: false
67
+ use_kl_loss: true
68
+ use_torch_compile: true
69
+ kl_loss_coef: 0.001
70
+ kl_loss_type: low_var_kl
71
+ ppo_epochs: 1
72
+ shuffle: false
73
+ checkpoint:
74
+ _target_: verl.trainer.config.CheckpointConfig
75
+ save_contents:
76
+ - model
77
+ - optimizer
78
+ - extra
79
+ load_contents: ${.save_contents}
80
+ async_save: false
81
+ use_fused_kernels: ${oc.select:actor_rollout_ref.model.use_fused_kernels,false}
82
+ profiler:
83
+ _target_: verl.utils.profiler.ProfilerConfig
84
+ tool: ${oc.select:global_profiler.tool,null}
85
+ enable: false
86
+ all_ranks: false
87
+ ranks: []
88
+ save_path: ${oc.select:global_profiler.save_path,null}
89
+ tool_config:
90
+ nsys:
91
+ _target_: verl.utils.profiler.config.NsightToolConfig
92
+ discrete: ${oc.select:global_profiler.global_tool_config.nsys.discrete}
93
+ npu:
94
+ _target_: verl.utils.profiler.config.NPUToolConfig
95
+ contents: []
96
+ level: level1
97
+ analysis: true
98
+ discrete: false
99
+ torch:
100
+ _target_: verl.utils.profiler.config.TorchProfilerToolConfig
101
+ step_start: 0
102
+ step_end: null
103
+ torch_memory:
104
+ _target_: verl.utils.profiler.config.TorchMemoryToolConfig
105
+ trace_alloc_max_entries: ${oc.select:global_profiler.global_tool_config.torch_memory.trace_alloc_max_entries,100000}
106
+ stack_depth: ${oc.select:global_profiler.global_tool_config.torch_memory.stack_depth,32}
107
+ router_replay:
108
+ _target_: verl.workers.config.RouterReplayConfig
109
+ mode: disabled
110
+ record_file: null
111
+ replay_file: null
112
+ grad_clip: 1.0
113
+ ulysses_sequence_parallel_size: 1
114
+ entropy_from_logits_with_chunking: false
115
+ entropy_checkpointing: false
116
+ use_remove_padding: ${oc.select:actor_rollout_ref.model.use_remove_padding,false}
117
+ ref:
118
+ rollout_n: ${oc.select:actor_rollout_ref.rollout.n,1}
119
+ strategy: ${actor_rollout_ref.actor.strategy}
120
+ use_torch_compile: ${oc.select:actor_rollout_ref.actor.use_torch_compile,true}
121
+ log_prob_micro_batch_size: null
122
+ log_prob_micro_batch_size_per_gpu: 32
123
+ log_prob_use_dynamic_bsz: ${oc.select:actor_rollout_ref.actor.use_dynamic_bsz,false}
124
+ log_prob_max_token_len_per_gpu: ${oc.select:actor_rollout_ref.actor.ppo_max_token_len_per_gpu,16384}
125
+ profiler:
126
+ _target_: verl.utils.profiler.ProfilerConfig
127
+ tool: ${oc.select:global_profiler.tool,null}
128
+ enable: false
129
+ all_ranks: false
130
+ ranks: []
131
+ save_path: ${oc.select:global_profiler.save_path,null}
132
+ tool_config:
133
+ nsys:
134
+ _target_: verl.utils.profiler.config.NsightToolConfig
135
+ discrete: ${oc.select:global_profiler.global_tool_config.nsys.discrete}
136
+ npu:
137
+ _target_: verl.utils.profiler.config.NPUToolConfig
138
+ contents: []
139
+ level: level1
140
+ analysis: true
141
+ discrete: false
142
+ torch:
143
+ _target_: verl.utils.profiler.config.TorchProfilerToolConfig
144
+ step_start: 0
145
+ step_end: null
146
+ torch_memory:
147
+ _target_: verl.utils.profiler.config.TorchMemoryToolConfig
148
+ trace_alloc_max_entries: ${oc.select:global_profiler.global_tool_config.torch_memory.trace_alloc_max_entries,100000}
149
+ stack_depth: ${oc.select:global_profiler.global_tool_config.torch_memory.stack_depth,32}
150
+ router_replay:
151
+ _target_: verl.workers.config.RouterReplayConfig
152
+ mode: disabled
153
+ record_file: null
154
+ replay_file: null
155
+ fsdp_config:
156
+ _target_: verl.workers.config.FSDPEngineConfig
157
+ wrap_policy:
158
+ min_num_params: 0
159
+ param_offload: false
160
+ optimizer_offload: false
161
+ offload_policy: false
162
+ reshard_after_forward: true
163
+ fsdp_size: -1
164
+ forward_prefetch: false
165
+ model_dtype: fp32
166
+ use_orig_params: false
167
+ seed: 42
168
+ full_determinism: false
169
+ ulysses_sequence_parallel_size: 1
170
+ entropy_from_logits_with_chunking: false
171
+ use_torch_compile: true
172
+ entropy_checkpointing: false
173
+ forward_only: true
174
+ strategy: fsdp
175
+ dtype: bfloat16
176
+ _target_: verl.workers.config.FSDPActorConfig
177
+ ulysses_sequence_parallel_size: ${oc.select:actor_rollout_ref.actor.ulysses_sequence_parallel_size,1}
178
+ entropy_from_logits_with_chunking: false
179
+ entropy_checkpointing: false
180
+ rollout:
181
+ _target_: verl.workers.config.RolloutConfig
182
+ name: vllm
183
+ mode: async
184
+ temperature: 1.0
185
+ top_k: -1
186
+ top_p: 1
187
+ prompt_length: 1536
188
+ response_length: 512
189
+ dtype: bfloat16
190
+ gpu_memory_utilization: 0.6
191
+ ignore_eos: false
192
+ enforce_eager: false
193
+ cudagraph_capture_sizes: null
194
+ free_cache_engine: true
195
+ tensor_model_parallel_size: 2
196
+ data_parallel_size: 1
197
+ expert_parallel_size: 1
198
+ pipeline_model_parallel_size: 1
199
+ max_num_batched_tokens: 8192
200
+ max_model_len: null
201
+ max_num_seqs: 1024
202
+ enable_chunked_prefill: true
203
+ enable_prefix_caching: true
204
+ load_format: dummy
205
+ log_prob_micro_batch_size: null
206
+ log_prob_micro_batch_size_per_gpu: 32
207
+ log_prob_use_dynamic_bsz: ${oc.select:actor_rollout_ref.actor.use_dynamic_bsz,false}
208
+ log_prob_max_token_len_per_gpu: ${oc.select:actor_rollout_ref.actor.ppo_max_token_len_per_gpu,16384}
209
+ disable_log_stats: true
210
+ do_sample: true
211
+ 'n': 1
212
+ over_sample_rate: 0
213
+ multi_stage_wake_up: false
214
+ engine_kwargs:
215
+ vllm: {}
216
+ sglang: {}
217
+ val_kwargs:
218
+ _target_: verl.workers.config.SamplingConfig
219
+ top_k: -1
220
+ top_p: 1.0
221
+ temperature: 0
222
+ 'n': 1
223
+ do_sample: false
224
+ multi_turn:
225
+ _target_: verl.workers.config.MultiTurnConfig
226
+ enable: false
227
+ max_assistant_turns: null
228
+ tool_config_path: null
229
+ max_user_turns: null
230
+ max_parallel_calls: 1
231
+ max_tool_response_length: 256
232
+ tool_response_truncate_side: middle
233
+ interaction_config_path: null
234
+ use_inference_chat_template: false
235
+ tokenization_sanity_check_mode: strict
236
+ format: hermes
237
+ num_repeat_rollouts: null
238
+ calculate_log_probs: false
239
+ agent:
240
+ _target_: verl.workers.config.AgentLoopConfig
241
+ num_workers: 8
242
+ default_agent_loop: single_turn_agent
243
+ agent_loop_config_path: null
244
+ custom_async_server:
245
+ _target_: verl.workers.config.CustomAsyncServerConfig
246
+ path: null
247
+ name: null
248
+ update_weights_bucket_megabytes: 512
249
+ trace:
250
+ _target_: verl.workers.config.TraceConfig
251
+ backend: null
252
+ token2text: false
253
+ max_samples_per_step_per_worker: null
254
+ skip_rollout: false
255
+ skip_dump_dir: /tmp/rollout_dump
256
+ skip_tokenizer_init: true
257
+ enable_rollout_routing_replay: false
258
+ profiler:
259
+ _target_: verl.utils.profiler.ProfilerConfig
260
+ tool: ${oc.select:global_profiler.tool,null}
261
+ enable: ${oc.select:actor_rollout_ref.actor.profiler.enable,false}
262
+ all_ranks: ${oc.select:actor_rollout_ref.actor.profiler.all_ranks,false}
263
+ ranks: ${oc.select:actor_rollout_ref.actor.profiler.ranks,[]}
264
+ save_path: ${oc.select:global_profiler.save_path,null}
265
+ tool_config: ${oc.select:actor_rollout_ref.actor.profiler.tool_config,null}
266
+ prometheus:
267
+ _target_: verl.workers.config.PrometheusConfig
268
+ enable: false
269
+ port: 9090
270
+ file: /tmp/ray/session_latest/metrics/prometheus/prometheus.yml
271
+ served_model_name: ${oc.select:actor_rollout_ref.model.path,null}
272
+ layered_summon: false
273
+ model:
274
+ _target_: verl.workers.config.HFModelConfig
275
+ path: /mnt/tidal-alsh01/dataset/redtrans/zhangruiqi/models/Qwen3-8B
276
+ hf_config_path: null
277
+ tokenizer_path: null
278
+ use_shm: false
279
+ trust_remote_code: false
280
+ custom_chat_template: null
281
+ external_lib: null
282
+ override_config: {}
283
+ enable_gradient_checkpointing: true
284
+ enable_activation_offload: false
285
+ use_remove_padding: true
286
+ lora_rank: 0
287
+ lora_alpha: 16
288
+ target_modules: all-linear
289
+ exclude_modules: null
290
+ lora_adapter_path: null
291
+ use_liger: false
292
+ use_fused_kernels: false
293
+ fused_kernel_options:
294
+ impl_backend: torch
295
+ hybrid_engine: true
296
+ nccl_timeout: 600
297
+ data:
298
+ tokenizer: null
299
+ use_shm: false
300
+ train_files: /mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/train.parquet
301
+ val_files: /mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/test.parquet
302
+ train_max_samples: -1
303
+ val_max_samples: -1
304
+ prompt_key: prompt
305
+ reward_fn_key: data_source
306
+ max_prompt_length: 512
307
+ max_response_length: 1024
308
+ train_batch_size: 1024
309
+ val_batch_size: null
310
+ tool_config_path: ${oc.select:actor_rollout_ref.rollout.multi_turn.tool_config_path,
311
+ null}
312
+ return_raw_input_ids: false
313
+ return_raw_chat: true
314
+ return_full_prompt: false
315
+ shuffle: true
316
+ seed: null
317
+ dataloader_num_workers: 8
318
+ image_patch_size: 14
319
+ validation_shuffle: false
320
+ filter_overlong_prompts: true
321
+ filter_overlong_prompts_workers: 1
322
+ truncation: error
323
+ image_key: images
324
+ video_key: videos
325
+ trust_remote_code: false
326
+ custom_cls:
327
+ path: null
328
+ name: null
329
+ return_multi_modal_inputs: true
330
+ sampler:
331
+ class_path: null
332
+ class_name: null
333
+ datagen:
334
+ path: null
335
+ name: null
336
+ apply_chat_template_kwargs: {}
337
+ reward_manager:
338
+ _target_: verl.trainer.config.config.RewardManagerConfig
339
+ source: register
340
+ name: ${oc.select:reward_model.reward_manager,naive}
341
+ module:
342
+ _target_: verl.trainer.config.config.ModuleConfig
343
+ path: null
344
+ name: custom_reward_manager
345
+ critic:
346
+ optim:
347
+ _target_: verl.workers.config.FSDPOptimizerConfig
348
+ optimizer: AdamW
349
+ optimizer_impl: torch.optim
350
+ lr: 1.0e-05
351
+ lr_warmup_steps_ratio: 0.0
352
+ total_training_steps: -1
353
+ weight_decay: 0.01
354
+ lr_warmup_steps: -1
355
+ betas:
356
+ - 0.9
357
+ - 0.999
358
+ clip_grad: 1.0
359
+ min_lr_ratio: 0.0
360
+ num_cycles: 0.5
361
+ lr_scheduler_type: constant
362
+ warmup_style: null
363
+ override_optimizer_config: null
364
+ model:
365
+ fsdp_config:
366
+ _target_: verl.workers.config.FSDPEngineConfig
367
+ wrap_policy:
368
+ min_num_params: 0
369
+ param_offload: false
370
+ optimizer_offload: false
371
+ offload_policy: false
372
+ reshard_after_forward: true
373
+ fsdp_size: -1
374
+ forward_prefetch: false
375
+ model_dtype: fp32
376
+ use_orig_params: false
377
+ seed: 42
378
+ full_determinism: false
379
+ ulysses_sequence_parallel_size: 1
380
+ entropy_from_logits_with_chunking: false
381
+ use_torch_compile: true
382
+ entropy_checkpointing: false
383
+ forward_only: false
384
+ strategy: fsdp
385
+ dtype: bfloat16
386
+ path: ~/models/deepseek-llm-7b-chat
387
+ tokenizer_path: ${oc.select:actor_rollout_ref.model.path,"~/models/deepseek-llm-7b-chat"}
388
+ override_config: {}
389
+ external_lib: ${oc.select:actor_rollout_ref.model.external_lib,null}
390
+ trust_remote_code: ${oc.select:actor_rollout_ref.model.trust_remote_code,false}
391
+ _target_: verl.workers.config.FSDPCriticModelCfg
392
+ use_shm: false
393
+ enable_gradient_checkpointing: true
394
+ enable_activation_offload: false
395
+ use_remove_padding: false
396
+ lora_rank: 0
397
+ lora_alpha: 16
398
+ target_modules: all-linear
399
+ _target_: verl.workers.config.FSDPCriticConfig
400
+ rollout_n: ${oc.select:actor_rollout_ref.rollout.n,1}
401
+ strategy: fsdp
402
+ enable: null
403
+ ppo_mini_batch_size: ${oc.select:actor_rollout_ref.actor.ppo_mini_batch_size,256}
404
+ ppo_micro_batch_size: null
405
+ ppo_micro_batch_size_per_gpu: ${oc.select:.ppo_micro_batch_size,null}
406
+ use_dynamic_bsz: ${oc.select:actor_rollout_ref.actor.use_dynamic_bsz,false}
407
+ ppo_max_token_len_per_gpu: 32768
408
+ forward_max_token_len_per_gpu: ${.ppo_max_token_len_per_gpu}
409
+ ppo_epochs: ${oc.select:actor_rollout_ref.actor.ppo_epochs,1}
410
+ shuffle: ${oc.select:actor_rollout_ref.actor.shuffle,false}
411
+ cliprange_value: 0.5
412
+ loss_agg_mode: ${oc.select:actor_rollout_ref.actor.loss_agg_mode,token-mean}
413
+ checkpoint:
414
+ _target_: verl.trainer.config.CheckpointConfig
415
+ save_contents:
416
+ - model
417
+ - optimizer
418
+ - extra
419
+ load_contents: ${.save_contents}
420
+ async_save: false
421
+ profiler:
422
+ _target_: verl.utils.profiler.ProfilerConfig
423
+ tool: ${oc.select:global_profiler.tool,null}
424
+ enable: false
425
+ all_ranks: false
426
+ ranks: []
427
+ save_path: ${oc.select:global_profiler.save_path,null}
428
+ tool_config:
429
+ nsys:
430
+ _target_: verl.utils.profiler.config.NsightToolConfig
431
+ discrete: ${oc.select:global_profiler.global_tool_config.nsys.discrete}
432
+ npu:
433
+ _target_: verl.utils.profiler.config.NPUToolConfig
434
+ contents: []
435
+ level: level1
436
+ analysis: true
437
+ discrete: false
438
+ torch:
439
+ _target_: verl.utils.profiler.config.TorchProfilerToolConfig
440
+ step_start: 0
441
+ step_end: null
442
+ torch_memory:
443
+ _target_: verl.utils.profiler.config.TorchMemoryToolConfig
444
+ trace_alloc_max_entries: ${oc.select:global_profiler.global_tool_config.torch_memory.trace_alloc_max_entries,100000}
445
+ stack_depth: ${oc.select:global_profiler.global_tool_config.torch_memory.stack_depth,32}
446
+ forward_micro_batch_size: ${oc.select:.ppo_micro_batch_size,null}
447
+ forward_micro_batch_size_per_gpu: ${oc.select:.ppo_micro_batch_size_per_gpu,null}
448
+ ulysses_sequence_parallel_size: 1
449
+ grad_clip: 1.0
450
+ reward_model:
451
+ enable: false
452
+ enable_resource_pool: false
453
+ n_gpus_per_node: 0
454
+ nnodes: 0
455
+ strategy: fsdp
456
+ model:
457
+ input_tokenizer: ${actor_rollout_ref.model.path}
458
+ path: ~/models/FsfairX-LLaMA3-RM-v0.1
459
+ external_lib: ${actor_rollout_ref.model.external_lib}
460
+ trust_remote_code: false
461
+ override_config: {}
462
+ use_shm: false
463
+ use_remove_padding: false
464
+ use_fused_kernels: ${actor_rollout_ref.model.use_fused_kernels}
465
+ fsdp_config:
466
+ _target_: verl.workers.config.FSDPEngineConfig
467
+ wrap_policy:
468
+ min_num_params: 0
469
+ param_offload: false
470
+ reshard_after_forward: true
471
+ fsdp_size: -1
472
+ forward_prefetch: false
473
+ micro_batch_size: null
474
+ micro_batch_size_per_gpu: null
475
+ max_length: null
476
+ use_dynamic_bsz: ${critic.use_dynamic_bsz}
477
+ forward_max_token_len_per_gpu: ${critic.forward_max_token_len_per_gpu}
478
+ reward_manager: naive
479
+ launch_reward_fn_async: false
480
+ sandbox_fusion:
481
+ url: null
482
+ max_concurrent: 64
483
+ memory_limit_mb: 1024
484
+ profiler:
485
+ _target_: verl.utils.profiler.ProfilerConfig
486
+ tool: ${oc.select:global_profiler.tool,null}
487
+ enable: false
488
+ all_ranks: false
489
+ ranks: []
490
+ save_path: ${oc.select:global_profiler.save_path,null}
491
+ tool_config: ${oc.select:actor_rollout_ref.actor.profiler.tool_config,null}
492
+ ulysses_sequence_parallel_size: 1
493
+ use_reward_loop: true
494
+ rollout:
495
+ _target_: verl.workers.config.RolloutConfig
496
+ name: ???
497
+ dtype: bfloat16
498
+ gpu_memory_utilization: 0.5
499
+ enforce_eager: true
500
+ cudagraph_capture_sizes: null
501
+ free_cache_engine: true
502
+ data_parallel_size: 1
503
+ expert_parallel_size: 1
504
+ tensor_model_parallel_size: 2
505
+ max_num_batched_tokens: 8192
506
+ max_model_len: null
507
+ max_num_seqs: 1024
508
+ load_format: auto
509
+ engine_kwargs: {}
510
+ limit_images: null
511
+ enable_chunked_prefill: true
512
+ enable_prefix_caching: true
513
+ disable_log_stats: true
514
+ skip_tokenizer_init: true
515
+ prompt_length: 512
516
+ response_length: 512
517
+ algorithm:
518
+ rollout_correction:
519
+ rollout_is: null
520
+ rollout_is_threshold: 2.0
521
+ rollout_rs: null
522
+ rollout_rs_threshold: null
523
+ rollout_rs_threshold_lower: null
524
+ rollout_token_veto_threshold: null
525
+ bypass_mode: false
526
+ use_policy_gradient: false
527
+ rollout_is_batch_normalize: false
528
+ _target_: verl.trainer.config.AlgoConfig
529
+ gamma: 1.0
530
+ lam: 1.0
531
+ adv_estimator: sgrpo
532
+ norm_adv_by_std_in_grpo: false
533
+ use_kl_in_reward: false
534
+ kl_penalty: kl
535
+ kl_ctrl:
536
+ _target_: verl.trainer.config.KLControlConfig
537
+ type: fixed
538
+ kl_coef: 0.001
539
+ horizon: 10000
540
+ target_kl: 0.1
541
+ use_pf_ppo: false
542
+ pf_ppo:
543
+ reweight_method: pow
544
+ weight_pow: 2.0
545
+ sgrpo:
546
+ num_truncations: 4
547
+ force_answer_prompt: Time is limited, stop thinking and start answering.\n</think>\n\n
548
+ answer_max_tokens: 256
549
+ min_truncation_ratio: 0.1
550
+ max_truncation_ratio: 0.9
551
+ custom_reward_function:
552
+ path: null
553
+ name: compute_score
554
+ trainer:
555
+ balance_batch: true
556
+ total_epochs: 15
557
+ total_training_steps: null
558
+ project_name: verl_sgrpo_example_gsm8k
559
+ experiment_name: qwen3_8b_sgrpo
560
+ logger:
561
+ - console
562
+ - wandb
563
+ log_val_generations: 0
564
+ rollout_data_dir: ./outputs/sgrpo_samples
565
+ rollout_data_dump_freq: 20
566
+ validation_data_dir: null
567
+ nnodes: 1
568
+ n_gpus_per_node: 8
569
+ save_freq: 20
570
+ esi_redundant_time: 0
571
+ resume_mode: disable
572
+ resume_from_path: null
573
+ val_before_train: true
574
+ val_only: false
575
+ test_freq: 5
576
+ critic_warmup: 0
577
+ default_hdfs_dir: null
578
+ del_local_ckpt_after_load: false
579
+ default_local_dir: checkpoints/${trainer.project_name}/${trainer.experiment_name}
580
+ max_actor_ckpt_to_keep: null
581
+ max_critic_ckpt_to_keep: null
582
+ ray_wait_register_center_timeout: 300
583
+ device: cuda
584
+ use_legacy_worker_impl: auto
585
+ global_profiler:
586
+ _target_: verl.utils.profiler.ProfilerConfig
587
+ tool: null
588
+ steps: null
589
+ profile_continuous_steps: false
590
+ save_path: outputs/profile
591
+ global_tool_config:
592
+ nsys:
593
+ _target_: verl.utils.profiler.config.NsightToolConfig
594
+ discrete: false
595
+ controller_nsight_options:
596
+ trace: cuda,nvtx,cublas,ucx
597
+ cuda-memory-usage: 'true'
598
+ cuda-graph-trace: graph
599
+ worker_nsight_options:
600
+ trace: cuda,nvtx,cublas,ucx
601
+ cuda-memory-usage: 'true'
602
+ cuda-graph-trace: graph
603
+ capture-range: cudaProfilerApi
604
+ capture-range-end: null
605
+ kill: none
606
+ torch_memory:
607
+ trace_alloc_max_entries: 100000
608
+ stack_depth: 32
609
+ context: all
610
+ stacks: all
611
+ kw_args: {}
612
+ transfer_queue:
613
+ enable: false
614
+ ray_kwargs:
615
+ ray_init:
616
+ num_cpus: null
617
+ timeline_json_file: null
outputs/2026-01-14/00-16-12/.hydra/overrides.yaml ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ - algorithm.adv_estimator=sgrpo
2
+ - algorithm.norm_adv_by_std_in_grpo=False
3
+ - algorithm.use_kl_in_reward=False
4
+ - data.train_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/train.parquet
5
+ - data.val_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/test.parquet
6
+ - data.train_batch_size=1024
7
+ - data.max_prompt_length=512
8
+ - data.max_response_length=1024
9
+ - data.filter_overlong_prompts=True
10
+ - data.truncation=error
11
+ - data.return_raw_chat=True
12
+ - actor_rollout_ref.model.path=/mnt/tidal-alsh01/dataset/redtrans/zhangruiqi/models/Qwen3-8B
13
+ - actor_rollout_ref.actor.optim.lr=1e-6
14
+ - actor_rollout_ref.model.use_remove_padding=True
15
+ - actor_rollout_ref.actor.ppo_mini_batch_size=256
16
+ - actor_rollout_ref.actor.ppo_micro_batch_size_per_gpu=32
17
+ - actor_rollout_ref.actor.use_kl_loss=True
18
+ - actor_rollout_ref.actor.kl_loss_coef=0.001
19
+ - actor_rollout_ref.actor.kl_loss_type=low_var_kl
20
+ - actor_rollout_ref.actor.entropy_coeff=0
21
+ - actor_rollout_ref.model.enable_gradient_checkpointing=True
22
+ - actor_rollout_ref.actor.fsdp_config.param_offload=False
23
+ - actor_rollout_ref.actor.fsdp_config.optimizer_offload=False
24
+ - actor_rollout_ref.rollout.log_prob_micro_batch_size_per_gpu=32
25
+ - actor_rollout_ref.rollout.tensor_model_parallel_size=2
26
+ - actor_rollout_ref.rollout.name=vllm
27
+ - actor_rollout_ref.rollout.gpu_memory_utilization=0.6
28
+ - actor_rollout_ref.rollout.n=1
29
+ - actor_rollout_ref.rollout.prompt_length=1536
30
+ - actor_rollout_ref.rollout.response_length=512
31
+ - actor_rollout_ref.ref.log_prob_micro_batch_size_per_gpu=32
32
+ - actor_rollout_ref.ref.fsdp_config.param_offload=False
33
+ - sgrpo.num_truncations=4
34
+ - sgrpo.force_answer_prompt="Time is limited, stop thinking and start answering.\n</think>\n\n"
35
+ - sgrpo.answer_max_tokens=256
36
+ - trainer.critic_warmup=0
37
+ - trainer.logger=["console","wandb"]
38
+ - trainer.project_name=verl_sgrpo_example_gsm8k
39
+ - trainer.experiment_name=qwen3_8b_sgrpo
40
+ - trainer.resume_mode=disable
41
+ - trainer.n_gpus_per_node=8
42
+ - trainer.nnodes=1
43
+ - trainer.save_freq=20
44
+ - trainer.test_freq=5
45
+ - trainer.total_epochs=15
46
+ - trainer.rollout_data_dir=./outputs/sgrpo_samples
47
+ - trainer.rollout_data_dump_freq=20
outputs/2026-01-14/00-17-48/.hydra/config.yaml ADDED
@@ -0,0 +1,617 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ actor_rollout_ref:
2
+ actor:
3
+ optim:
4
+ _target_: verl.workers.config.FSDPOptimizerConfig
5
+ optimizer: AdamW
6
+ optimizer_impl: torch.optim
7
+ lr: 1.0e-06
8
+ lr_warmup_steps_ratio: 0.0
9
+ total_training_steps: -1
10
+ weight_decay: 0.01
11
+ lr_warmup_steps: -1
12
+ betas:
13
+ - 0.9
14
+ - 0.999
15
+ clip_grad: 1.0
16
+ min_lr_ratio: 0.0
17
+ num_cycles: 0.5
18
+ lr_scheduler_type: constant
19
+ warmup_style: null
20
+ override_optimizer_config: null
21
+ fsdp_config:
22
+ _target_: verl.workers.config.FSDPEngineConfig
23
+ wrap_policy:
24
+ min_num_params: 0
25
+ param_offload: false
26
+ optimizer_offload: false
27
+ offload_policy: false
28
+ reshard_after_forward: true
29
+ fsdp_size: -1
30
+ forward_prefetch: false
31
+ model_dtype: fp32
32
+ use_orig_params: false
33
+ seed: 42
34
+ full_determinism: false
35
+ ulysses_sequence_parallel_size: 1
36
+ entropy_from_logits_with_chunking: false
37
+ use_torch_compile: true
38
+ entropy_checkpointing: false
39
+ forward_only: false
40
+ strategy: fsdp
41
+ dtype: bfloat16
42
+ _target_: verl.workers.config.FSDPActorConfig
43
+ rollout_n: ${oc.select:actor_rollout_ref.rollout.n,1}
44
+ strategy: fsdp
45
+ ppo_mini_batch_size: 256
46
+ ppo_micro_batch_size: null
47
+ ppo_micro_batch_size_per_gpu: 32
48
+ use_dynamic_bsz: false
49
+ ppo_max_token_len_per_gpu: 16384
50
+ clip_ratio: 0.2
51
+ clip_ratio_low: 0.2
52
+ clip_ratio_high: 0.2
53
+ freeze_vision_tower: false
54
+ policy_loss:
55
+ _target_: verl.workers.config.PolicyLossConfig
56
+ loss_mode: vanilla
57
+ clip_cov_ratio: 0.0002
58
+ clip_cov_lb: 1.0
59
+ clip_cov_ub: 5.0
60
+ kl_cov_ratio: 0.0002
61
+ ppo_kl_coef: 0.1
62
+ clip_ratio_c: 3.0
63
+ loss_agg_mode: token-mean
64
+ loss_scale_factor: null
65
+ entropy_coeff: 0
66
+ calculate_entropy: false
67
+ use_kl_loss: true
68
+ use_torch_compile: true
69
+ kl_loss_coef: 0.001
70
+ kl_loss_type: low_var_kl
71
+ ppo_epochs: 1
72
+ shuffle: false
73
+ checkpoint:
74
+ _target_: verl.trainer.config.CheckpointConfig
75
+ save_contents:
76
+ - model
77
+ - optimizer
78
+ - extra
79
+ load_contents: ${.save_contents}
80
+ async_save: false
81
+ use_fused_kernels: ${oc.select:actor_rollout_ref.model.use_fused_kernels,false}
82
+ profiler:
83
+ _target_: verl.utils.profiler.ProfilerConfig
84
+ tool: ${oc.select:global_profiler.tool,null}
85
+ enable: false
86
+ all_ranks: false
87
+ ranks: []
88
+ save_path: ${oc.select:global_profiler.save_path,null}
89
+ tool_config:
90
+ nsys:
91
+ _target_: verl.utils.profiler.config.NsightToolConfig
92
+ discrete: ${oc.select:global_profiler.global_tool_config.nsys.discrete}
93
+ npu:
94
+ _target_: verl.utils.profiler.config.NPUToolConfig
95
+ contents: []
96
+ level: level1
97
+ analysis: true
98
+ discrete: false
99
+ torch:
100
+ _target_: verl.utils.profiler.config.TorchProfilerToolConfig
101
+ step_start: 0
102
+ step_end: null
103
+ torch_memory:
104
+ _target_: verl.utils.profiler.config.TorchMemoryToolConfig
105
+ trace_alloc_max_entries: ${oc.select:global_profiler.global_tool_config.torch_memory.trace_alloc_max_entries,100000}
106
+ stack_depth: ${oc.select:global_profiler.global_tool_config.torch_memory.stack_depth,32}
107
+ router_replay:
108
+ _target_: verl.workers.config.RouterReplayConfig
109
+ mode: disabled
110
+ record_file: null
111
+ replay_file: null
112
+ grad_clip: 1.0
113
+ ulysses_sequence_parallel_size: 1
114
+ entropy_from_logits_with_chunking: false
115
+ entropy_checkpointing: false
116
+ use_remove_padding: ${oc.select:actor_rollout_ref.model.use_remove_padding,false}
117
+ ref:
118
+ rollout_n: ${oc.select:actor_rollout_ref.rollout.n,1}
119
+ strategy: ${actor_rollout_ref.actor.strategy}
120
+ use_torch_compile: ${oc.select:actor_rollout_ref.actor.use_torch_compile,true}
121
+ log_prob_micro_batch_size: null
122
+ log_prob_micro_batch_size_per_gpu: 32
123
+ log_prob_use_dynamic_bsz: ${oc.select:actor_rollout_ref.actor.use_dynamic_bsz,false}
124
+ log_prob_max_token_len_per_gpu: ${oc.select:actor_rollout_ref.actor.ppo_max_token_len_per_gpu,16384}
125
+ profiler:
126
+ _target_: verl.utils.profiler.ProfilerConfig
127
+ tool: ${oc.select:global_profiler.tool,null}
128
+ enable: false
129
+ all_ranks: false
130
+ ranks: []
131
+ save_path: ${oc.select:global_profiler.save_path,null}
132
+ tool_config:
133
+ nsys:
134
+ _target_: verl.utils.profiler.config.NsightToolConfig
135
+ discrete: ${oc.select:global_profiler.global_tool_config.nsys.discrete}
136
+ npu:
137
+ _target_: verl.utils.profiler.config.NPUToolConfig
138
+ contents: []
139
+ level: level1
140
+ analysis: true
141
+ discrete: false
142
+ torch:
143
+ _target_: verl.utils.profiler.config.TorchProfilerToolConfig
144
+ step_start: 0
145
+ step_end: null
146
+ torch_memory:
147
+ _target_: verl.utils.profiler.config.TorchMemoryToolConfig
148
+ trace_alloc_max_entries: ${oc.select:global_profiler.global_tool_config.torch_memory.trace_alloc_max_entries,100000}
149
+ stack_depth: ${oc.select:global_profiler.global_tool_config.torch_memory.stack_depth,32}
150
+ router_replay:
151
+ _target_: verl.workers.config.RouterReplayConfig
152
+ mode: disabled
153
+ record_file: null
154
+ replay_file: null
155
+ fsdp_config:
156
+ _target_: verl.workers.config.FSDPEngineConfig
157
+ wrap_policy:
158
+ min_num_params: 0
159
+ param_offload: false
160
+ optimizer_offload: false
161
+ offload_policy: false
162
+ reshard_after_forward: true
163
+ fsdp_size: -1
164
+ forward_prefetch: false
165
+ model_dtype: fp32
166
+ use_orig_params: false
167
+ seed: 42
168
+ full_determinism: false
169
+ ulysses_sequence_parallel_size: 1
170
+ entropy_from_logits_with_chunking: false
171
+ use_torch_compile: true
172
+ entropy_checkpointing: false
173
+ forward_only: true
174
+ strategy: fsdp
175
+ dtype: bfloat16
176
+ _target_: verl.workers.config.FSDPActorConfig
177
+ ulysses_sequence_parallel_size: ${oc.select:actor_rollout_ref.actor.ulysses_sequence_parallel_size,1}
178
+ entropy_from_logits_with_chunking: false
179
+ entropy_checkpointing: false
180
+ rollout:
181
+ _target_: verl.workers.config.RolloutConfig
182
+ name: vllm
183
+ mode: async
184
+ temperature: 1.0
185
+ top_k: -1
186
+ top_p: 1
187
+ prompt_length: 1536
188
+ response_length: 512
189
+ dtype: bfloat16
190
+ gpu_memory_utilization: 0.6
191
+ ignore_eos: false
192
+ enforce_eager: false
193
+ cudagraph_capture_sizes: null
194
+ free_cache_engine: true
195
+ tensor_model_parallel_size: 2
196
+ data_parallel_size: 1
197
+ expert_parallel_size: 1
198
+ pipeline_model_parallel_size: 1
199
+ max_num_batched_tokens: 8192
200
+ max_model_len: null
201
+ max_num_seqs: 1024
202
+ enable_chunked_prefill: true
203
+ enable_prefix_caching: true
204
+ load_format: dummy
205
+ log_prob_micro_batch_size: null
206
+ log_prob_micro_batch_size_per_gpu: 32
207
+ log_prob_use_dynamic_bsz: ${oc.select:actor_rollout_ref.actor.use_dynamic_bsz,false}
208
+ log_prob_max_token_len_per_gpu: ${oc.select:actor_rollout_ref.actor.ppo_max_token_len_per_gpu,16384}
209
+ disable_log_stats: true
210
+ do_sample: true
211
+ 'n': 1
212
+ over_sample_rate: 0
213
+ multi_stage_wake_up: false
214
+ engine_kwargs:
215
+ vllm: {}
216
+ sglang: {}
217
+ val_kwargs:
218
+ _target_: verl.workers.config.SamplingConfig
219
+ top_k: -1
220
+ top_p: 1.0
221
+ temperature: 0
222
+ 'n': 1
223
+ do_sample: false
224
+ multi_turn:
225
+ _target_: verl.workers.config.MultiTurnConfig
226
+ enable: false
227
+ max_assistant_turns: null
228
+ tool_config_path: null
229
+ max_user_turns: null
230
+ max_parallel_calls: 1
231
+ max_tool_response_length: 256
232
+ tool_response_truncate_side: middle
233
+ interaction_config_path: null
234
+ use_inference_chat_template: false
235
+ tokenization_sanity_check_mode: strict
236
+ format: hermes
237
+ num_repeat_rollouts: null
238
+ calculate_log_probs: false
239
+ agent:
240
+ _target_: verl.workers.config.AgentLoopConfig
241
+ num_workers: 8
242
+ default_agent_loop: single_turn_agent
243
+ agent_loop_config_path: null
244
+ custom_async_server:
245
+ _target_: verl.workers.config.CustomAsyncServerConfig
246
+ path: null
247
+ name: null
248
+ update_weights_bucket_megabytes: 512
249
+ trace:
250
+ _target_: verl.workers.config.TraceConfig
251
+ backend: null
252
+ token2text: false
253
+ max_samples_per_step_per_worker: null
254
+ skip_rollout: false
255
+ skip_dump_dir: /tmp/rollout_dump
256
+ skip_tokenizer_init: true
257
+ enable_rollout_routing_replay: false
258
+ profiler:
259
+ _target_: verl.utils.profiler.ProfilerConfig
260
+ tool: ${oc.select:global_profiler.tool,null}
261
+ enable: ${oc.select:actor_rollout_ref.actor.profiler.enable,false}
262
+ all_ranks: ${oc.select:actor_rollout_ref.actor.profiler.all_ranks,false}
263
+ ranks: ${oc.select:actor_rollout_ref.actor.profiler.ranks,[]}
264
+ save_path: ${oc.select:global_profiler.save_path,null}
265
+ tool_config: ${oc.select:actor_rollout_ref.actor.profiler.tool_config,null}
266
+ prometheus:
267
+ _target_: verl.workers.config.PrometheusConfig
268
+ enable: false
269
+ port: 9090
270
+ file: /tmp/ray/session_latest/metrics/prometheus/prometheus.yml
271
+ served_model_name: ${oc.select:actor_rollout_ref.model.path,null}
272
+ layered_summon: false
273
+ model:
274
+ _target_: verl.workers.config.HFModelConfig
275
+ path: /mnt/tidal-alsh01/dataset/redtrans/zhangruiqi/models/Qwen3-4B
276
+ hf_config_path: null
277
+ tokenizer_path: null
278
+ use_shm: false
279
+ trust_remote_code: false
280
+ custom_chat_template: null
281
+ external_lib: null
282
+ override_config: {}
283
+ enable_gradient_checkpointing: true
284
+ enable_activation_offload: false
285
+ use_remove_padding: true
286
+ lora_rank: 0
287
+ lora_alpha: 16
288
+ target_modules: all-linear
289
+ exclude_modules: null
290
+ lora_adapter_path: null
291
+ use_liger: false
292
+ use_fused_kernels: false
293
+ fused_kernel_options:
294
+ impl_backend: torch
295
+ hybrid_engine: true
296
+ nccl_timeout: 600
297
+ data:
298
+ tokenizer: null
299
+ use_shm: false
300
+ train_files: /mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/train.parquet
301
+ val_files: /mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/test.parquet
302
+ train_max_samples: -1
303
+ val_max_samples: -1
304
+ prompt_key: prompt
305
+ reward_fn_key: data_source
306
+ max_prompt_length: 512
307
+ max_response_length: 1024
308
+ train_batch_size: 1024
309
+ val_batch_size: null
310
+ tool_config_path: ${oc.select:actor_rollout_ref.rollout.multi_turn.tool_config_path,
311
+ null}
312
+ return_raw_input_ids: false
313
+ return_raw_chat: true
314
+ return_full_prompt: false
315
+ shuffle: true
316
+ seed: null
317
+ dataloader_num_workers: 8
318
+ image_patch_size: 14
319
+ validation_shuffle: false
320
+ filter_overlong_prompts: true
321
+ filter_overlong_prompts_workers: 1
322
+ truncation: error
323
+ image_key: images
324
+ video_key: videos
325
+ trust_remote_code: false
326
+ custom_cls:
327
+ path: null
328
+ name: null
329
+ return_multi_modal_inputs: true
330
+ sampler:
331
+ class_path: null
332
+ class_name: null
333
+ datagen:
334
+ path: null
335
+ name: null
336
+ apply_chat_template_kwargs: {}
337
+ reward_manager:
338
+ _target_: verl.trainer.config.config.RewardManagerConfig
339
+ source: register
340
+ name: ${oc.select:reward_model.reward_manager,naive}
341
+ module:
342
+ _target_: verl.trainer.config.config.ModuleConfig
343
+ path: null
344
+ name: custom_reward_manager
345
+ critic:
346
+ optim:
347
+ _target_: verl.workers.config.FSDPOptimizerConfig
348
+ optimizer: AdamW
349
+ optimizer_impl: torch.optim
350
+ lr: 1.0e-05
351
+ lr_warmup_steps_ratio: 0.0
352
+ total_training_steps: -1
353
+ weight_decay: 0.01
354
+ lr_warmup_steps: -1
355
+ betas:
356
+ - 0.9
357
+ - 0.999
358
+ clip_grad: 1.0
359
+ min_lr_ratio: 0.0
360
+ num_cycles: 0.5
361
+ lr_scheduler_type: constant
362
+ warmup_style: null
363
+ override_optimizer_config: null
364
+ model:
365
+ fsdp_config:
366
+ _target_: verl.workers.config.FSDPEngineConfig
367
+ wrap_policy:
368
+ min_num_params: 0
369
+ param_offload: false
370
+ optimizer_offload: false
371
+ offload_policy: false
372
+ reshard_after_forward: true
373
+ fsdp_size: -1
374
+ forward_prefetch: false
375
+ model_dtype: fp32
376
+ use_orig_params: false
377
+ seed: 42
378
+ full_determinism: false
379
+ ulysses_sequence_parallel_size: 1
380
+ entropy_from_logits_with_chunking: false
381
+ use_torch_compile: true
382
+ entropy_checkpointing: false
383
+ forward_only: false
384
+ strategy: fsdp
385
+ dtype: bfloat16
386
+ path: ~/models/deepseek-llm-7b-chat
387
+ tokenizer_path: ${oc.select:actor_rollout_ref.model.path,"~/models/deepseek-llm-7b-chat"}
388
+ override_config: {}
389
+ external_lib: ${oc.select:actor_rollout_ref.model.external_lib,null}
390
+ trust_remote_code: ${oc.select:actor_rollout_ref.model.trust_remote_code,false}
391
+ _target_: verl.workers.config.FSDPCriticModelCfg
392
+ use_shm: false
393
+ enable_gradient_checkpointing: true
394
+ enable_activation_offload: false
395
+ use_remove_padding: false
396
+ lora_rank: 0
397
+ lora_alpha: 16
398
+ target_modules: all-linear
399
+ _target_: verl.workers.config.FSDPCriticConfig
400
+ rollout_n: ${oc.select:actor_rollout_ref.rollout.n,1}
401
+ strategy: fsdp
402
+ enable: null
403
+ ppo_mini_batch_size: ${oc.select:actor_rollout_ref.actor.ppo_mini_batch_size,256}
404
+ ppo_micro_batch_size: null
405
+ ppo_micro_batch_size_per_gpu: ${oc.select:.ppo_micro_batch_size,null}
406
+ use_dynamic_bsz: ${oc.select:actor_rollout_ref.actor.use_dynamic_bsz,false}
407
+ ppo_max_token_len_per_gpu: 32768
408
+ forward_max_token_len_per_gpu: ${.ppo_max_token_len_per_gpu}
409
+ ppo_epochs: ${oc.select:actor_rollout_ref.actor.ppo_epochs,1}
410
+ shuffle: ${oc.select:actor_rollout_ref.actor.shuffle,false}
411
+ cliprange_value: 0.5
412
+ loss_agg_mode: ${oc.select:actor_rollout_ref.actor.loss_agg_mode,token-mean}
413
+ checkpoint:
414
+ _target_: verl.trainer.config.CheckpointConfig
415
+ save_contents:
416
+ - model
417
+ - optimizer
418
+ - extra
419
+ load_contents: ${.save_contents}
420
+ async_save: false
421
+ profiler:
422
+ _target_: verl.utils.profiler.ProfilerConfig
423
+ tool: ${oc.select:global_profiler.tool,null}
424
+ enable: false
425
+ all_ranks: false
426
+ ranks: []
427
+ save_path: ${oc.select:global_profiler.save_path,null}
428
+ tool_config:
429
+ nsys:
430
+ _target_: verl.utils.profiler.config.NsightToolConfig
431
+ discrete: ${oc.select:global_profiler.global_tool_config.nsys.discrete}
432
+ npu:
433
+ _target_: verl.utils.profiler.config.NPUToolConfig
434
+ contents: []
435
+ level: level1
436
+ analysis: true
437
+ discrete: false
438
+ torch:
439
+ _target_: verl.utils.profiler.config.TorchProfilerToolConfig
440
+ step_start: 0
441
+ step_end: null
442
+ torch_memory:
443
+ _target_: verl.utils.profiler.config.TorchMemoryToolConfig
444
+ trace_alloc_max_entries: ${oc.select:global_profiler.global_tool_config.torch_memory.trace_alloc_max_entries,100000}
445
+ stack_depth: ${oc.select:global_profiler.global_tool_config.torch_memory.stack_depth,32}
446
+ forward_micro_batch_size: ${oc.select:.ppo_micro_batch_size,null}
447
+ forward_micro_batch_size_per_gpu: ${oc.select:.ppo_micro_batch_size_per_gpu,null}
448
+ ulysses_sequence_parallel_size: 1
449
+ grad_clip: 1.0
450
+ reward_model:
451
+ enable: false
452
+ enable_resource_pool: false
453
+ n_gpus_per_node: 0
454
+ nnodes: 0
455
+ strategy: fsdp
456
+ model:
457
+ input_tokenizer: ${actor_rollout_ref.model.path}
458
+ path: ~/models/FsfairX-LLaMA3-RM-v0.1
459
+ external_lib: ${actor_rollout_ref.model.external_lib}
460
+ trust_remote_code: false
461
+ override_config: {}
462
+ use_shm: false
463
+ use_remove_padding: false
464
+ use_fused_kernels: ${actor_rollout_ref.model.use_fused_kernels}
465
+ fsdp_config:
466
+ _target_: verl.workers.config.FSDPEngineConfig
467
+ wrap_policy:
468
+ min_num_params: 0
469
+ param_offload: false
470
+ reshard_after_forward: true
471
+ fsdp_size: -1
472
+ forward_prefetch: false
473
+ micro_batch_size: null
474
+ micro_batch_size_per_gpu: null
475
+ max_length: null
476
+ use_dynamic_bsz: ${critic.use_dynamic_bsz}
477
+ forward_max_token_len_per_gpu: ${critic.forward_max_token_len_per_gpu}
478
+ reward_manager: naive
479
+ launch_reward_fn_async: false
480
+ sandbox_fusion:
481
+ url: null
482
+ max_concurrent: 64
483
+ memory_limit_mb: 1024
484
+ profiler:
485
+ _target_: verl.utils.profiler.ProfilerConfig
486
+ tool: ${oc.select:global_profiler.tool,null}
487
+ enable: false
488
+ all_ranks: false
489
+ ranks: []
490
+ save_path: ${oc.select:global_profiler.save_path,null}
491
+ tool_config: ${oc.select:actor_rollout_ref.actor.profiler.tool_config,null}
492
+ ulysses_sequence_parallel_size: 1
493
+ use_reward_loop: true
494
+ rollout:
495
+ _target_: verl.workers.config.RolloutConfig
496
+ name: ???
497
+ dtype: bfloat16
498
+ gpu_memory_utilization: 0.5
499
+ enforce_eager: true
500
+ cudagraph_capture_sizes: null
501
+ free_cache_engine: true
502
+ data_parallel_size: 1
503
+ expert_parallel_size: 1
504
+ tensor_model_parallel_size: 2
505
+ max_num_batched_tokens: 8192
506
+ max_model_len: null
507
+ max_num_seqs: 1024
508
+ load_format: auto
509
+ engine_kwargs: {}
510
+ limit_images: null
511
+ enable_chunked_prefill: true
512
+ enable_prefix_caching: true
513
+ disable_log_stats: true
514
+ skip_tokenizer_init: true
515
+ prompt_length: 512
516
+ response_length: 512
517
+ algorithm:
518
+ rollout_correction:
519
+ rollout_is: null
520
+ rollout_is_threshold: 2.0
521
+ rollout_rs: null
522
+ rollout_rs_threshold: null
523
+ rollout_rs_threshold_lower: null
524
+ rollout_token_veto_threshold: null
525
+ bypass_mode: false
526
+ use_policy_gradient: false
527
+ rollout_is_batch_normalize: false
528
+ _target_: verl.trainer.config.AlgoConfig
529
+ gamma: 1.0
530
+ lam: 1.0
531
+ adv_estimator: sgrpo
532
+ norm_adv_by_std_in_grpo: false
533
+ use_kl_in_reward: false
534
+ kl_penalty: kl
535
+ kl_ctrl:
536
+ _target_: verl.trainer.config.KLControlConfig
537
+ type: fixed
538
+ kl_coef: 0.001
539
+ horizon: 10000
540
+ target_kl: 0.1
541
+ use_pf_ppo: false
542
+ pf_ppo:
543
+ reweight_method: pow
544
+ weight_pow: 2.0
545
+ sgrpo:
546
+ num_truncations: 4
547
+ force_answer_prompt: Time is limited, stop thinking and start answering.\n</think>\n\n
548
+ answer_max_tokens: 256
549
+ min_truncation_ratio: 0.1
550
+ max_truncation_ratio: 0.9
551
+ custom_reward_function:
552
+ path: null
553
+ name: compute_score
554
+ trainer:
555
+ balance_batch: true
556
+ total_epochs: 15
557
+ total_training_steps: null
558
+ project_name: verl_sgrpo_example_gsm8k
559
+ experiment_name: qwen3_8b_sgrpo
560
+ logger:
561
+ - console
562
+ - wandb
563
+ log_val_generations: 0
564
+ rollout_data_dir: ./outputs/sgrpo_samples
565
+ rollout_data_dump_freq: 20
566
+ validation_data_dir: null
567
+ nnodes: 1
568
+ n_gpus_per_node: 8
569
+ save_freq: 20
570
+ esi_redundant_time: 0
571
+ resume_mode: disable
572
+ resume_from_path: null
573
+ val_before_train: true
574
+ val_only: false
575
+ test_freq: 5
576
+ critic_warmup: 0
577
+ default_hdfs_dir: null
578
+ del_local_ckpt_after_load: false
579
+ default_local_dir: checkpoints/${trainer.project_name}/${trainer.experiment_name}
580
+ max_actor_ckpt_to_keep: null
581
+ max_critic_ckpt_to_keep: null
582
+ ray_wait_register_center_timeout: 300
583
+ device: cuda
584
+ use_legacy_worker_impl: auto
585
+ global_profiler:
586
+ _target_: verl.utils.profiler.ProfilerConfig
587
+ tool: null
588
+ steps: null
589
+ profile_continuous_steps: false
590
+ save_path: outputs/profile
591
+ global_tool_config:
592
+ nsys:
593
+ _target_: verl.utils.profiler.config.NsightToolConfig
594
+ discrete: false
595
+ controller_nsight_options:
596
+ trace: cuda,nvtx,cublas,ucx
597
+ cuda-memory-usage: 'true'
598
+ cuda-graph-trace: graph
599
+ worker_nsight_options:
600
+ trace: cuda,nvtx,cublas,ucx
601
+ cuda-memory-usage: 'true'
602
+ cuda-graph-trace: graph
603
+ capture-range: cudaProfilerApi
604
+ capture-range-end: null
605
+ kill: none
606
+ torch_memory:
607
+ trace_alloc_max_entries: 100000
608
+ stack_depth: 32
609
+ context: all
610
+ stacks: all
611
+ kw_args: {}
612
+ transfer_queue:
613
+ enable: false
614
+ ray_kwargs:
615
+ ray_init:
616
+ num_cpus: null
617
+ timeline_json_file: null
outputs/2026-01-14/09-28-49/.hydra/overrides.yaml ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ - algorithm.adv_estimator=sgrpo
2
+ - algorithm.norm_adv_by_std_in_grpo=False
3
+ - algorithm.use_kl_in_reward=False
4
+ - data.train_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/train.parquet
5
+ - data.val_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/test.parquet
6
+ - data.train_batch_size=1024
7
+ - data.max_prompt_length=512
8
+ - data.max_response_length=1024
9
+ - data.filter_overlong_prompts=True
10
+ - data.truncation=error
11
+ - data.return_raw_chat=True
12
+ - actor_rollout_ref.model.path=/mnt/tidal-alsh01/dataset/redtrans/zhangruiqi/models/Qwen3-4B
13
+ - actor_rollout_ref.actor.optim.lr=1e-6
14
+ - actor_rollout_ref.model.use_remove_padding=True
15
+ - actor_rollout_ref.actor.ppo_mini_batch_size=256
16
+ - actor_rollout_ref.actor.ppo_micro_batch_size_per_gpu=32
17
+ - actor_rollout_ref.actor.use_kl_loss=True
18
+ - actor_rollout_ref.actor.kl_loss_coef=0.001
19
+ - actor_rollout_ref.actor.kl_loss_type=low_var_kl
20
+ - actor_rollout_ref.actor.entropy_coeff=0
21
+ - actor_rollout_ref.model.enable_gradient_checkpointing=True
22
+ - actor_rollout_ref.actor.fsdp_config.param_offload=False
23
+ - actor_rollout_ref.actor.fsdp_config.optimizer_offload=False
24
+ - actor_rollout_ref.rollout.log_prob_micro_batch_size_per_gpu=32
25
+ - actor_rollout_ref.rollout.tensor_model_parallel_size=2
26
+ - actor_rollout_ref.rollout.name=vllm
27
+ - actor_rollout_ref.rollout.gpu_memory_utilization=0.6
28
+ - actor_rollout_ref.rollout.n=1
29
+ - actor_rollout_ref.rollout.prompt_length=1536
30
+ - actor_rollout_ref.rollout.response_length=512
31
+ - actor_rollout_ref.ref.log_prob_micro_batch_size_per_gpu=32
32
+ - actor_rollout_ref.ref.fsdp_config.param_offload=False
33
+ - sgrpo.num_truncations=4
34
+ - sgrpo.force_answer_prompt="Time is limited, stop thinking and start answering.\n</think>\n\n"
35
+ - sgrpo.answer_max_tokens=256
36
+ - trainer.critic_warmup=0
37
+ - trainer.logger=["console","wandb"]
38
+ - trainer.project_name=verl_sgrpo_example_gsm8k
39
+ - trainer.experiment_name=qwen3_8b_sgrpo
40
+ - trainer.resume_mode=disable
41
+ - trainer.n_gpus_per_node=8
42
+ - trainer.nnodes=1
43
+ - trainer.save_freq=20
44
+ - trainer.test_freq=5
45
+ - trainer.total_epochs=15
46
+ - trainer.rollout_data_dir=./outputs/sgrpo_samples
47
+ - trainer.rollout_data_dump_freq=20
outputs/2026-01-14/09-30-16/.hydra/overrides.yaml ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ - algorithm.adv_estimator=sgrpo
2
+ - algorithm.norm_adv_by_std_in_grpo=False
3
+ - algorithm.use_kl_in_reward=False
4
+ - data.train_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/train.parquet
5
+ - data.val_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/test.parquet
6
+ - data.train_batch_size=1024
7
+ - data.max_prompt_length=512
8
+ - data.max_response_length=1024
9
+ - data.filter_overlong_prompts=True
10
+ - data.truncation=error
11
+ - data.return_raw_chat=True
12
+ - actor_rollout_ref.model.path=/mnt/tidal-alsh01/dataset/redtrans/zhangruiqi/models/Qwen3-4B
13
+ - actor_rollout_ref.actor.optim.lr=1e-6
14
+ - actor_rollout_ref.model.use_remove_padding=True
15
+ - actor_rollout_ref.actor.ppo_mini_batch_size=256
16
+ - actor_rollout_ref.actor.ppo_micro_batch_size_per_gpu=32
17
+ - actor_rollout_ref.actor.use_kl_loss=True
18
+ - actor_rollout_ref.actor.kl_loss_coef=0.001
19
+ - actor_rollout_ref.actor.kl_loss_type=low_var_kl
20
+ - actor_rollout_ref.actor.entropy_coeff=0
21
+ - actor_rollout_ref.model.enable_gradient_checkpointing=True
22
+ - actor_rollout_ref.actor.fsdp_config.param_offload=False
23
+ - actor_rollout_ref.actor.fsdp_config.optimizer_offload=False
24
+ - actor_rollout_ref.rollout.log_prob_micro_batch_size_per_gpu=32
25
+ - actor_rollout_ref.rollout.tensor_model_parallel_size=2
26
+ - actor_rollout_ref.rollout.name=vllm
27
+ - actor_rollout_ref.rollout.gpu_memory_utilization=0.6
28
+ - actor_rollout_ref.rollout.n=1
29
+ - actor_rollout_ref.rollout.prompt_length=1536
30
+ - actor_rollout_ref.rollout.response_length=512
31
+ - actor_rollout_ref.ref.log_prob_micro_batch_size_per_gpu=32
32
+ - actor_rollout_ref.ref.fsdp_config.param_offload=False
33
+ - sgrpo.num_truncations=4
34
+ - sgrpo.force_answer_prompt="Time is limited, stop thinking and start answering.\n</think>\n\n"
35
+ - sgrpo.answer_max_tokens=256
36
+ - trainer.critic_warmup=0
37
+ - trainer.logger=["console"]
38
+ - trainer.project_name=verl_sgrpo_example_gsm8k
39
+ - trainer.experiment_name=qwen3_8b_sgrpo
40
+ - trainer.resume_mode=disable
41
+ - trainer.n_gpus_per_node=8
42
+ - trainer.nnodes=1
43
+ - trainer.save_freq=20
44
+ - trainer.test_freq=5
45
+ - trainer.total_epochs=15
46
+ - trainer.rollout_data_dir=./outputs/sgrpo_samples
47
+ - trainer.rollout_data_dump_freq=20
outputs/2026-01-14/09-35-00/.hydra/overrides.yaml ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ - algorithm.adv_estimator=sgrpo
2
+ - algorithm.norm_adv_by_std_in_grpo=False
3
+ - algorithm.use_kl_in_reward=False
4
+ - data.train_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/train.parquet
5
+ - data.val_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/test.parquet
6
+ - data.train_batch_size=1024
7
+ - data.max_prompt_length=512
8
+ - data.max_response_length=1024
9
+ - data.filter_overlong_prompts=True
10
+ - data.truncation=error
11
+ - data.return_raw_chat=True
12
+ - actor_rollout_ref.model.path=/mnt/tidal-alsh01/dataset/redtrans/zhangruiqi/models/Qwen3-4B
13
+ - actor_rollout_ref.actor.optim.lr=1e-6
14
+ - actor_rollout_ref.model.use_remove_padding=True
15
+ - actor_rollout_ref.actor.ppo_mini_batch_size=256
16
+ - actor_rollout_ref.actor.ppo_micro_batch_size_per_gpu=32
17
+ - actor_rollout_ref.actor.use_kl_loss=True
18
+ - actor_rollout_ref.actor.kl_loss_coef=0.001
19
+ - actor_rollout_ref.actor.kl_loss_type=low_var_kl
20
+ - actor_rollout_ref.actor.entropy_coeff=0
21
+ - actor_rollout_ref.model.enable_gradient_checkpointing=True
22
+ - actor_rollout_ref.actor.fsdp_config.param_offload=False
23
+ - actor_rollout_ref.actor.fsdp_config.optimizer_offload=False
24
+ - actor_rollout_ref.rollout.log_prob_micro_batch_size_per_gpu=32
25
+ - actor_rollout_ref.rollout.tensor_model_parallel_size=2
26
+ - actor_rollout_ref.rollout.name=vllm
27
+ - actor_rollout_ref.rollout.gpu_memory_utilization=0.6
28
+ - actor_rollout_ref.rollout.n=1
29
+ - actor_rollout_ref.rollout.prompt_length=1536
30
+ - actor_rollout_ref.rollout.response_length=512
31
+ - actor_rollout_ref.ref.log_prob_micro_batch_size_per_gpu=32
32
+ - actor_rollout_ref.ref.fsdp_config.param_offload=False
33
+ - sgrpo.num_truncations=4
34
+ - sgrpo.force_answer_prompt="Time is limited, stop thinking and start answering.\n</think>\n\n"
35
+ - sgrpo.answer_max_tokens=256
36
+ - trainer.critic_warmup=0
37
+ - trainer.logger=["console"]
38
+ - trainer.project_name=verl_sgrpo_example_gsm8k
39
+ - trainer.experiment_name=qwen3_8b_sgrpo
40
+ - trainer.resume_mode=disable
41
+ - trainer.n_gpus_per_node=8
42
+ - trainer.nnodes=1
43
+ - trainer.save_freq=20
44
+ - trainer.test_freq=5
45
+ - trainer.total_epochs=15
46
+ - trainer.rollout_data_dir=./outputs/sgrpo_samples
47
+ - trainer.rollout_data_dump_freq=20
outputs/2026-01-14/14-26-40/.hydra/hydra.yaml ADDED
@@ -0,0 +1,218 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
4
+ sweep:
5
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.mode=RUN
114
+ task:
115
+ - algorithm.adv_estimator=sgrpo
116
+ - algorithm.norm_adv_by_std_in_grpo=False
117
+ - algorithm.use_kl_in_reward=False
118
+ - data.train_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/train.parquet
119
+ - data.val_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/test.parquet
120
+ - data.train_batch_size=1024
121
+ - data.max_prompt_length=512
122
+ - data.max_response_length=1024
123
+ - data.filter_overlong_prompts=True
124
+ - data.truncation=error
125
+ - data.return_raw_chat=True
126
+ - actor_rollout_ref.model.path=/mnt/tidal-alsh01/dataset/redtrans/zhangruiqi/models/Qwen3-4B
127
+ - actor_rollout_ref.actor.optim.lr=1e-6
128
+ - actor_rollout_ref.model.use_remove_padding=True
129
+ - actor_rollout_ref.actor.ppo_mini_batch_size=256
130
+ - actor_rollout_ref.actor.ppo_micro_batch_size_per_gpu=32
131
+ - actor_rollout_ref.actor.use_kl_loss=True
132
+ - actor_rollout_ref.actor.kl_loss_coef=0.001
133
+ - actor_rollout_ref.actor.kl_loss_type=low_var_kl
134
+ - actor_rollout_ref.actor.entropy_coeff=0
135
+ - actor_rollout_ref.model.enable_gradient_checkpointing=True
136
+ - actor_rollout_ref.actor.fsdp_config.param_offload=False
137
+ - actor_rollout_ref.actor.fsdp_config.optimizer_offload=False
138
+ - actor_rollout_ref.rollout.log_prob_micro_batch_size_per_gpu=32
139
+ - actor_rollout_ref.rollout.tensor_model_parallel_size=2
140
+ - actor_rollout_ref.rollout.name=vllm
141
+ - actor_rollout_ref.rollout.gpu_memory_utilization=0.6
142
+ - actor_rollout_ref.rollout.n=1
143
+ - actor_rollout_ref.rollout.prompt_length=1536
144
+ - actor_rollout_ref.rollout.response_length=512
145
+ - actor_rollout_ref.ref.log_prob_micro_batch_size_per_gpu=32
146
+ - actor_rollout_ref.ref.fsdp_config.param_offload=False
147
+ - sgrpo.num_truncations=4
148
+ - sgrpo.force_answer_prompt="Time is limited, stop thinking and start answering.\n</think>\n\n"
149
+ - sgrpo.answer_max_tokens=256
150
+ - trainer.critic_warmup=0
151
+ - trainer.logger=["console"]
152
+ - trainer.project_name=verl_sgrpo_example_gsm8k
153
+ - trainer.experiment_name=qwen3_8b_sgrpo
154
+ - trainer.resume_mode=disable
155
+ - trainer.n_gpus_per_node=8
156
+ - trainer.nnodes=1
157
+ - trainer.save_freq=20
158
+ - trainer.test_freq=5
159
+ - trainer.total_epochs=15
160
+ - trainer.rollout_data_dir=./outputs/sgrpo_samples
161
+ - trainer.rollout_data_dump_freq=20
162
+ job:
163
+ name: main_sgrpo
164
+ chdir: null
165
+ override_dirname: actor_rollout_ref.actor.entropy_coeff=0,actor_rollout_ref.actor.fsdp_config.optimizer_offload=False,actor_rollout_ref.actor.fsdp_config.param_offload=False,actor_rollout_ref.actor.kl_loss_coef=0.001,actor_rollout_ref.actor.kl_loss_type=low_var_kl,actor_rollout_ref.actor.optim.lr=1e-6,actor_rollout_ref.actor.ppo_micro_batch_size_per_gpu=32,actor_rollout_ref.actor.ppo_mini_batch_size=256,actor_rollout_ref.actor.use_kl_loss=True,actor_rollout_ref.model.enable_gradient_checkpointing=True,actor_rollout_ref.model.path=/mnt/tidal-alsh01/dataset/redtrans/zhangruiqi/models/Qwen3-4B,actor_rollout_ref.model.use_remove_padding=True,actor_rollout_ref.ref.fsdp_config.param_offload=False,actor_rollout_ref.ref.log_prob_micro_batch_size_per_gpu=32,actor_rollout_ref.rollout.gpu_memory_utilization=0.6,actor_rollout_ref.rollout.log_prob_micro_batch_size_per_gpu=32,actor_rollout_ref.rollout.n=1,actor_rollout_ref.rollout.name=vllm,actor_rollout_ref.rollout.prompt_length=1536,actor_rollout_ref.rollout.response_length=512,actor_rollout_ref.rollout.tensor_model_parallel_size=2,algorithm.adv_estimator=sgrpo,algorithm.norm_adv_by_std_in_grpo=False,algorithm.use_kl_in_reward=False,data.filter_overlong_prompts=True,data.max_prompt_length=512,data.max_response_length=1024,data.return_raw_chat=True,data.train_batch_size=1024,data.train_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/train.parquet,data.truncation=error,data.val_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/test.parquet,sgrpo.answer_max_tokens=256,sgrpo.force_answer_prompt="Time
166
+ is limited, stop thinking and start answering.\n</think>\n\n",sgrpo.num_truncations=4,trainer.critic_warmup=0,trainer.experiment_name=qwen3_8b_sgrpo,trainer.logger=["console"],trainer.n_gpus_per_node=8,trainer.nnodes=1,trainer.project_name=verl_sgrpo_example_gsm8k,trainer.resume_mode=disable,trainer.rollout_data_dir=./outputs/sgrpo_samples,trainer.rollout_data_dump_freq=20,trainer.save_freq=20,trainer.test_freq=5,trainer.total_epochs=15
167
+ id: ???
168
+ num: ???
169
+ config_name: sgrpo_trainer
170
+ env_set: {}
171
+ env_copy: []
172
+ config:
173
+ override_dirname:
174
+ kv_sep: '='
175
+ item_sep: ','
176
+ exclude_keys: []
177
+ runtime:
178
+ version: 1.3.2
179
+ version_base: '1.3'
180
+ cwd: /mnt/tidal-alsh01/usr/zhangruiqi1/my/verl
181
+ config_sources:
182
+ - path: hydra.conf
183
+ schema: pkg
184
+ provider: hydra
185
+ - path: /mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/config
186
+ schema: file
187
+ provider: main
188
+ - path: verl.trainer.config
189
+ schema: pkg
190
+ provider: verl-trainer-config
191
+ - path: ''
192
+ schema: structured
193
+ provider: schema
194
+ output_dir: /mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/outputs/2026-01-14/14-26-40
195
+ choices:
196
+ algorithm@algorithm.rollout_correction: rollout_correction
197
+ reward_model: dp_reward_loop
198
+ critic: dp_critic
199
+ critic/../engine@critic.model.fsdp_config: fsdp
200
+ critic/../optim@critic.optim: fsdp
201
+ model@actor_rollout_ref.model: hf_model
202
+ rollout@actor_rollout_ref.rollout: rollout
203
+ ref@actor_rollout_ref.ref: dp_ref
204
+ ref/../engine@actor_rollout_ref.ref.fsdp_config: fsdp
205
+ data: legacy_data
206
+ actor@actor_rollout_ref.actor: dp_actor
207
+ actor/../engine@actor_rollout_ref.actor.fsdp_config: fsdp
208
+ actor/../optim@actor_rollout_ref.actor.optim: fsdp
209
+ hydra/env: default
210
+ hydra/callbacks: null
211
+ hydra/job_logging: default
212
+ hydra/hydra_logging: default
213
+ hydra/hydra_help: default
214
+ hydra/help: default
215
+ hydra/sweeper: basic
216
+ hydra/launcher: basic
217
+ hydra/output: default
218
+ verbose: false
outputs/2026-01-14/14-26-40/.hydra/overrides.yaml ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ - algorithm.adv_estimator=sgrpo
2
+ - algorithm.norm_adv_by_std_in_grpo=False
3
+ - algorithm.use_kl_in_reward=False
4
+ - data.train_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/train.parquet
5
+ - data.val_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/test.parquet
6
+ - data.train_batch_size=1024
7
+ - data.max_prompt_length=512
8
+ - data.max_response_length=1024
9
+ - data.filter_overlong_prompts=True
10
+ - data.truncation=error
11
+ - data.return_raw_chat=True
12
+ - actor_rollout_ref.model.path=/mnt/tidal-alsh01/dataset/redtrans/zhangruiqi/models/Qwen3-4B
13
+ - actor_rollout_ref.actor.optim.lr=1e-6
14
+ - actor_rollout_ref.model.use_remove_padding=True
15
+ - actor_rollout_ref.actor.ppo_mini_batch_size=256
16
+ - actor_rollout_ref.actor.ppo_micro_batch_size_per_gpu=32
17
+ - actor_rollout_ref.actor.use_kl_loss=True
18
+ - actor_rollout_ref.actor.kl_loss_coef=0.001
19
+ - actor_rollout_ref.actor.kl_loss_type=low_var_kl
20
+ - actor_rollout_ref.actor.entropy_coeff=0
21
+ - actor_rollout_ref.model.enable_gradient_checkpointing=True
22
+ - actor_rollout_ref.actor.fsdp_config.param_offload=False
23
+ - actor_rollout_ref.actor.fsdp_config.optimizer_offload=False
24
+ - actor_rollout_ref.rollout.log_prob_micro_batch_size_per_gpu=32
25
+ - actor_rollout_ref.rollout.tensor_model_parallel_size=2
26
+ - actor_rollout_ref.rollout.name=vllm
27
+ - actor_rollout_ref.rollout.gpu_memory_utilization=0.6
28
+ - actor_rollout_ref.rollout.n=1
29
+ - actor_rollout_ref.rollout.prompt_length=1536
30
+ - actor_rollout_ref.rollout.response_length=512
31
+ - actor_rollout_ref.ref.log_prob_micro_batch_size_per_gpu=32
32
+ - actor_rollout_ref.ref.fsdp_config.param_offload=False
33
+ - sgrpo.num_truncations=4
34
+ - sgrpo.force_answer_prompt="Time is limited, stop thinking and start answering.\n</think>\n\n"
35
+ - sgrpo.answer_max_tokens=256
36
+ - trainer.critic_warmup=0
37
+ - trainer.logger=["console"]
38
+ - trainer.project_name=verl_sgrpo_example_gsm8k
39
+ - trainer.experiment_name=qwen3_8b_sgrpo
40
+ - trainer.resume_mode=disable
41
+ - trainer.n_gpus_per_node=8
42
+ - trainer.nnodes=1
43
+ - trainer.save_freq=20
44
+ - trainer.test_freq=5
45
+ - trainer.total_epochs=15
46
+ - trainer.rollout_data_dir=./outputs/sgrpo_samples
47
+ - trainer.rollout_data_dump_freq=20
outputs/2026-01-14/15-08-55/.hydra/config.yaml ADDED
@@ -0,0 +1,616 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ actor_rollout_ref:
2
+ actor:
3
+ optim:
4
+ _target_: verl.workers.config.FSDPOptimizerConfig
5
+ optimizer: AdamW
6
+ optimizer_impl: torch.optim
7
+ lr: 1.0e-06
8
+ lr_warmup_steps_ratio: 0.0
9
+ total_training_steps: -1
10
+ weight_decay: 0.01
11
+ lr_warmup_steps: -1
12
+ betas:
13
+ - 0.9
14
+ - 0.999
15
+ clip_grad: 1.0
16
+ min_lr_ratio: 0.0
17
+ num_cycles: 0.5
18
+ lr_scheduler_type: constant
19
+ warmup_style: null
20
+ override_optimizer_config: null
21
+ fsdp_config:
22
+ _target_: verl.workers.config.FSDPEngineConfig
23
+ wrap_policy:
24
+ min_num_params: 0
25
+ param_offload: false
26
+ optimizer_offload: false
27
+ offload_policy: false
28
+ reshard_after_forward: true
29
+ fsdp_size: -1
30
+ forward_prefetch: false
31
+ model_dtype: fp32
32
+ use_orig_params: false
33
+ seed: 42
34
+ full_determinism: false
35
+ ulysses_sequence_parallel_size: 1
36
+ entropy_from_logits_with_chunking: false
37
+ use_torch_compile: true
38
+ entropy_checkpointing: false
39
+ forward_only: false
40
+ strategy: fsdp
41
+ dtype: bfloat16
42
+ _target_: verl.workers.config.FSDPActorConfig
43
+ rollout_n: ${oc.select:actor_rollout_ref.rollout.n,1}
44
+ strategy: fsdp
45
+ ppo_mini_batch_size: 256
46
+ ppo_micro_batch_size: null
47
+ ppo_micro_batch_size_per_gpu: 32
48
+ use_dynamic_bsz: false
49
+ ppo_max_token_len_per_gpu: 16384
50
+ clip_ratio: 0.2
51
+ clip_ratio_low: 0.2
52
+ clip_ratio_high: 0.2
53
+ freeze_vision_tower: false
54
+ policy_loss:
55
+ _target_: verl.workers.config.PolicyLossConfig
56
+ loss_mode: vanilla
57
+ clip_cov_ratio: 0.0002
58
+ clip_cov_lb: 1.0
59
+ clip_cov_ub: 5.0
60
+ kl_cov_ratio: 0.0002
61
+ ppo_kl_coef: 0.1
62
+ clip_ratio_c: 3.0
63
+ loss_agg_mode: token-mean
64
+ loss_scale_factor: null
65
+ entropy_coeff: 0
66
+ calculate_entropy: false
67
+ use_kl_loss: true
68
+ use_torch_compile: true
69
+ kl_loss_coef: 0.001
70
+ kl_loss_type: low_var_kl
71
+ ppo_epochs: 1
72
+ shuffle: false
73
+ checkpoint:
74
+ _target_: verl.trainer.config.CheckpointConfig
75
+ save_contents:
76
+ - model
77
+ - optimizer
78
+ - extra
79
+ load_contents: ${.save_contents}
80
+ async_save: false
81
+ use_fused_kernels: ${oc.select:actor_rollout_ref.model.use_fused_kernels,false}
82
+ profiler:
83
+ _target_: verl.utils.profiler.ProfilerConfig
84
+ tool: ${oc.select:global_profiler.tool,null}
85
+ enable: false
86
+ all_ranks: false
87
+ ranks: []
88
+ save_path: ${oc.select:global_profiler.save_path,null}
89
+ tool_config:
90
+ nsys:
91
+ _target_: verl.utils.profiler.config.NsightToolConfig
92
+ discrete: ${oc.select:global_profiler.global_tool_config.nsys.discrete}
93
+ npu:
94
+ _target_: verl.utils.profiler.config.NPUToolConfig
95
+ contents: []
96
+ level: level1
97
+ analysis: true
98
+ discrete: false
99
+ torch:
100
+ _target_: verl.utils.profiler.config.TorchProfilerToolConfig
101
+ step_start: 0
102
+ step_end: null
103
+ torch_memory:
104
+ _target_: verl.utils.profiler.config.TorchMemoryToolConfig
105
+ trace_alloc_max_entries: ${oc.select:global_profiler.global_tool_config.torch_memory.trace_alloc_max_entries,100000}
106
+ stack_depth: ${oc.select:global_profiler.global_tool_config.torch_memory.stack_depth,32}
107
+ router_replay:
108
+ _target_: verl.workers.config.RouterReplayConfig
109
+ mode: disabled
110
+ record_file: null
111
+ replay_file: null
112
+ grad_clip: 1.0
113
+ ulysses_sequence_parallel_size: 1
114
+ entropy_from_logits_with_chunking: false
115
+ entropy_checkpointing: false
116
+ use_remove_padding: ${oc.select:actor_rollout_ref.model.use_remove_padding,false}
117
+ ref:
118
+ rollout_n: ${oc.select:actor_rollout_ref.rollout.n,1}
119
+ strategy: ${actor_rollout_ref.actor.strategy}
120
+ use_torch_compile: ${oc.select:actor_rollout_ref.actor.use_torch_compile,true}
121
+ log_prob_micro_batch_size: null
122
+ log_prob_micro_batch_size_per_gpu: 32
123
+ log_prob_use_dynamic_bsz: ${oc.select:actor_rollout_ref.actor.use_dynamic_bsz,false}
124
+ log_prob_max_token_len_per_gpu: ${oc.select:actor_rollout_ref.actor.ppo_max_token_len_per_gpu,16384}
125
+ profiler:
126
+ _target_: verl.utils.profiler.ProfilerConfig
127
+ tool: ${oc.select:global_profiler.tool,null}
128
+ enable: false
129
+ all_ranks: false
130
+ ranks: []
131
+ save_path: ${oc.select:global_profiler.save_path,null}
132
+ tool_config:
133
+ nsys:
134
+ _target_: verl.utils.profiler.config.NsightToolConfig
135
+ discrete: ${oc.select:global_profiler.global_tool_config.nsys.discrete}
136
+ npu:
137
+ _target_: verl.utils.profiler.config.NPUToolConfig
138
+ contents: []
139
+ level: level1
140
+ analysis: true
141
+ discrete: false
142
+ torch:
143
+ _target_: verl.utils.profiler.config.TorchProfilerToolConfig
144
+ step_start: 0
145
+ step_end: null
146
+ torch_memory:
147
+ _target_: verl.utils.profiler.config.TorchMemoryToolConfig
148
+ trace_alloc_max_entries: ${oc.select:global_profiler.global_tool_config.torch_memory.trace_alloc_max_entries,100000}
149
+ stack_depth: ${oc.select:global_profiler.global_tool_config.torch_memory.stack_depth,32}
150
+ router_replay:
151
+ _target_: verl.workers.config.RouterReplayConfig
152
+ mode: disabled
153
+ record_file: null
154
+ replay_file: null
155
+ fsdp_config:
156
+ _target_: verl.workers.config.FSDPEngineConfig
157
+ wrap_policy:
158
+ min_num_params: 0
159
+ param_offload: false
160
+ optimizer_offload: false
161
+ offload_policy: false
162
+ reshard_after_forward: true
163
+ fsdp_size: -1
164
+ forward_prefetch: false
165
+ model_dtype: fp32
166
+ use_orig_params: false
167
+ seed: 42
168
+ full_determinism: false
169
+ ulysses_sequence_parallel_size: 1
170
+ entropy_from_logits_with_chunking: false
171
+ use_torch_compile: true
172
+ entropy_checkpointing: false
173
+ forward_only: true
174
+ strategy: fsdp
175
+ dtype: bfloat16
176
+ _target_: verl.workers.config.FSDPActorConfig
177
+ ulysses_sequence_parallel_size: ${oc.select:actor_rollout_ref.actor.ulysses_sequence_parallel_size,1}
178
+ entropy_from_logits_with_chunking: false
179
+ entropy_checkpointing: false
180
+ rollout:
181
+ _target_: verl.workers.config.RolloutConfig
182
+ name: vllm
183
+ mode: async
184
+ temperature: 1.0
185
+ top_k: -1
186
+ top_p: 1
187
+ prompt_length: 1536
188
+ response_length: 512
189
+ dtype: bfloat16
190
+ gpu_memory_utilization: 0.6
191
+ ignore_eos: false
192
+ enforce_eager: false
193
+ cudagraph_capture_sizes: null
194
+ free_cache_engine: true
195
+ tensor_model_parallel_size: 2
196
+ data_parallel_size: 1
197
+ expert_parallel_size: 1
198
+ pipeline_model_parallel_size: 1
199
+ max_num_batched_tokens: 8192
200
+ max_model_len: null
201
+ max_num_seqs: 1024
202
+ enable_chunked_prefill: true
203
+ enable_prefix_caching: true
204
+ load_format: dummy
205
+ log_prob_micro_batch_size: null
206
+ log_prob_micro_batch_size_per_gpu: 32
207
+ log_prob_use_dynamic_bsz: ${oc.select:actor_rollout_ref.actor.use_dynamic_bsz,false}
208
+ log_prob_max_token_len_per_gpu: ${oc.select:actor_rollout_ref.actor.ppo_max_token_len_per_gpu,16384}
209
+ disable_log_stats: true
210
+ do_sample: true
211
+ 'n': 1
212
+ over_sample_rate: 0
213
+ multi_stage_wake_up: false
214
+ engine_kwargs:
215
+ vllm: {}
216
+ sglang: {}
217
+ val_kwargs:
218
+ _target_: verl.workers.config.SamplingConfig
219
+ top_k: -1
220
+ top_p: 1.0
221
+ temperature: 0
222
+ 'n': 1
223
+ do_sample: false
224
+ multi_turn:
225
+ _target_: verl.workers.config.MultiTurnConfig
226
+ enable: false
227
+ max_assistant_turns: null
228
+ tool_config_path: null
229
+ max_user_turns: null
230
+ max_parallel_calls: 1
231
+ max_tool_response_length: 256
232
+ tool_response_truncate_side: middle
233
+ interaction_config_path: null
234
+ use_inference_chat_template: false
235
+ tokenization_sanity_check_mode: strict
236
+ format: hermes
237
+ num_repeat_rollouts: null
238
+ calculate_log_probs: false
239
+ agent:
240
+ _target_: verl.workers.config.AgentLoopConfig
241
+ num_workers: 8
242
+ default_agent_loop: single_turn_agent
243
+ agent_loop_config_path: null
244
+ custom_async_server:
245
+ _target_: verl.workers.config.CustomAsyncServerConfig
246
+ path: null
247
+ name: null
248
+ update_weights_bucket_megabytes: 512
249
+ trace:
250
+ _target_: verl.workers.config.TraceConfig
251
+ backend: null
252
+ token2text: false
253
+ max_samples_per_step_per_worker: null
254
+ skip_rollout: false
255
+ skip_dump_dir: /tmp/rollout_dump
256
+ skip_tokenizer_init: true
257
+ enable_rollout_routing_replay: false
258
+ profiler:
259
+ _target_: verl.utils.profiler.ProfilerConfig
260
+ tool: ${oc.select:global_profiler.tool,null}
261
+ enable: ${oc.select:actor_rollout_ref.actor.profiler.enable,false}
262
+ all_ranks: ${oc.select:actor_rollout_ref.actor.profiler.all_ranks,false}
263
+ ranks: ${oc.select:actor_rollout_ref.actor.profiler.ranks,[]}
264
+ save_path: ${oc.select:global_profiler.save_path,null}
265
+ tool_config: ${oc.select:actor_rollout_ref.actor.profiler.tool_config,null}
266
+ prometheus:
267
+ _target_: verl.workers.config.PrometheusConfig
268
+ enable: false
269
+ port: 9090
270
+ file: /tmp/ray/session_latest/metrics/prometheus/prometheus.yml
271
+ served_model_name: ${oc.select:actor_rollout_ref.model.path,null}
272
+ layered_summon: false
273
+ model:
274
+ _target_: verl.workers.config.HFModelConfig
275
+ path: /mnt/tidal-alsh01/dataset/redtrans/zhangruiqi/models/Qwen3-4B
276
+ hf_config_path: null
277
+ tokenizer_path: null
278
+ use_shm: false
279
+ trust_remote_code: false
280
+ custom_chat_template: null
281
+ external_lib: null
282
+ override_config: {}
283
+ enable_gradient_checkpointing: true
284
+ enable_activation_offload: false
285
+ use_remove_padding: true
286
+ lora_rank: 0
287
+ lora_alpha: 16
288
+ target_modules: all-linear
289
+ exclude_modules: null
290
+ lora_adapter_path: null
291
+ use_liger: false
292
+ use_fused_kernels: false
293
+ fused_kernel_options:
294
+ impl_backend: torch
295
+ hybrid_engine: true
296
+ nccl_timeout: 600
297
+ data:
298
+ tokenizer: null
299
+ use_shm: false
300
+ train_files: /mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/train.parquet
301
+ val_files: /mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/test.parquet
302
+ train_max_samples: -1
303
+ val_max_samples: -1
304
+ prompt_key: prompt
305
+ reward_fn_key: data_source
306
+ max_prompt_length: 512
307
+ max_response_length: 1024
308
+ train_batch_size: 1024
309
+ val_batch_size: null
310
+ tool_config_path: ${oc.select:actor_rollout_ref.rollout.multi_turn.tool_config_path,
311
+ null}
312
+ return_raw_input_ids: false
313
+ return_raw_chat: true
314
+ return_full_prompt: false
315
+ shuffle: true
316
+ seed: null
317
+ dataloader_num_workers: 8
318
+ image_patch_size: 14
319
+ validation_shuffle: false
320
+ filter_overlong_prompts: true
321
+ filter_overlong_prompts_workers: 1
322
+ truncation: error
323
+ image_key: images
324
+ video_key: videos
325
+ trust_remote_code: false
326
+ custom_cls:
327
+ path: null
328
+ name: null
329
+ return_multi_modal_inputs: true
330
+ sampler:
331
+ class_path: null
332
+ class_name: null
333
+ datagen:
334
+ path: null
335
+ name: null
336
+ apply_chat_template_kwargs: {}
337
+ reward_manager:
338
+ _target_: verl.trainer.config.config.RewardManagerConfig
339
+ source: register
340
+ name: ${oc.select:reward_model.reward_manager,naive}
341
+ module:
342
+ _target_: verl.trainer.config.config.ModuleConfig
343
+ path: null
344
+ name: custom_reward_manager
345
+ critic:
346
+ optim:
347
+ _target_: verl.workers.config.FSDPOptimizerConfig
348
+ optimizer: AdamW
349
+ optimizer_impl: torch.optim
350
+ lr: 1.0e-05
351
+ lr_warmup_steps_ratio: 0.0
352
+ total_training_steps: -1
353
+ weight_decay: 0.01
354
+ lr_warmup_steps: -1
355
+ betas:
356
+ - 0.9
357
+ - 0.999
358
+ clip_grad: 1.0
359
+ min_lr_ratio: 0.0
360
+ num_cycles: 0.5
361
+ lr_scheduler_type: constant
362
+ warmup_style: null
363
+ override_optimizer_config: null
364
+ model:
365
+ fsdp_config:
366
+ _target_: verl.workers.config.FSDPEngineConfig
367
+ wrap_policy:
368
+ min_num_params: 0
369
+ param_offload: false
370
+ optimizer_offload: false
371
+ offload_policy: false
372
+ reshard_after_forward: true
373
+ fsdp_size: -1
374
+ forward_prefetch: false
375
+ model_dtype: fp32
376
+ use_orig_params: false
377
+ seed: 42
378
+ full_determinism: false
379
+ ulysses_sequence_parallel_size: 1
380
+ entropy_from_logits_with_chunking: false
381
+ use_torch_compile: true
382
+ entropy_checkpointing: false
383
+ forward_only: false
384
+ strategy: fsdp
385
+ dtype: bfloat16
386
+ path: ~/models/deepseek-llm-7b-chat
387
+ tokenizer_path: ${oc.select:actor_rollout_ref.model.path,"~/models/deepseek-llm-7b-chat"}
388
+ override_config: {}
389
+ external_lib: ${oc.select:actor_rollout_ref.model.external_lib,null}
390
+ trust_remote_code: ${oc.select:actor_rollout_ref.model.trust_remote_code,false}
391
+ _target_: verl.workers.config.FSDPCriticModelCfg
392
+ use_shm: false
393
+ enable_gradient_checkpointing: true
394
+ enable_activation_offload: false
395
+ use_remove_padding: false
396
+ lora_rank: 0
397
+ lora_alpha: 16
398
+ target_modules: all-linear
399
+ _target_: verl.workers.config.FSDPCriticConfig
400
+ rollout_n: ${oc.select:actor_rollout_ref.rollout.n,1}
401
+ strategy: fsdp
402
+ enable: null
403
+ ppo_mini_batch_size: ${oc.select:actor_rollout_ref.actor.ppo_mini_batch_size,256}
404
+ ppo_micro_batch_size: null
405
+ ppo_micro_batch_size_per_gpu: ${oc.select:.ppo_micro_batch_size,null}
406
+ use_dynamic_bsz: ${oc.select:actor_rollout_ref.actor.use_dynamic_bsz,false}
407
+ ppo_max_token_len_per_gpu: 32768
408
+ forward_max_token_len_per_gpu: ${.ppo_max_token_len_per_gpu}
409
+ ppo_epochs: ${oc.select:actor_rollout_ref.actor.ppo_epochs,1}
410
+ shuffle: ${oc.select:actor_rollout_ref.actor.shuffle,false}
411
+ cliprange_value: 0.5
412
+ loss_agg_mode: ${oc.select:actor_rollout_ref.actor.loss_agg_mode,token-mean}
413
+ checkpoint:
414
+ _target_: verl.trainer.config.CheckpointConfig
415
+ save_contents:
416
+ - model
417
+ - optimizer
418
+ - extra
419
+ load_contents: ${.save_contents}
420
+ async_save: false
421
+ profiler:
422
+ _target_: verl.utils.profiler.ProfilerConfig
423
+ tool: ${oc.select:global_profiler.tool,null}
424
+ enable: false
425
+ all_ranks: false
426
+ ranks: []
427
+ save_path: ${oc.select:global_profiler.save_path,null}
428
+ tool_config:
429
+ nsys:
430
+ _target_: verl.utils.profiler.config.NsightToolConfig
431
+ discrete: ${oc.select:global_profiler.global_tool_config.nsys.discrete}
432
+ npu:
433
+ _target_: verl.utils.profiler.config.NPUToolConfig
434
+ contents: []
435
+ level: level1
436
+ analysis: true
437
+ discrete: false
438
+ torch:
439
+ _target_: verl.utils.profiler.config.TorchProfilerToolConfig
440
+ step_start: 0
441
+ step_end: null
442
+ torch_memory:
443
+ _target_: verl.utils.profiler.config.TorchMemoryToolConfig
444
+ trace_alloc_max_entries: ${oc.select:global_profiler.global_tool_config.torch_memory.trace_alloc_max_entries,100000}
445
+ stack_depth: ${oc.select:global_profiler.global_tool_config.torch_memory.stack_depth,32}
446
+ forward_micro_batch_size: ${oc.select:.ppo_micro_batch_size,null}
447
+ forward_micro_batch_size_per_gpu: ${oc.select:.ppo_micro_batch_size_per_gpu,null}
448
+ ulysses_sequence_parallel_size: 1
449
+ grad_clip: 1.0
450
+ reward_model:
451
+ enable: false
452
+ enable_resource_pool: false
453
+ n_gpus_per_node: 0
454
+ nnodes: 0
455
+ strategy: fsdp
456
+ model:
457
+ input_tokenizer: ${actor_rollout_ref.model.path}
458
+ path: ~/models/FsfairX-LLaMA3-RM-v0.1
459
+ external_lib: ${actor_rollout_ref.model.external_lib}
460
+ trust_remote_code: false
461
+ override_config: {}
462
+ use_shm: false
463
+ use_remove_padding: false
464
+ use_fused_kernels: ${actor_rollout_ref.model.use_fused_kernels}
465
+ fsdp_config:
466
+ _target_: verl.workers.config.FSDPEngineConfig
467
+ wrap_policy:
468
+ min_num_params: 0
469
+ param_offload: false
470
+ reshard_after_forward: true
471
+ fsdp_size: -1
472
+ forward_prefetch: false
473
+ micro_batch_size: null
474
+ micro_batch_size_per_gpu: null
475
+ max_length: null
476
+ use_dynamic_bsz: ${critic.use_dynamic_bsz}
477
+ forward_max_token_len_per_gpu: ${critic.forward_max_token_len_per_gpu}
478
+ reward_manager: naive
479
+ launch_reward_fn_async: false
480
+ sandbox_fusion:
481
+ url: null
482
+ max_concurrent: 64
483
+ memory_limit_mb: 1024
484
+ profiler:
485
+ _target_: verl.utils.profiler.ProfilerConfig
486
+ tool: ${oc.select:global_profiler.tool,null}
487
+ enable: false
488
+ all_ranks: false
489
+ ranks: []
490
+ save_path: ${oc.select:global_profiler.save_path,null}
491
+ tool_config: ${oc.select:actor_rollout_ref.actor.profiler.tool_config,null}
492
+ ulysses_sequence_parallel_size: 1
493
+ use_reward_loop: true
494
+ rollout:
495
+ _target_: verl.workers.config.RolloutConfig
496
+ name: ???
497
+ dtype: bfloat16
498
+ gpu_memory_utilization: 0.5
499
+ enforce_eager: true
500
+ cudagraph_capture_sizes: null
501
+ free_cache_engine: true
502
+ data_parallel_size: 1
503
+ expert_parallel_size: 1
504
+ tensor_model_parallel_size: 2
505
+ max_num_batched_tokens: 8192
506
+ max_model_len: null
507
+ max_num_seqs: 1024
508
+ load_format: auto
509
+ engine_kwargs: {}
510
+ limit_images: null
511
+ enable_chunked_prefill: true
512
+ enable_prefix_caching: true
513
+ disable_log_stats: true
514
+ skip_tokenizer_init: true
515
+ prompt_length: 512
516
+ response_length: 512
517
+ algorithm:
518
+ rollout_correction:
519
+ rollout_is: null
520
+ rollout_is_threshold: 2.0
521
+ rollout_rs: null
522
+ rollout_rs_threshold: null
523
+ rollout_rs_threshold_lower: null
524
+ rollout_token_veto_threshold: null
525
+ bypass_mode: false
526
+ use_policy_gradient: false
527
+ rollout_is_batch_normalize: false
528
+ _target_: verl.trainer.config.AlgoConfig
529
+ gamma: 1.0
530
+ lam: 1.0
531
+ adv_estimator: sgrpo
532
+ norm_adv_by_std_in_grpo: false
533
+ use_kl_in_reward: false
534
+ kl_penalty: kl
535
+ kl_ctrl:
536
+ _target_: verl.trainer.config.KLControlConfig
537
+ type: fixed
538
+ kl_coef: 0.001
539
+ horizon: 10000
540
+ target_kl: 0.1
541
+ use_pf_ppo: false
542
+ pf_ppo:
543
+ reweight_method: pow
544
+ weight_pow: 2.0
545
+ sgrpo:
546
+ num_truncations: 4
547
+ force_answer_prompt: Time is limited, stop thinking and start answering.\n</think>\n\n
548
+ answer_max_tokens: 256
549
+ min_truncation_ratio: 0.1
550
+ max_truncation_ratio: 0.9
551
+ custom_reward_function:
552
+ path: null
553
+ name: compute_score
554
+ trainer:
555
+ balance_batch: true
556
+ total_epochs: 15
557
+ total_training_steps: null
558
+ project_name: verl_sgrpo_example_gsm8k
559
+ experiment_name: qwen3_8b_sgrpo
560
+ logger:
561
+ - console
562
+ log_val_generations: 0
563
+ rollout_data_dir: ./outputs/sgrpo_samples
564
+ rollout_data_dump_freq: 20
565
+ validation_data_dir: null
566
+ nnodes: 1
567
+ n_gpus_per_node: 8
568
+ save_freq: 20
569
+ esi_redundant_time: 0
570
+ resume_mode: disable
571
+ resume_from_path: null
572
+ val_before_train: true
573
+ val_only: false
574
+ test_freq: 5
575
+ critic_warmup: 0
576
+ default_hdfs_dir: null
577
+ del_local_ckpt_after_load: false
578
+ default_local_dir: checkpoints/${trainer.project_name}/${trainer.experiment_name}
579
+ max_actor_ckpt_to_keep: null
580
+ max_critic_ckpt_to_keep: null
581
+ ray_wait_register_center_timeout: 300
582
+ device: cuda
583
+ use_legacy_worker_impl: auto
584
+ global_profiler:
585
+ _target_: verl.utils.profiler.ProfilerConfig
586
+ tool: null
587
+ steps: null
588
+ profile_continuous_steps: false
589
+ save_path: outputs/profile
590
+ global_tool_config:
591
+ nsys:
592
+ _target_: verl.utils.profiler.config.NsightToolConfig
593
+ discrete: false
594
+ controller_nsight_options:
595
+ trace: cuda,nvtx,cublas,ucx
596
+ cuda-memory-usage: 'true'
597
+ cuda-graph-trace: graph
598
+ worker_nsight_options:
599
+ trace: cuda,nvtx,cublas,ucx
600
+ cuda-memory-usage: 'true'
601
+ cuda-graph-trace: graph
602
+ capture-range: cudaProfilerApi
603
+ capture-range-end: null
604
+ kill: none
605
+ torch_memory:
606
+ trace_alloc_max_entries: 100000
607
+ stack_depth: 32
608
+ context: all
609
+ stacks: all
610
+ kw_args: {}
611
+ transfer_queue:
612
+ enable: false
613
+ ray_kwargs:
614
+ ray_init:
615
+ num_cpus: null
616
+ timeline_json_file: null
outputs/2026-01-14/15-08-55/.hydra/overrides.yaml ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ - algorithm.adv_estimator=sgrpo
2
+ - algorithm.norm_adv_by_std_in_grpo=False
3
+ - algorithm.use_kl_in_reward=False
4
+ - data.train_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/train.parquet
5
+ - data.val_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/test.parquet
6
+ - data.train_batch_size=1024
7
+ - data.max_prompt_length=512
8
+ - data.max_response_length=1024
9
+ - data.filter_overlong_prompts=True
10
+ - data.truncation=error
11
+ - data.return_raw_chat=True
12
+ - actor_rollout_ref.model.path=/mnt/tidal-alsh01/dataset/redtrans/zhangruiqi/models/Qwen3-4B
13
+ - actor_rollout_ref.actor.optim.lr=1e-6
14
+ - actor_rollout_ref.model.use_remove_padding=True
15
+ - actor_rollout_ref.actor.ppo_mini_batch_size=256
16
+ - actor_rollout_ref.actor.ppo_micro_batch_size_per_gpu=32
17
+ - actor_rollout_ref.actor.use_kl_loss=True
18
+ - actor_rollout_ref.actor.kl_loss_coef=0.001
19
+ - actor_rollout_ref.actor.kl_loss_type=low_var_kl
20
+ - actor_rollout_ref.actor.entropy_coeff=0
21
+ - actor_rollout_ref.model.enable_gradient_checkpointing=True
22
+ - actor_rollout_ref.actor.fsdp_config.param_offload=False
23
+ - actor_rollout_ref.actor.fsdp_config.optimizer_offload=False
24
+ - actor_rollout_ref.rollout.log_prob_micro_batch_size_per_gpu=32
25
+ - actor_rollout_ref.rollout.tensor_model_parallel_size=2
26
+ - actor_rollout_ref.rollout.name=vllm
27
+ - actor_rollout_ref.rollout.gpu_memory_utilization=0.6
28
+ - actor_rollout_ref.rollout.n=1
29
+ - actor_rollout_ref.rollout.prompt_length=1536
30
+ - actor_rollout_ref.rollout.response_length=512
31
+ - actor_rollout_ref.ref.log_prob_micro_batch_size_per_gpu=32
32
+ - actor_rollout_ref.ref.fsdp_config.param_offload=False
33
+ - sgrpo.num_truncations=4
34
+ - sgrpo.force_answer_prompt="Time is limited, stop thinking and start answering.\n</think>\n\n"
35
+ - sgrpo.answer_max_tokens=256
36
+ - trainer.critic_warmup=0
37
+ - trainer.logger=["console"]
38
+ - trainer.project_name=verl_sgrpo_example_gsm8k
39
+ - trainer.experiment_name=qwen3_8b_sgrpo
40
+ - trainer.resume_mode=disable
41
+ - trainer.n_gpus_per_node=8
42
+ - trainer.nnodes=1
43
+ - trainer.save_freq=20
44
+ - trainer.test_freq=5
45
+ - trainer.total_epochs=15
46
+ - trainer.rollout_data_dir=./outputs/sgrpo_samples
47
+ - trainer.rollout_data_dump_freq=20
outputs/2026-01-14/15-16-37/.hydra/config.yaml ADDED
@@ -0,0 +1,616 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ actor_rollout_ref:
2
+ actor:
3
+ optim:
4
+ _target_: verl.workers.config.FSDPOptimizerConfig
5
+ optimizer: AdamW
6
+ optimizer_impl: torch.optim
7
+ lr: 1.0e-06
8
+ lr_warmup_steps_ratio: 0.0
9
+ total_training_steps: -1
10
+ weight_decay: 0.01
11
+ lr_warmup_steps: -1
12
+ betas:
13
+ - 0.9
14
+ - 0.999
15
+ clip_grad: 1.0
16
+ min_lr_ratio: 0.0
17
+ num_cycles: 0.5
18
+ lr_scheduler_type: constant
19
+ warmup_style: null
20
+ override_optimizer_config: null
21
+ fsdp_config:
22
+ _target_: verl.workers.config.FSDPEngineConfig
23
+ wrap_policy:
24
+ min_num_params: 0
25
+ param_offload: false
26
+ optimizer_offload: false
27
+ offload_policy: false
28
+ reshard_after_forward: true
29
+ fsdp_size: -1
30
+ forward_prefetch: false
31
+ model_dtype: fp32
32
+ use_orig_params: false
33
+ seed: 42
34
+ full_determinism: false
35
+ ulysses_sequence_parallel_size: 1
36
+ entropy_from_logits_with_chunking: false
37
+ use_torch_compile: true
38
+ entropy_checkpointing: false
39
+ forward_only: false
40
+ strategy: fsdp
41
+ dtype: bfloat16
42
+ _target_: verl.workers.config.FSDPActorConfig
43
+ rollout_n: ${oc.select:actor_rollout_ref.rollout.n,1}
44
+ strategy: fsdp
45
+ ppo_mini_batch_size: 256
46
+ ppo_micro_batch_size: null
47
+ ppo_micro_batch_size_per_gpu: 32
48
+ use_dynamic_bsz: false
49
+ ppo_max_token_len_per_gpu: 16384
50
+ clip_ratio: 0.2
51
+ clip_ratio_low: 0.2
52
+ clip_ratio_high: 0.2
53
+ freeze_vision_tower: false
54
+ policy_loss:
55
+ _target_: verl.workers.config.PolicyLossConfig
56
+ loss_mode: vanilla
57
+ clip_cov_ratio: 0.0002
58
+ clip_cov_lb: 1.0
59
+ clip_cov_ub: 5.0
60
+ kl_cov_ratio: 0.0002
61
+ ppo_kl_coef: 0.1
62
+ clip_ratio_c: 3.0
63
+ loss_agg_mode: token-mean
64
+ loss_scale_factor: null
65
+ entropy_coeff: 0
66
+ calculate_entropy: false
67
+ use_kl_loss: true
68
+ use_torch_compile: true
69
+ kl_loss_coef: 0.001
70
+ kl_loss_type: low_var_kl
71
+ ppo_epochs: 1
72
+ shuffle: false
73
+ checkpoint:
74
+ _target_: verl.trainer.config.CheckpointConfig
75
+ save_contents:
76
+ - model
77
+ - optimizer
78
+ - extra
79
+ load_contents: ${.save_contents}
80
+ async_save: false
81
+ use_fused_kernels: ${oc.select:actor_rollout_ref.model.use_fused_kernels,false}
82
+ profiler:
83
+ _target_: verl.utils.profiler.ProfilerConfig
84
+ tool: ${oc.select:global_profiler.tool,null}
85
+ enable: false
86
+ all_ranks: false
87
+ ranks: []
88
+ save_path: ${oc.select:global_profiler.save_path,null}
89
+ tool_config:
90
+ nsys:
91
+ _target_: verl.utils.profiler.config.NsightToolConfig
92
+ discrete: ${oc.select:global_profiler.global_tool_config.nsys.discrete}
93
+ npu:
94
+ _target_: verl.utils.profiler.config.NPUToolConfig
95
+ contents: []
96
+ level: level1
97
+ analysis: true
98
+ discrete: false
99
+ torch:
100
+ _target_: verl.utils.profiler.config.TorchProfilerToolConfig
101
+ step_start: 0
102
+ step_end: null
103
+ torch_memory:
104
+ _target_: verl.utils.profiler.config.TorchMemoryToolConfig
105
+ trace_alloc_max_entries: ${oc.select:global_profiler.global_tool_config.torch_memory.trace_alloc_max_entries,100000}
106
+ stack_depth: ${oc.select:global_profiler.global_tool_config.torch_memory.stack_depth,32}
107
+ router_replay:
108
+ _target_: verl.workers.config.RouterReplayConfig
109
+ mode: disabled
110
+ record_file: null
111
+ replay_file: null
112
+ grad_clip: 1.0
113
+ ulysses_sequence_parallel_size: 1
114
+ entropy_from_logits_with_chunking: false
115
+ entropy_checkpointing: false
116
+ use_remove_padding: ${oc.select:actor_rollout_ref.model.use_remove_padding,false}
117
+ ref:
118
+ rollout_n: ${oc.select:actor_rollout_ref.rollout.n,1}
119
+ strategy: ${actor_rollout_ref.actor.strategy}
120
+ use_torch_compile: ${oc.select:actor_rollout_ref.actor.use_torch_compile,true}
121
+ log_prob_micro_batch_size: null
122
+ log_prob_micro_batch_size_per_gpu: 32
123
+ log_prob_use_dynamic_bsz: ${oc.select:actor_rollout_ref.actor.use_dynamic_bsz,false}
124
+ log_prob_max_token_len_per_gpu: ${oc.select:actor_rollout_ref.actor.ppo_max_token_len_per_gpu,16384}
125
+ profiler:
126
+ _target_: verl.utils.profiler.ProfilerConfig
127
+ tool: ${oc.select:global_profiler.tool,null}
128
+ enable: false
129
+ all_ranks: false
130
+ ranks: []
131
+ save_path: ${oc.select:global_profiler.save_path,null}
132
+ tool_config:
133
+ nsys:
134
+ _target_: verl.utils.profiler.config.NsightToolConfig
135
+ discrete: ${oc.select:global_profiler.global_tool_config.nsys.discrete}
136
+ npu:
137
+ _target_: verl.utils.profiler.config.NPUToolConfig
138
+ contents: []
139
+ level: level1
140
+ analysis: true
141
+ discrete: false
142
+ torch:
143
+ _target_: verl.utils.profiler.config.TorchProfilerToolConfig
144
+ step_start: 0
145
+ step_end: null
146
+ torch_memory:
147
+ _target_: verl.utils.profiler.config.TorchMemoryToolConfig
148
+ trace_alloc_max_entries: ${oc.select:global_profiler.global_tool_config.torch_memory.trace_alloc_max_entries,100000}
149
+ stack_depth: ${oc.select:global_profiler.global_tool_config.torch_memory.stack_depth,32}
150
+ router_replay:
151
+ _target_: verl.workers.config.RouterReplayConfig
152
+ mode: disabled
153
+ record_file: null
154
+ replay_file: null
155
+ fsdp_config:
156
+ _target_: verl.workers.config.FSDPEngineConfig
157
+ wrap_policy:
158
+ min_num_params: 0
159
+ param_offload: false
160
+ optimizer_offload: false
161
+ offload_policy: false
162
+ reshard_after_forward: true
163
+ fsdp_size: -1
164
+ forward_prefetch: false
165
+ model_dtype: fp32
166
+ use_orig_params: false
167
+ seed: 42
168
+ full_determinism: false
169
+ ulysses_sequence_parallel_size: 1
170
+ entropy_from_logits_with_chunking: false
171
+ use_torch_compile: true
172
+ entropy_checkpointing: false
173
+ forward_only: true
174
+ strategy: fsdp
175
+ dtype: bfloat16
176
+ _target_: verl.workers.config.FSDPActorConfig
177
+ ulysses_sequence_parallel_size: ${oc.select:actor_rollout_ref.actor.ulysses_sequence_parallel_size,1}
178
+ entropy_from_logits_with_chunking: false
179
+ entropy_checkpointing: false
180
+ rollout:
181
+ _target_: verl.workers.config.RolloutConfig
182
+ name: vllm
183
+ mode: async
184
+ temperature: 1.0
185
+ top_k: -1
186
+ top_p: 1
187
+ prompt_length: 1536
188
+ response_length: 512
189
+ dtype: bfloat16
190
+ gpu_memory_utilization: 0.6
191
+ ignore_eos: false
192
+ enforce_eager: false
193
+ cudagraph_capture_sizes: null
194
+ free_cache_engine: true
195
+ tensor_model_parallel_size: 2
196
+ data_parallel_size: 1
197
+ expert_parallel_size: 1
198
+ pipeline_model_parallel_size: 1
199
+ max_num_batched_tokens: 8192
200
+ max_model_len: null
201
+ max_num_seqs: 1024
202
+ enable_chunked_prefill: true
203
+ enable_prefix_caching: true
204
+ load_format: dummy
205
+ log_prob_micro_batch_size: null
206
+ log_prob_micro_batch_size_per_gpu: 32
207
+ log_prob_use_dynamic_bsz: ${oc.select:actor_rollout_ref.actor.use_dynamic_bsz,false}
208
+ log_prob_max_token_len_per_gpu: ${oc.select:actor_rollout_ref.actor.ppo_max_token_len_per_gpu,16384}
209
+ disable_log_stats: true
210
+ do_sample: true
211
+ 'n': 1
212
+ over_sample_rate: 0
213
+ multi_stage_wake_up: false
214
+ engine_kwargs:
215
+ vllm: {}
216
+ sglang: {}
217
+ val_kwargs:
218
+ _target_: verl.workers.config.SamplingConfig
219
+ top_k: -1
220
+ top_p: 1.0
221
+ temperature: 0
222
+ 'n': 1
223
+ do_sample: false
224
+ multi_turn:
225
+ _target_: verl.workers.config.MultiTurnConfig
226
+ enable: false
227
+ max_assistant_turns: null
228
+ tool_config_path: null
229
+ max_user_turns: null
230
+ max_parallel_calls: 1
231
+ max_tool_response_length: 256
232
+ tool_response_truncate_side: middle
233
+ interaction_config_path: null
234
+ use_inference_chat_template: false
235
+ tokenization_sanity_check_mode: strict
236
+ format: hermes
237
+ num_repeat_rollouts: null
238
+ calculate_log_probs: false
239
+ agent:
240
+ _target_: verl.workers.config.AgentLoopConfig
241
+ num_workers: 8
242
+ default_agent_loop: single_turn_agent
243
+ agent_loop_config_path: null
244
+ custom_async_server:
245
+ _target_: verl.workers.config.CustomAsyncServerConfig
246
+ path: null
247
+ name: null
248
+ update_weights_bucket_megabytes: 512
249
+ trace:
250
+ _target_: verl.workers.config.TraceConfig
251
+ backend: null
252
+ token2text: false
253
+ max_samples_per_step_per_worker: null
254
+ skip_rollout: false
255
+ skip_dump_dir: /tmp/rollout_dump
256
+ skip_tokenizer_init: true
257
+ enable_rollout_routing_replay: false
258
+ profiler:
259
+ _target_: verl.utils.profiler.ProfilerConfig
260
+ tool: ${oc.select:global_profiler.tool,null}
261
+ enable: ${oc.select:actor_rollout_ref.actor.profiler.enable,false}
262
+ all_ranks: ${oc.select:actor_rollout_ref.actor.profiler.all_ranks,false}
263
+ ranks: ${oc.select:actor_rollout_ref.actor.profiler.ranks,[]}
264
+ save_path: ${oc.select:global_profiler.save_path,null}
265
+ tool_config: ${oc.select:actor_rollout_ref.actor.profiler.tool_config,null}
266
+ prometheus:
267
+ _target_: verl.workers.config.PrometheusConfig
268
+ enable: false
269
+ port: 9090
270
+ file: /tmp/ray/session_latest/metrics/prometheus/prometheus.yml
271
+ served_model_name: ${oc.select:actor_rollout_ref.model.path,null}
272
+ layered_summon: false
273
+ model:
274
+ _target_: verl.workers.config.HFModelConfig
275
+ path: /mnt/tidal-alsh01/dataset/redtrans/zhangruiqi/models/Qwen3-4B
276
+ hf_config_path: null
277
+ tokenizer_path: null
278
+ use_shm: false
279
+ trust_remote_code: false
280
+ custom_chat_template: null
281
+ external_lib: null
282
+ override_config: {}
283
+ enable_gradient_checkpointing: true
284
+ enable_activation_offload: false
285
+ use_remove_padding: true
286
+ lora_rank: 0
287
+ lora_alpha: 16
288
+ target_modules: all-linear
289
+ exclude_modules: null
290
+ lora_adapter_path: null
291
+ use_liger: false
292
+ use_fused_kernels: false
293
+ fused_kernel_options:
294
+ impl_backend: torch
295
+ hybrid_engine: true
296
+ nccl_timeout: 600
297
+ data:
298
+ tokenizer: null
299
+ use_shm: false
300
+ train_files: /mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/train.parquet
301
+ val_files: /mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/test.parquet
302
+ train_max_samples: -1
303
+ val_max_samples: -1
304
+ prompt_key: prompt
305
+ reward_fn_key: data_source
306
+ max_prompt_length: 512
307
+ max_response_length: 1024
308
+ train_batch_size: 1024
309
+ val_batch_size: null
310
+ tool_config_path: ${oc.select:actor_rollout_ref.rollout.multi_turn.tool_config_path,
311
+ null}
312
+ return_raw_input_ids: false
313
+ return_raw_chat: true
314
+ return_full_prompt: false
315
+ shuffle: true
316
+ seed: null
317
+ dataloader_num_workers: 8
318
+ image_patch_size: 14
319
+ validation_shuffle: false
320
+ filter_overlong_prompts: true
321
+ filter_overlong_prompts_workers: 1
322
+ truncation: error
323
+ image_key: images
324
+ video_key: videos
325
+ trust_remote_code: false
326
+ custom_cls:
327
+ path: null
328
+ name: null
329
+ return_multi_modal_inputs: true
330
+ sampler:
331
+ class_path: null
332
+ class_name: null
333
+ datagen:
334
+ path: null
335
+ name: null
336
+ apply_chat_template_kwargs: {}
337
+ reward_manager:
338
+ _target_: verl.trainer.config.config.RewardManagerConfig
339
+ source: register
340
+ name: ${oc.select:reward_model.reward_manager,naive}
341
+ module:
342
+ _target_: verl.trainer.config.config.ModuleConfig
343
+ path: null
344
+ name: custom_reward_manager
345
+ critic:
346
+ optim:
347
+ _target_: verl.workers.config.FSDPOptimizerConfig
348
+ optimizer: AdamW
349
+ optimizer_impl: torch.optim
350
+ lr: 1.0e-05
351
+ lr_warmup_steps_ratio: 0.0
352
+ total_training_steps: -1
353
+ weight_decay: 0.01
354
+ lr_warmup_steps: -1
355
+ betas:
356
+ - 0.9
357
+ - 0.999
358
+ clip_grad: 1.0
359
+ min_lr_ratio: 0.0
360
+ num_cycles: 0.5
361
+ lr_scheduler_type: constant
362
+ warmup_style: null
363
+ override_optimizer_config: null
364
+ model:
365
+ fsdp_config:
366
+ _target_: verl.workers.config.FSDPEngineConfig
367
+ wrap_policy:
368
+ min_num_params: 0
369
+ param_offload: false
370
+ optimizer_offload: false
371
+ offload_policy: false
372
+ reshard_after_forward: true
373
+ fsdp_size: -1
374
+ forward_prefetch: false
375
+ model_dtype: fp32
376
+ use_orig_params: false
377
+ seed: 42
378
+ full_determinism: false
379
+ ulysses_sequence_parallel_size: 1
380
+ entropy_from_logits_with_chunking: false
381
+ use_torch_compile: true
382
+ entropy_checkpointing: false
383
+ forward_only: false
384
+ strategy: fsdp
385
+ dtype: bfloat16
386
+ path: ~/models/deepseek-llm-7b-chat
387
+ tokenizer_path: ${oc.select:actor_rollout_ref.model.path,"~/models/deepseek-llm-7b-chat"}
388
+ override_config: {}
389
+ external_lib: ${oc.select:actor_rollout_ref.model.external_lib,null}
390
+ trust_remote_code: ${oc.select:actor_rollout_ref.model.trust_remote_code,false}
391
+ _target_: verl.workers.config.FSDPCriticModelCfg
392
+ use_shm: false
393
+ enable_gradient_checkpointing: true
394
+ enable_activation_offload: false
395
+ use_remove_padding: false
396
+ lora_rank: 0
397
+ lora_alpha: 16
398
+ target_modules: all-linear
399
+ _target_: verl.workers.config.FSDPCriticConfig
400
+ rollout_n: ${oc.select:actor_rollout_ref.rollout.n,1}
401
+ strategy: fsdp
402
+ enable: null
403
+ ppo_mini_batch_size: ${oc.select:actor_rollout_ref.actor.ppo_mini_batch_size,256}
404
+ ppo_micro_batch_size: null
405
+ ppo_micro_batch_size_per_gpu: ${oc.select:.ppo_micro_batch_size,null}
406
+ use_dynamic_bsz: ${oc.select:actor_rollout_ref.actor.use_dynamic_bsz,false}
407
+ ppo_max_token_len_per_gpu: 32768
408
+ forward_max_token_len_per_gpu: ${.ppo_max_token_len_per_gpu}
409
+ ppo_epochs: ${oc.select:actor_rollout_ref.actor.ppo_epochs,1}
410
+ shuffle: ${oc.select:actor_rollout_ref.actor.shuffle,false}
411
+ cliprange_value: 0.5
412
+ loss_agg_mode: ${oc.select:actor_rollout_ref.actor.loss_agg_mode,token-mean}
413
+ checkpoint:
414
+ _target_: verl.trainer.config.CheckpointConfig
415
+ save_contents:
416
+ - model
417
+ - optimizer
418
+ - extra
419
+ load_contents: ${.save_contents}
420
+ async_save: false
421
+ profiler:
422
+ _target_: verl.utils.profiler.ProfilerConfig
423
+ tool: ${oc.select:global_profiler.tool,null}
424
+ enable: false
425
+ all_ranks: false
426
+ ranks: []
427
+ save_path: ${oc.select:global_profiler.save_path,null}
428
+ tool_config:
429
+ nsys:
430
+ _target_: verl.utils.profiler.config.NsightToolConfig
431
+ discrete: ${oc.select:global_profiler.global_tool_config.nsys.discrete}
432
+ npu:
433
+ _target_: verl.utils.profiler.config.NPUToolConfig
434
+ contents: []
435
+ level: level1
436
+ analysis: true
437
+ discrete: false
438
+ torch:
439
+ _target_: verl.utils.profiler.config.TorchProfilerToolConfig
440
+ step_start: 0
441
+ step_end: null
442
+ torch_memory:
443
+ _target_: verl.utils.profiler.config.TorchMemoryToolConfig
444
+ trace_alloc_max_entries: ${oc.select:global_profiler.global_tool_config.torch_memory.trace_alloc_max_entries,100000}
445
+ stack_depth: ${oc.select:global_profiler.global_tool_config.torch_memory.stack_depth,32}
446
+ forward_micro_batch_size: ${oc.select:.ppo_micro_batch_size,null}
447
+ forward_micro_batch_size_per_gpu: ${oc.select:.ppo_micro_batch_size_per_gpu,null}
448
+ ulysses_sequence_parallel_size: 1
449
+ grad_clip: 1.0
450
+ reward_model:
451
+ enable: false
452
+ enable_resource_pool: false
453
+ n_gpus_per_node: 0
454
+ nnodes: 0
455
+ strategy: fsdp
456
+ model:
457
+ input_tokenizer: ${actor_rollout_ref.model.path}
458
+ path: ~/models/FsfairX-LLaMA3-RM-v0.1
459
+ external_lib: ${actor_rollout_ref.model.external_lib}
460
+ trust_remote_code: false
461
+ override_config: {}
462
+ use_shm: false
463
+ use_remove_padding: false
464
+ use_fused_kernels: ${actor_rollout_ref.model.use_fused_kernels}
465
+ fsdp_config:
466
+ _target_: verl.workers.config.FSDPEngineConfig
467
+ wrap_policy:
468
+ min_num_params: 0
469
+ param_offload: false
470
+ reshard_after_forward: true
471
+ fsdp_size: -1
472
+ forward_prefetch: false
473
+ micro_batch_size: null
474
+ micro_batch_size_per_gpu: null
475
+ max_length: null
476
+ use_dynamic_bsz: ${critic.use_dynamic_bsz}
477
+ forward_max_token_len_per_gpu: ${critic.forward_max_token_len_per_gpu}
478
+ reward_manager: naive
479
+ launch_reward_fn_async: false
480
+ sandbox_fusion:
481
+ url: null
482
+ max_concurrent: 64
483
+ memory_limit_mb: 1024
484
+ profiler:
485
+ _target_: verl.utils.profiler.ProfilerConfig
486
+ tool: ${oc.select:global_profiler.tool,null}
487
+ enable: false
488
+ all_ranks: false
489
+ ranks: []
490
+ save_path: ${oc.select:global_profiler.save_path,null}
491
+ tool_config: ${oc.select:actor_rollout_ref.actor.profiler.tool_config,null}
492
+ ulysses_sequence_parallel_size: 1
493
+ use_reward_loop: true
494
+ rollout:
495
+ _target_: verl.workers.config.RolloutConfig
496
+ name: ???
497
+ dtype: bfloat16
498
+ gpu_memory_utilization: 0.5
499
+ enforce_eager: true
500
+ cudagraph_capture_sizes: null
501
+ free_cache_engine: true
502
+ data_parallel_size: 1
503
+ expert_parallel_size: 1
504
+ tensor_model_parallel_size: 2
505
+ max_num_batched_tokens: 8192
506
+ max_model_len: null
507
+ max_num_seqs: 1024
508
+ load_format: auto
509
+ engine_kwargs: {}
510
+ limit_images: null
511
+ enable_chunked_prefill: true
512
+ enable_prefix_caching: true
513
+ disable_log_stats: true
514
+ skip_tokenizer_init: true
515
+ prompt_length: 512
516
+ response_length: 512
517
+ algorithm:
518
+ rollout_correction:
519
+ rollout_is: null
520
+ rollout_is_threshold: 2.0
521
+ rollout_rs: null
522
+ rollout_rs_threshold: null
523
+ rollout_rs_threshold_lower: null
524
+ rollout_token_veto_threshold: null
525
+ bypass_mode: false
526
+ use_policy_gradient: false
527
+ rollout_is_batch_normalize: false
528
+ _target_: verl.trainer.config.AlgoConfig
529
+ gamma: 1.0
530
+ lam: 1.0
531
+ adv_estimator: sgrpo
532
+ norm_adv_by_std_in_grpo: false
533
+ use_kl_in_reward: false
534
+ kl_penalty: kl
535
+ kl_ctrl:
536
+ _target_: verl.trainer.config.KLControlConfig
537
+ type: fixed
538
+ kl_coef: 0.001
539
+ horizon: 10000
540
+ target_kl: 0.1
541
+ use_pf_ppo: false
542
+ pf_ppo:
543
+ reweight_method: pow
544
+ weight_pow: 2.0
545
+ sgrpo:
546
+ num_truncations: 4
547
+ force_answer_prompt: Time is limited, stop thinking and start answering.\n</think>\n\n
548
+ answer_max_tokens: 256
549
+ min_truncation_ratio: 0.1
550
+ max_truncation_ratio: 0.9
551
+ custom_reward_function:
552
+ path: null
553
+ name: compute_score
554
+ trainer:
555
+ balance_batch: true
556
+ total_epochs: 15
557
+ total_training_steps: null
558
+ project_name: verl_sgrpo_example_gsm8k
559
+ experiment_name: qwen3_8b_sgrpo
560
+ logger:
561
+ - console
562
+ log_val_generations: 0
563
+ rollout_data_dir: ./outputs/sgrpo_samples
564
+ rollout_data_dump_freq: 20
565
+ validation_data_dir: null
566
+ nnodes: 1
567
+ n_gpus_per_node: 8
568
+ save_freq: 20
569
+ esi_redundant_time: 0
570
+ resume_mode: disable
571
+ resume_from_path: null
572
+ val_before_train: true
573
+ val_only: false
574
+ test_freq: 5
575
+ critic_warmup: 0
576
+ default_hdfs_dir: null
577
+ del_local_ckpt_after_load: false
578
+ default_local_dir: checkpoints/${trainer.project_name}/${trainer.experiment_name}
579
+ max_actor_ckpt_to_keep: null
580
+ max_critic_ckpt_to_keep: null
581
+ ray_wait_register_center_timeout: 300
582
+ device: cuda
583
+ use_legacy_worker_impl: auto
584
+ global_profiler:
585
+ _target_: verl.utils.profiler.ProfilerConfig
586
+ tool: null
587
+ steps: null
588
+ profile_continuous_steps: false
589
+ save_path: outputs/profile
590
+ global_tool_config:
591
+ nsys:
592
+ _target_: verl.utils.profiler.config.NsightToolConfig
593
+ discrete: false
594
+ controller_nsight_options:
595
+ trace: cuda,nvtx,cublas,ucx
596
+ cuda-memory-usage: 'true'
597
+ cuda-graph-trace: graph
598
+ worker_nsight_options:
599
+ trace: cuda,nvtx,cublas,ucx
600
+ cuda-memory-usage: 'true'
601
+ cuda-graph-trace: graph
602
+ capture-range: cudaProfilerApi
603
+ capture-range-end: null
604
+ kill: none
605
+ torch_memory:
606
+ trace_alloc_max_entries: 100000
607
+ stack_depth: 32
608
+ context: all
609
+ stacks: all
610
+ kw_args: {}
611
+ transfer_queue:
612
+ enable: false
613
+ ray_kwargs:
614
+ ray_init:
615
+ num_cpus: null
616
+ timeline_json_file: null
outputs/2026-01-14/15-16-37/.hydra/hydra.yaml ADDED
@@ -0,0 +1,218 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
4
+ sweep:
5
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.mode=RUN
114
+ task:
115
+ - algorithm.adv_estimator=sgrpo
116
+ - algorithm.norm_adv_by_std_in_grpo=False
117
+ - algorithm.use_kl_in_reward=False
118
+ - data.train_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/train.parquet
119
+ - data.val_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/test.parquet
120
+ - data.train_batch_size=1024
121
+ - data.max_prompt_length=512
122
+ - data.max_response_length=1024
123
+ - data.filter_overlong_prompts=True
124
+ - data.truncation=error
125
+ - data.return_raw_chat=True
126
+ - actor_rollout_ref.model.path=/mnt/tidal-alsh01/dataset/redtrans/zhangruiqi/models/Qwen3-4B
127
+ - actor_rollout_ref.actor.optim.lr=1e-6
128
+ - actor_rollout_ref.model.use_remove_padding=True
129
+ - actor_rollout_ref.actor.ppo_mini_batch_size=256
130
+ - actor_rollout_ref.actor.ppo_micro_batch_size_per_gpu=32
131
+ - actor_rollout_ref.actor.use_kl_loss=True
132
+ - actor_rollout_ref.actor.kl_loss_coef=0.001
133
+ - actor_rollout_ref.actor.kl_loss_type=low_var_kl
134
+ - actor_rollout_ref.actor.entropy_coeff=0
135
+ - actor_rollout_ref.model.enable_gradient_checkpointing=True
136
+ - actor_rollout_ref.actor.fsdp_config.param_offload=False
137
+ - actor_rollout_ref.actor.fsdp_config.optimizer_offload=False
138
+ - actor_rollout_ref.rollout.log_prob_micro_batch_size_per_gpu=32
139
+ - actor_rollout_ref.rollout.tensor_model_parallel_size=2
140
+ - actor_rollout_ref.rollout.name=vllm
141
+ - actor_rollout_ref.rollout.gpu_memory_utilization=0.6
142
+ - actor_rollout_ref.rollout.n=1
143
+ - actor_rollout_ref.rollout.prompt_length=1536
144
+ - actor_rollout_ref.rollout.response_length=512
145
+ - actor_rollout_ref.ref.log_prob_micro_batch_size_per_gpu=32
146
+ - actor_rollout_ref.ref.fsdp_config.param_offload=False
147
+ - sgrpo.num_truncations=4
148
+ - sgrpo.force_answer_prompt="Time is limited, stop thinking and start answering.\n</think>\n\n"
149
+ - sgrpo.answer_max_tokens=256
150
+ - trainer.critic_warmup=0
151
+ - trainer.logger=["console"]
152
+ - trainer.project_name=verl_sgrpo_example_gsm8k
153
+ - trainer.experiment_name=qwen3_8b_sgrpo
154
+ - trainer.resume_mode=disable
155
+ - trainer.n_gpus_per_node=8
156
+ - trainer.nnodes=1
157
+ - trainer.save_freq=20
158
+ - trainer.test_freq=5
159
+ - trainer.total_epochs=15
160
+ - trainer.rollout_data_dir=./outputs/sgrpo_samples
161
+ - trainer.rollout_data_dump_freq=20
162
+ job:
163
+ name: main_sgrpo
164
+ chdir: null
165
+ override_dirname: actor_rollout_ref.actor.entropy_coeff=0,actor_rollout_ref.actor.fsdp_config.optimizer_offload=False,actor_rollout_ref.actor.fsdp_config.param_offload=False,actor_rollout_ref.actor.kl_loss_coef=0.001,actor_rollout_ref.actor.kl_loss_type=low_var_kl,actor_rollout_ref.actor.optim.lr=1e-6,actor_rollout_ref.actor.ppo_micro_batch_size_per_gpu=32,actor_rollout_ref.actor.ppo_mini_batch_size=256,actor_rollout_ref.actor.use_kl_loss=True,actor_rollout_ref.model.enable_gradient_checkpointing=True,actor_rollout_ref.model.path=/mnt/tidal-alsh01/dataset/redtrans/zhangruiqi/models/Qwen3-4B,actor_rollout_ref.model.use_remove_padding=True,actor_rollout_ref.ref.fsdp_config.param_offload=False,actor_rollout_ref.ref.log_prob_micro_batch_size_per_gpu=32,actor_rollout_ref.rollout.gpu_memory_utilization=0.6,actor_rollout_ref.rollout.log_prob_micro_batch_size_per_gpu=32,actor_rollout_ref.rollout.n=1,actor_rollout_ref.rollout.name=vllm,actor_rollout_ref.rollout.prompt_length=1536,actor_rollout_ref.rollout.response_length=512,actor_rollout_ref.rollout.tensor_model_parallel_size=2,algorithm.adv_estimator=sgrpo,algorithm.norm_adv_by_std_in_grpo=False,algorithm.use_kl_in_reward=False,data.filter_overlong_prompts=True,data.max_prompt_length=512,data.max_response_length=1024,data.return_raw_chat=True,data.train_batch_size=1024,data.train_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/train.parquet,data.truncation=error,data.val_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/test.parquet,sgrpo.answer_max_tokens=256,sgrpo.force_answer_prompt="Time
166
+ is limited, stop thinking and start answering.\n</think>\n\n",sgrpo.num_truncations=4,trainer.critic_warmup=0,trainer.experiment_name=qwen3_8b_sgrpo,trainer.logger=["console"],trainer.n_gpus_per_node=8,trainer.nnodes=1,trainer.project_name=verl_sgrpo_example_gsm8k,trainer.resume_mode=disable,trainer.rollout_data_dir=./outputs/sgrpo_samples,trainer.rollout_data_dump_freq=20,trainer.save_freq=20,trainer.test_freq=5,trainer.total_epochs=15
167
+ id: ???
168
+ num: ???
169
+ config_name: sgrpo_trainer
170
+ env_set: {}
171
+ env_copy: []
172
+ config:
173
+ override_dirname:
174
+ kv_sep: '='
175
+ item_sep: ','
176
+ exclude_keys: []
177
+ runtime:
178
+ version: 1.3.2
179
+ version_base: '1.3'
180
+ cwd: /mnt/tidal-alsh01/usr/zhangruiqi1/my/verl
181
+ config_sources:
182
+ - path: hydra.conf
183
+ schema: pkg
184
+ provider: hydra
185
+ - path: /mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/config
186
+ schema: file
187
+ provider: main
188
+ - path: verl.trainer.config
189
+ schema: pkg
190
+ provider: verl-trainer-config
191
+ - path: ''
192
+ schema: structured
193
+ provider: schema
194
+ output_dir: /mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/outputs/2026-01-14/15-16-37
195
+ choices:
196
+ algorithm@algorithm.rollout_correction: rollout_correction
197
+ reward_model: dp_reward_loop
198
+ critic: dp_critic
199
+ critic/../engine@critic.model.fsdp_config: fsdp
200
+ critic/../optim@critic.optim: fsdp
201
+ model@actor_rollout_ref.model: hf_model
202
+ rollout@actor_rollout_ref.rollout: rollout
203
+ ref@actor_rollout_ref.ref: dp_ref
204
+ ref/../engine@actor_rollout_ref.ref.fsdp_config: fsdp
205
+ data: legacy_data
206
+ actor@actor_rollout_ref.actor: dp_actor
207
+ actor/../engine@actor_rollout_ref.actor.fsdp_config: fsdp
208
+ actor/../optim@actor_rollout_ref.actor.optim: fsdp
209
+ hydra/env: default
210
+ hydra/callbacks: null
211
+ hydra/job_logging: default
212
+ hydra/hydra_logging: default
213
+ hydra/hydra_help: default
214
+ hydra/help: default
215
+ hydra/sweeper: basic
216
+ hydra/launcher: basic
217
+ hydra/output: default
218
+ verbose: false
outputs/2026-01-14/15-31-37/.hydra/config.yaml ADDED
@@ -0,0 +1,617 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ actor_rollout_ref:
2
+ actor:
3
+ optim:
4
+ _target_: verl.workers.config.FSDPOptimizerConfig
5
+ optimizer: AdamW
6
+ optimizer_impl: torch.optim
7
+ lr: 1.0e-06
8
+ lr_warmup_steps_ratio: 0.0
9
+ total_training_steps: -1
10
+ weight_decay: 0.01
11
+ lr_warmup_steps: -1
12
+ betas:
13
+ - 0.9
14
+ - 0.999
15
+ clip_grad: 1.0
16
+ min_lr_ratio: 0.0
17
+ num_cycles: 0.5
18
+ lr_scheduler_type: constant
19
+ warmup_style: null
20
+ override_optimizer_config: null
21
+ fsdp_config:
22
+ _target_: verl.workers.config.FSDPEngineConfig
23
+ wrap_policy:
24
+ min_num_params: 0
25
+ param_offload: false
26
+ optimizer_offload: false
27
+ offload_policy: false
28
+ reshard_after_forward: true
29
+ fsdp_size: -1
30
+ forward_prefetch: false
31
+ model_dtype: fp32
32
+ use_orig_params: false
33
+ seed: 42
34
+ full_determinism: false
35
+ ulysses_sequence_parallel_size: 1
36
+ entropy_from_logits_with_chunking: false
37
+ use_torch_compile: true
38
+ entropy_checkpointing: false
39
+ forward_only: false
40
+ strategy: fsdp
41
+ dtype: bfloat16
42
+ _target_: verl.workers.config.FSDPActorConfig
43
+ rollout_n: ${oc.select:actor_rollout_ref.rollout.n,1}
44
+ strategy: fsdp
45
+ ppo_mini_batch_size: 256
46
+ ppo_micro_batch_size: null
47
+ ppo_micro_batch_size_per_gpu: 32
48
+ use_dynamic_bsz: false
49
+ ppo_max_token_len_per_gpu: 16384
50
+ clip_ratio: 0.2
51
+ clip_ratio_low: 0.2
52
+ clip_ratio_high: 0.2
53
+ freeze_vision_tower: false
54
+ policy_loss:
55
+ _target_: verl.workers.config.PolicyLossConfig
56
+ loss_mode: vanilla
57
+ clip_cov_ratio: 0.0002
58
+ clip_cov_lb: 1.0
59
+ clip_cov_ub: 5.0
60
+ kl_cov_ratio: 0.0002
61
+ ppo_kl_coef: 0.1
62
+ clip_ratio_c: 3.0
63
+ loss_agg_mode: token-mean
64
+ loss_scale_factor: null
65
+ entropy_coeff: 0
66
+ calculate_entropy: false
67
+ use_kl_loss: true
68
+ use_torch_compile: true
69
+ kl_loss_coef: 0.001
70
+ kl_loss_type: low_var_kl
71
+ ppo_epochs: 1
72
+ shuffle: false
73
+ checkpoint:
74
+ _target_: verl.trainer.config.CheckpointConfig
75
+ save_contents:
76
+ - model
77
+ - optimizer
78
+ - extra
79
+ load_contents: ${.save_contents}
80
+ async_save: false
81
+ use_fused_kernels: ${oc.select:actor_rollout_ref.model.use_fused_kernels,false}
82
+ profiler:
83
+ _target_: verl.utils.profiler.ProfilerConfig
84
+ tool: ${oc.select:global_profiler.tool,null}
85
+ enable: false
86
+ all_ranks: false
87
+ ranks: []
88
+ save_path: ${oc.select:global_profiler.save_path,null}
89
+ tool_config:
90
+ nsys:
91
+ _target_: verl.utils.profiler.config.NsightToolConfig
92
+ discrete: ${oc.select:global_profiler.global_tool_config.nsys.discrete}
93
+ npu:
94
+ _target_: verl.utils.profiler.config.NPUToolConfig
95
+ contents: []
96
+ level: level1
97
+ analysis: true
98
+ discrete: false
99
+ torch:
100
+ _target_: verl.utils.profiler.config.TorchProfilerToolConfig
101
+ step_start: 0
102
+ step_end: null
103
+ torch_memory:
104
+ _target_: verl.utils.profiler.config.TorchMemoryToolConfig
105
+ trace_alloc_max_entries: ${oc.select:global_profiler.global_tool_config.torch_memory.trace_alloc_max_entries,100000}
106
+ stack_depth: ${oc.select:global_profiler.global_tool_config.torch_memory.stack_depth,32}
107
+ router_replay:
108
+ _target_: verl.workers.config.RouterReplayConfig
109
+ mode: disabled
110
+ record_file: null
111
+ replay_file: null
112
+ grad_clip: 1.0
113
+ ulysses_sequence_parallel_size: 1
114
+ entropy_from_logits_with_chunking: false
115
+ entropy_checkpointing: false
116
+ use_remove_padding: ${oc.select:actor_rollout_ref.model.use_remove_padding,false}
117
+ ref:
118
+ rollout_n: ${oc.select:actor_rollout_ref.rollout.n,1}
119
+ strategy: ${actor_rollout_ref.actor.strategy}
120
+ use_torch_compile: ${oc.select:actor_rollout_ref.actor.use_torch_compile,true}
121
+ log_prob_micro_batch_size: null
122
+ log_prob_micro_batch_size_per_gpu: 32
123
+ log_prob_use_dynamic_bsz: ${oc.select:actor_rollout_ref.actor.use_dynamic_bsz,false}
124
+ log_prob_max_token_len_per_gpu: ${oc.select:actor_rollout_ref.actor.ppo_max_token_len_per_gpu,16384}
125
+ profiler:
126
+ _target_: verl.utils.profiler.ProfilerConfig
127
+ tool: ${oc.select:global_profiler.tool,null}
128
+ enable: false
129
+ all_ranks: false
130
+ ranks: []
131
+ save_path: ${oc.select:global_profiler.save_path,null}
132
+ tool_config:
133
+ nsys:
134
+ _target_: verl.utils.profiler.config.NsightToolConfig
135
+ discrete: ${oc.select:global_profiler.global_tool_config.nsys.discrete}
136
+ npu:
137
+ _target_: verl.utils.profiler.config.NPUToolConfig
138
+ contents: []
139
+ level: level1
140
+ analysis: true
141
+ discrete: false
142
+ torch:
143
+ _target_: verl.utils.profiler.config.TorchProfilerToolConfig
144
+ step_start: 0
145
+ step_end: null
146
+ torch_memory:
147
+ _target_: verl.utils.profiler.config.TorchMemoryToolConfig
148
+ trace_alloc_max_entries: ${oc.select:global_profiler.global_tool_config.torch_memory.trace_alloc_max_entries,100000}
149
+ stack_depth: ${oc.select:global_profiler.global_tool_config.torch_memory.stack_depth,32}
150
+ router_replay:
151
+ _target_: verl.workers.config.RouterReplayConfig
152
+ mode: disabled
153
+ record_file: null
154
+ replay_file: null
155
+ fsdp_config:
156
+ _target_: verl.workers.config.FSDPEngineConfig
157
+ wrap_policy:
158
+ min_num_params: 0
159
+ param_offload: false
160
+ optimizer_offload: false
161
+ offload_policy: false
162
+ reshard_after_forward: true
163
+ fsdp_size: -1
164
+ forward_prefetch: false
165
+ model_dtype: fp32
166
+ use_orig_params: false
167
+ seed: 42
168
+ full_determinism: false
169
+ ulysses_sequence_parallel_size: 1
170
+ entropy_from_logits_with_chunking: false
171
+ use_torch_compile: true
172
+ entropy_checkpointing: false
173
+ forward_only: true
174
+ strategy: fsdp
175
+ dtype: bfloat16
176
+ _target_: verl.workers.config.FSDPActorConfig
177
+ ulysses_sequence_parallel_size: ${oc.select:actor_rollout_ref.actor.ulysses_sequence_parallel_size,1}
178
+ entropy_from_logits_with_chunking: false
179
+ entropy_checkpointing: false
180
+ rollout:
181
+ _target_: verl.workers.config.RolloutConfig
182
+ name: vllm
183
+ mode: async
184
+ temperature: 1.0
185
+ top_k: -1
186
+ top_p: 1
187
+ prompt_length: 1536
188
+ response_length: 512
189
+ dtype: bfloat16
190
+ gpu_memory_utilization: 0.6
191
+ ignore_eos: false
192
+ enforce_eager: false
193
+ cudagraph_capture_sizes: null
194
+ free_cache_engine: true
195
+ tensor_model_parallel_size: 2
196
+ data_parallel_size: 1
197
+ expert_parallel_size: 1
198
+ pipeline_model_parallel_size: 1
199
+ max_num_batched_tokens: 8192
200
+ max_model_len: null
201
+ max_num_seqs: 1024
202
+ enable_chunked_prefill: true
203
+ enable_prefix_caching: true
204
+ load_format: dummy
205
+ log_prob_micro_batch_size: null
206
+ log_prob_micro_batch_size_per_gpu: 32
207
+ log_prob_use_dynamic_bsz: ${oc.select:actor_rollout_ref.actor.use_dynamic_bsz,false}
208
+ log_prob_max_token_len_per_gpu: ${oc.select:actor_rollout_ref.actor.ppo_max_token_len_per_gpu,16384}
209
+ disable_log_stats: true
210
+ do_sample: true
211
+ 'n': 1
212
+ over_sample_rate: 0
213
+ multi_stage_wake_up: false
214
+ engine_kwargs:
215
+ vllm: {}
216
+ sglang: {}
217
+ val_kwargs:
218
+ _target_: verl.workers.config.SamplingConfig
219
+ top_k: -1
220
+ top_p: 1.0
221
+ temperature: 0
222
+ 'n': 1
223
+ do_sample: false
224
+ multi_turn:
225
+ _target_: verl.workers.config.MultiTurnConfig
226
+ enable: false
227
+ max_assistant_turns: null
228
+ tool_config_path: null
229
+ max_user_turns: null
230
+ max_parallel_calls: 1
231
+ max_tool_response_length: 256
232
+ tool_response_truncate_side: middle
233
+ interaction_config_path: null
234
+ use_inference_chat_template: false
235
+ tokenization_sanity_check_mode: strict
236
+ format: hermes
237
+ num_repeat_rollouts: null
238
+ calculate_log_probs: false
239
+ agent:
240
+ _target_: verl.workers.config.AgentLoopConfig
241
+ num_workers: 8
242
+ default_agent_loop: single_turn_agent
243
+ agent_loop_config_path: null
244
+ custom_async_server:
245
+ _target_: verl.workers.config.CustomAsyncServerConfig
246
+ path: null
247
+ name: null
248
+ update_weights_bucket_megabytes: 512
249
+ trace:
250
+ _target_: verl.workers.config.TraceConfig
251
+ backend: null
252
+ token2text: false
253
+ max_samples_per_step_per_worker: null
254
+ skip_rollout: false
255
+ skip_dump_dir: /tmp/rollout_dump
256
+ skip_tokenizer_init: true
257
+ enable_rollout_routing_replay: false
258
+ profiler:
259
+ _target_: verl.utils.profiler.ProfilerConfig
260
+ tool: ${oc.select:global_profiler.tool,null}
261
+ enable: ${oc.select:actor_rollout_ref.actor.profiler.enable,false}
262
+ all_ranks: ${oc.select:actor_rollout_ref.actor.profiler.all_ranks,false}
263
+ ranks: ${oc.select:actor_rollout_ref.actor.profiler.ranks,[]}
264
+ save_path: ${oc.select:global_profiler.save_path,null}
265
+ tool_config: ${oc.select:actor_rollout_ref.actor.profiler.tool_config,null}
266
+ prometheus:
267
+ _target_: verl.workers.config.PrometheusConfig
268
+ enable: false
269
+ port: 9090
270
+ file: /tmp/ray/session_latest/metrics/prometheus/prometheus.yml
271
+ served_model_name: ${oc.select:actor_rollout_ref.model.path,null}
272
+ layered_summon: false
273
+ model:
274
+ _target_: verl.workers.config.HFModelConfig
275
+ path: /mnt/tidal-alsh01/dataset/redtrans/zhangruiqi/models/Qwen3-4B
276
+ hf_config_path: null
277
+ tokenizer_path: null
278
+ use_shm: false
279
+ trust_remote_code: false
280
+ custom_chat_template: null
281
+ external_lib: null
282
+ override_config: {}
283
+ enable_gradient_checkpointing: true
284
+ enable_activation_offload: false
285
+ use_remove_padding: true
286
+ lora_rank: 0
287
+ lora_alpha: 16
288
+ target_modules: all-linear
289
+ exclude_modules: null
290
+ lora_adapter_path: null
291
+ use_liger: false
292
+ use_fused_kernels: false
293
+ fused_kernel_options:
294
+ impl_backend: torch
295
+ hybrid_engine: true
296
+ nccl_timeout: 600
297
+ data:
298
+ tokenizer: null
299
+ use_shm: false
300
+ train_files: /mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/train.parquet
301
+ val_files: /mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/test.parquet
302
+ train_max_samples: -1
303
+ val_max_samples: -1
304
+ prompt_key: prompt
305
+ reward_fn_key: data_source
306
+ max_prompt_length: 512
307
+ max_response_length: 1024
308
+ train_batch_size: 1024
309
+ val_batch_size: null
310
+ tool_config_path: ${oc.select:actor_rollout_ref.rollout.multi_turn.tool_config_path,
311
+ null}
312
+ return_raw_input_ids: false
313
+ return_raw_chat: true
314
+ return_full_prompt: false
315
+ shuffle: true
316
+ seed: null
317
+ dataloader_num_workers: 8
318
+ image_patch_size: 14
319
+ validation_shuffle: false
320
+ filter_overlong_prompts: true
321
+ filter_overlong_prompts_workers: 1
322
+ truncation: error
323
+ image_key: images
324
+ video_key: videos
325
+ trust_remote_code: false
326
+ custom_cls:
327
+ path: null
328
+ name: null
329
+ return_multi_modal_inputs: true
330
+ sampler:
331
+ class_path: null
332
+ class_name: null
333
+ datagen:
334
+ path: null
335
+ name: null
336
+ apply_chat_template_kwargs: {}
337
+ reward_manager:
338
+ _target_: verl.trainer.config.config.RewardManagerConfig
339
+ source: register
340
+ name: ${oc.select:reward_model.reward_manager,naive}
341
+ module:
342
+ _target_: verl.trainer.config.config.ModuleConfig
343
+ path: null
344
+ name: custom_reward_manager
345
+ critic:
346
+ optim:
347
+ _target_: verl.workers.config.FSDPOptimizerConfig
348
+ optimizer: AdamW
349
+ optimizer_impl: torch.optim
350
+ lr: 1.0e-05
351
+ lr_warmup_steps_ratio: 0.0
352
+ total_training_steps: -1
353
+ weight_decay: 0.01
354
+ lr_warmup_steps: -1
355
+ betas:
356
+ - 0.9
357
+ - 0.999
358
+ clip_grad: 1.0
359
+ min_lr_ratio: 0.0
360
+ num_cycles: 0.5
361
+ lr_scheduler_type: constant
362
+ warmup_style: null
363
+ override_optimizer_config: null
364
+ model:
365
+ fsdp_config:
366
+ _target_: verl.workers.config.FSDPEngineConfig
367
+ wrap_policy:
368
+ min_num_params: 0
369
+ param_offload: false
370
+ optimizer_offload: false
371
+ offload_policy: false
372
+ reshard_after_forward: true
373
+ fsdp_size: -1
374
+ forward_prefetch: false
375
+ model_dtype: fp32
376
+ use_orig_params: false
377
+ seed: 42
378
+ full_determinism: false
379
+ ulysses_sequence_parallel_size: 1
380
+ entropy_from_logits_with_chunking: false
381
+ use_torch_compile: true
382
+ entropy_checkpointing: false
383
+ forward_only: false
384
+ strategy: fsdp
385
+ dtype: bfloat16
386
+ path: ~/models/deepseek-llm-7b-chat
387
+ tokenizer_path: ${oc.select:actor_rollout_ref.model.path,"~/models/deepseek-llm-7b-chat"}
388
+ override_config: {}
389
+ external_lib: ${oc.select:actor_rollout_ref.model.external_lib,null}
390
+ trust_remote_code: ${oc.select:actor_rollout_ref.model.trust_remote_code,false}
391
+ _target_: verl.workers.config.FSDPCriticModelCfg
392
+ use_shm: false
393
+ enable_gradient_checkpointing: true
394
+ enable_activation_offload: false
395
+ use_remove_padding: false
396
+ lora_rank: 0
397
+ lora_alpha: 16
398
+ target_modules: all-linear
399
+ _target_: verl.workers.config.FSDPCriticConfig
400
+ rollout_n: ${oc.select:actor_rollout_ref.rollout.n,1}
401
+ strategy: fsdp
402
+ enable: null
403
+ ppo_mini_batch_size: ${oc.select:actor_rollout_ref.actor.ppo_mini_batch_size,256}
404
+ ppo_micro_batch_size: null
405
+ ppo_micro_batch_size_per_gpu: ${oc.select:.ppo_micro_batch_size,null}
406
+ use_dynamic_bsz: ${oc.select:actor_rollout_ref.actor.use_dynamic_bsz,false}
407
+ ppo_max_token_len_per_gpu: 32768
408
+ forward_max_token_len_per_gpu: ${.ppo_max_token_len_per_gpu}
409
+ ppo_epochs: ${oc.select:actor_rollout_ref.actor.ppo_epochs,1}
410
+ shuffle: ${oc.select:actor_rollout_ref.actor.shuffle,false}
411
+ cliprange_value: 0.5
412
+ loss_agg_mode: ${oc.select:actor_rollout_ref.actor.loss_agg_mode,token-mean}
413
+ checkpoint:
414
+ _target_: verl.trainer.config.CheckpointConfig
415
+ save_contents:
416
+ - model
417
+ - optimizer
418
+ - extra
419
+ load_contents: ${.save_contents}
420
+ async_save: false
421
+ profiler:
422
+ _target_: verl.utils.profiler.ProfilerConfig
423
+ tool: ${oc.select:global_profiler.tool,null}
424
+ enable: false
425
+ all_ranks: false
426
+ ranks: []
427
+ save_path: ${oc.select:global_profiler.save_path,null}
428
+ tool_config:
429
+ nsys:
430
+ _target_: verl.utils.profiler.config.NsightToolConfig
431
+ discrete: ${oc.select:global_profiler.global_tool_config.nsys.discrete}
432
+ npu:
433
+ _target_: verl.utils.profiler.config.NPUToolConfig
434
+ contents: []
435
+ level: level1
436
+ analysis: true
437
+ discrete: false
438
+ torch:
439
+ _target_: verl.utils.profiler.config.TorchProfilerToolConfig
440
+ step_start: 0
441
+ step_end: null
442
+ torch_memory:
443
+ _target_: verl.utils.profiler.config.TorchMemoryToolConfig
444
+ trace_alloc_max_entries: ${oc.select:global_profiler.global_tool_config.torch_memory.trace_alloc_max_entries,100000}
445
+ stack_depth: ${oc.select:global_profiler.global_tool_config.torch_memory.stack_depth,32}
446
+ forward_micro_batch_size: ${oc.select:.ppo_micro_batch_size,null}
447
+ forward_micro_batch_size_per_gpu: ${oc.select:.ppo_micro_batch_size_per_gpu,null}
448
+ ulysses_sequence_parallel_size: 1
449
+ grad_clip: 1.0
450
+ reward_model:
451
+ enable: false
452
+ enable_resource_pool: false
453
+ n_gpus_per_node: 0
454
+ nnodes: 0
455
+ strategy: fsdp
456
+ model:
457
+ input_tokenizer: ${actor_rollout_ref.model.path}
458
+ path: ~/models/FsfairX-LLaMA3-RM-v0.1
459
+ external_lib: ${actor_rollout_ref.model.external_lib}
460
+ trust_remote_code: false
461
+ override_config: {}
462
+ use_shm: false
463
+ use_remove_padding: false
464
+ use_fused_kernels: ${actor_rollout_ref.model.use_fused_kernels}
465
+ fsdp_config:
466
+ _target_: verl.workers.config.FSDPEngineConfig
467
+ wrap_policy:
468
+ min_num_params: 0
469
+ param_offload: false
470
+ reshard_after_forward: true
471
+ fsdp_size: -1
472
+ forward_prefetch: false
473
+ micro_batch_size: null
474
+ micro_batch_size_per_gpu: null
475
+ max_length: null
476
+ use_dynamic_bsz: ${critic.use_dynamic_bsz}
477
+ forward_max_token_len_per_gpu: ${critic.forward_max_token_len_per_gpu}
478
+ reward_manager: naive
479
+ launch_reward_fn_async: false
480
+ sandbox_fusion:
481
+ url: null
482
+ max_concurrent: 64
483
+ memory_limit_mb: 1024
484
+ profiler:
485
+ _target_: verl.utils.profiler.ProfilerConfig
486
+ tool: ${oc.select:global_profiler.tool,null}
487
+ enable: false
488
+ all_ranks: false
489
+ ranks: []
490
+ save_path: ${oc.select:global_profiler.save_path,null}
491
+ tool_config: ${oc.select:actor_rollout_ref.actor.profiler.tool_config,null}
492
+ ulysses_sequence_parallel_size: 1
493
+ use_reward_loop: true
494
+ rollout:
495
+ _target_: verl.workers.config.RolloutConfig
496
+ name: ???
497
+ dtype: bfloat16
498
+ gpu_memory_utilization: 0.5
499
+ enforce_eager: true
500
+ cudagraph_capture_sizes: null
501
+ free_cache_engine: true
502
+ data_parallel_size: 1
503
+ expert_parallel_size: 1
504
+ tensor_model_parallel_size: 2
505
+ max_num_batched_tokens: 8192
506
+ max_model_len: null
507
+ max_num_seqs: 1024
508
+ load_format: auto
509
+ engine_kwargs: {}
510
+ limit_images: null
511
+ enable_chunked_prefill: true
512
+ enable_prefix_caching: true
513
+ disable_log_stats: true
514
+ skip_tokenizer_init: true
515
+ prompt_length: 512
516
+ response_length: 512
517
+ algorithm:
518
+ rollout_correction:
519
+ rollout_is: null
520
+ rollout_is_threshold: 2.0
521
+ rollout_rs: null
522
+ rollout_rs_threshold: null
523
+ rollout_rs_threshold_lower: null
524
+ rollout_token_veto_threshold: null
525
+ bypass_mode: false
526
+ use_policy_gradient: false
527
+ rollout_is_batch_normalize: false
528
+ _target_: verl.trainer.config.AlgoConfig
529
+ gamma: 1.0
530
+ lam: 1.0
531
+ adv_estimator: sgrpo
532
+ norm_adv_by_std_in_grpo: false
533
+ use_kl_in_reward: false
534
+ kl_penalty: kl
535
+ kl_ctrl:
536
+ _target_: verl.trainer.config.KLControlConfig
537
+ type: fixed
538
+ kl_coef: 0.001
539
+ horizon: 10000
540
+ target_kl: 0.1
541
+ use_pf_ppo: false
542
+ pf_ppo:
543
+ reweight_method: pow
544
+ weight_pow: 2.0
545
+ sgrpo:
546
+ num_truncations: 4
547
+ force_answer_prompt: Time is limited, stop thinking and start answering.\n</think>\n\n
548
+ answer_max_tokens: 256
549
+ min_truncation_ratio: 0.1
550
+ max_truncation_ratio: 0.9
551
+ custom_reward_function:
552
+ path: null
553
+ name: compute_score
554
+ trainer:
555
+ balance_batch: true
556
+ total_epochs: 15
557
+ total_training_steps: null
558
+ project_name: verl_sgrpo_example_gsm8k
559
+ experiment_name: qwen3_8b_sgrpo
560
+ logger:
561
+ - console
562
+ - wandb
563
+ log_val_generations: 0
564
+ rollout_data_dir: ./outputs/sgrpo_samples
565
+ rollout_data_dump_freq: 20
566
+ validation_data_dir: null
567
+ nnodes: 1
568
+ n_gpus_per_node: 8
569
+ save_freq: 20
570
+ esi_redundant_time: 0
571
+ resume_mode: disable
572
+ resume_from_path: null
573
+ val_before_train: true
574
+ val_only: false
575
+ test_freq: 5
576
+ critic_warmup: 0
577
+ default_hdfs_dir: null
578
+ del_local_ckpt_after_load: false
579
+ default_local_dir: checkpoints/${trainer.project_name}/${trainer.experiment_name}
580
+ max_actor_ckpt_to_keep: null
581
+ max_critic_ckpt_to_keep: null
582
+ ray_wait_register_center_timeout: 300
583
+ device: cuda
584
+ use_legacy_worker_impl: auto
585
+ global_profiler:
586
+ _target_: verl.utils.profiler.ProfilerConfig
587
+ tool: null
588
+ steps: null
589
+ profile_continuous_steps: false
590
+ save_path: outputs/profile
591
+ global_tool_config:
592
+ nsys:
593
+ _target_: verl.utils.profiler.config.NsightToolConfig
594
+ discrete: false
595
+ controller_nsight_options:
596
+ trace: cuda,nvtx,cublas,ucx
597
+ cuda-memory-usage: 'true'
598
+ cuda-graph-trace: graph
599
+ worker_nsight_options:
600
+ trace: cuda,nvtx,cublas,ucx
601
+ cuda-memory-usage: 'true'
602
+ cuda-graph-trace: graph
603
+ capture-range: cudaProfilerApi
604
+ capture-range-end: null
605
+ kill: none
606
+ torch_memory:
607
+ trace_alloc_max_entries: 100000
608
+ stack_depth: 32
609
+ context: all
610
+ stacks: all
611
+ kw_args: {}
612
+ transfer_queue:
613
+ enable: false
614
+ ray_kwargs:
615
+ ray_init:
616
+ num_cpus: null
617
+ timeline_json_file: null
outputs/2026-01-14/15-31-37/.hydra/overrides.yaml ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ - algorithm.adv_estimator=sgrpo
2
+ - algorithm.norm_adv_by_std_in_grpo=False
3
+ - algorithm.use_kl_in_reward=False
4
+ - data.train_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/train.parquet
5
+ - data.val_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/test.parquet
6
+ - data.train_batch_size=1024
7
+ - data.max_prompt_length=512
8
+ - data.max_response_length=1024
9
+ - data.filter_overlong_prompts=True
10
+ - data.truncation=error
11
+ - data.return_raw_chat=True
12
+ - actor_rollout_ref.model.path=/mnt/tidal-alsh01/dataset/redtrans/zhangruiqi/models/Qwen3-4B
13
+ - actor_rollout_ref.actor.optim.lr=1e-6
14
+ - actor_rollout_ref.model.use_remove_padding=True
15
+ - actor_rollout_ref.actor.ppo_mini_batch_size=256
16
+ - actor_rollout_ref.actor.ppo_micro_batch_size_per_gpu=32
17
+ - actor_rollout_ref.actor.use_kl_loss=True
18
+ - actor_rollout_ref.actor.kl_loss_coef=0.001
19
+ - actor_rollout_ref.actor.kl_loss_type=low_var_kl
20
+ - actor_rollout_ref.actor.entropy_coeff=0
21
+ - actor_rollout_ref.model.enable_gradient_checkpointing=True
22
+ - actor_rollout_ref.actor.fsdp_config.param_offload=False
23
+ - actor_rollout_ref.actor.fsdp_config.optimizer_offload=False
24
+ - actor_rollout_ref.rollout.log_prob_micro_batch_size_per_gpu=32
25
+ - actor_rollout_ref.rollout.tensor_model_parallel_size=2
26
+ - actor_rollout_ref.rollout.name=vllm
27
+ - actor_rollout_ref.rollout.gpu_memory_utilization=0.6
28
+ - actor_rollout_ref.rollout.n=1
29
+ - actor_rollout_ref.rollout.prompt_length=1536
30
+ - actor_rollout_ref.rollout.response_length=512
31
+ - actor_rollout_ref.ref.log_prob_micro_batch_size_per_gpu=32
32
+ - actor_rollout_ref.ref.fsdp_config.param_offload=False
33
+ - sgrpo.num_truncations=4
34
+ - sgrpo.force_answer_prompt="Time is limited, stop thinking and start answering.\n</think>\n\n"
35
+ - sgrpo.answer_max_tokens=256
36
+ - trainer.critic_warmup=0
37
+ - trainer.logger=["console","wandb"]
38
+ - trainer.project_name=verl_sgrpo_example_gsm8k
39
+ - trainer.experiment_name=qwen3_8b_sgrpo
40
+ - trainer.resume_mode=disable
41
+ - trainer.n_gpus_per_node=8
42
+ - trainer.nnodes=1
43
+ - trainer.save_freq=20
44
+ - trainer.test_freq=5
45
+ - trainer.total_epochs=15
46
+ - trainer.rollout_data_dir=./outputs/sgrpo_samples
47
+ - trainer.rollout_data_dump_freq=20
outputs/2026-01-14/15-31-37/main_sgrpo.log ADDED
File without changes
outputs/2026-01-14/15-45-38/.hydra/hydra.yaml ADDED
@@ -0,0 +1,218 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
4
+ sweep:
5
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.mode=RUN
114
+ task:
115
+ - algorithm.adv_estimator=sgrpo
116
+ - algorithm.norm_adv_by_std_in_grpo=False
117
+ - algorithm.use_kl_in_reward=False
118
+ - data.train_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/train.parquet
119
+ - data.val_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/test.parquet
120
+ - data.train_batch_size=1024
121
+ - data.max_prompt_length=512
122
+ - data.max_response_length=1024
123
+ - data.filter_overlong_prompts=True
124
+ - data.truncation=error
125
+ - data.return_raw_chat=True
126
+ - actor_rollout_ref.model.path=/mnt/tidal-alsh01/dataset/redtrans/zhangruiqi/models/Qwen3-4B
127
+ - actor_rollout_ref.actor.optim.lr=1e-6
128
+ - actor_rollout_ref.model.use_remove_padding=True
129
+ - actor_rollout_ref.actor.ppo_mini_batch_size=256
130
+ - actor_rollout_ref.actor.ppo_micro_batch_size_per_gpu=32
131
+ - actor_rollout_ref.actor.use_kl_loss=True
132
+ - actor_rollout_ref.actor.kl_loss_coef=0.001
133
+ - actor_rollout_ref.actor.kl_loss_type=low_var_kl
134
+ - actor_rollout_ref.actor.entropy_coeff=0
135
+ - actor_rollout_ref.model.enable_gradient_checkpointing=True
136
+ - actor_rollout_ref.actor.fsdp_config.param_offload=False
137
+ - actor_rollout_ref.actor.fsdp_config.optimizer_offload=False
138
+ - actor_rollout_ref.rollout.log_prob_micro_batch_size_per_gpu=32
139
+ - actor_rollout_ref.rollout.tensor_model_parallel_size=2
140
+ - actor_rollout_ref.rollout.name=vllm
141
+ - actor_rollout_ref.rollout.gpu_memory_utilization=0.6
142
+ - actor_rollout_ref.rollout.n=1
143
+ - actor_rollout_ref.rollout.prompt_length=1536
144
+ - actor_rollout_ref.rollout.response_length=512
145
+ - actor_rollout_ref.ref.log_prob_micro_batch_size_per_gpu=32
146
+ - actor_rollout_ref.ref.fsdp_config.param_offload=False
147
+ - sgrpo.num_truncations=4
148
+ - sgrpo.force_answer_prompt="Time is limited, stop thinking and start answering.\n</think>\n\n"
149
+ - sgrpo.answer_max_tokens=256
150
+ - trainer.critic_warmup=0
151
+ - trainer.logger=["console"]
152
+ - trainer.project_name=verl_sgrpo_example_gsm8k
153
+ - trainer.experiment_name=qwen3_8b_sgrpo
154
+ - trainer.resume_mode=disable
155
+ - trainer.n_gpus_per_node=8
156
+ - trainer.nnodes=1
157
+ - trainer.save_freq=20
158
+ - trainer.test_freq=5
159
+ - trainer.total_epochs=15
160
+ - trainer.rollout_data_dir=./outputs/sgrpo_samples
161
+ - trainer.rollout_data_dump_freq=20
162
+ job:
163
+ name: main_sgrpo
164
+ chdir: null
165
+ override_dirname: actor_rollout_ref.actor.entropy_coeff=0,actor_rollout_ref.actor.fsdp_config.optimizer_offload=False,actor_rollout_ref.actor.fsdp_config.param_offload=False,actor_rollout_ref.actor.kl_loss_coef=0.001,actor_rollout_ref.actor.kl_loss_type=low_var_kl,actor_rollout_ref.actor.optim.lr=1e-6,actor_rollout_ref.actor.ppo_micro_batch_size_per_gpu=32,actor_rollout_ref.actor.ppo_mini_batch_size=256,actor_rollout_ref.actor.use_kl_loss=True,actor_rollout_ref.model.enable_gradient_checkpointing=True,actor_rollout_ref.model.path=/mnt/tidal-alsh01/dataset/redtrans/zhangruiqi/models/Qwen3-4B,actor_rollout_ref.model.use_remove_padding=True,actor_rollout_ref.ref.fsdp_config.param_offload=False,actor_rollout_ref.ref.log_prob_micro_batch_size_per_gpu=32,actor_rollout_ref.rollout.gpu_memory_utilization=0.6,actor_rollout_ref.rollout.log_prob_micro_batch_size_per_gpu=32,actor_rollout_ref.rollout.n=1,actor_rollout_ref.rollout.name=vllm,actor_rollout_ref.rollout.prompt_length=1536,actor_rollout_ref.rollout.response_length=512,actor_rollout_ref.rollout.tensor_model_parallel_size=2,algorithm.adv_estimator=sgrpo,algorithm.norm_adv_by_std_in_grpo=False,algorithm.use_kl_in_reward=False,data.filter_overlong_prompts=True,data.max_prompt_length=512,data.max_response_length=1024,data.return_raw_chat=True,data.train_batch_size=1024,data.train_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/train.parquet,data.truncation=error,data.val_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/test.parquet,sgrpo.answer_max_tokens=256,sgrpo.force_answer_prompt="Time
166
+ is limited, stop thinking and start answering.\n</think>\n\n",sgrpo.num_truncations=4,trainer.critic_warmup=0,trainer.experiment_name=qwen3_8b_sgrpo,trainer.logger=["console"],trainer.n_gpus_per_node=8,trainer.nnodes=1,trainer.project_name=verl_sgrpo_example_gsm8k,trainer.resume_mode=disable,trainer.rollout_data_dir=./outputs/sgrpo_samples,trainer.rollout_data_dump_freq=20,trainer.save_freq=20,trainer.test_freq=5,trainer.total_epochs=15
167
+ id: ???
168
+ num: ???
169
+ config_name: sgrpo_trainer
170
+ env_set: {}
171
+ env_copy: []
172
+ config:
173
+ override_dirname:
174
+ kv_sep: '='
175
+ item_sep: ','
176
+ exclude_keys: []
177
+ runtime:
178
+ version: 1.3.2
179
+ version_base: '1.3'
180
+ cwd: /mnt/tidal-alsh01/usr/zhangruiqi1/my/verl
181
+ config_sources:
182
+ - path: hydra.conf
183
+ schema: pkg
184
+ provider: hydra
185
+ - path: /mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/config
186
+ schema: file
187
+ provider: main
188
+ - path: verl.trainer.config
189
+ schema: pkg
190
+ provider: verl-trainer-config
191
+ - path: ''
192
+ schema: structured
193
+ provider: schema
194
+ output_dir: /mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/outputs/2026-01-14/15-45-38
195
+ choices:
196
+ algorithm@algorithm.rollout_correction: rollout_correction
197
+ reward_model: dp_reward_loop
198
+ critic: dp_critic
199
+ critic/../engine@critic.model.fsdp_config: fsdp
200
+ critic/../optim@critic.optim: fsdp
201
+ model@actor_rollout_ref.model: hf_model
202
+ rollout@actor_rollout_ref.rollout: rollout
203
+ ref@actor_rollout_ref.ref: dp_ref
204
+ ref/../engine@actor_rollout_ref.ref.fsdp_config: fsdp
205
+ data: legacy_data
206
+ actor@actor_rollout_ref.actor: dp_actor
207
+ actor/../engine@actor_rollout_ref.actor.fsdp_config: fsdp
208
+ actor/../optim@actor_rollout_ref.actor.optim: fsdp
209
+ hydra/env: default
210
+ hydra/callbacks: null
211
+ hydra/job_logging: default
212
+ hydra/hydra_logging: default
213
+ hydra/hydra_help: default
214
+ hydra/help: default
215
+ hydra/sweeper: basic
216
+ hydra/launcher: basic
217
+ hydra/output: default
218
+ verbose: false
outputs/2026-01-14/15-45-38/main_sgrpo.log ADDED
File without changes
outputs/2026-01-14/16-00-30/.hydra/config.yaml ADDED
@@ -0,0 +1,616 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ actor_rollout_ref:
2
+ actor:
3
+ optim:
4
+ _target_: verl.workers.config.FSDPOptimizerConfig
5
+ optimizer: AdamW
6
+ optimizer_impl: torch.optim
7
+ lr: 1.0e-06
8
+ lr_warmup_steps_ratio: 0.0
9
+ total_training_steps: -1
10
+ weight_decay: 0.01
11
+ lr_warmup_steps: -1
12
+ betas:
13
+ - 0.9
14
+ - 0.999
15
+ clip_grad: 1.0
16
+ min_lr_ratio: 0.0
17
+ num_cycles: 0.5
18
+ lr_scheduler_type: constant
19
+ warmup_style: null
20
+ override_optimizer_config: null
21
+ fsdp_config:
22
+ _target_: verl.workers.config.FSDPEngineConfig
23
+ wrap_policy:
24
+ min_num_params: 0
25
+ param_offload: false
26
+ optimizer_offload: false
27
+ offload_policy: false
28
+ reshard_after_forward: true
29
+ fsdp_size: -1
30
+ forward_prefetch: false
31
+ model_dtype: fp32
32
+ use_orig_params: false
33
+ seed: 42
34
+ full_determinism: false
35
+ ulysses_sequence_parallel_size: 1
36
+ entropy_from_logits_with_chunking: false
37
+ use_torch_compile: true
38
+ entropy_checkpointing: false
39
+ forward_only: false
40
+ strategy: fsdp
41
+ dtype: bfloat16
42
+ _target_: verl.workers.config.FSDPActorConfig
43
+ rollout_n: ${oc.select:actor_rollout_ref.rollout.n,1}
44
+ strategy: fsdp
45
+ ppo_mini_batch_size: 256
46
+ ppo_micro_batch_size: null
47
+ ppo_micro_batch_size_per_gpu: 32
48
+ use_dynamic_bsz: false
49
+ ppo_max_token_len_per_gpu: 16384
50
+ clip_ratio: 0.2
51
+ clip_ratio_low: 0.2
52
+ clip_ratio_high: 0.2
53
+ freeze_vision_tower: false
54
+ policy_loss:
55
+ _target_: verl.workers.config.PolicyLossConfig
56
+ loss_mode: vanilla
57
+ clip_cov_ratio: 0.0002
58
+ clip_cov_lb: 1.0
59
+ clip_cov_ub: 5.0
60
+ kl_cov_ratio: 0.0002
61
+ ppo_kl_coef: 0.1
62
+ clip_ratio_c: 3.0
63
+ loss_agg_mode: token-mean
64
+ loss_scale_factor: null
65
+ entropy_coeff: 0
66
+ calculate_entropy: false
67
+ use_kl_loss: true
68
+ use_torch_compile: true
69
+ kl_loss_coef: 0.001
70
+ kl_loss_type: low_var_kl
71
+ ppo_epochs: 1
72
+ shuffle: false
73
+ checkpoint:
74
+ _target_: verl.trainer.config.CheckpointConfig
75
+ save_contents:
76
+ - model
77
+ - optimizer
78
+ - extra
79
+ load_contents: ${.save_contents}
80
+ async_save: false
81
+ use_fused_kernels: ${oc.select:actor_rollout_ref.model.use_fused_kernels,false}
82
+ profiler:
83
+ _target_: verl.utils.profiler.ProfilerConfig
84
+ tool: ${oc.select:global_profiler.tool,null}
85
+ enable: false
86
+ all_ranks: false
87
+ ranks: []
88
+ save_path: ${oc.select:global_profiler.save_path,null}
89
+ tool_config:
90
+ nsys:
91
+ _target_: verl.utils.profiler.config.NsightToolConfig
92
+ discrete: ${oc.select:global_profiler.global_tool_config.nsys.discrete}
93
+ npu:
94
+ _target_: verl.utils.profiler.config.NPUToolConfig
95
+ contents: []
96
+ level: level1
97
+ analysis: true
98
+ discrete: false
99
+ torch:
100
+ _target_: verl.utils.profiler.config.TorchProfilerToolConfig
101
+ step_start: 0
102
+ step_end: null
103
+ torch_memory:
104
+ _target_: verl.utils.profiler.config.TorchMemoryToolConfig
105
+ trace_alloc_max_entries: ${oc.select:global_profiler.global_tool_config.torch_memory.trace_alloc_max_entries,100000}
106
+ stack_depth: ${oc.select:global_profiler.global_tool_config.torch_memory.stack_depth,32}
107
+ router_replay:
108
+ _target_: verl.workers.config.RouterReplayConfig
109
+ mode: disabled
110
+ record_file: null
111
+ replay_file: null
112
+ grad_clip: 1.0
113
+ ulysses_sequence_parallel_size: 1
114
+ entropy_from_logits_with_chunking: false
115
+ entropy_checkpointing: false
116
+ use_remove_padding: ${oc.select:actor_rollout_ref.model.use_remove_padding,false}
117
+ ref:
118
+ rollout_n: ${oc.select:actor_rollout_ref.rollout.n,1}
119
+ strategy: ${actor_rollout_ref.actor.strategy}
120
+ use_torch_compile: ${oc.select:actor_rollout_ref.actor.use_torch_compile,true}
121
+ log_prob_micro_batch_size: null
122
+ log_prob_micro_batch_size_per_gpu: 32
123
+ log_prob_use_dynamic_bsz: ${oc.select:actor_rollout_ref.actor.use_dynamic_bsz,false}
124
+ log_prob_max_token_len_per_gpu: ${oc.select:actor_rollout_ref.actor.ppo_max_token_len_per_gpu,16384}
125
+ profiler:
126
+ _target_: verl.utils.profiler.ProfilerConfig
127
+ tool: ${oc.select:global_profiler.tool,null}
128
+ enable: false
129
+ all_ranks: false
130
+ ranks: []
131
+ save_path: ${oc.select:global_profiler.save_path,null}
132
+ tool_config:
133
+ nsys:
134
+ _target_: verl.utils.profiler.config.NsightToolConfig
135
+ discrete: ${oc.select:global_profiler.global_tool_config.nsys.discrete}
136
+ npu:
137
+ _target_: verl.utils.profiler.config.NPUToolConfig
138
+ contents: []
139
+ level: level1
140
+ analysis: true
141
+ discrete: false
142
+ torch:
143
+ _target_: verl.utils.profiler.config.TorchProfilerToolConfig
144
+ step_start: 0
145
+ step_end: null
146
+ torch_memory:
147
+ _target_: verl.utils.profiler.config.TorchMemoryToolConfig
148
+ trace_alloc_max_entries: ${oc.select:global_profiler.global_tool_config.torch_memory.trace_alloc_max_entries,100000}
149
+ stack_depth: ${oc.select:global_profiler.global_tool_config.torch_memory.stack_depth,32}
150
+ router_replay:
151
+ _target_: verl.workers.config.RouterReplayConfig
152
+ mode: disabled
153
+ record_file: null
154
+ replay_file: null
155
+ fsdp_config:
156
+ _target_: verl.workers.config.FSDPEngineConfig
157
+ wrap_policy:
158
+ min_num_params: 0
159
+ param_offload: false
160
+ optimizer_offload: false
161
+ offload_policy: false
162
+ reshard_after_forward: true
163
+ fsdp_size: -1
164
+ forward_prefetch: false
165
+ model_dtype: fp32
166
+ use_orig_params: false
167
+ seed: 42
168
+ full_determinism: false
169
+ ulysses_sequence_parallel_size: 1
170
+ entropy_from_logits_with_chunking: false
171
+ use_torch_compile: true
172
+ entropy_checkpointing: false
173
+ forward_only: true
174
+ strategy: fsdp
175
+ dtype: bfloat16
176
+ _target_: verl.workers.config.FSDPActorConfig
177
+ ulysses_sequence_parallel_size: ${oc.select:actor_rollout_ref.actor.ulysses_sequence_parallel_size,1}
178
+ entropy_from_logits_with_chunking: false
179
+ entropy_checkpointing: false
180
+ rollout:
181
+ _target_: verl.workers.config.RolloutConfig
182
+ name: vllm
183
+ mode: async
184
+ temperature: 1.0
185
+ top_k: -1
186
+ top_p: 1
187
+ prompt_length: 1536
188
+ response_length: 512
189
+ dtype: bfloat16
190
+ gpu_memory_utilization: 0.6
191
+ ignore_eos: false
192
+ enforce_eager: false
193
+ cudagraph_capture_sizes: null
194
+ free_cache_engine: true
195
+ tensor_model_parallel_size: 2
196
+ data_parallel_size: 1
197
+ expert_parallel_size: 1
198
+ pipeline_model_parallel_size: 1
199
+ max_num_batched_tokens: 8192
200
+ max_model_len: null
201
+ max_num_seqs: 1024
202
+ enable_chunked_prefill: true
203
+ enable_prefix_caching: true
204
+ load_format: dummy
205
+ log_prob_micro_batch_size: null
206
+ log_prob_micro_batch_size_per_gpu: 32
207
+ log_prob_use_dynamic_bsz: ${oc.select:actor_rollout_ref.actor.use_dynamic_bsz,false}
208
+ log_prob_max_token_len_per_gpu: ${oc.select:actor_rollout_ref.actor.ppo_max_token_len_per_gpu,16384}
209
+ disable_log_stats: true
210
+ do_sample: true
211
+ 'n': 1
212
+ over_sample_rate: 0
213
+ multi_stage_wake_up: false
214
+ engine_kwargs:
215
+ vllm: {}
216
+ sglang: {}
217
+ val_kwargs:
218
+ _target_: verl.workers.config.SamplingConfig
219
+ top_k: -1
220
+ top_p: 1.0
221
+ temperature: 0
222
+ 'n': 1
223
+ do_sample: false
224
+ multi_turn:
225
+ _target_: verl.workers.config.MultiTurnConfig
226
+ enable: false
227
+ max_assistant_turns: null
228
+ tool_config_path: null
229
+ max_user_turns: null
230
+ max_parallel_calls: 1
231
+ max_tool_response_length: 256
232
+ tool_response_truncate_side: middle
233
+ interaction_config_path: null
234
+ use_inference_chat_template: false
235
+ tokenization_sanity_check_mode: strict
236
+ format: hermes
237
+ num_repeat_rollouts: null
238
+ calculate_log_probs: false
239
+ agent:
240
+ _target_: verl.workers.config.AgentLoopConfig
241
+ num_workers: 8
242
+ default_agent_loop: single_turn_agent
243
+ agent_loop_config_path: null
244
+ custom_async_server:
245
+ _target_: verl.workers.config.CustomAsyncServerConfig
246
+ path: null
247
+ name: null
248
+ update_weights_bucket_megabytes: 512
249
+ trace:
250
+ _target_: verl.workers.config.TraceConfig
251
+ backend: null
252
+ token2text: false
253
+ max_samples_per_step_per_worker: null
254
+ skip_rollout: false
255
+ skip_dump_dir: /tmp/rollout_dump
256
+ skip_tokenizer_init: true
257
+ enable_rollout_routing_replay: false
258
+ profiler:
259
+ _target_: verl.utils.profiler.ProfilerConfig
260
+ tool: ${oc.select:global_profiler.tool,null}
261
+ enable: ${oc.select:actor_rollout_ref.actor.profiler.enable,false}
262
+ all_ranks: ${oc.select:actor_rollout_ref.actor.profiler.all_ranks,false}
263
+ ranks: ${oc.select:actor_rollout_ref.actor.profiler.ranks,[]}
264
+ save_path: ${oc.select:global_profiler.save_path,null}
265
+ tool_config: ${oc.select:actor_rollout_ref.actor.profiler.tool_config,null}
266
+ prometheus:
267
+ _target_: verl.workers.config.PrometheusConfig
268
+ enable: false
269
+ port: 9090
270
+ file: /tmp/ray/session_latest/metrics/prometheus/prometheus.yml
271
+ served_model_name: ${oc.select:actor_rollout_ref.model.path,null}
272
+ layered_summon: false
273
+ model:
274
+ _target_: verl.workers.config.HFModelConfig
275
+ path: /mnt/tidal-alsh01/dataset/redtrans/zhangruiqi/models/Qwen3-4B
276
+ hf_config_path: null
277
+ tokenizer_path: null
278
+ use_shm: false
279
+ trust_remote_code: false
280
+ custom_chat_template: null
281
+ external_lib: null
282
+ override_config: {}
283
+ enable_gradient_checkpointing: true
284
+ enable_activation_offload: false
285
+ use_remove_padding: true
286
+ lora_rank: 0
287
+ lora_alpha: 16
288
+ target_modules: all-linear
289
+ exclude_modules: null
290
+ lora_adapter_path: null
291
+ use_liger: false
292
+ use_fused_kernels: false
293
+ fused_kernel_options:
294
+ impl_backend: torch
295
+ hybrid_engine: true
296
+ nccl_timeout: 600
297
+ data:
298
+ tokenizer: null
299
+ use_shm: false
300
+ train_files: /mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/train.parquet
301
+ val_files: /mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/test.parquet
302
+ train_max_samples: -1
303
+ val_max_samples: -1
304
+ prompt_key: prompt
305
+ reward_fn_key: data_source
306
+ max_prompt_length: 512
307
+ max_response_length: 1024
308
+ train_batch_size: 1024
309
+ val_batch_size: null
310
+ tool_config_path: ${oc.select:actor_rollout_ref.rollout.multi_turn.tool_config_path,
311
+ null}
312
+ return_raw_input_ids: false
313
+ return_raw_chat: true
314
+ return_full_prompt: false
315
+ shuffle: true
316
+ seed: null
317
+ dataloader_num_workers: 8
318
+ image_patch_size: 14
319
+ validation_shuffle: false
320
+ filter_overlong_prompts: true
321
+ filter_overlong_prompts_workers: 1
322
+ truncation: error
323
+ image_key: images
324
+ video_key: videos
325
+ trust_remote_code: false
326
+ custom_cls:
327
+ path: null
328
+ name: null
329
+ return_multi_modal_inputs: true
330
+ sampler:
331
+ class_path: null
332
+ class_name: null
333
+ datagen:
334
+ path: null
335
+ name: null
336
+ apply_chat_template_kwargs: {}
337
+ reward_manager:
338
+ _target_: verl.trainer.config.config.RewardManagerConfig
339
+ source: register
340
+ name: ${oc.select:reward_model.reward_manager,naive}
341
+ module:
342
+ _target_: verl.trainer.config.config.ModuleConfig
343
+ path: null
344
+ name: custom_reward_manager
345
+ critic:
346
+ optim:
347
+ _target_: verl.workers.config.FSDPOptimizerConfig
348
+ optimizer: AdamW
349
+ optimizer_impl: torch.optim
350
+ lr: 1.0e-05
351
+ lr_warmup_steps_ratio: 0.0
352
+ total_training_steps: -1
353
+ weight_decay: 0.01
354
+ lr_warmup_steps: -1
355
+ betas:
356
+ - 0.9
357
+ - 0.999
358
+ clip_grad: 1.0
359
+ min_lr_ratio: 0.0
360
+ num_cycles: 0.5
361
+ lr_scheduler_type: constant
362
+ warmup_style: null
363
+ override_optimizer_config: null
364
+ model:
365
+ fsdp_config:
366
+ _target_: verl.workers.config.FSDPEngineConfig
367
+ wrap_policy:
368
+ min_num_params: 0
369
+ param_offload: false
370
+ optimizer_offload: false
371
+ offload_policy: false
372
+ reshard_after_forward: true
373
+ fsdp_size: -1
374
+ forward_prefetch: false
375
+ model_dtype: fp32
376
+ use_orig_params: false
377
+ seed: 42
378
+ full_determinism: false
379
+ ulysses_sequence_parallel_size: 1
380
+ entropy_from_logits_with_chunking: false
381
+ use_torch_compile: true
382
+ entropy_checkpointing: false
383
+ forward_only: false
384
+ strategy: fsdp
385
+ dtype: bfloat16
386
+ path: ~/models/deepseek-llm-7b-chat
387
+ tokenizer_path: ${oc.select:actor_rollout_ref.model.path,"~/models/deepseek-llm-7b-chat"}
388
+ override_config: {}
389
+ external_lib: ${oc.select:actor_rollout_ref.model.external_lib,null}
390
+ trust_remote_code: ${oc.select:actor_rollout_ref.model.trust_remote_code,false}
391
+ _target_: verl.workers.config.FSDPCriticModelCfg
392
+ use_shm: false
393
+ enable_gradient_checkpointing: true
394
+ enable_activation_offload: false
395
+ use_remove_padding: false
396
+ lora_rank: 0
397
+ lora_alpha: 16
398
+ target_modules: all-linear
399
+ _target_: verl.workers.config.FSDPCriticConfig
400
+ rollout_n: ${oc.select:actor_rollout_ref.rollout.n,1}
401
+ strategy: fsdp
402
+ enable: null
403
+ ppo_mini_batch_size: ${oc.select:actor_rollout_ref.actor.ppo_mini_batch_size,256}
404
+ ppo_micro_batch_size: null
405
+ ppo_micro_batch_size_per_gpu: ${oc.select:.ppo_micro_batch_size,null}
406
+ use_dynamic_bsz: ${oc.select:actor_rollout_ref.actor.use_dynamic_bsz,false}
407
+ ppo_max_token_len_per_gpu: 32768
408
+ forward_max_token_len_per_gpu: ${.ppo_max_token_len_per_gpu}
409
+ ppo_epochs: ${oc.select:actor_rollout_ref.actor.ppo_epochs,1}
410
+ shuffle: ${oc.select:actor_rollout_ref.actor.shuffle,false}
411
+ cliprange_value: 0.5
412
+ loss_agg_mode: ${oc.select:actor_rollout_ref.actor.loss_agg_mode,token-mean}
413
+ checkpoint:
414
+ _target_: verl.trainer.config.CheckpointConfig
415
+ save_contents:
416
+ - model
417
+ - optimizer
418
+ - extra
419
+ load_contents: ${.save_contents}
420
+ async_save: false
421
+ profiler:
422
+ _target_: verl.utils.profiler.ProfilerConfig
423
+ tool: ${oc.select:global_profiler.tool,null}
424
+ enable: false
425
+ all_ranks: false
426
+ ranks: []
427
+ save_path: ${oc.select:global_profiler.save_path,null}
428
+ tool_config:
429
+ nsys:
430
+ _target_: verl.utils.profiler.config.NsightToolConfig
431
+ discrete: ${oc.select:global_profiler.global_tool_config.nsys.discrete}
432
+ npu:
433
+ _target_: verl.utils.profiler.config.NPUToolConfig
434
+ contents: []
435
+ level: level1
436
+ analysis: true
437
+ discrete: false
438
+ torch:
439
+ _target_: verl.utils.profiler.config.TorchProfilerToolConfig
440
+ step_start: 0
441
+ step_end: null
442
+ torch_memory:
443
+ _target_: verl.utils.profiler.config.TorchMemoryToolConfig
444
+ trace_alloc_max_entries: ${oc.select:global_profiler.global_tool_config.torch_memory.trace_alloc_max_entries,100000}
445
+ stack_depth: ${oc.select:global_profiler.global_tool_config.torch_memory.stack_depth,32}
446
+ forward_micro_batch_size: ${oc.select:.ppo_micro_batch_size,null}
447
+ forward_micro_batch_size_per_gpu: ${oc.select:.ppo_micro_batch_size_per_gpu,null}
448
+ ulysses_sequence_parallel_size: 1
449
+ grad_clip: 1.0
450
+ reward_model:
451
+ enable: false
452
+ enable_resource_pool: false
453
+ n_gpus_per_node: 0
454
+ nnodes: 0
455
+ strategy: fsdp
456
+ model:
457
+ input_tokenizer: ${actor_rollout_ref.model.path}
458
+ path: ~/models/FsfairX-LLaMA3-RM-v0.1
459
+ external_lib: ${actor_rollout_ref.model.external_lib}
460
+ trust_remote_code: false
461
+ override_config: {}
462
+ use_shm: false
463
+ use_remove_padding: false
464
+ use_fused_kernels: ${actor_rollout_ref.model.use_fused_kernels}
465
+ fsdp_config:
466
+ _target_: verl.workers.config.FSDPEngineConfig
467
+ wrap_policy:
468
+ min_num_params: 0
469
+ param_offload: false
470
+ reshard_after_forward: true
471
+ fsdp_size: -1
472
+ forward_prefetch: false
473
+ micro_batch_size: null
474
+ micro_batch_size_per_gpu: null
475
+ max_length: null
476
+ use_dynamic_bsz: ${critic.use_dynamic_bsz}
477
+ forward_max_token_len_per_gpu: ${critic.forward_max_token_len_per_gpu}
478
+ reward_manager: naive
479
+ launch_reward_fn_async: false
480
+ sandbox_fusion:
481
+ url: null
482
+ max_concurrent: 64
483
+ memory_limit_mb: 1024
484
+ profiler:
485
+ _target_: verl.utils.profiler.ProfilerConfig
486
+ tool: ${oc.select:global_profiler.tool,null}
487
+ enable: false
488
+ all_ranks: false
489
+ ranks: []
490
+ save_path: ${oc.select:global_profiler.save_path,null}
491
+ tool_config: ${oc.select:actor_rollout_ref.actor.profiler.tool_config,null}
492
+ ulysses_sequence_parallel_size: 1
493
+ use_reward_loop: true
494
+ rollout:
495
+ _target_: verl.workers.config.RolloutConfig
496
+ name: ???
497
+ dtype: bfloat16
498
+ gpu_memory_utilization: 0.5
499
+ enforce_eager: true
500
+ cudagraph_capture_sizes: null
501
+ free_cache_engine: true
502
+ data_parallel_size: 1
503
+ expert_parallel_size: 1
504
+ tensor_model_parallel_size: 2
505
+ max_num_batched_tokens: 8192
506
+ max_model_len: null
507
+ max_num_seqs: 1024
508
+ load_format: auto
509
+ engine_kwargs: {}
510
+ limit_images: null
511
+ enable_chunked_prefill: true
512
+ enable_prefix_caching: true
513
+ disable_log_stats: true
514
+ skip_tokenizer_init: true
515
+ prompt_length: 512
516
+ response_length: 512
517
+ algorithm:
518
+ rollout_correction:
519
+ rollout_is: null
520
+ rollout_is_threshold: 2.0
521
+ rollout_rs: null
522
+ rollout_rs_threshold: null
523
+ rollout_rs_threshold_lower: null
524
+ rollout_token_veto_threshold: null
525
+ bypass_mode: false
526
+ use_policy_gradient: false
527
+ rollout_is_batch_normalize: false
528
+ _target_: verl.trainer.config.AlgoConfig
529
+ gamma: 1.0
530
+ lam: 1.0
531
+ adv_estimator: sgrpo
532
+ norm_adv_by_std_in_grpo: false
533
+ use_kl_in_reward: false
534
+ kl_penalty: kl
535
+ kl_ctrl:
536
+ _target_: verl.trainer.config.KLControlConfig
537
+ type: fixed
538
+ kl_coef: 0.001
539
+ horizon: 10000
540
+ target_kl: 0.1
541
+ use_pf_ppo: false
542
+ pf_ppo:
543
+ reweight_method: pow
544
+ weight_pow: 2.0
545
+ sgrpo:
546
+ num_truncations: 4
547
+ force_answer_prompt: Time is limited, stop thinking and start answering.\n</think>\n\n
548
+ answer_max_tokens: 256
549
+ min_truncation_ratio: 0.1
550
+ max_truncation_ratio: 0.9
551
+ custom_reward_function:
552
+ path: null
553
+ name: compute_score
554
+ trainer:
555
+ balance_batch: true
556
+ total_epochs: 15
557
+ total_training_steps: null
558
+ project_name: verl_sgrpo_example_gsm8k
559
+ experiment_name: qwen3_8b_sgrpo
560
+ logger:
561
+ - console
562
+ log_val_generations: 0
563
+ rollout_data_dir: ./outputs/sgrpo_samples
564
+ rollout_data_dump_freq: 20
565
+ validation_data_dir: null
566
+ nnodes: 1
567
+ n_gpus_per_node: 8
568
+ save_freq: 20
569
+ esi_redundant_time: 0
570
+ resume_mode: disable
571
+ resume_from_path: null
572
+ val_before_train: true
573
+ val_only: false
574
+ test_freq: 5
575
+ critic_warmup: 0
576
+ default_hdfs_dir: null
577
+ del_local_ckpt_after_load: false
578
+ default_local_dir: checkpoints/${trainer.project_name}/${trainer.experiment_name}
579
+ max_actor_ckpt_to_keep: null
580
+ max_critic_ckpt_to_keep: null
581
+ ray_wait_register_center_timeout: 300
582
+ device: cuda
583
+ use_legacy_worker_impl: auto
584
+ global_profiler:
585
+ _target_: verl.utils.profiler.ProfilerConfig
586
+ tool: null
587
+ steps: null
588
+ profile_continuous_steps: false
589
+ save_path: outputs/profile
590
+ global_tool_config:
591
+ nsys:
592
+ _target_: verl.utils.profiler.config.NsightToolConfig
593
+ discrete: false
594
+ controller_nsight_options:
595
+ trace: cuda,nvtx,cublas,ucx
596
+ cuda-memory-usage: 'true'
597
+ cuda-graph-trace: graph
598
+ worker_nsight_options:
599
+ trace: cuda,nvtx,cublas,ucx
600
+ cuda-memory-usage: 'true'
601
+ cuda-graph-trace: graph
602
+ capture-range: cudaProfilerApi
603
+ capture-range-end: null
604
+ kill: none
605
+ torch_memory:
606
+ trace_alloc_max_entries: 100000
607
+ stack_depth: 32
608
+ context: all
609
+ stacks: all
610
+ kw_args: {}
611
+ transfer_queue:
612
+ enable: false
613
+ ray_kwargs:
614
+ ray_init:
615
+ num_cpus: null
616
+ timeline_json_file: null
outputs/2026-01-14/16-00-30/.hydra/overrides.yaml ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ - algorithm.adv_estimator=sgrpo
2
+ - algorithm.norm_adv_by_std_in_grpo=False
3
+ - algorithm.use_kl_in_reward=False
4
+ - data.train_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/train.parquet
5
+ - data.val_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/test.parquet
6
+ - data.train_batch_size=1024
7
+ - data.max_prompt_length=512
8
+ - data.max_response_length=1024
9
+ - data.filter_overlong_prompts=True
10
+ - data.truncation=error
11
+ - data.return_raw_chat=True
12
+ - actor_rollout_ref.model.path=/mnt/tidal-alsh01/dataset/redtrans/zhangruiqi/models/Qwen3-4B
13
+ - actor_rollout_ref.actor.optim.lr=1e-6
14
+ - actor_rollout_ref.model.use_remove_padding=True
15
+ - actor_rollout_ref.actor.ppo_mini_batch_size=256
16
+ - actor_rollout_ref.actor.ppo_micro_batch_size_per_gpu=32
17
+ - actor_rollout_ref.actor.use_kl_loss=True
18
+ - actor_rollout_ref.actor.kl_loss_coef=0.001
19
+ - actor_rollout_ref.actor.kl_loss_type=low_var_kl
20
+ - actor_rollout_ref.actor.entropy_coeff=0
21
+ - actor_rollout_ref.model.enable_gradient_checkpointing=True
22
+ - actor_rollout_ref.actor.fsdp_config.param_offload=False
23
+ - actor_rollout_ref.actor.fsdp_config.optimizer_offload=False
24
+ - actor_rollout_ref.rollout.log_prob_micro_batch_size_per_gpu=32
25
+ - actor_rollout_ref.rollout.tensor_model_parallel_size=2
26
+ - actor_rollout_ref.rollout.name=vllm
27
+ - actor_rollout_ref.rollout.gpu_memory_utilization=0.6
28
+ - actor_rollout_ref.rollout.n=1
29
+ - actor_rollout_ref.rollout.prompt_length=1536
30
+ - actor_rollout_ref.rollout.response_length=512
31
+ - actor_rollout_ref.ref.log_prob_micro_batch_size_per_gpu=32
32
+ - actor_rollout_ref.ref.fsdp_config.param_offload=False
33
+ - sgrpo.num_truncations=4
34
+ - sgrpo.force_answer_prompt="Time is limited, stop thinking and start answering.\n</think>\n\n"
35
+ - sgrpo.answer_max_tokens=256
36
+ - trainer.critic_warmup=0
37
+ - trainer.logger=["console"]
38
+ - trainer.project_name=verl_sgrpo_example_gsm8k
39
+ - trainer.experiment_name=qwen3_8b_sgrpo
40
+ - trainer.resume_mode=disable
41
+ - trainer.n_gpus_per_node=8
42
+ - trainer.nnodes=1
43
+ - trainer.save_freq=20
44
+ - trainer.test_freq=5
45
+ - trainer.total_epochs=15
46
+ - trainer.rollout_data_dir=./outputs/sgrpo_samples
47
+ - trainer.rollout_data_dump_freq=20
outputs/2026-01-14/16-00-30/main_sgrpo.log ADDED
File without changes
outputs/2026-01-14/16-50-33/.hydra/config.yaml ADDED
@@ -0,0 +1,616 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ actor_rollout_ref:
2
+ actor:
3
+ optim:
4
+ _target_: verl.workers.config.FSDPOptimizerConfig
5
+ optimizer: AdamW
6
+ optimizer_impl: torch.optim
7
+ lr: 1.0e-06
8
+ lr_warmup_steps_ratio: 0.0
9
+ total_training_steps: -1
10
+ weight_decay: 0.01
11
+ lr_warmup_steps: -1
12
+ betas:
13
+ - 0.9
14
+ - 0.999
15
+ clip_grad: 1.0
16
+ min_lr_ratio: 0.0
17
+ num_cycles: 0.5
18
+ lr_scheduler_type: constant
19
+ warmup_style: null
20
+ override_optimizer_config: null
21
+ fsdp_config:
22
+ _target_: verl.workers.config.FSDPEngineConfig
23
+ wrap_policy:
24
+ min_num_params: 0
25
+ param_offload: false
26
+ optimizer_offload: false
27
+ offload_policy: false
28
+ reshard_after_forward: true
29
+ fsdp_size: -1
30
+ forward_prefetch: false
31
+ model_dtype: fp32
32
+ use_orig_params: false
33
+ seed: 42
34
+ full_determinism: false
35
+ ulysses_sequence_parallel_size: 1
36
+ entropy_from_logits_with_chunking: false
37
+ use_torch_compile: true
38
+ entropy_checkpointing: false
39
+ forward_only: false
40
+ strategy: fsdp
41
+ dtype: bfloat16
42
+ _target_: verl.workers.config.FSDPActorConfig
43
+ rollout_n: ${oc.select:actor_rollout_ref.rollout.n,1}
44
+ strategy: fsdp
45
+ ppo_mini_batch_size: 256
46
+ ppo_micro_batch_size: null
47
+ ppo_micro_batch_size_per_gpu: 32
48
+ use_dynamic_bsz: false
49
+ ppo_max_token_len_per_gpu: 16384
50
+ clip_ratio: 0.2
51
+ clip_ratio_low: 0.2
52
+ clip_ratio_high: 0.2
53
+ freeze_vision_tower: false
54
+ policy_loss:
55
+ _target_: verl.workers.config.PolicyLossConfig
56
+ loss_mode: vanilla
57
+ clip_cov_ratio: 0.0002
58
+ clip_cov_lb: 1.0
59
+ clip_cov_ub: 5.0
60
+ kl_cov_ratio: 0.0002
61
+ ppo_kl_coef: 0.1
62
+ clip_ratio_c: 3.0
63
+ loss_agg_mode: token-mean
64
+ loss_scale_factor: null
65
+ entropy_coeff: 0
66
+ calculate_entropy: false
67
+ use_kl_loss: true
68
+ use_torch_compile: true
69
+ kl_loss_coef: 0.001
70
+ kl_loss_type: low_var_kl
71
+ ppo_epochs: 1
72
+ shuffle: false
73
+ checkpoint:
74
+ _target_: verl.trainer.config.CheckpointConfig
75
+ save_contents:
76
+ - model
77
+ - optimizer
78
+ - extra
79
+ load_contents: ${.save_contents}
80
+ async_save: false
81
+ use_fused_kernels: ${oc.select:actor_rollout_ref.model.use_fused_kernels,false}
82
+ profiler:
83
+ _target_: verl.utils.profiler.ProfilerConfig
84
+ tool: ${oc.select:global_profiler.tool,null}
85
+ enable: false
86
+ all_ranks: false
87
+ ranks: []
88
+ save_path: ${oc.select:global_profiler.save_path,null}
89
+ tool_config:
90
+ nsys:
91
+ _target_: verl.utils.profiler.config.NsightToolConfig
92
+ discrete: ${oc.select:global_profiler.global_tool_config.nsys.discrete}
93
+ npu:
94
+ _target_: verl.utils.profiler.config.NPUToolConfig
95
+ contents: []
96
+ level: level1
97
+ analysis: true
98
+ discrete: false
99
+ torch:
100
+ _target_: verl.utils.profiler.config.TorchProfilerToolConfig
101
+ step_start: 0
102
+ step_end: null
103
+ torch_memory:
104
+ _target_: verl.utils.profiler.config.TorchMemoryToolConfig
105
+ trace_alloc_max_entries: ${oc.select:global_profiler.global_tool_config.torch_memory.trace_alloc_max_entries,100000}
106
+ stack_depth: ${oc.select:global_profiler.global_tool_config.torch_memory.stack_depth,32}
107
+ router_replay:
108
+ _target_: verl.workers.config.RouterReplayConfig
109
+ mode: disabled
110
+ record_file: null
111
+ replay_file: null
112
+ grad_clip: 1.0
113
+ ulysses_sequence_parallel_size: 1
114
+ entropy_from_logits_with_chunking: false
115
+ entropy_checkpointing: false
116
+ use_remove_padding: ${oc.select:actor_rollout_ref.model.use_remove_padding,false}
117
+ ref:
118
+ rollout_n: ${oc.select:actor_rollout_ref.rollout.n,1}
119
+ strategy: ${actor_rollout_ref.actor.strategy}
120
+ use_torch_compile: ${oc.select:actor_rollout_ref.actor.use_torch_compile,true}
121
+ log_prob_micro_batch_size: null
122
+ log_prob_micro_batch_size_per_gpu: 32
123
+ log_prob_use_dynamic_bsz: ${oc.select:actor_rollout_ref.actor.use_dynamic_bsz,false}
124
+ log_prob_max_token_len_per_gpu: ${oc.select:actor_rollout_ref.actor.ppo_max_token_len_per_gpu,16384}
125
+ profiler:
126
+ _target_: verl.utils.profiler.ProfilerConfig
127
+ tool: ${oc.select:global_profiler.tool,null}
128
+ enable: false
129
+ all_ranks: false
130
+ ranks: []
131
+ save_path: ${oc.select:global_profiler.save_path,null}
132
+ tool_config:
133
+ nsys:
134
+ _target_: verl.utils.profiler.config.NsightToolConfig
135
+ discrete: ${oc.select:global_profiler.global_tool_config.nsys.discrete}
136
+ npu:
137
+ _target_: verl.utils.profiler.config.NPUToolConfig
138
+ contents: []
139
+ level: level1
140
+ analysis: true
141
+ discrete: false
142
+ torch:
143
+ _target_: verl.utils.profiler.config.TorchProfilerToolConfig
144
+ step_start: 0
145
+ step_end: null
146
+ torch_memory:
147
+ _target_: verl.utils.profiler.config.TorchMemoryToolConfig
148
+ trace_alloc_max_entries: ${oc.select:global_profiler.global_tool_config.torch_memory.trace_alloc_max_entries,100000}
149
+ stack_depth: ${oc.select:global_profiler.global_tool_config.torch_memory.stack_depth,32}
150
+ router_replay:
151
+ _target_: verl.workers.config.RouterReplayConfig
152
+ mode: disabled
153
+ record_file: null
154
+ replay_file: null
155
+ fsdp_config:
156
+ _target_: verl.workers.config.FSDPEngineConfig
157
+ wrap_policy:
158
+ min_num_params: 0
159
+ param_offload: false
160
+ optimizer_offload: false
161
+ offload_policy: false
162
+ reshard_after_forward: true
163
+ fsdp_size: -1
164
+ forward_prefetch: false
165
+ model_dtype: fp32
166
+ use_orig_params: false
167
+ seed: 42
168
+ full_determinism: false
169
+ ulysses_sequence_parallel_size: 1
170
+ entropy_from_logits_with_chunking: false
171
+ use_torch_compile: true
172
+ entropy_checkpointing: false
173
+ forward_only: true
174
+ strategy: fsdp
175
+ dtype: bfloat16
176
+ _target_: verl.workers.config.FSDPActorConfig
177
+ ulysses_sequence_parallel_size: ${oc.select:actor_rollout_ref.actor.ulysses_sequence_parallel_size,1}
178
+ entropy_from_logits_with_chunking: false
179
+ entropy_checkpointing: false
180
+ rollout:
181
+ _target_: verl.workers.config.RolloutConfig
182
+ name: vllm
183
+ mode: async
184
+ temperature: 1.0
185
+ top_k: -1
186
+ top_p: 1
187
+ prompt_length: 1536
188
+ response_length: 512
189
+ dtype: bfloat16
190
+ gpu_memory_utilization: 0.6
191
+ ignore_eos: false
192
+ enforce_eager: false
193
+ cudagraph_capture_sizes: null
194
+ free_cache_engine: true
195
+ tensor_model_parallel_size: 2
196
+ data_parallel_size: 1
197
+ expert_parallel_size: 1
198
+ pipeline_model_parallel_size: 1
199
+ max_num_batched_tokens: 8192
200
+ max_model_len: null
201
+ max_num_seqs: 1024
202
+ enable_chunked_prefill: true
203
+ enable_prefix_caching: true
204
+ load_format: dummy
205
+ log_prob_micro_batch_size: null
206
+ log_prob_micro_batch_size_per_gpu: 32
207
+ log_prob_use_dynamic_bsz: ${oc.select:actor_rollout_ref.actor.use_dynamic_bsz,false}
208
+ log_prob_max_token_len_per_gpu: ${oc.select:actor_rollout_ref.actor.ppo_max_token_len_per_gpu,16384}
209
+ disable_log_stats: true
210
+ do_sample: true
211
+ 'n': 1
212
+ over_sample_rate: 0
213
+ multi_stage_wake_up: false
214
+ engine_kwargs:
215
+ vllm: {}
216
+ sglang: {}
217
+ val_kwargs:
218
+ _target_: verl.workers.config.SamplingConfig
219
+ top_k: -1
220
+ top_p: 1.0
221
+ temperature: 0
222
+ 'n': 1
223
+ do_sample: false
224
+ multi_turn:
225
+ _target_: verl.workers.config.MultiTurnConfig
226
+ enable: false
227
+ max_assistant_turns: null
228
+ tool_config_path: null
229
+ max_user_turns: null
230
+ max_parallel_calls: 1
231
+ max_tool_response_length: 256
232
+ tool_response_truncate_side: middle
233
+ interaction_config_path: null
234
+ use_inference_chat_template: false
235
+ tokenization_sanity_check_mode: strict
236
+ format: hermes
237
+ num_repeat_rollouts: null
238
+ calculate_log_probs: false
239
+ agent:
240
+ _target_: verl.workers.config.AgentLoopConfig
241
+ num_workers: 8
242
+ default_agent_loop: single_turn_agent
243
+ agent_loop_config_path: null
244
+ custom_async_server:
245
+ _target_: verl.workers.config.CustomAsyncServerConfig
246
+ path: null
247
+ name: null
248
+ update_weights_bucket_megabytes: 512
249
+ trace:
250
+ _target_: verl.workers.config.TraceConfig
251
+ backend: null
252
+ token2text: false
253
+ max_samples_per_step_per_worker: null
254
+ skip_rollout: false
255
+ skip_dump_dir: /tmp/rollout_dump
256
+ skip_tokenizer_init: true
257
+ enable_rollout_routing_replay: false
258
+ profiler:
259
+ _target_: verl.utils.profiler.ProfilerConfig
260
+ tool: ${oc.select:global_profiler.tool,null}
261
+ enable: ${oc.select:actor_rollout_ref.actor.profiler.enable,false}
262
+ all_ranks: ${oc.select:actor_rollout_ref.actor.profiler.all_ranks,false}
263
+ ranks: ${oc.select:actor_rollout_ref.actor.profiler.ranks,[]}
264
+ save_path: ${oc.select:global_profiler.save_path,null}
265
+ tool_config: ${oc.select:actor_rollout_ref.actor.profiler.tool_config,null}
266
+ prometheus:
267
+ _target_: verl.workers.config.PrometheusConfig
268
+ enable: false
269
+ port: 9090
270
+ file: /tmp/ray/session_latest/metrics/prometheus/prometheus.yml
271
+ served_model_name: ${oc.select:actor_rollout_ref.model.path,null}
272
+ layered_summon: false
273
+ model:
274
+ _target_: verl.workers.config.HFModelConfig
275
+ path: /mnt/tidal-alsh01/dataset/redtrans/zhangruiqi/models/Qwen3-4B
276
+ hf_config_path: null
277
+ tokenizer_path: null
278
+ use_shm: false
279
+ trust_remote_code: false
280
+ custom_chat_template: null
281
+ external_lib: null
282
+ override_config: {}
283
+ enable_gradient_checkpointing: true
284
+ enable_activation_offload: false
285
+ use_remove_padding: true
286
+ lora_rank: 0
287
+ lora_alpha: 16
288
+ target_modules: all-linear
289
+ exclude_modules: null
290
+ lora_adapter_path: null
291
+ use_liger: false
292
+ use_fused_kernels: false
293
+ fused_kernel_options:
294
+ impl_backend: torch
295
+ hybrid_engine: true
296
+ nccl_timeout: 600
297
+ data:
298
+ tokenizer: null
299
+ use_shm: false
300
+ train_files: /mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/train.parquet
301
+ val_files: /mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/test.parquet
302
+ train_max_samples: -1
303
+ val_max_samples: -1
304
+ prompt_key: prompt
305
+ reward_fn_key: data_source
306
+ max_prompt_length: 512
307
+ max_response_length: 1024
308
+ train_batch_size: 1024
309
+ val_batch_size: null
310
+ tool_config_path: ${oc.select:actor_rollout_ref.rollout.multi_turn.tool_config_path,
311
+ null}
312
+ return_raw_input_ids: false
313
+ return_raw_chat: true
314
+ return_full_prompt: false
315
+ shuffle: true
316
+ seed: null
317
+ dataloader_num_workers: 8
318
+ image_patch_size: 14
319
+ validation_shuffle: false
320
+ filter_overlong_prompts: true
321
+ filter_overlong_prompts_workers: 1
322
+ truncation: error
323
+ image_key: images
324
+ video_key: videos
325
+ trust_remote_code: false
326
+ custom_cls:
327
+ path: null
328
+ name: null
329
+ return_multi_modal_inputs: true
330
+ sampler:
331
+ class_path: null
332
+ class_name: null
333
+ datagen:
334
+ path: null
335
+ name: null
336
+ apply_chat_template_kwargs: {}
337
+ reward_manager:
338
+ _target_: verl.trainer.config.config.RewardManagerConfig
339
+ source: register
340
+ name: ${oc.select:reward_model.reward_manager,naive}
341
+ module:
342
+ _target_: verl.trainer.config.config.ModuleConfig
343
+ path: null
344
+ name: custom_reward_manager
345
+ critic:
346
+ optim:
347
+ _target_: verl.workers.config.FSDPOptimizerConfig
348
+ optimizer: AdamW
349
+ optimizer_impl: torch.optim
350
+ lr: 1.0e-05
351
+ lr_warmup_steps_ratio: 0.0
352
+ total_training_steps: -1
353
+ weight_decay: 0.01
354
+ lr_warmup_steps: -1
355
+ betas:
356
+ - 0.9
357
+ - 0.999
358
+ clip_grad: 1.0
359
+ min_lr_ratio: 0.0
360
+ num_cycles: 0.5
361
+ lr_scheduler_type: constant
362
+ warmup_style: null
363
+ override_optimizer_config: null
364
+ model:
365
+ fsdp_config:
366
+ _target_: verl.workers.config.FSDPEngineConfig
367
+ wrap_policy:
368
+ min_num_params: 0
369
+ param_offload: false
370
+ optimizer_offload: false
371
+ offload_policy: false
372
+ reshard_after_forward: true
373
+ fsdp_size: -1
374
+ forward_prefetch: false
375
+ model_dtype: fp32
376
+ use_orig_params: false
377
+ seed: 42
378
+ full_determinism: false
379
+ ulysses_sequence_parallel_size: 1
380
+ entropy_from_logits_with_chunking: false
381
+ use_torch_compile: true
382
+ entropy_checkpointing: false
383
+ forward_only: false
384
+ strategy: fsdp
385
+ dtype: bfloat16
386
+ path: ~/models/deepseek-llm-7b-chat
387
+ tokenizer_path: ${oc.select:actor_rollout_ref.model.path,"~/models/deepseek-llm-7b-chat"}
388
+ override_config: {}
389
+ external_lib: ${oc.select:actor_rollout_ref.model.external_lib,null}
390
+ trust_remote_code: ${oc.select:actor_rollout_ref.model.trust_remote_code,false}
391
+ _target_: verl.workers.config.FSDPCriticModelCfg
392
+ use_shm: false
393
+ enable_gradient_checkpointing: true
394
+ enable_activation_offload: false
395
+ use_remove_padding: false
396
+ lora_rank: 0
397
+ lora_alpha: 16
398
+ target_modules: all-linear
399
+ _target_: verl.workers.config.FSDPCriticConfig
400
+ rollout_n: ${oc.select:actor_rollout_ref.rollout.n,1}
401
+ strategy: fsdp
402
+ enable: null
403
+ ppo_mini_batch_size: ${oc.select:actor_rollout_ref.actor.ppo_mini_batch_size,256}
404
+ ppo_micro_batch_size: null
405
+ ppo_micro_batch_size_per_gpu: ${oc.select:.ppo_micro_batch_size,null}
406
+ use_dynamic_bsz: ${oc.select:actor_rollout_ref.actor.use_dynamic_bsz,false}
407
+ ppo_max_token_len_per_gpu: 32768
408
+ forward_max_token_len_per_gpu: ${.ppo_max_token_len_per_gpu}
409
+ ppo_epochs: ${oc.select:actor_rollout_ref.actor.ppo_epochs,1}
410
+ shuffle: ${oc.select:actor_rollout_ref.actor.shuffle,false}
411
+ cliprange_value: 0.5
412
+ loss_agg_mode: ${oc.select:actor_rollout_ref.actor.loss_agg_mode,token-mean}
413
+ checkpoint:
414
+ _target_: verl.trainer.config.CheckpointConfig
415
+ save_contents:
416
+ - model
417
+ - optimizer
418
+ - extra
419
+ load_contents: ${.save_contents}
420
+ async_save: false
421
+ profiler:
422
+ _target_: verl.utils.profiler.ProfilerConfig
423
+ tool: ${oc.select:global_profiler.tool,null}
424
+ enable: false
425
+ all_ranks: false
426
+ ranks: []
427
+ save_path: ${oc.select:global_profiler.save_path,null}
428
+ tool_config:
429
+ nsys:
430
+ _target_: verl.utils.profiler.config.NsightToolConfig
431
+ discrete: ${oc.select:global_profiler.global_tool_config.nsys.discrete}
432
+ npu:
433
+ _target_: verl.utils.profiler.config.NPUToolConfig
434
+ contents: []
435
+ level: level1
436
+ analysis: true
437
+ discrete: false
438
+ torch:
439
+ _target_: verl.utils.profiler.config.TorchProfilerToolConfig
440
+ step_start: 0
441
+ step_end: null
442
+ torch_memory:
443
+ _target_: verl.utils.profiler.config.TorchMemoryToolConfig
444
+ trace_alloc_max_entries: ${oc.select:global_profiler.global_tool_config.torch_memory.trace_alloc_max_entries,100000}
445
+ stack_depth: ${oc.select:global_profiler.global_tool_config.torch_memory.stack_depth,32}
446
+ forward_micro_batch_size: ${oc.select:.ppo_micro_batch_size,null}
447
+ forward_micro_batch_size_per_gpu: ${oc.select:.ppo_micro_batch_size_per_gpu,null}
448
+ ulysses_sequence_parallel_size: 1
449
+ grad_clip: 1.0
450
+ reward_model:
451
+ enable: false
452
+ enable_resource_pool: false
453
+ n_gpus_per_node: 0
454
+ nnodes: 0
455
+ strategy: fsdp
456
+ model:
457
+ input_tokenizer: ${actor_rollout_ref.model.path}
458
+ path: ~/models/FsfairX-LLaMA3-RM-v0.1
459
+ external_lib: ${actor_rollout_ref.model.external_lib}
460
+ trust_remote_code: false
461
+ override_config: {}
462
+ use_shm: false
463
+ use_remove_padding: false
464
+ use_fused_kernels: ${actor_rollout_ref.model.use_fused_kernels}
465
+ fsdp_config:
466
+ _target_: verl.workers.config.FSDPEngineConfig
467
+ wrap_policy:
468
+ min_num_params: 0
469
+ param_offload: false
470
+ reshard_after_forward: true
471
+ fsdp_size: -1
472
+ forward_prefetch: false
473
+ micro_batch_size: null
474
+ micro_batch_size_per_gpu: null
475
+ max_length: null
476
+ use_dynamic_bsz: ${critic.use_dynamic_bsz}
477
+ forward_max_token_len_per_gpu: ${critic.forward_max_token_len_per_gpu}
478
+ reward_manager: naive
479
+ launch_reward_fn_async: false
480
+ sandbox_fusion:
481
+ url: null
482
+ max_concurrent: 64
483
+ memory_limit_mb: 1024
484
+ profiler:
485
+ _target_: verl.utils.profiler.ProfilerConfig
486
+ tool: ${oc.select:global_profiler.tool,null}
487
+ enable: false
488
+ all_ranks: false
489
+ ranks: []
490
+ save_path: ${oc.select:global_profiler.save_path,null}
491
+ tool_config: ${oc.select:actor_rollout_ref.actor.profiler.tool_config,null}
492
+ ulysses_sequence_parallel_size: 1
493
+ use_reward_loop: true
494
+ rollout:
495
+ _target_: verl.workers.config.RolloutConfig
496
+ name: ???
497
+ dtype: bfloat16
498
+ gpu_memory_utilization: 0.5
499
+ enforce_eager: true
500
+ cudagraph_capture_sizes: null
501
+ free_cache_engine: true
502
+ data_parallel_size: 1
503
+ expert_parallel_size: 1
504
+ tensor_model_parallel_size: 2
505
+ max_num_batched_tokens: 8192
506
+ max_model_len: null
507
+ max_num_seqs: 1024
508
+ load_format: auto
509
+ engine_kwargs: {}
510
+ limit_images: null
511
+ enable_chunked_prefill: true
512
+ enable_prefix_caching: true
513
+ disable_log_stats: true
514
+ skip_tokenizer_init: true
515
+ prompt_length: 512
516
+ response_length: 512
517
+ algorithm:
518
+ rollout_correction:
519
+ rollout_is: null
520
+ rollout_is_threshold: 2.0
521
+ rollout_rs: null
522
+ rollout_rs_threshold: null
523
+ rollout_rs_threshold_lower: null
524
+ rollout_token_veto_threshold: null
525
+ bypass_mode: false
526
+ use_policy_gradient: false
527
+ rollout_is_batch_normalize: false
528
+ _target_: verl.trainer.config.AlgoConfig
529
+ gamma: 1.0
530
+ lam: 1.0
531
+ adv_estimator: sgrpo
532
+ norm_adv_by_std_in_grpo: false
533
+ use_kl_in_reward: false
534
+ kl_penalty: kl
535
+ kl_ctrl:
536
+ _target_: verl.trainer.config.KLControlConfig
537
+ type: fixed
538
+ kl_coef: 0.001
539
+ horizon: 10000
540
+ target_kl: 0.1
541
+ use_pf_ppo: false
542
+ pf_ppo:
543
+ reweight_method: pow
544
+ weight_pow: 2.0
545
+ sgrpo:
546
+ num_truncations: 4
547
+ force_answer_prompt: Time is limited, stop thinking and start answering.\n</think>\n\n
548
+ answer_max_tokens: 256
549
+ min_truncation_ratio: 0.1
550
+ max_truncation_ratio: 0.9
551
+ custom_reward_function:
552
+ path: null
553
+ name: compute_score
554
+ trainer:
555
+ balance_batch: true
556
+ total_epochs: 15
557
+ total_training_steps: null
558
+ project_name: verl_sgrpo_example_gsm8k
559
+ experiment_name: qwen3_8b_sgrpo
560
+ logger:
561
+ - console
562
+ log_val_generations: 0
563
+ rollout_data_dir: ./outputs/sgrpo_samples
564
+ rollout_data_dump_freq: 20
565
+ validation_data_dir: null
566
+ nnodes: 1
567
+ n_gpus_per_node: 8
568
+ save_freq: 20
569
+ esi_redundant_time: 0
570
+ resume_mode: disable
571
+ resume_from_path: null
572
+ val_before_train: true
573
+ val_only: false
574
+ test_freq: 5
575
+ critic_warmup: 0
576
+ default_hdfs_dir: null
577
+ del_local_ckpt_after_load: false
578
+ default_local_dir: checkpoints/${trainer.project_name}/${trainer.experiment_name}
579
+ max_actor_ckpt_to_keep: null
580
+ max_critic_ckpt_to_keep: null
581
+ ray_wait_register_center_timeout: 300
582
+ device: cuda
583
+ use_legacy_worker_impl: auto
584
+ global_profiler:
585
+ _target_: verl.utils.profiler.ProfilerConfig
586
+ tool: null
587
+ steps: null
588
+ profile_continuous_steps: false
589
+ save_path: outputs/profile
590
+ global_tool_config:
591
+ nsys:
592
+ _target_: verl.utils.profiler.config.NsightToolConfig
593
+ discrete: false
594
+ controller_nsight_options:
595
+ trace: cuda,nvtx,cublas,ucx
596
+ cuda-memory-usage: 'true'
597
+ cuda-graph-trace: graph
598
+ worker_nsight_options:
599
+ trace: cuda,nvtx,cublas,ucx
600
+ cuda-memory-usage: 'true'
601
+ cuda-graph-trace: graph
602
+ capture-range: cudaProfilerApi
603
+ capture-range-end: null
604
+ kill: none
605
+ torch_memory:
606
+ trace_alloc_max_entries: 100000
607
+ stack_depth: 32
608
+ context: all
609
+ stacks: all
610
+ kw_args: {}
611
+ transfer_queue:
612
+ enable: false
613
+ ray_kwargs:
614
+ ray_init:
615
+ num_cpus: null
616
+ timeline_json_file: null
outputs/2026-01-14/16-54-53/.hydra/overrides.yaml ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ - algorithm.adv_estimator=sgrpo
2
+ - algorithm.norm_adv_by_std_in_grpo=False
3
+ - algorithm.use_kl_in_reward=False
4
+ - data.train_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/train.parquet
5
+ - data.val_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/test.parquet
6
+ - data.train_batch_size=1024
7
+ - data.max_prompt_length=512
8
+ - data.max_response_length=1024
9
+ - data.filter_overlong_prompts=True
10
+ - data.truncation=error
11
+ - data.return_raw_chat=True
12
+ - actor_rollout_ref.model.path=/mnt/tidal-alsh01/dataset/redtrans/zhangruiqi/models/Qwen3-4B
13
+ - actor_rollout_ref.actor.optim.lr=1e-6
14
+ - actor_rollout_ref.model.use_remove_padding=True
15
+ - actor_rollout_ref.actor.ppo_mini_batch_size=256
16
+ - actor_rollout_ref.actor.ppo_micro_batch_size_per_gpu=32
17
+ - actor_rollout_ref.actor.use_kl_loss=True
18
+ - actor_rollout_ref.actor.kl_loss_coef=0.001
19
+ - actor_rollout_ref.actor.kl_loss_type=low_var_kl
20
+ - actor_rollout_ref.actor.entropy_coeff=0
21
+ - actor_rollout_ref.model.enable_gradient_checkpointing=True
22
+ - actor_rollout_ref.actor.fsdp_config.param_offload=False
23
+ - actor_rollout_ref.actor.fsdp_config.optimizer_offload=False
24
+ - actor_rollout_ref.rollout.log_prob_micro_batch_size_per_gpu=32
25
+ - actor_rollout_ref.rollout.tensor_model_parallel_size=2
26
+ - actor_rollout_ref.rollout.name=vllm
27
+ - actor_rollout_ref.rollout.gpu_memory_utilization=0.6
28
+ - actor_rollout_ref.rollout.n=1
29
+ - actor_rollout_ref.rollout.prompt_length=1536
30
+ - actor_rollout_ref.rollout.response_length=512
31
+ - actor_rollout_ref.ref.log_prob_micro_batch_size_per_gpu=32
32
+ - actor_rollout_ref.ref.fsdp_config.param_offload=False
33
+ - sgrpo.num_truncations=4
34
+ - sgrpo.force_answer_prompt="Time is limited, stop thinking and start answering.\n</think>\n\n"
35
+ - sgrpo.answer_max_tokens=256
36
+ - trainer.critic_warmup=0
37
+ - trainer.logger=["console"]
38
+ - trainer.project_name=verl_sgrpo_example_gsm8k
39
+ - trainer.experiment_name=qwen3_8b_sgrpo
40
+ - trainer.resume_mode=disable
41
+ - trainer.n_gpus_per_node=8
42
+ - trainer.nnodes=1
43
+ - trainer.save_freq=20
44
+ - trainer.test_freq=5
45
+ - trainer.total_epochs=15
46
+ - trainer.rollout_data_dir=./outputs/sgrpo_samples
47
+ - trainer.rollout_data_dump_freq=20
outputs/2026-01-14/16-54-53/main_sgrpo.log ADDED
File without changes
outputs/2026-01-14/17-46-32/.hydra/hydra.yaml ADDED
@@ -0,0 +1,218 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
4
+ sweep:
5
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.mode=RUN
114
+ task:
115
+ - algorithm.adv_estimator=sgrpo
116
+ - algorithm.norm_adv_by_std_in_grpo=False
117
+ - algorithm.use_kl_in_reward=False
118
+ - data.train_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/train.parquet
119
+ - data.val_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/test.parquet
120
+ - data.train_batch_size=1024
121
+ - data.max_prompt_length=512
122
+ - data.max_response_length=1024
123
+ - data.filter_overlong_prompts=True
124
+ - data.truncation=error
125
+ - data.return_raw_chat=True
126
+ - actor_rollout_ref.model.path=/mnt/tidal-alsh01/dataset/redtrans/zhangruiqi/models/Qwen3-4B
127
+ - actor_rollout_ref.actor.optim.lr=1e-6
128
+ - actor_rollout_ref.model.use_remove_padding=True
129
+ - actor_rollout_ref.actor.ppo_mini_batch_size=256
130
+ - actor_rollout_ref.actor.ppo_micro_batch_size_per_gpu=32
131
+ - actor_rollout_ref.actor.use_kl_loss=True
132
+ - actor_rollout_ref.actor.kl_loss_coef=0.001
133
+ - actor_rollout_ref.actor.kl_loss_type=low_var_kl
134
+ - actor_rollout_ref.actor.entropy_coeff=0
135
+ - actor_rollout_ref.model.enable_gradient_checkpointing=True
136
+ - actor_rollout_ref.actor.fsdp_config.param_offload=False
137
+ - actor_rollout_ref.actor.fsdp_config.optimizer_offload=False
138
+ - actor_rollout_ref.rollout.log_prob_micro_batch_size_per_gpu=32
139
+ - actor_rollout_ref.rollout.tensor_model_parallel_size=2
140
+ - actor_rollout_ref.rollout.name=vllm
141
+ - actor_rollout_ref.rollout.gpu_memory_utilization=0.6
142
+ - actor_rollout_ref.rollout.n=1
143
+ - actor_rollout_ref.rollout.prompt_length=1536
144
+ - actor_rollout_ref.rollout.response_length=512
145
+ - actor_rollout_ref.ref.log_prob_micro_batch_size_per_gpu=32
146
+ - actor_rollout_ref.ref.fsdp_config.param_offload=False
147
+ - sgrpo.num_truncations=4
148
+ - sgrpo.force_answer_prompt="Time is limited, stop thinking and start answering.\n</think>\n\n"
149
+ - sgrpo.answer_max_tokens=256
150
+ - trainer.critic_warmup=0
151
+ - trainer.logger=["console"]
152
+ - trainer.project_name=verl_sgrpo_example_gsm8k
153
+ - trainer.experiment_name=qwen3_8b_sgrpo
154
+ - trainer.resume_mode=disable
155
+ - trainer.n_gpus_per_node=8
156
+ - trainer.nnodes=1
157
+ - trainer.save_freq=20
158
+ - trainer.test_freq=5
159
+ - trainer.total_epochs=15
160
+ - trainer.rollout_data_dir=./outputs/sgrpo_samples
161
+ - trainer.rollout_data_dump_freq=20
162
+ job:
163
+ name: main_sgrpo
164
+ chdir: null
165
+ override_dirname: actor_rollout_ref.actor.entropy_coeff=0,actor_rollout_ref.actor.fsdp_config.optimizer_offload=False,actor_rollout_ref.actor.fsdp_config.param_offload=False,actor_rollout_ref.actor.kl_loss_coef=0.001,actor_rollout_ref.actor.kl_loss_type=low_var_kl,actor_rollout_ref.actor.optim.lr=1e-6,actor_rollout_ref.actor.ppo_micro_batch_size_per_gpu=32,actor_rollout_ref.actor.ppo_mini_batch_size=256,actor_rollout_ref.actor.use_kl_loss=True,actor_rollout_ref.model.enable_gradient_checkpointing=True,actor_rollout_ref.model.path=/mnt/tidal-alsh01/dataset/redtrans/zhangruiqi/models/Qwen3-4B,actor_rollout_ref.model.use_remove_padding=True,actor_rollout_ref.ref.fsdp_config.param_offload=False,actor_rollout_ref.ref.log_prob_micro_batch_size_per_gpu=32,actor_rollout_ref.rollout.gpu_memory_utilization=0.6,actor_rollout_ref.rollout.log_prob_micro_batch_size_per_gpu=32,actor_rollout_ref.rollout.n=1,actor_rollout_ref.rollout.name=vllm,actor_rollout_ref.rollout.prompt_length=1536,actor_rollout_ref.rollout.response_length=512,actor_rollout_ref.rollout.tensor_model_parallel_size=2,algorithm.adv_estimator=sgrpo,algorithm.norm_adv_by_std_in_grpo=False,algorithm.use_kl_in_reward=False,data.filter_overlong_prompts=True,data.max_prompt_length=512,data.max_response_length=1024,data.return_raw_chat=True,data.train_batch_size=1024,data.train_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/train.parquet,data.truncation=error,data.val_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/test.parquet,sgrpo.answer_max_tokens=256,sgrpo.force_answer_prompt="Time
166
+ is limited, stop thinking and start answering.\n</think>\n\n",sgrpo.num_truncations=4,trainer.critic_warmup=0,trainer.experiment_name=qwen3_8b_sgrpo,trainer.logger=["console"],trainer.n_gpus_per_node=8,trainer.nnodes=1,trainer.project_name=verl_sgrpo_example_gsm8k,trainer.resume_mode=disable,trainer.rollout_data_dir=./outputs/sgrpo_samples,trainer.rollout_data_dump_freq=20,trainer.save_freq=20,trainer.test_freq=5,trainer.total_epochs=15
167
+ id: ???
168
+ num: ???
169
+ config_name: sgrpo_trainer
170
+ env_set: {}
171
+ env_copy: []
172
+ config:
173
+ override_dirname:
174
+ kv_sep: '='
175
+ item_sep: ','
176
+ exclude_keys: []
177
+ runtime:
178
+ version: 1.3.2
179
+ version_base: '1.3'
180
+ cwd: /mnt/tidal-alsh01/usr/zhangruiqi1/my/verl
181
+ config_sources:
182
+ - path: hydra.conf
183
+ schema: pkg
184
+ provider: hydra
185
+ - path: /mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/config
186
+ schema: file
187
+ provider: main
188
+ - path: verl.trainer.config
189
+ schema: pkg
190
+ provider: verl-trainer-config
191
+ - path: ''
192
+ schema: structured
193
+ provider: schema
194
+ output_dir: /mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/outputs/2026-01-14/17-46-32
195
+ choices:
196
+ algorithm@algorithm.rollout_correction: rollout_correction
197
+ reward_model: dp_reward_loop
198
+ critic: dp_critic
199
+ critic/../engine@critic.model.fsdp_config: fsdp
200
+ critic/../optim@critic.optim: fsdp
201
+ model@actor_rollout_ref.model: hf_model
202
+ rollout@actor_rollout_ref.rollout: rollout
203
+ ref@actor_rollout_ref.ref: dp_ref
204
+ ref/../engine@actor_rollout_ref.ref.fsdp_config: fsdp
205
+ data: legacy_data
206
+ actor@actor_rollout_ref.actor: dp_actor
207
+ actor/../engine@actor_rollout_ref.actor.fsdp_config: fsdp
208
+ actor/../optim@actor_rollout_ref.actor.optim: fsdp
209
+ hydra/env: default
210
+ hydra/callbacks: null
211
+ hydra/job_logging: default
212
+ hydra/hydra_logging: default
213
+ hydra/hydra_help: default
214
+ hydra/help: default
215
+ hydra/sweeper: basic
216
+ hydra/launcher: basic
217
+ hydra/output: default
218
+ verbose: false
outputs/2026-01-14/17-46-32/.hydra/overrides.yaml ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ - algorithm.adv_estimator=sgrpo
2
+ - algorithm.norm_adv_by_std_in_grpo=False
3
+ - algorithm.use_kl_in_reward=False
4
+ - data.train_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/train.parquet
5
+ - data.val_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/test.parquet
6
+ - data.train_batch_size=1024
7
+ - data.max_prompt_length=512
8
+ - data.max_response_length=1024
9
+ - data.filter_overlong_prompts=True
10
+ - data.truncation=error
11
+ - data.return_raw_chat=True
12
+ - actor_rollout_ref.model.path=/mnt/tidal-alsh01/dataset/redtrans/zhangruiqi/models/Qwen3-4B
13
+ - actor_rollout_ref.actor.optim.lr=1e-6
14
+ - actor_rollout_ref.model.use_remove_padding=True
15
+ - actor_rollout_ref.actor.ppo_mini_batch_size=256
16
+ - actor_rollout_ref.actor.ppo_micro_batch_size_per_gpu=32
17
+ - actor_rollout_ref.actor.use_kl_loss=True
18
+ - actor_rollout_ref.actor.kl_loss_coef=0.001
19
+ - actor_rollout_ref.actor.kl_loss_type=low_var_kl
20
+ - actor_rollout_ref.actor.entropy_coeff=0
21
+ - actor_rollout_ref.model.enable_gradient_checkpointing=True
22
+ - actor_rollout_ref.actor.fsdp_config.param_offload=False
23
+ - actor_rollout_ref.actor.fsdp_config.optimizer_offload=False
24
+ - actor_rollout_ref.rollout.log_prob_micro_batch_size_per_gpu=32
25
+ - actor_rollout_ref.rollout.tensor_model_parallel_size=2
26
+ - actor_rollout_ref.rollout.name=vllm
27
+ - actor_rollout_ref.rollout.gpu_memory_utilization=0.6
28
+ - actor_rollout_ref.rollout.n=1
29
+ - actor_rollout_ref.rollout.prompt_length=1536
30
+ - actor_rollout_ref.rollout.response_length=512
31
+ - actor_rollout_ref.ref.log_prob_micro_batch_size_per_gpu=32
32
+ - actor_rollout_ref.ref.fsdp_config.param_offload=False
33
+ - sgrpo.num_truncations=4
34
+ - sgrpo.force_answer_prompt="Time is limited, stop thinking and start answering.\n</think>\n\n"
35
+ - sgrpo.answer_max_tokens=256
36
+ - trainer.critic_warmup=0
37
+ - trainer.logger=["console"]
38
+ - trainer.project_name=verl_sgrpo_example_gsm8k
39
+ - trainer.experiment_name=qwen3_8b_sgrpo
40
+ - trainer.resume_mode=disable
41
+ - trainer.n_gpus_per_node=8
42
+ - trainer.nnodes=1
43
+ - trainer.save_freq=20
44
+ - trainer.test_freq=5
45
+ - trainer.total_epochs=15
46
+ - trainer.rollout_data_dir=./outputs/sgrpo_samples
47
+ - trainer.rollout_data_dump_freq=20
outputs/2026-01-14/17-46-32/main_sgrpo.log ADDED
File without changes
outputs/2026-01-14/19-57-10/.hydra/overrides.yaml ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ - algorithm.adv_estimator=sgrpo
2
+ - algorithm.norm_adv_by_std_in_grpo=False
3
+ - algorithm.use_kl_in_reward=False
4
+ - data.train_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/train.parquet
5
+ - data.val_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/test.parquet
6
+ - data.train_batch_size=1024
7
+ - data.max_prompt_length=512
8
+ - data.max_response_length=1024
9
+ - data.filter_overlong_prompts=True
10
+ - data.truncation=error
11
+ - data.return_raw_chat=True
12
+ - actor_rollout_ref.model.path=/mnt/tidal-alsh01/dataset/redtrans/zhangruiqi/models/Qwen3-4B
13
+ - actor_rollout_ref.actor.optim.lr=1e-6
14
+ - actor_rollout_ref.model.use_remove_padding=True
15
+ - actor_rollout_ref.actor.ppo_mini_batch_size=256
16
+ - actor_rollout_ref.actor.ppo_micro_batch_size_per_gpu=32
17
+ - actor_rollout_ref.actor.use_kl_loss=True
18
+ - actor_rollout_ref.actor.kl_loss_coef=0.001
19
+ - actor_rollout_ref.actor.kl_loss_type=low_var_kl
20
+ - actor_rollout_ref.actor.entropy_coeff=0
21
+ - actor_rollout_ref.model.enable_gradient_checkpointing=True
22
+ - actor_rollout_ref.actor.fsdp_config.param_offload=False
23
+ - actor_rollout_ref.actor.fsdp_config.optimizer_offload=False
24
+ - actor_rollout_ref.rollout.log_prob_micro_batch_size_per_gpu=32
25
+ - actor_rollout_ref.rollout.tensor_model_parallel_size=2
26
+ - actor_rollout_ref.rollout.name=vllm
27
+ - actor_rollout_ref.rollout.gpu_memory_utilization=0.6
28
+ - actor_rollout_ref.rollout.n=1
29
+ - actor_rollout_ref.rollout.prompt_length=1536
30
+ - actor_rollout_ref.rollout.response_length=512
31
+ - actor_rollout_ref.ref.log_prob_micro_batch_size_per_gpu=32
32
+ - actor_rollout_ref.ref.fsdp_config.param_offload=False
33
+ - sgrpo.num_truncations=4
34
+ - sgrpo.force_answer_prompt="Time is limited, stop thinking and start answering.\n</think>\n\n"
35
+ - sgrpo.answer_max_tokens=256
36
+ - trainer.critic_warmup=0
37
+ - trainer.logger=["console"]
38
+ - trainer.project_name=verl_sgrpo_example_gsm8k
39
+ - trainer.experiment_name=qwen3_8b_sgrpo
40
+ - trainer.resume_mode=disable
41
+ - trainer.n_gpus_per_node=8
42
+ - trainer.nnodes=1
43
+ - trainer.save_freq=20
44
+ - trainer.test_freq=5
45
+ - trainer.total_epochs=15
46
+ - trainer.rollout_data_dir=./outputs/sgrpo_samples
47
+ - trainer.rollout_data_dump_freq=20
outputs/2026-01-14/19-57-10/main_sgrpo.log ADDED
File without changes
outputs/2026-01-14/19-57-55/.hydra/overrides.yaml ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ - algorithm.adv_estimator=sgrpo
2
+ - algorithm.norm_adv_by_std_in_grpo=False
3
+ - algorithm.use_kl_in_reward=False
4
+ - data.train_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/train.parquet
5
+ - data.val_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/test.parquet
6
+ - data.train_batch_size=1024
7
+ - data.max_prompt_length=512
8
+ - data.max_response_length=1024
9
+ - data.filter_overlong_prompts=True
10
+ - data.truncation=error
11
+ - data.return_raw_chat=True
12
+ - actor_rollout_ref.model.path=/mnt/tidal-alsh01/dataset/redtrans/zhangruiqi/models/Qwen3-4B
13
+ - actor_rollout_ref.actor.optim.lr=1e-6
14
+ - actor_rollout_ref.model.use_remove_padding=True
15
+ - actor_rollout_ref.actor.ppo_mini_batch_size=256
16
+ - actor_rollout_ref.actor.ppo_micro_batch_size_per_gpu=32
17
+ - actor_rollout_ref.actor.use_kl_loss=True
18
+ - actor_rollout_ref.actor.kl_loss_coef=0.001
19
+ - actor_rollout_ref.actor.kl_loss_type=low_var_kl
20
+ - actor_rollout_ref.actor.entropy_coeff=0
21
+ - actor_rollout_ref.model.enable_gradient_checkpointing=True
22
+ - actor_rollout_ref.actor.fsdp_config.param_offload=False
23
+ - actor_rollout_ref.actor.fsdp_config.optimizer_offload=False
24
+ - actor_rollout_ref.rollout.log_prob_micro_batch_size_per_gpu=32
25
+ - actor_rollout_ref.rollout.tensor_model_parallel_size=2
26
+ - actor_rollout_ref.rollout.name=vllm
27
+ - actor_rollout_ref.rollout.gpu_memory_utilization=0.6
28
+ - actor_rollout_ref.rollout.n=1
29
+ - actor_rollout_ref.rollout.prompt_length=1536
30
+ - actor_rollout_ref.rollout.response_length=512
31
+ - actor_rollout_ref.ref.log_prob_micro_batch_size_per_gpu=32
32
+ - actor_rollout_ref.ref.fsdp_config.param_offload=False
33
+ - sgrpo.num_truncations=4
34
+ - sgrpo.force_answer_prompt="Time is limited, stop thinking and start answering.\n</think>\n\n"
35
+ - sgrpo.answer_max_tokens=256
36
+ - trainer.critic_warmup=0
37
+ - trainer.logger=["console"]
38
+ - trainer.project_name=verl_sgrpo_example_gsm8k
39
+ - trainer.experiment_name=qwen3_8b_sgrpo
40
+ - trainer.resume_mode=disable
41
+ - trainer.n_gpus_per_node=8
42
+ - trainer.nnodes=1
43
+ - trainer.save_freq=20
44
+ - trainer.test_freq=5
45
+ - trainer.total_epochs=15
46
+ - trainer.rollout_data_dir=./outputs/sgrpo_samples
47
+ - trainer.rollout_data_dump_freq=20
outputs/2026-01-14/19-57-55/main_sgrpo.log ADDED
File without changes
outputs/2026-01-14/20-02-28/.hydra/config.yaml ADDED
@@ -0,0 +1,616 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ actor_rollout_ref:
2
+ actor:
3
+ optim:
4
+ _target_: verl.workers.config.FSDPOptimizerConfig
5
+ optimizer: AdamW
6
+ optimizer_impl: torch.optim
7
+ lr: 1.0e-06
8
+ lr_warmup_steps_ratio: 0.0
9
+ total_training_steps: -1
10
+ weight_decay: 0.01
11
+ lr_warmup_steps: -1
12
+ betas:
13
+ - 0.9
14
+ - 0.999
15
+ clip_grad: 1.0
16
+ min_lr_ratio: 0.0
17
+ num_cycles: 0.5
18
+ lr_scheduler_type: constant
19
+ warmup_style: null
20
+ override_optimizer_config: null
21
+ fsdp_config:
22
+ _target_: verl.workers.config.FSDPEngineConfig
23
+ wrap_policy:
24
+ min_num_params: 0
25
+ param_offload: false
26
+ optimizer_offload: false
27
+ offload_policy: false
28
+ reshard_after_forward: true
29
+ fsdp_size: -1
30
+ forward_prefetch: false
31
+ model_dtype: fp32
32
+ use_orig_params: false
33
+ seed: 42
34
+ full_determinism: false
35
+ ulysses_sequence_parallel_size: 1
36
+ entropy_from_logits_with_chunking: false
37
+ use_torch_compile: true
38
+ entropy_checkpointing: false
39
+ forward_only: false
40
+ strategy: fsdp
41
+ dtype: bfloat16
42
+ _target_: verl.workers.config.FSDPActorConfig
43
+ rollout_n: ${oc.select:actor_rollout_ref.rollout.n,1}
44
+ strategy: fsdp
45
+ ppo_mini_batch_size: 256
46
+ ppo_micro_batch_size: null
47
+ ppo_micro_batch_size_per_gpu: 32
48
+ use_dynamic_bsz: false
49
+ ppo_max_token_len_per_gpu: 16384
50
+ clip_ratio: 0.2
51
+ clip_ratio_low: 0.2
52
+ clip_ratio_high: 0.2
53
+ freeze_vision_tower: false
54
+ policy_loss:
55
+ _target_: verl.workers.config.PolicyLossConfig
56
+ loss_mode: vanilla
57
+ clip_cov_ratio: 0.0002
58
+ clip_cov_lb: 1.0
59
+ clip_cov_ub: 5.0
60
+ kl_cov_ratio: 0.0002
61
+ ppo_kl_coef: 0.1
62
+ clip_ratio_c: 3.0
63
+ loss_agg_mode: token-mean
64
+ loss_scale_factor: null
65
+ entropy_coeff: 0
66
+ calculate_entropy: false
67
+ use_kl_loss: true
68
+ use_torch_compile: true
69
+ kl_loss_coef: 0.001
70
+ kl_loss_type: low_var_kl
71
+ ppo_epochs: 1
72
+ shuffle: false
73
+ checkpoint:
74
+ _target_: verl.trainer.config.CheckpointConfig
75
+ save_contents:
76
+ - model
77
+ - optimizer
78
+ - extra
79
+ load_contents: ${.save_contents}
80
+ async_save: false
81
+ use_fused_kernels: ${oc.select:actor_rollout_ref.model.use_fused_kernels,false}
82
+ profiler:
83
+ _target_: verl.utils.profiler.ProfilerConfig
84
+ tool: ${oc.select:global_profiler.tool,null}
85
+ enable: false
86
+ all_ranks: false
87
+ ranks: []
88
+ save_path: ${oc.select:global_profiler.save_path,null}
89
+ tool_config:
90
+ nsys:
91
+ _target_: verl.utils.profiler.config.NsightToolConfig
92
+ discrete: ${oc.select:global_profiler.global_tool_config.nsys.discrete}
93
+ npu:
94
+ _target_: verl.utils.profiler.config.NPUToolConfig
95
+ contents: []
96
+ level: level1
97
+ analysis: true
98
+ discrete: false
99
+ torch:
100
+ _target_: verl.utils.profiler.config.TorchProfilerToolConfig
101
+ step_start: 0
102
+ step_end: null
103
+ torch_memory:
104
+ _target_: verl.utils.profiler.config.TorchMemoryToolConfig
105
+ trace_alloc_max_entries: ${oc.select:global_profiler.global_tool_config.torch_memory.trace_alloc_max_entries,100000}
106
+ stack_depth: ${oc.select:global_profiler.global_tool_config.torch_memory.stack_depth,32}
107
+ router_replay:
108
+ _target_: verl.workers.config.RouterReplayConfig
109
+ mode: disabled
110
+ record_file: null
111
+ replay_file: null
112
+ grad_clip: 1.0
113
+ ulysses_sequence_parallel_size: 1
114
+ entropy_from_logits_with_chunking: false
115
+ entropy_checkpointing: false
116
+ use_remove_padding: ${oc.select:actor_rollout_ref.model.use_remove_padding,false}
117
+ ref:
118
+ rollout_n: ${oc.select:actor_rollout_ref.rollout.n,1}
119
+ strategy: ${actor_rollout_ref.actor.strategy}
120
+ use_torch_compile: ${oc.select:actor_rollout_ref.actor.use_torch_compile,true}
121
+ log_prob_micro_batch_size: null
122
+ log_prob_micro_batch_size_per_gpu: 32
123
+ log_prob_use_dynamic_bsz: ${oc.select:actor_rollout_ref.actor.use_dynamic_bsz,false}
124
+ log_prob_max_token_len_per_gpu: ${oc.select:actor_rollout_ref.actor.ppo_max_token_len_per_gpu,16384}
125
+ profiler:
126
+ _target_: verl.utils.profiler.ProfilerConfig
127
+ tool: ${oc.select:global_profiler.tool,null}
128
+ enable: false
129
+ all_ranks: false
130
+ ranks: []
131
+ save_path: ${oc.select:global_profiler.save_path,null}
132
+ tool_config:
133
+ nsys:
134
+ _target_: verl.utils.profiler.config.NsightToolConfig
135
+ discrete: ${oc.select:global_profiler.global_tool_config.nsys.discrete}
136
+ npu:
137
+ _target_: verl.utils.profiler.config.NPUToolConfig
138
+ contents: []
139
+ level: level1
140
+ analysis: true
141
+ discrete: false
142
+ torch:
143
+ _target_: verl.utils.profiler.config.TorchProfilerToolConfig
144
+ step_start: 0
145
+ step_end: null
146
+ torch_memory:
147
+ _target_: verl.utils.profiler.config.TorchMemoryToolConfig
148
+ trace_alloc_max_entries: ${oc.select:global_profiler.global_tool_config.torch_memory.trace_alloc_max_entries,100000}
149
+ stack_depth: ${oc.select:global_profiler.global_tool_config.torch_memory.stack_depth,32}
150
+ router_replay:
151
+ _target_: verl.workers.config.RouterReplayConfig
152
+ mode: disabled
153
+ record_file: null
154
+ replay_file: null
155
+ fsdp_config:
156
+ _target_: verl.workers.config.FSDPEngineConfig
157
+ wrap_policy:
158
+ min_num_params: 0
159
+ param_offload: false
160
+ optimizer_offload: false
161
+ offload_policy: false
162
+ reshard_after_forward: true
163
+ fsdp_size: -1
164
+ forward_prefetch: false
165
+ model_dtype: fp32
166
+ use_orig_params: false
167
+ seed: 42
168
+ full_determinism: false
169
+ ulysses_sequence_parallel_size: 1
170
+ entropy_from_logits_with_chunking: false
171
+ use_torch_compile: true
172
+ entropy_checkpointing: false
173
+ forward_only: true
174
+ strategy: fsdp
175
+ dtype: bfloat16
176
+ _target_: verl.workers.config.FSDPActorConfig
177
+ ulysses_sequence_parallel_size: ${oc.select:actor_rollout_ref.actor.ulysses_sequence_parallel_size,1}
178
+ entropy_from_logits_with_chunking: false
179
+ entropy_checkpointing: false
180
+ rollout:
181
+ _target_: verl.workers.config.RolloutConfig
182
+ name: vllm
183
+ mode: async
184
+ temperature: 1.0
185
+ top_k: -1
186
+ top_p: 1
187
+ prompt_length: 1536
188
+ response_length: 512
189
+ dtype: bfloat16
190
+ gpu_memory_utilization: 0.6
191
+ ignore_eos: false
192
+ enforce_eager: false
193
+ cudagraph_capture_sizes: null
194
+ free_cache_engine: true
195
+ tensor_model_parallel_size: 2
196
+ data_parallel_size: 1
197
+ expert_parallel_size: 1
198
+ pipeline_model_parallel_size: 1
199
+ max_num_batched_tokens: 8192
200
+ max_model_len: null
201
+ max_num_seqs: 1024
202
+ enable_chunked_prefill: true
203
+ enable_prefix_caching: true
204
+ load_format: dummy
205
+ log_prob_micro_batch_size: null
206
+ log_prob_micro_batch_size_per_gpu: 32
207
+ log_prob_use_dynamic_bsz: ${oc.select:actor_rollout_ref.actor.use_dynamic_bsz,false}
208
+ log_prob_max_token_len_per_gpu: ${oc.select:actor_rollout_ref.actor.ppo_max_token_len_per_gpu,16384}
209
+ disable_log_stats: true
210
+ do_sample: true
211
+ 'n': 1
212
+ over_sample_rate: 0
213
+ multi_stage_wake_up: false
214
+ engine_kwargs:
215
+ vllm: {}
216
+ sglang: {}
217
+ val_kwargs:
218
+ _target_: verl.workers.config.SamplingConfig
219
+ top_k: -1
220
+ top_p: 1.0
221
+ temperature: 0
222
+ 'n': 1
223
+ do_sample: false
224
+ multi_turn:
225
+ _target_: verl.workers.config.MultiTurnConfig
226
+ enable: false
227
+ max_assistant_turns: null
228
+ tool_config_path: null
229
+ max_user_turns: null
230
+ max_parallel_calls: 1
231
+ max_tool_response_length: 256
232
+ tool_response_truncate_side: middle
233
+ interaction_config_path: null
234
+ use_inference_chat_template: false
235
+ tokenization_sanity_check_mode: strict
236
+ format: hermes
237
+ num_repeat_rollouts: null
238
+ calculate_log_probs: false
239
+ agent:
240
+ _target_: verl.workers.config.AgentLoopConfig
241
+ num_workers: 8
242
+ default_agent_loop: single_turn_agent
243
+ agent_loop_config_path: null
244
+ custom_async_server:
245
+ _target_: verl.workers.config.CustomAsyncServerConfig
246
+ path: null
247
+ name: null
248
+ update_weights_bucket_megabytes: 512
249
+ trace:
250
+ _target_: verl.workers.config.TraceConfig
251
+ backend: null
252
+ token2text: false
253
+ max_samples_per_step_per_worker: null
254
+ skip_rollout: false
255
+ skip_dump_dir: /tmp/rollout_dump
256
+ skip_tokenizer_init: true
257
+ enable_rollout_routing_replay: false
258
+ profiler:
259
+ _target_: verl.utils.profiler.ProfilerConfig
260
+ tool: ${oc.select:global_profiler.tool,null}
261
+ enable: ${oc.select:actor_rollout_ref.actor.profiler.enable,false}
262
+ all_ranks: ${oc.select:actor_rollout_ref.actor.profiler.all_ranks,false}
263
+ ranks: ${oc.select:actor_rollout_ref.actor.profiler.ranks,[]}
264
+ save_path: ${oc.select:global_profiler.save_path,null}
265
+ tool_config: ${oc.select:actor_rollout_ref.actor.profiler.tool_config,null}
266
+ prometheus:
267
+ _target_: verl.workers.config.PrometheusConfig
268
+ enable: false
269
+ port: 9090
270
+ file: /tmp/ray/session_latest/metrics/prometheus/prometheus.yml
271
+ served_model_name: ${oc.select:actor_rollout_ref.model.path,null}
272
+ layered_summon: false
273
+ model:
274
+ _target_: verl.workers.config.HFModelConfig
275
+ path: /mnt/tidal-alsh01/dataset/redtrans/zhangruiqi/models/Qwen3-4B
276
+ hf_config_path: null
277
+ tokenizer_path: null
278
+ use_shm: false
279
+ trust_remote_code: false
280
+ custom_chat_template: null
281
+ external_lib: null
282
+ override_config: {}
283
+ enable_gradient_checkpointing: true
284
+ enable_activation_offload: false
285
+ use_remove_padding: true
286
+ lora_rank: 0
287
+ lora_alpha: 16
288
+ target_modules: all-linear
289
+ exclude_modules: null
290
+ lora_adapter_path: null
291
+ use_liger: false
292
+ use_fused_kernels: false
293
+ fused_kernel_options:
294
+ impl_backend: torch
295
+ hybrid_engine: true
296
+ nccl_timeout: 600
297
+ data:
298
+ tokenizer: null
299
+ use_shm: false
300
+ train_files: /mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/train.parquet
301
+ val_files: /mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/test.parquet
302
+ train_max_samples: -1
303
+ val_max_samples: -1
304
+ prompt_key: prompt
305
+ reward_fn_key: data_source
306
+ max_prompt_length: 512
307
+ max_response_length: 1024
308
+ train_batch_size: 1024
309
+ val_batch_size: null
310
+ tool_config_path: ${oc.select:actor_rollout_ref.rollout.multi_turn.tool_config_path,
311
+ null}
312
+ return_raw_input_ids: false
313
+ return_raw_chat: true
314
+ return_full_prompt: false
315
+ shuffle: true
316
+ seed: null
317
+ dataloader_num_workers: 8
318
+ image_patch_size: 14
319
+ validation_shuffle: false
320
+ filter_overlong_prompts: true
321
+ filter_overlong_prompts_workers: 1
322
+ truncation: error
323
+ image_key: images
324
+ video_key: videos
325
+ trust_remote_code: false
326
+ custom_cls:
327
+ path: null
328
+ name: null
329
+ return_multi_modal_inputs: true
330
+ sampler:
331
+ class_path: null
332
+ class_name: null
333
+ datagen:
334
+ path: null
335
+ name: null
336
+ apply_chat_template_kwargs: {}
337
+ reward_manager:
338
+ _target_: verl.trainer.config.config.RewardManagerConfig
339
+ source: register
340
+ name: ${oc.select:reward_model.reward_manager,naive}
341
+ module:
342
+ _target_: verl.trainer.config.config.ModuleConfig
343
+ path: null
344
+ name: custom_reward_manager
345
+ critic:
346
+ optim:
347
+ _target_: verl.workers.config.FSDPOptimizerConfig
348
+ optimizer: AdamW
349
+ optimizer_impl: torch.optim
350
+ lr: 1.0e-05
351
+ lr_warmup_steps_ratio: 0.0
352
+ total_training_steps: -1
353
+ weight_decay: 0.01
354
+ lr_warmup_steps: -1
355
+ betas:
356
+ - 0.9
357
+ - 0.999
358
+ clip_grad: 1.0
359
+ min_lr_ratio: 0.0
360
+ num_cycles: 0.5
361
+ lr_scheduler_type: constant
362
+ warmup_style: null
363
+ override_optimizer_config: null
364
+ model:
365
+ fsdp_config:
366
+ _target_: verl.workers.config.FSDPEngineConfig
367
+ wrap_policy:
368
+ min_num_params: 0
369
+ param_offload: false
370
+ optimizer_offload: false
371
+ offload_policy: false
372
+ reshard_after_forward: true
373
+ fsdp_size: -1
374
+ forward_prefetch: false
375
+ model_dtype: fp32
376
+ use_orig_params: false
377
+ seed: 42
378
+ full_determinism: false
379
+ ulysses_sequence_parallel_size: 1
380
+ entropy_from_logits_with_chunking: false
381
+ use_torch_compile: true
382
+ entropy_checkpointing: false
383
+ forward_only: false
384
+ strategy: fsdp
385
+ dtype: bfloat16
386
+ path: ~/models/deepseek-llm-7b-chat
387
+ tokenizer_path: ${oc.select:actor_rollout_ref.model.path,"~/models/deepseek-llm-7b-chat"}
388
+ override_config: {}
389
+ external_lib: ${oc.select:actor_rollout_ref.model.external_lib,null}
390
+ trust_remote_code: ${oc.select:actor_rollout_ref.model.trust_remote_code,false}
391
+ _target_: verl.workers.config.FSDPCriticModelCfg
392
+ use_shm: false
393
+ enable_gradient_checkpointing: true
394
+ enable_activation_offload: false
395
+ use_remove_padding: false
396
+ lora_rank: 0
397
+ lora_alpha: 16
398
+ target_modules: all-linear
399
+ _target_: verl.workers.config.FSDPCriticConfig
400
+ rollout_n: ${oc.select:actor_rollout_ref.rollout.n,1}
401
+ strategy: fsdp
402
+ enable: null
403
+ ppo_mini_batch_size: ${oc.select:actor_rollout_ref.actor.ppo_mini_batch_size,256}
404
+ ppo_micro_batch_size: null
405
+ ppo_micro_batch_size_per_gpu: ${oc.select:.ppo_micro_batch_size,null}
406
+ use_dynamic_bsz: ${oc.select:actor_rollout_ref.actor.use_dynamic_bsz,false}
407
+ ppo_max_token_len_per_gpu: 32768
408
+ forward_max_token_len_per_gpu: ${.ppo_max_token_len_per_gpu}
409
+ ppo_epochs: ${oc.select:actor_rollout_ref.actor.ppo_epochs,1}
410
+ shuffle: ${oc.select:actor_rollout_ref.actor.shuffle,false}
411
+ cliprange_value: 0.5
412
+ loss_agg_mode: ${oc.select:actor_rollout_ref.actor.loss_agg_mode,token-mean}
413
+ checkpoint:
414
+ _target_: verl.trainer.config.CheckpointConfig
415
+ save_contents:
416
+ - model
417
+ - optimizer
418
+ - extra
419
+ load_contents: ${.save_contents}
420
+ async_save: false
421
+ profiler:
422
+ _target_: verl.utils.profiler.ProfilerConfig
423
+ tool: ${oc.select:global_profiler.tool,null}
424
+ enable: false
425
+ all_ranks: false
426
+ ranks: []
427
+ save_path: ${oc.select:global_profiler.save_path,null}
428
+ tool_config:
429
+ nsys:
430
+ _target_: verl.utils.profiler.config.NsightToolConfig
431
+ discrete: ${oc.select:global_profiler.global_tool_config.nsys.discrete}
432
+ npu:
433
+ _target_: verl.utils.profiler.config.NPUToolConfig
434
+ contents: []
435
+ level: level1
436
+ analysis: true
437
+ discrete: false
438
+ torch:
439
+ _target_: verl.utils.profiler.config.TorchProfilerToolConfig
440
+ step_start: 0
441
+ step_end: null
442
+ torch_memory:
443
+ _target_: verl.utils.profiler.config.TorchMemoryToolConfig
444
+ trace_alloc_max_entries: ${oc.select:global_profiler.global_tool_config.torch_memory.trace_alloc_max_entries,100000}
445
+ stack_depth: ${oc.select:global_profiler.global_tool_config.torch_memory.stack_depth,32}
446
+ forward_micro_batch_size: ${oc.select:.ppo_micro_batch_size,null}
447
+ forward_micro_batch_size_per_gpu: ${oc.select:.ppo_micro_batch_size_per_gpu,null}
448
+ ulysses_sequence_parallel_size: 1
449
+ grad_clip: 1.0
450
+ reward_model:
451
+ enable: false
452
+ enable_resource_pool: false
453
+ n_gpus_per_node: 0
454
+ nnodes: 0
455
+ strategy: fsdp
456
+ model:
457
+ input_tokenizer: ${actor_rollout_ref.model.path}
458
+ path: ~/models/FsfairX-LLaMA3-RM-v0.1
459
+ external_lib: ${actor_rollout_ref.model.external_lib}
460
+ trust_remote_code: false
461
+ override_config: {}
462
+ use_shm: false
463
+ use_remove_padding: false
464
+ use_fused_kernels: ${actor_rollout_ref.model.use_fused_kernels}
465
+ fsdp_config:
466
+ _target_: verl.workers.config.FSDPEngineConfig
467
+ wrap_policy:
468
+ min_num_params: 0
469
+ param_offload: false
470
+ reshard_after_forward: true
471
+ fsdp_size: -1
472
+ forward_prefetch: false
473
+ micro_batch_size: null
474
+ micro_batch_size_per_gpu: null
475
+ max_length: null
476
+ use_dynamic_bsz: ${critic.use_dynamic_bsz}
477
+ forward_max_token_len_per_gpu: ${critic.forward_max_token_len_per_gpu}
478
+ reward_manager: naive
479
+ launch_reward_fn_async: false
480
+ sandbox_fusion:
481
+ url: null
482
+ max_concurrent: 64
483
+ memory_limit_mb: 1024
484
+ profiler:
485
+ _target_: verl.utils.profiler.ProfilerConfig
486
+ tool: ${oc.select:global_profiler.tool,null}
487
+ enable: false
488
+ all_ranks: false
489
+ ranks: []
490
+ save_path: ${oc.select:global_profiler.save_path,null}
491
+ tool_config: ${oc.select:actor_rollout_ref.actor.profiler.tool_config,null}
492
+ ulysses_sequence_parallel_size: 1
493
+ use_reward_loop: true
494
+ rollout:
495
+ _target_: verl.workers.config.RolloutConfig
496
+ name: ???
497
+ dtype: bfloat16
498
+ gpu_memory_utilization: 0.5
499
+ enforce_eager: true
500
+ cudagraph_capture_sizes: null
501
+ free_cache_engine: true
502
+ data_parallel_size: 1
503
+ expert_parallel_size: 1
504
+ tensor_model_parallel_size: 2
505
+ max_num_batched_tokens: 8192
506
+ max_model_len: null
507
+ max_num_seqs: 1024
508
+ load_format: auto
509
+ engine_kwargs: {}
510
+ limit_images: null
511
+ enable_chunked_prefill: true
512
+ enable_prefix_caching: true
513
+ disable_log_stats: true
514
+ skip_tokenizer_init: true
515
+ prompt_length: 512
516
+ response_length: 512
517
+ algorithm:
518
+ rollout_correction:
519
+ rollout_is: null
520
+ rollout_is_threshold: 2.0
521
+ rollout_rs: null
522
+ rollout_rs_threshold: null
523
+ rollout_rs_threshold_lower: null
524
+ rollout_token_veto_threshold: null
525
+ bypass_mode: false
526
+ use_policy_gradient: false
527
+ rollout_is_batch_normalize: false
528
+ _target_: verl.trainer.config.AlgoConfig
529
+ gamma: 1.0
530
+ lam: 1.0
531
+ adv_estimator: sgrpo
532
+ norm_adv_by_std_in_grpo: false
533
+ use_kl_in_reward: false
534
+ kl_penalty: kl
535
+ kl_ctrl:
536
+ _target_: verl.trainer.config.KLControlConfig
537
+ type: fixed
538
+ kl_coef: 0.001
539
+ horizon: 10000
540
+ target_kl: 0.1
541
+ use_pf_ppo: false
542
+ pf_ppo:
543
+ reweight_method: pow
544
+ weight_pow: 2.0
545
+ sgrpo:
546
+ num_truncations: 4
547
+ force_answer_prompt: Time is limited, stop thinking and start answering.\n</think>\n\n
548
+ answer_max_tokens: 256
549
+ min_truncation_ratio: 0.1
550
+ max_truncation_ratio: 0.9
551
+ custom_reward_function:
552
+ path: null
553
+ name: compute_score
554
+ trainer:
555
+ balance_batch: true
556
+ total_epochs: 15
557
+ total_training_steps: null
558
+ project_name: verl_sgrpo_example_gsm8k
559
+ experiment_name: qwen3_8b_sgrpo
560
+ logger:
561
+ - console
562
+ log_val_generations: 0
563
+ rollout_data_dir: ./outputs/sgrpo_samples
564
+ rollout_data_dump_freq: 20
565
+ validation_data_dir: null
566
+ nnodes: 1
567
+ n_gpus_per_node: 8
568
+ save_freq: 20
569
+ esi_redundant_time: 0
570
+ resume_mode: disable
571
+ resume_from_path: null
572
+ val_before_train: true
573
+ val_only: false
574
+ test_freq: 5
575
+ critic_warmup: 0
576
+ default_hdfs_dir: null
577
+ del_local_ckpt_after_load: false
578
+ default_local_dir: checkpoints/${trainer.project_name}/${trainer.experiment_name}
579
+ max_actor_ckpt_to_keep: null
580
+ max_critic_ckpt_to_keep: null
581
+ ray_wait_register_center_timeout: 300
582
+ device: cuda
583
+ use_legacy_worker_impl: auto
584
+ global_profiler:
585
+ _target_: verl.utils.profiler.ProfilerConfig
586
+ tool: null
587
+ steps: null
588
+ profile_continuous_steps: false
589
+ save_path: outputs/profile
590
+ global_tool_config:
591
+ nsys:
592
+ _target_: verl.utils.profiler.config.NsightToolConfig
593
+ discrete: false
594
+ controller_nsight_options:
595
+ trace: cuda,nvtx,cublas,ucx
596
+ cuda-memory-usage: 'true'
597
+ cuda-graph-trace: graph
598
+ worker_nsight_options:
599
+ trace: cuda,nvtx,cublas,ucx
600
+ cuda-memory-usage: 'true'
601
+ cuda-graph-trace: graph
602
+ capture-range: cudaProfilerApi
603
+ capture-range-end: null
604
+ kill: none
605
+ torch_memory:
606
+ trace_alloc_max_entries: 100000
607
+ stack_depth: 32
608
+ context: all
609
+ stacks: all
610
+ kw_args: {}
611
+ transfer_queue:
612
+ enable: false
613
+ ray_kwargs:
614
+ ray_init:
615
+ num_cpus: null
616
+ timeline_json_file: null
outputs/2026-01-14/20-02-28/main_sgrpo.log ADDED
File without changes
outputs/2026-01-14/20-19-23/.hydra/config.yaml ADDED
@@ -0,0 +1,616 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ actor_rollout_ref:
2
+ actor:
3
+ optim:
4
+ _target_: verl.workers.config.FSDPOptimizerConfig
5
+ optimizer: AdamW
6
+ optimizer_impl: torch.optim
7
+ lr: 1.0e-06
8
+ lr_warmup_steps_ratio: 0.0
9
+ total_training_steps: -1
10
+ weight_decay: 0.01
11
+ lr_warmup_steps: -1
12
+ betas:
13
+ - 0.9
14
+ - 0.999
15
+ clip_grad: 1.0
16
+ min_lr_ratio: 0.0
17
+ num_cycles: 0.5
18
+ lr_scheduler_type: constant
19
+ warmup_style: null
20
+ override_optimizer_config: null
21
+ fsdp_config:
22
+ _target_: verl.workers.config.FSDPEngineConfig
23
+ wrap_policy:
24
+ min_num_params: 0
25
+ param_offload: false
26
+ optimizer_offload: false
27
+ offload_policy: false
28
+ reshard_after_forward: true
29
+ fsdp_size: -1
30
+ forward_prefetch: false
31
+ model_dtype: fp32
32
+ use_orig_params: false
33
+ seed: 42
34
+ full_determinism: false
35
+ ulysses_sequence_parallel_size: 1
36
+ entropy_from_logits_with_chunking: false
37
+ use_torch_compile: true
38
+ entropy_checkpointing: false
39
+ forward_only: false
40
+ strategy: fsdp
41
+ dtype: bfloat16
42
+ _target_: verl.workers.config.FSDPActorConfig
43
+ rollout_n: ${oc.select:actor_rollout_ref.rollout.n,1}
44
+ strategy: fsdp
45
+ ppo_mini_batch_size: 256
46
+ ppo_micro_batch_size: null
47
+ ppo_micro_batch_size_per_gpu: 32
48
+ use_dynamic_bsz: false
49
+ ppo_max_token_len_per_gpu: 16384
50
+ clip_ratio: 0.2
51
+ clip_ratio_low: 0.2
52
+ clip_ratio_high: 0.2
53
+ freeze_vision_tower: false
54
+ policy_loss:
55
+ _target_: verl.workers.config.PolicyLossConfig
56
+ loss_mode: vanilla
57
+ clip_cov_ratio: 0.0002
58
+ clip_cov_lb: 1.0
59
+ clip_cov_ub: 5.0
60
+ kl_cov_ratio: 0.0002
61
+ ppo_kl_coef: 0.1
62
+ clip_ratio_c: 3.0
63
+ loss_agg_mode: token-mean
64
+ loss_scale_factor: null
65
+ entropy_coeff: 0
66
+ calculate_entropy: false
67
+ use_kl_loss: true
68
+ use_torch_compile: true
69
+ kl_loss_coef: 0.001
70
+ kl_loss_type: low_var_kl
71
+ ppo_epochs: 1
72
+ shuffle: false
73
+ checkpoint:
74
+ _target_: verl.trainer.config.CheckpointConfig
75
+ save_contents:
76
+ - model
77
+ - optimizer
78
+ - extra
79
+ load_contents: ${.save_contents}
80
+ async_save: false
81
+ use_fused_kernels: ${oc.select:actor_rollout_ref.model.use_fused_kernels,false}
82
+ profiler:
83
+ _target_: verl.utils.profiler.ProfilerConfig
84
+ tool: ${oc.select:global_profiler.tool,null}
85
+ enable: false
86
+ all_ranks: false
87
+ ranks: []
88
+ save_path: ${oc.select:global_profiler.save_path,null}
89
+ tool_config:
90
+ nsys:
91
+ _target_: verl.utils.profiler.config.NsightToolConfig
92
+ discrete: ${oc.select:global_profiler.global_tool_config.nsys.discrete}
93
+ npu:
94
+ _target_: verl.utils.profiler.config.NPUToolConfig
95
+ contents: []
96
+ level: level1
97
+ analysis: true
98
+ discrete: false
99
+ torch:
100
+ _target_: verl.utils.profiler.config.TorchProfilerToolConfig
101
+ step_start: 0
102
+ step_end: null
103
+ torch_memory:
104
+ _target_: verl.utils.profiler.config.TorchMemoryToolConfig
105
+ trace_alloc_max_entries: ${oc.select:global_profiler.global_tool_config.torch_memory.trace_alloc_max_entries,100000}
106
+ stack_depth: ${oc.select:global_profiler.global_tool_config.torch_memory.stack_depth,32}
107
+ router_replay:
108
+ _target_: verl.workers.config.RouterReplayConfig
109
+ mode: disabled
110
+ record_file: null
111
+ replay_file: null
112
+ grad_clip: 1.0
113
+ ulysses_sequence_parallel_size: 1
114
+ entropy_from_logits_with_chunking: false
115
+ entropy_checkpointing: false
116
+ use_remove_padding: ${oc.select:actor_rollout_ref.model.use_remove_padding,false}
117
+ ref:
118
+ rollout_n: ${oc.select:actor_rollout_ref.rollout.n,1}
119
+ strategy: ${actor_rollout_ref.actor.strategy}
120
+ use_torch_compile: ${oc.select:actor_rollout_ref.actor.use_torch_compile,true}
121
+ log_prob_micro_batch_size: null
122
+ log_prob_micro_batch_size_per_gpu: 32
123
+ log_prob_use_dynamic_bsz: ${oc.select:actor_rollout_ref.actor.use_dynamic_bsz,false}
124
+ log_prob_max_token_len_per_gpu: ${oc.select:actor_rollout_ref.actor.ppo_max_token_len_per_gpu,16384}
125
+ profiler:
126
+ _target_: verl.utils.profiler.ProfilerConfig
127
+ tool: ${oc.select:global_profiler.tool,null}
128
+ enable: false
129
+ all_ranks: false
130
+ ranks: []
131
+ save_path: ${oc.select:global_profiler.save_path,null}
132
+ tool_config:
133
+ nsys:
134
+ _target_: verl.utils.profiler.config.NsightToolConfig
135
+ discrete: ${oc.select:global_profiler.global_tool_config.nsys.discrete}
136
+ npu:
137
+ _target_: verl.utils.profiler.config.NPUToolConfig
138
+ contents: []
139
+ level: level1
140
+ analysis: true
141
+ discrete: false
142
+ torch:
143
+ _target_: verl.utils.profiler.config.TorchProfilerToolConfig
144
+ step_start: 0
145
+ step_end: null
146
+ torch_memory:
147
+ _target_: verl.utils.profiler.config.TorchMemoryToolConfig
148
+ trace_alloc_max_entries: ${oc.select:global_profiler.global_tool_config.torch_memory.trace_alloc_max_entries,100000}
149
+ stack_depth: ${oc.select:global_profiler.global_tool_config.torch_memory.stack_depth,32}
150
+ router_replay:
151
+ _target_: verl.workers.config.RouterReplayConfig
152
+ mode: disabled
153
+ record_file: null
154
+ replay_file: null
155
+ fsdp_config:
156
+ _target_: verl.workers.config.FSDPEngineConfig
157
+ wrap_policy:
158
+ min_num_params: 0
159
+ param_offload: false
160
+ optimizer_offload: false
161
+ offload_policy: false
162
+ reshard_after_forward: true
163
+ fsdp_size: -1
164
+ forward_prefetch: false
165
+ model_dtype: fp32
166
+ use_orig_params: false
167
+ seed: 42
168
+ full_determinism: false
169
+ ulysses_sequence_parallel_size: 1
170
+ entropy_from_logits_with_chunking: false
171
+ use_torch_compile: true
172
+ entropy_checkpointing: false
173
+ forward_only: true
174
+ strategy: fsdp
175
+ dtype: bfloat16
176
+ _target_: verl.workers.config.FSDPActorConfig
177
+ ulysses_sequence_parallel_size: ${oc.select:actor_rollout_ref.actor.ulysses_sequence_parallel_size,1}
178
+ entropy_from_logits_with_chunking: false
179
+ entropy_checkpointing: false
180
+ rollout:
181
+ _target_: verl.workers.config.RolloutConfig
182
+ name: vllm
183
+ mode: async
184
+ temperature: 1.0
185
+ top_k: -1
186
+ top_p: 1
187
+ prompt_length: 1536
188
+ response_length: 512
189
+ dtype: bfloat16
190
+ gpu_memory_utilization: 0.6
191
+ ignore_eos: false
192
+ enforce_eager: false
193
+ cudagraph_capture_sizes: null
194
+ free_cache_engine: true
195
+ tensor_model_parallel_size: 2
196
+ data_parallel_size: 1
197
+ expert_parallel_size: 1
198
+ pipeline_model_parallel_size: 1
199
+ max_num_batched_tokens: 8192
200
+ max_model_len: null
201
+ max_num_seqs: 1024
202
+ enable_chunked_prefill: true
203
+ enable_prefix_caching: true
204
+ load_format: dummy
205
+ log_prob_micro_batch_size: null
206
+ log_prob_micro_batch_size_per_gpu: 32
207
+ log_prob_use_dynamic_bsz: ${oc.select:actor_rollout_ref.actor.use_dynamic_bsz,false}
208
+ log_prob_max_token_len_per_gpu: ${oc.select:actor_rollout_ref.actor.ppo_max_token_len_per_gpu,16384}
209
+ disable_log_stats: true
210
+ do_sample: true
211
+ 'n': 1
212
+ over_sample_rate: 0
213
+ multi_stage_wake_up: false
214
+ engine_kwargs:
215
+ vllm: {}
216
+ sglang: {}
217
+ val_kwargs:
218
+ _target_: verl.workers.config.SamplingConfig
219
+ top_k: -1
220
+ top_p: 1.0
221
+ temperature: 0
222
+ 'n': 1
223
+ do_sample: false
224
+ multi_turn:
225
+ _target_: verl.workers.config.MultiTurnConfig
226
+ enable: false
227
+ max_assistant_turns: null
228
+ tool_config_path: null
229
+ max_user_turns: null
230
+ max_parallel_calls: 1
231
+ max_tool_response_length: 256
232
+ tool_response_truncate_side: middle
233
+ interaction_config_path: null
234
+ use_inference_chat_template: false
235
+ tokenization_sanity_check_mode: strict
236
+ format: hermes
237
+ num_repeat_rollouts: null
238
+ calculate_log_probs: false
239
+ agent:
240
+ _target_: verl.workers.config.AgentLoopConfig
241
+ num_workers: 8
242
+ default_agent_loop: single_turn_agent
243
+ agent_loop_config_path: null
244
+ custom_async_server:
245
+ _target_: verl.workers.config.CustomAsyncServerConfig
246
+ path: null
247
+ name: null
248
+ update_weights_bucket_megabytes: 512
249
+ trace:
250
+ _target_: verl.workers.config.TraceConfig
251
+ backend: null
252
+ token2text: false
253
+ max_samples_per_step_per_worker: null
254
+ skip_rollout: false
255
+ skip_dump_dir: /tmp/rollout_dump
256
+ skip_tokenizer_init: true
257
+ enable_rollout_routing_replay: false
258
+ profiler:
259
+ _target_: verl.utils.profiler.ProfilerConfig
260
+ tool: ${oc.select:global_profiler.tool,null}
261
+ enable: ${oc.select:actor_rollout_ref.actor.profiler.enable,false}
262
+ all_ranks: ${oc.select:actor_rollout_ref.actor.profiler.all_ranks,false}
263
+ ranks: ${oc.select:actor_rollout_ref.actor.profiler.ranks,[]}
264
+ save_path: ${oc.select:global_profiler.save_path,null}
265
+ tool_config: ${oc.select:actor_rollout_ref.actor.profiler.tool_config,null}
266
+ prometheus:
267
+ _target_: verl.workers.config.PrometheusConfig
268
+ enable: false
269
+ port: 9090
270
+ file: /tmp/ray/session_latest/metrics/prometheus/prometheus.yml
271
+ served_model_name: ${oc.select:actor_rollout_ref.model.path,null}
272
+ layered_summon: false
273
+ model:
274
+ _target_: verl.workers.config.HFModelConfig
275
+ path: /mnt/tidal-alsh01/dataset/redtrans/zhangruiqi/models/Qwen3-4B
276
+ hf_config_path: null
277
+ tokenizer_path: null
278
+ use_shm: false
279
+ trust_remote_code: false
280
+ custom_chat_template: null
281
+ external_lib: null
282
+ override_config: {}
283
+ enable_gradient_checkpointing: true
284
+ enable_activation_offload: false
285
+ use_remove_padding: true
286
+ lora_rank: 0
287
+ lora_alpha: 16
288
+ target_modules: all-linear
289
+ exclude_modules: null
290
+ lora_adapter_path: null
291
+ use_liger: false
292
+ use_fused_kernels: false
293
+ fused_kernel_options:
294
+ impl_backend: torch
295
+ hybrid_engine: true
296
+ nccl_timeout: 600
297
+ data:
298
+ tokenizer: null
299
+ use_shm: false
300
+ train_files: /mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/train.parquet
301
+ val_files: /mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/test.parquet
302
+ train_max_samples: -1
303
+ val_max_samples: -1
304
+ prompt_key: prompt
305
+ reward_fn_key: data_source
306
+ max_prompt_length: 512
307
+ max_response_length: 1024
308
+ train_batch_size: 1024
309
+ val_batch_size: null
310
+ tool_config_path: ${oc.select:actor_rollout_ref.rollout.multi_turn.tool_config_path,
311
+ null}
312
+ return_raw_input_ids: false
313
+ return_raw_chat: true
314
+ return_full_prompt: false
315
+ shuffle: true
316
+ seed: null
317
+ dataloader_num_workers: 8
318
+ image_patch_size: 14
319
+ validation_shuffle: false
320
+ filter_overlong_prompts: true
321
+ filter_overlong_prompts_workers: 1
322
+ truncation: error
323
+ image_key: images
324
+ video_key: videos
325
+ trust_remote_code: false
326
+ custom_cls:
327
+ path: null
328
+ name: null
329
+ return_multi_modal_inputs: true
330
+ sampler:
331
+ class_path: null
332
+ class_name: null
333
+ datagen:
334
+ path: null
335
+ name: null
336
+ apply_chat_template_kwargs: {}
337
+ reward_manager:
338
+ _target_: verl.trainer.config.config.RewardManagerConfig
339
+ source: register
340
+ name: ${oc.select:reward_model.reward_manager,naive}
341
+ module:
342
+ _target_: verl.trainer.config.config.ModuleConfig
343
+ path: null
344
+ name: custom_reward_manager
345
+ critic:
346
+ optim:
347
+ _target_: verl.workers.config.FSDPOptimizerConfig
348
+ optimizer: AdamW
349
+ optimizer_impl: torch.optim
350
+ lr: 1.0e-05
351
+ lr_warmup_steps_ratio: 0.0
352
+ total_training_steps: -1
353
+ weight_decay: 0.01
354
+ lr_warmup_steps: -1
355
+ betas:
356
+ - 0.9
357
+ - 0.999
358
+ clip_grad: 1.0
359
+ min_lr_ratio: 0.0
360
+ num_cycles: 0.5
361
+ lr_scheduler_type: constant
362
+ warmup_style: null
363
+ override_optimizer_config: null
364
+ model:
365
+ fsdp_config:
366
+ _target_: verl.workers.config.FSDPEngineConfig
367
+ wrap_policy:
368
+ min_num_params: 0
369
+ param_offload: false
370
+ optimizer_offload: false
371
+ offload_policy: false
372
+ reshard_after_forward: true
373
+ fsdp_size: -1
374
+ forward_prefetch: false
375
+ model_dtype: fp32
376
+ use_orig_params: false
377
+ seed: 42
378
+ full_determinism: false
379
+ ulysses_sequence_parallel_size: 1
380
+ entropy_from_logits_with_chunking: false
381
+ use_torch_compile: true
382
+ entropy_checkpointing: false
383
+ forward_only: false
384
+ strategy: fsdp
385
+ dtype: bfloat16
386
+ path: ~/models/deepseek-llm-7b-chat
387
+ tokenizer_path: ${oc.select:actor_rollout_ref.model.path,"~/models/deepseek-llm-7b-chat"}
388
+ override_config: {}
389
+ external_lib: ${oc.select:actor_rollout_ref.model.external_lib,null}
390
+ trust_remote_code: ${oc.select:actor_rollout_ref.model.trust_remote_code,false}
391
+ _target_: verl.workers.config.FSDPCriticModelCfg
392
+ use_shm: false
393
+ enable_gradient_checkpointing: true
394
+ enable_activation_offload: false
395
+ use_remove_padding: false
396
+ lora_rank: 0
397
+ lora_alpha: 16
398
+ target_modules: all-linear
399
+ _target_: verl.workers.config.FSDPCriticConfig
400
+ rollout_n: ${oc.select:actor_rollout_ref.rollout.n,1}
401
+ strategy: fsdp
402
+ enable: null
403
+ ppo_mini_batch_size: ${oc.select:actor_rollout_ref.actor.ppo_mini_batch_size,256}
404
+ ppo_micro_batch_size: null
405
+ ppo_micro_batch_size_per_gpu: ${oc.select:.ppo_micro_batch_size,null}
406
+ use_dynamic_bsz: ${oc.select:actor_rollout_ref.actor.use_dynamic_bsz,false}
407
+ ppo_max_token_len_per_gpu: 32768
408
+ forward_max_token_len_per_gpu: ${.ppo_max_token_len_per_gpu}
409
+ ppo_epochs: ${oc.select:actor_rollout_ref.actor.ppo_epochs,1}
410
+ shuffle: ${oc.select:actor_rollout_ref.actor.shuffle,false}
411
+ cliprange_value: 0.5
412
+ loss_agg_mode: ${oc.select:actor_rollout_ref.actor.loss_agg_mode,token-mean}
413
+ checkpoint:
414
+ _target_: verl.trainer.config.CheckpointConfig
415
+ save_contents:
416
+ - model
417
+ - optimizer
418
+ - extra
419
+ load_contents: ${.save_contents}
420
+ async_save: false
421
+ profiler:
422
+ _target_: verl.utils.profiler.ProfilerConfig
423
+ tool: ${oc.select:global_profiler.tool,null}
424
+ enable: false
425
+ all_ranks: false
426
+ ranks: []
427
+ save_path: ${oc.select:global_profiler.save_path,null}
428
+ tool_config:
429
+ nsys:
430
+ _target_: verl.utils.profiler.config.NsightToolConfig
431
+ discrete: ${oc.select:global_profiler.global_tool_config.nsys.discrete}
432
+ npu:
433
+ _target_: verl.utils.profiler.config.NPUToolConfig
434
+ contents: []
435
+ level: level1
436
+ analysis: true
437
+ discrete: false
438
+ torch:
439
+ _target_: verl.utils.profiler.config.TorchProfilerToolConfig
440
+ step_start: 0
441
+ step_end: null
442
+ torch_memory:
443
+ _target_: verl.utils.profiler.config.TorchMemoryToolConfig
444
+ trace_alloc_max_entries: ${oc.select:global_profiler.global_tool_config.torch_memory.trace_alloc_max_entries,100000}
445
+ stack_depth: ${oc.select:global_profiler.global_tool_config.torch_memory.stack_depth,32}
446
+ forward_micro_batch_size: ${oc.select:.ppo_micro_batch_size,null}
447
+ forward_micro_batch_size_per_gpu: ${oc.select:.ppo_micro_batch_size_per_gpu,null}
448
+ ulysses_sequence_parallel_size: 1
449
+ grad_clip: 1.0
450
+ reward_model:
451
+ enable: false
452
+ enable_resource_pool: false
453
+ n_gpus_per_node: 0
454
+ nnodes: 0
455
+ strategy: fsdp
456
+ model:
457
+ input_tokenizer: ${actor_rollout_ref.model.path}
458
+ path: ~/models/FsfairX-LLaMA3-RM-v0.1
459
+ external_lib: ${actor_rollout_ref.model.external_lib}
460
+ trust_remote_code: false
461
+ override_config: {}
462
+ use_shm: false
463
+ use_remove_padding: false
464
+ use_fused_kernels: ${actor_rollout_ref.model.use_fused_kernels}
465
+ fsdp_config:
466
+ _target_: verl.workers.config.FSDPEngineConfig
467
+ wrap_policy:
468
+ min_num_params: 0
469
+ param_offload: false
470
+ reshard_after_forward: true
471
+ fsdp_size: -1
472
+ forward_prefetch: false
473
+ micro_batch_size: null
474
+ micro_batch_size_per_gpu: null
475
+ max_length: null
476
+ use_dynamic_bsz: ${critic.use_dynamic_bsz}
477
+ forward_max_token_len_per_gpu: ${critic.forward_max_token_len_per_gpu}
478
+ reward_manager: naive
479
+ launch_reward_fn_async: false
480
+ sandbox_fusion:
481
+ url: null
482
+ max_concurrent: 64
483
+ memory_limit_mb: 1024
484
+ profiler:
485
+ _target_: verl.utils.profiler.ProfilerConfig
486
+ tool: ${oc.select:global_profiler.tool,null}
487
+ enable: false
488
+ all_ranks: false
489
+ ranks: []
490
+ save_path: ${oc.select:global_profiler.save_path,null}
491
+ tool_config: ${oc.select:actor_rollout_ref.actor.profiler.tool_config,null}
492
+ ulysses_sequence_parallel_size: 1
493
+ use_reward_loop: true
494
+ rollout:
495
+ _target_: verl.workers.config.RolloutConfig
496
+ name: ???
497
+ dtype: bfloat16
498
+ gpu_memory_utilization: 0.5
499
+ enforce_eager: true
500
+ cudagraph_capture_sizes: null
501
+ free_cache_engine: true
502
+ data_parallel_size: 1
503
+ expert_parallel_size: 1
504
+ tensor_model_parallel_size: 2
505
+ max_num_batched_tokens: 8192
506
+ max_model_len: null
507
+ max_num_seqs: 1024
508
+ load_format: auto
509
+ engine_kwargs: {}
510
+ limit_images: null
511
+ enable_chunked_prefill: true
512
+ enable_prefix_caching: true
513
+ disable_log_stats: true
514
+ skip_tokenizer_init: true
515
+ prompt_length: 512
516
+ response_length: 512
517
+ algorithm:
518
+ rollout_correction:
519
+ rollout_is: null
520
+ rollout_is_threshold: 2.0
521
+ rollout_rs: null
522
+ rollout_rs_threshold: null
523
+ rollout_rs_threshold_lower: null
524
+ rollout_token_veto_threshold: null
525
+ bypass_mode: false
526
+ use_policy_gradient: false
527
+ rollout_is_batch_normalize: false
528
+ _target_: verl.trainer.config.AlgoConfig
529
+ gamma: 1.0
530
+ lam: 1.0
531
+ adv_estimator: sgrpo
532
+ norm_adv_by_std_in_grpo: false
533
+ use_kl_in_reward: false
534
+ kl_penalty: kl
535
+ kl_ctrl:
536
+ _target_: verl.trainer.config.KLControlConfig
537
+ type: fixed
538
+ kl_coef: 0.001
539
+ horizon: 10000
540
+ target_kl: 0.1
541
+ use_pf_ppo: false
542
+ pf_ppo:
543
+ reweight_method: pow
544
+ weight_pow: 2.0
545
+ sgrpo:
546
+ num_truncations: 4
547
+ force_answer_prompt: Time is limited, stop thinking and start answering.\n</think>\n\n
548
+ answer_max_tokens: 256
549
+ min_truncation_ratio: 0.1
550
+ max_truncation_ratio: 0.9
551
+ custom_reward_function:
552
+ path: null
553
+ name: compute_score
554
+ trainer:
555
+ balance_batch: true
556
+ total_epochs: 15
557
+ total_training_steps: null
558
+ project_name: verl_sgrpo_example_gsm8k
559
+ experiment_name: qwen3_8b_sgrpo
560
+ logger:
561
+ - console
562
+ log_val_generations: 0
563
+ rollout_data_dir: ./outputs/sgrpo_samples
564
+ rollout_data_dump_freq: 20
565
+ validation_data_dir: null
566
+ nnodes: 1
567
+ n_gpus_per_node: 8
568
+ save_freq: 20
569
+ esi_redundant_time: 0
570
+ resume_mode: disable
571
+ resume_from_path: null
572
+ val_before_train: true
573
+ val_only: false
574
+ test_freq: 5
575
+ critic_warmup: 0
576
+ default_hdfs_dir: null
577
+ del_local_ckpt_after_load: false
578
+ default_local_dir: checkpoints/${trainer.project_name}/${trainer.experiment_name}
579
+ max_actor_ckpt_to_keep: null
580
+ max_critic_ckpt_to_keep: null
581
+ ray_wait_register_center_timeout: 300
582
+ device: cuda
583
+ use_legacy_worker_impl: auto
584
+ global_profiler:
585
+ _target_: verl.utils.profiler.ProfilerConfig
586
+ tool: null
587
+ steps: null
588
+ profile_continuous_steps: false
589
+ save_path: outputs/profile
590
+ global_tool_config:
591
+ nsys:
592
+ _target_: verl.utils.profiler.config.NsightToolConfig
593
+ discrete: false
594
+ controller_nsight_options:
595
+ trace: cuda,nvtx,cublas,ucx
596
+ cuda-memory-usage: 'true'
597
+ cuda-graph-trace: graph
598
+ worker_nsight_options:
599
+ trace: cuda,nvtx,cublas,ucx
600
+ cuda-memory-usage: 'true'
601
+ cuda-graph-trace: graph
602
+ capture-range: cudaProfilerApi
603
+ capture-range-end: null
604
+ kill: none
605
+ torch_memory:
606
+ trace_alloc_max_entries: 100000
607
+ stack_depth: 32
608
+ context: all
609
+ stacks: all
610
+ kw_args: {}
611
+ transfer_queue:
612
+ enable: false
613
+ ray_kwargs:
614
+ ray_init:
615
+ num_cpus: null
616
+ timeline_json_file: null
outputs/2026-01-14/20-19-23/main_sgrpo.log ADDED
File without changes
outputs/2026-01-14/20-27-06/.hydra/hydra.yaml ADDED
@@ -0,0 +1,218 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
4
+ sweep:
5
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.mode=RUN
114
+ task:
115
+ - algorithm.adv_estimator=sgrpo
116
+ - algorithm.norm_adv_by_std_in_grpo=False
117
+ - algorithm.use_kl_in_reward=False
118
+ - data.train_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/train.parquet
119
+ - data.val_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/test.parquet
120
+ - data.train_batch_size=1024
121
+ - data.max_prompt_length=512
122
+ - data.max_response_length=1024
123
+ - data.filter_overlong_prompts=True
124
+ - data.truncation=error
125
+ - data.return_raw_chat=True
126
+ - actor_rollout_ref.model.path=/mnt/tidal-alsh01/dataset/redtrans/zhangruiqi/models/Qwen3-4B
127
+ - actor_rollout_ref.actor.optim.lr=1e-6
128
+ - actor_rollout_ref.model.use_remove_padding=True
129
+ - actor_rollout_ref.actor.ppo_mini_batch_size=256
130
+ - actor_rollout_ref.actor.ppo_micro_batch_size_per_gpu=32
131
+ - actor_rollout_ref.actor.use_kl_loss=True
132
+ - actor_rollout_ref.actor.kl_loss_coef=0.001
133
+ - actor_rollout_ref.actor.kl_loss_type=low_var_kl
134
+ - actor_rollout_ref.actor.entropy_coeff=0
135
+ - actor_rollout_ref.model.enable_gradient_checkpointing=True
136
+ - actor_rollout_ref.actor.fsdp_config.param_offload=False
137
+ - actor_rollout_ref.actor.fsdp_config.optimizer_offload=False
138
+ - actor_rollout_ref.rollout.log_prob_micro_batch_size_per_gpu=32
139
+ - actor_rollout_ref.rollout.tensor_model_parallel_size=2
140
+ - actor_rollout_ref.rollout.name=vllm
141
+ - actor_rollout_ref.rollout.gpu_memory_utilization=0.6
142
+ - actor_rollout_ref.rollout.n=1
143
+ - actor_rollout_ref.rollout.prompt_length=1536
144
+ - actor_rollout_ref.rollout.response_length=512
145
+ - actor_rollout_ref.ref.log_prob_micro_batch_size_per_gpu=32
146
+ - actor_rollout_ref.ref.fsdp_config.param_offload=False
147
+ - sgrpo.num_truncations=4
148
+ - sgrpo.force_answer_prompt="Time is limited, stop thinking and start answering.\n</think>\n\n"
149
+ - sgrpo.answer_max_tokens=256
150
+ - trainer.critic_warmup=0
151
+ - trainer.logger=["console"]
152
+ - trainer.project_name=verl_sgrpo_example_gsm8k
153
+ - trainer.experiment_name=qwen3_8b_sgrpo
154
+ - trainer.resume_mode=disable
155
+ - trainer.n_gpus_per_node=8
156
+ - trainer.nnodes=1
157
+ - trainer.save_freq=20
158
+ - trainer.test_freq=5
159
+ - trainer.total_epochs=15
160
+ - trainer.rollout_data_dir=./outputs/sgrpo_samples
161
+ - trainer.rollout_data_dump_freq=20
162
+ job:
163
+ name: main_sgrpo
164
+ chdir: null
165
+ override_dirname: actor_rollout_ref.actor.entropy_coeff=0,actor_rollout_ref.actor.fsdp_config.optimizer_offload=False,actor_rollout_ref.actor.fsdp_config.param_offload=False,actor_rollout_ref.actor.kl_loss_coef=0.001,actor_rollout_ref.actor.kl_loss_type=low_var_kl,actor_rollout_ref.actor.optim.lr=1e-6,actor_rollout_ref.actor.ppo_micro_batch_size_per_gpu=32,actor_rollout_ref.actor.ppo_mini_batch_size=256,actor_rollout_ref.actor.use_kl_loss=True,actor_rollout_ref.model.enable_gradient_checkpointing=True,actor_rollout_ref.model.path=/mnt/tidal-alsh01/dataset/redtrans/zhangruiqi/models/Qwen3-4B,actor_rollout_ref.model.use_remove_padding=True,actor_rollout_ref.ref.fsdp_config.param_offload=False,actor_rollout_ref.ref.log_prob_micro_batch_size_per_gpu=32,actor_rollout_ref.rollout.gpu_memory_utilization=0.6,actor_rollout_ref.rollout.log_prob_micro_batch_size_per_gpu=32,actor_rollout_ref.rollout.n=1,actor_rollout_ref.rollout.name=vllm,actor_rollout_ref.rollout.prompt_length=1536,actor_rollout_ref.rollout.response_length=512,actor_rollout_ref.rollout.tensor_model_parallel_size=2,algorithm.adv_estimator=sgrpo,algorithm.norm_adv_by_std_in_grpo=False,algorithm.use_kl_in_reward=False,data.filter_overlong_prompts=True,data.max_prompt_length=512,data.max_response_length=1024,data.return_raw_chat=True,data.train_batch_size=1024,data.train_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/train.parquet,data.truncation=error,data.val_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/test.parquet,sgrpo.answer_max_tokens=256,sgrpo.force_answer_prompt="Time
166
+ is limited, stop thinking and start answering.\n</think>\n\n",sgrpo.num_truncations=4,trainer.critic_warmup=0,trainer.experiment_name=qwen3_8b_sgrpo,trainer.logger=["console"],trainer.n_gpus_per_node=8,trainer.nnodes=1,trainer.project_name=verl_sgrpo_example_gsm8k,trainer.resume_mode=disable,trainer.rollout_data_dir=./outputs/sgrpo_samples,trainer.rollout_data_dump_freq=20,trainer.save_freq=20,trainer.test_freq=5,trainer.total_epochs=15
167
+ id: ???
168
+ num: ???
169
+ config_name: sgrpo_trainer
170
+ env_set: {}
171
+ env_copy: []
172
+ config:
173
+ override_dirname:
174
+ kv_sep: '='
175
+ item_sep: ','
176
+ exclude_keys: []
177
+ runtime:
178
+ version: 1.3.2
179
+ version_base: '1.3'
180
+ cwd: /mnt/tidal-alsh01/usr/zhangruiqi1/my/verl
181
+ config_sources:
182
+ - path: hydra.conf
183
+ schema: pkg
184
+ provider: hydra
185
+ - path: /mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/config
186
+ schema: file
187
+ provider: main
188
+ - path: verl.trainer.config
189
+ schema: pkg
190
+ provider: verl-trainer-config
191
+ - path: ''
192
+ schema: structured
193
+ provider: schema
194
+ output_dir: /mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/outputs/2026-01-14/20-27-06
195
+ choices:
196
+ algorithm@algorithm.rollout_correction: rollout_correction
197
+ reward_model: dp_reward_loop
198
+ critic: dp_critic
199
+ critic/../engine@critic.model.fsdp_config: fsdp
200
+ critic/../optim@critic.optim: fsdp
201
+ model@actor_rollout_ref.model: hf_model
202
+ rollout@actor_rollout_ref.rollout: rollout
203
+ ref@actor_rollout_ref.ref: dp_ref
204
+ ref/../engine@actor_rollout_ref.ref.fsdp_config: fsdp
205
+ data: legacy_data
206
+ actor@actor_rollout_ref.actor: dp_actor
207
+ actor/../engine@actor_rollout_ref.actor.fsdp_config: fsdp
208
+ actor/../optim@actor_rollout_ref.actor.optim: fsdp
209
+ hydra/env: default
210
+ hydra/callbacks: null
211
+ hydra/job_logging: default
212
+ hydra/hydra_logging: default
213
+ hydra/hydra_help: default
214
+ hydra/help: default
215
+ hydra/sweeper: basic
216
+ hydra/launcher: basic
217
+ hydra/output: default
218
+ verbose: false
outputs/2026-01-14/20-54-11/.hydra/hydra.yaml ADDED
@@ -0,0 +1,218 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
4
+ sweep:
5
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.mode=RUN
114
+ task:
115
+ - algorithm.adv_estimator=sgrpo
116
+ - algorithm.norm_adv_by_std_in_grpo=False
117
+ - algorithm.use_kl_in_reward=False
118
+ - data.train_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/train.parquet
119
+ - data.val_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/test.parquet
120
+ - data.train_batch_size=1024
121
+ - data.max_prompt_length=512
122
+ - data.max_response_length=1024
123
+ - data.filter_overlong_prompts=True
124
+ - data.truncation=error
125
+ - data.return_raw_chat=True
126
+ - actor_rollout_ref.model.path=/mnt/tidal-alsh01/dataset/redtrans/zhangruiqi/models/Qwen3-4B
127
+ - actor_rollout_ref.actor.optim.lr=1e-6
128
+ - actor_rollout_ref.model.use_remove_padding=True
129
+ - actor_rollout_ref.actor.ppo_mini_batch_size=256
130
+ - actor_rollout_ref.actor.ppo_micro_batch_size_per_gpu=32
131
+ - actor_rollout_ref.actor.use_kl_loss=True
132
+ - actor_rollout_ref.actor.kl_loss_coef=0.001
133
+ - actor_rollout_ref.actor.kl_loss_type=low_var_kl
134
+ - actor_rollout_ref.actor.entropy_coeff=0
135
+ - actor_rollout_ref.model.enable_gradient_checkpointing=True
136
+ - actor_rollout_ref.actor.fsdp_config.param_offload=False
137
+ - actor_rollout_ref.actor.fsdp_config.optimizer_offload=False
138
+ - actor_rollout_ref.rollout.log_prob_micro_batch_size_per_gpu=32
139
+ - actor_rollout_ref.rollout.tensor_model_parallel_size=2
140
+ - actor_rollout_ref.rollout.name=vllm
141
+ - actor_rollout_ref.rollout.gpu_memory_utilization=0.6
142
+ - actor_rollout_ref.rollout.n=1
143
+ - actor_rollout_ref.rollout.prompt_length=1536
144
+ - actor_rollout_ref.rollout.response_length=512
145
+ - actor_rollout_ref.ref.log_prob_micro_batch_size_per_gpu=32
146
+ - actor_rollout_ref.ref.fsdp_config.param_offload=False
147
+ - sgrpo.num_truncations=4
148
+ - sgrpo.force_answer_prompt="Time is limited, stop thinking and start answering.\n</think>\n\n"
149
+ - sgrpo.answer_max_tokens=256
150
+ - trainer.critic_warmup=0
151
+ - trainer.logger=["console"]
152
+ - trainer.project_name=verl_sgrpo_example_gsm8k
153
+ - trainer.experiment_name=qwen3_8b_sgrpo
154
+ - trainer.resume_mode=disable
155
+ - trainer.n_gpus_per_node=8
156
+ - trainer.nnodes=1
157
+ - trainer.save_freq=20
158
+ - trainer.test_freq=5
159
+ - trainer.total_epochs=15
160
+ - trainer.rollout_data_dir=./outputs/sgrpo_samples
161
+ - trainer.rollout_data_dump_freq=20
162
+ job:
163
+ name: main_sgrpo
164
+ chdir: null
165
+ override_dirname: actor_rollout_ref.actor.entropy_coeff=0,actor_rollout_ref.actor.fsdp_config.optimizer_offload=False,actor_rollout_ref.actor.fsdp_config.param_offload=False,actor_rollout_ref.actor.kl_loss_coef=0.001,actor_rollout_ref.actor.kl_loss_type=low_var_kl,actor_rollout_ref.actor.optim.lr=1e-6,actor_rollout_ref.actor.ppo_micro_batch_size_per_gpu=32,actor_rollout_ref.actor.ppo_mini_batch_size=256,actor_rollout_ref.actor.use_kl_loss=True,actor_rollout_ref.model.enable_gradient_checkpointing=True,actor_rollout_ref.model.path=/mnt/tidal-alsh01/dataset/redtrans/zhangruiqi/models/Qwen3-4B,actor_rollout_ref.model.use_remove_padding=True,actor_rollout_ref.ref.fsdp_config.param_offload=False,actor_rollout_ref.ref.log_prob_micro_batch_size_per_gpu=32,actor_rollout_ref.rollout.gpu_memory_utilization=0.6,actor_rollout_ref.rollout.log_prob_micro_batch_size_per_gpu=32,actor_rollout_ref.rollout.n=1,actor_rollout_ref.rollout.name=vllm,actor_rollout_ref.rollout.prompt_length=1536,actor_rollout_ref.rollout.response_length=512,actor_rollout_ref.rollout.tensor_model_parallel_size=2,algorithm.adv_estimator=sgrpo,algorithm.norm_adv_by_std_in_grpo=False,algorithm.use_kl_in_reward=False,data.filter_overlong_prompts=True,data.max_prompt_length=512,data.max_response_length=1024,data.return_raw_chat=True,data.train_batch_size=1024,data.train_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/train.parquet,data.truncation=error,data.val_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/test.parquet,sgrpo.answer_max_tokens=256,sgrpo.force_answer_prompt="Time
166
+ is limited, stop thinking and start answering.\n</think>\n\n",sgrpo.num_truncations=4,trainer.critic_warmup=0,trainer.experiment_name=qwen3_8b_sgrpo,trainer.logger=["console"],trainer.n_gpus_per_node=8,trainer.nnodes=1,trainer.project_name=verl_sgrpo_example_gsm8k,trainer.resume_mode=disable,trainer.rollout_data_dir=./outputs/sgrpo_samples,trainer.rollout_data_dump_freq=20,trainer.save_freq=20,trainer.test_freq=5,trainer.total_epochs=15
167
+ id: ???
168
+ num: ???
169
+ config_name: sgrpo_trainer
170
+ env_set: {}
171
+ env_copy: []
172
+ config:
173
+ override_dirname:
174
+ kv_sep: '='
175
+ item_sep: ','
176
+ exclude_keys: []
177
+ runtime:
178
+ version: 1.3.2
179
+ version_base: '1.3'
180
+ cwd: /mnt/tidal-alsh01/usr/zhangruiqi1/my/verl
181
+ config_sources:
182
+ - path: hydra.conf
183
+ schema: pkg
184
+ provider: hydra
185
+ - path: /mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/config
186
+ schema: file
187
+ provider: main
188
+ - path: verl.trainer.config
189
+ schema: pkg
190
+ provider: verl-trainer-config
191
+ - path: ''
192
+ schema: structured
193
+ provider: schema
194
+ output_dir: /mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/outputs/2026-01-14/20-54-11
195
+ choices:
196
+ algorithm@algorithm.rollout_correction: rollout_correction
197
+ reward_model: dp_reward_loop
198
+ critic: dp_critic
199
+ critic/../engine@critic.model.fsdp_config: fsdp
200
+ critic/../optim@critic.optim: fsdp
201
+ model@actor_rollout_ref.model: hf_model
202
+ rollout@actor_rollout_ref.rollout: rollout
203
+ ref@actor_rollout_ref.ref: dp_ref
204
+ ref/../engine@actor_rollout_ref.ref.fsdp_config: fsdp
205
+ data: legacy_data
206
+ actor@actor_rollout_ref.actor: dp_actor
207
+ actor/../engine@actor_rollout_ref.actor.fsdp_config: fsdp
208
+ actor/../optim@actor_rollout_ref.actor.optim: fsdp
209
+ hydra/env: default
210
+ hydra/callbacks: null
211
+ hydra/job_logging: default
212
+ hydra/hydra_logging: default
213
+ hydra/hydra_help: default
214
+ hydra/help: default
215
+ hydra/sweeper: basic
216
+ hydra/launcher: basic
217
+ hydra/output: default
218
+ verbose: false
outputs/2026-01-14/21-01-00/.hydra/config.yaml ADDED
@@ -0,0 +1,616 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ actor_rollout_ref:
2
+ actor:
3
+ optim:
4
+ _target_: verl.workers.config.FSDPOptimizerConfig
5
+ optimizer: AdamW
6
+ optimizer_impl: torch.optim
7
+ lr: 1.0e-06
8
+ lr_warmup_steps_ratio: 0.0
9
+ total_training_steps: -1
10
+ weight_decay: 0.01
11
+ lr_warmup_steps: -1
12
+ betas:
13
+ - 0.9
14
+ - 0.999
15
+ clip_grad: 1.0
16
+ min_lr_ratio: 0.0
17
+ num_cycles: 0.5
18
+ lr_scheduler_type: constant
19
+ warmup_style: null
20
+ override_optimizer_config: null
21
+ fsdp_config:
22
+ _target_: verl.workers.config.FSDPEngineConfig
23
+ wrap_policy:
24
+ min_num_params: 0
25
+ param_offload: false
26
+ optimizer_offload: false
27
+ offload_policy: false
28
+ reshard_after_forward: true
29
+ fsdp_size: -1
30
+ forward_prefetch: false
31
+ model_dtype: fp32
32
+ use_orig_params: false
33
+ seed: 42
34
+ full_determinism: false
35
+ ulysses_sequence_parallel_size: 1
36
+ entropy_from_logits_with_chunking: false
37
+ use_torch_compile: true
38
+ entropy_checkpointing: false
39
+ forward_only: false
40
+ strategy: fsdp
41
+ dtype: bfloat16
42
+ _target_: verl.workers.config.FSDPActorConfig
43
+ rollout_n: ${oc.select:actor_rollout_ref.rollout.n,1}
44
+ strategy: fsdp
45
+ ppo_mini_batch_size: 256
46
+ ppo_micro_batch_size: null
47
+ ppo_micro_batch_size_per_gpu: 32
48
+ use_dynamic_bsz: false
49
+ ppo_max_token_len_per_gpu: 16384
50
+ clip_ratio: 0.2
51
+ clip_ratio_low: 0.2
52
+ clip_ratio_high: 0.2
53
+ freeze_vision_tower: false
54
+ policy_loss:
55
+ _target_: verl.workers.config.PolicyLossConfig
56
+ loss_mode: vanilla
57
+ clip_cov_ratio: 0.0002
58
+ clip_cov_lb: 1.0
59
+ clip_cov_ub: 5.0
60
+ kl_cov_ratio: 0.0002
61
+ ppo_kl_coef: 0.1
62
+ clip_ratio_c: 3.0
63
+ loss_agg_mode: token-mean
64
+ loss_scale_factor: null
65
+ entropy_coeff: 0
66
+ calculate_entropy: false
67
+ use_kl_loss: true
68
+ use_torch_compile: true
69
+ kl_loss_coef: 0.001
70
+ kl_loss_type: low_var_kl
71
+ ppo_epochs: 1
72
+ shuffle: false
73
+ checkpoint:
74
+ _target_: verl.trainer.config.CheckpointConfig
75
+ save_contents:
76
+ - model
77
+ - optimizer
78
+ - extra
79
+ load_contents: ${.save_contents}
80
+ async_save: false
81
+ use_fused_kernels: ${oc.select:actor_rollout_ref.model.use_fused_kernels,false}
82
+ profiler:
83
+ _target_: verl.utils.profiler.ProfilerConfig
84
+ tool: ${oc.select:global_profiler.tool,null}
85
+ enable: false
86
+ all_ranks: false
87
+ ranks: []
88
+ save_path: ${oc.select:global_profiler.save_path,null}
89
+ tool_config:
90
+ nsys:
91
+ _target_: verl.utils.profiler.config.NsightToolConfig
92
+ discrete: ${oc.select:global_profiler.global_tool_config.nsys.discrete}
93
+ npu:
94
+ _target_: verl.utils.profiler.config.NPUToolConfig
95
+ contents: []
96
+ level: level1
97
+ analysis: true
98
+ discrete: false
99
+ torch:
100
+ _target_: verl.utils.profiler.config.TorchProfilerToolConfig
101
+ step_start: 0
102
+ step_end: null
103
+ torch_memory:
104
+ _target_: verl.utils.profiler.config.TorchMemoryToolConfig
105
+ trace_alloc_max_entries: ${oc.select:global_profiler.global_tool_config.torch_memory.trace_alloc_max_entries,100000}
106
+ stack_depth: ${oc.select:global_profiler.global_tool_config.torch_memory.stack_depth,32}
107
+ router_replay:
108
+ _target_: verl.workers.config.RouterReplayConfig
109
+ mode: disabled
110
+ record_file: null
111
+ replay_file: null
112
+ grad_clip: 1.0
113
+ ulysses_sequence_parallel_size: 1
114
+ entropy_from_logits_with_chunking: false
115
+ entropy_checkpointing: false
116
+ use_remove_padding: ${oc.select:actor_rollout_ref.model.use_remove_padding,false}
117
+ ref:
118
+ rollout_n: ${oc.select:actor_rollout_ref.rollout.n,1}
119
+ strategy: ${actor_rollout_ref.actor.strategy}
120
+ use_torch_compile: ${oc.select:actor_rollout_ref.actor.use_torch_compile,true}
121
+ log_prob_micro_batch_size: null
122
+ log_prob_micro_batch_size_per_gpu: 32
123
+ log_prob_use_dynamic_bsz: ${oc.select:actor_rollout_ref.actor.use_dynamic_bsz,false}
124
+ log_prob_max_token_len_per_gpu: ${oc.select:actor_rollout_ref.actor.ppo_max_token_len_per_gpu,16384}
125
+ profiler:
126
+ _target_: verl.utils.profiler.ProfilerConfig
127
+ tool: ${oc.select:global_profiler.tool,null}
128
+ enable: false
129
+ all_ranks: false
130
+ ranks: []
131
+ save_path: ${oc.select:global_profiler.save_path,null}
132
+ tool_config:
133
+ nsys:
134
+ _target_: verl.utils.profiler.config.NsightToolConfig
135
+ discrete: ${oc.select:global_profiler.global_tool_config.nsys.discrete}
136
+ npu:
137
+ _target_: verl.utils.profiler.config.NPUToolConfig
138
+ contents: []
139
+ level: level1
140
+ analysis: true
141
+ discrete: false
142
+ torch:
143
+ _target_: verl.utils.profiler.config.TorchProfilerToolConfig
144
+ step_start: 0
145
+ step_end: null
146
+ torch_memory:
147
+ _target_: verl.utils.profiler.config.TorchMemoryToolConfig
148
+ trace_alloc_max_entries: ${oc.select:global_profiler.global_tool_config.torch_memory.trace_alloc_max_entries,100000}
149
+ stack_depth: ${oc.select:global_profiler.global_tool_config.torch_memory.stack_depth,32}
150
+ router_replay:
151
+ _target_: verl.workers.config.RouterReplayConfig
152
+ mode: disabled
153
+ record_file: null
154
+ replay_file: null
155
+ fsdp_config:
156
+ _target_: verl.workers.config.FSDPEngineConfig
157
+ wrap_policy:
158
+ min_num_params: 0
159
+ param_offload: false
160
+ optimizer_offload: false
161
+ offload_policy: false
162
+ reshard_after_forward: true
163
+ fsdp_size: -1
164
+ forward_prefetch: false
165
+ model_dtype: fp32
166
+ use_orig_params: false
167
+ seed: 42
168
+ full_determinism: false
169
+ ulysses_sequence_parallel_size: 1
170
+ entropy_from_logits_with_chunking: false
171
+ use_torch_compile: true
172
+ entropy_checkpointing: false
173
+ forward_only: true
174
+ strategy: fsdp
175
+ dtype: bfloat16
176
+ _target_: verl.workers.config.FSDPActorConfig
177
+ ulysses_sequence_parallel_size: ${oc.select:actor_rollout_ref.actor.ulysses_sequence_parallel_size,1}
178
+ entropy_from_logits_with_chunking: false
179
+ entropy_checkpointing: false
180
+ rollout:
181
+ _target_: verl.workers.config.RolloutConfig
182
+ name: vllm
183
+ mode: async
184
+ temperature: 1.0
185
+ top_k: -1
186
+ top_p: 1
187
+ prompt_length: 1536
188
+ response_length: 512
189
+ dtype: bfloat16
190
+ gpu_memory_utilization: 0.6
191
+ ignore_eos: false
192
+ enforce_eager: false
193
+ cudagraph_capture_sizes: null
194
+ free_cache_engine: true
195
+ tensor_model_parallel_size: 2
196
+ data_parallel_size: 1
197
+ expert_parallel_size: 1
198
+ pipeline_model_parallel_size: 1
199
+ max_num_batched_tokens: 8192
200
+ max_model_len: null
201
+ max_num_seqs: 1024
202
+ enable_chunked_prefill: true
203
+ enable_prefix_caching: true
204
+ load_format: dummy
205
+ log_prob_micro_batch_size: null
206
+ log_prob_micro_batch_size_per_gpu: 32
207
+ log_prob_use_dynamic_bsz: ${oc.select:actor_rollout_ref.actor.use_dynamic_bsz,false}
208
+ log_prob_max_token_len_per_gpu: ${oc.select:actor_rollout_ref.actor.ppo_max_token_len_per_gpu,16384}
209
+ disable_log_stats: true
210
+ do_sample: true
211
+ 'n': 1
212
+ over_sample_rate: 0
213
+ multi_stage_wake_up: false
214
+ engine_kwargs:
215
+ vllm: {}
216
+ sglang: {}
217
+ val_kwargs:
218
+ _target_: verl.workers.config.SamplingConfig
219
+ top_k: -1
220
+ top_p: 1.0
221
+ temperature: 0
222
+ 'n': 1
223
+ do_sample: false
224
+ multi_turn:
225
+ _target_: verl.workers.config.MultiTurnConfig
226
+ enable: false
227
+ max_assistant_turns: null
228
+ tool_config_path: null
229
+ max_user_turns: null
230
+ max_parallel_calls: 1
231
+ max_tool_response_length: 256
232
+ tool_response_truncate_side: middle
233
+ interaction_config_path: null
234
+ use_inference_chat_template: false
235
+ tokenization_sanity_check_mode: strict
236
+ format: hermes
237
+ num_repeat_rollouts: null
238
+ calculate_log_probs: false
239
+ agent:
240
+ _target_: verl.workers.config.AgentLoopConfig
241
+ num_workers: 8
242
+ default_agent_loop: single_turn_agent
243
+ agent_loop_config_path: null
244
+ custom_async_server:
245
+ _target_: verl.workers.config.CustomAsyncServerConfig
246
+ path: null
247
+ name: null
248
+ update_weights_bucket_megabytes: 512
249
+ trace:
250
+ _target_: verl.workers.config.TraceConfig
251
+ backend: null
252
+ token2text: false
253
+ max_samples_per_step_per_worker: null
254
+ skip_rollout: false
255
+ skip_dump_dir: /tmp/rollout_dump
256
+ skip_tokenizer_init: true
257
+ enable_rollout_routing_replay: false
258
+ profiler:
259
+ _target_: verl.utils.profiler.ProfilerConfig
260
+ tool: ${oc.select:global_profiler.tool,null}
261
+ enable: ${oc.select:actor_rollout_ref.actor.profiler.enable,false}
262
+ all_ranks: ${oc.select:actor_rollout_ref.actor.profiler.all_ranks,false}
263
+ ranks: ${oc.select:actor_rollout_ref.actor.profiler.ranks,[]}
264
+ save_path: ${oc.select:global_profiler.save_path,null}
265
+ tool_config: ${oc.select:actor_rollout_ref.actor.profiler.tool_config,null}
266
+ prometheus:
267
+ _target_: verl.workers.config.PrometheusConfig
268
+ enable: false
269
+ port: 9090
270
+ file: /tmp/ray/session_latest/metrics/prometheus/prometheus.yml
271
+ served_model_name: ${oc.select:actor_rollout_ref.model.path,null}
272
+ layered_summon: false
273
+ model:
274
+ _target_: verl.workers.config.HFModelConfig
275
+ path: /mnt/tidal-alsh01/dataset/redtrans/zhangruiqi/models/Qwen3-4B
276
+ hf_config_path: null
277
+ tokenizer_path: null
278
+ use_shm: false
279
+ trust_remote_code: false
280
+ custom_chat_template: null
281
+ external_lib: null
282
+ override_config: {}
283
+ enable_gradient_checkpointing: true
284
+ enable_activation_offload: false
285
+ use_remove_padding: true
286
+ lora_rank: 0
287
+ lora_alpha: 16
288
+ target_modules: all-linear
289
+ exclude_modules: null
290
+ lora_adapter_path: null
291
+ use_liger: false
292
+ use_fused_kernels: false
293
+ fused_kernel_options:
294
+ impl_backend: torch
295
+ hybrid_engine: true
296
+ nccl_timeout: 600
297
+ data:
298
+ tokenizer: null
299
+ use_shm: false
300
+ train_files: /mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/train.parquet
301
+ val_files: /mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/test.parquet
302
+ train_max_samples: -1
303
+ val_max_samples: -1
304
+ prompt_key: prompt
305
+ reward_fn_key: data_source
306
+ max_prompt_length: 512
307
+ max_response_length: 1024
308
+ train_batch_size: 1024
309
+ val_batch_size: null
310
+ tool_config_path: ${oc.select:actor_rollout_ref.rollout.multi_turn.tool_config_path,
311
+ null}
312
+ return_raw_input_ids: false
313
+ return_raw_chat: true
314
+ return_full_prompt: false
315
+ shuffle: true
316
+ seed: null
317
+ dataloader_num_workers: 8
318
+ image_patch_size: 14
319
+ validation_shuffle: false
320
+ filter_overlong_prompts: true
321
+ filter_overlong_prompts_workers: 1
322
+ truncation: error
323
+ image_key: images
324
+ video_key: videos
325
+ trust_remote_code: false
326
+ custom_cls:
327
+ path: null
328
+ name: null
329
+ return_multi_modal_inputs: true
330
+ sampler:
331
+ class_path: null
332
+ class_name: null
333
+ datagen:
334
+ path: null
335
+ name: null
336
+ apply_chat_template_kwargs: {}
337
+ reward_manager:
338
+ _target_: verl.trainer.config.config.RewardManagerConfig
339
+ source: register
340
+ name: ${oc.select:reward_model.reward_manager,naive}
341
+ module:
342
+ _target_: verl.trainer.config.config.ModuleConfig
343
+ path: null
344
+ name: custom_reward_manager
345
+ critic:
346
+ optim:
347
+ _target_: verl.workers.config.FSDPOptimizerConfig
348
+ optimizer: AdamW
349
+ optimizer_impl: torch.optim
350
+ lr: 1.0e-05
351
+ lr_warmup_steps_ratio: 0.0
352
+ total_training_steps: -1
353
+ weight_decay: 0.01
354
+ lr_warmup_steps: -1
355
+ betas:
356
+ - 0.9
357
+ - 0.999
358
+ clip_grad: 1.0
359
+ min_lr_ratio: 0.0
360
+ num_cycles: 0.5
361
+ lr_scheduler_type: constant
362
+ warmup_style: null
363
+ override_optimizer_config: null
364
+ model:
365
+ fsdp_config:
366
+ _target_: verl.workers.config.FSDPEngineConfig
367
+ wrap_policy:
368
+ min_num_params: 0
369
+ param_offload: false
370
+ optimizer_offload: false
371
+ offload_policy: false
372
+ reshard_after_forward: true
373
+ fsdp_size: -1
374
+ forward_prefetch: false
375
+ model_dtype: fp32
376
+ use_orig_params: false
377
+ seed: 42
378
+ full_determinism: false
379
+ ulysses_sequence_parallel_size: 1
380
+ entropy_from_logits_with_chunking: false
381
+ use_torch_compile: true
382
+ entropy_checkpointing: false
383
+ forward_only: false
384
+ strategy: fsdp
385
+ dtype: bfloat16
386
+ path: ~/models/deepseek-llm-7b-chat
387
+ tokenizer_path: ${oc.select:actor_rollout_ref.model.path,"~/models/deepseek-llm-7b-chat"}
388
+ override_config: {}
389
+ external_lib: ${oc.select:actor_rollout_ref.model.external_lib,null}
390
+ trust_remote_code: ${oc.select:actor_rollout_ref.model.trust_remote_code,false}
391
+ _target_: verl.workers.config.FSDPCriticModelCfg
392
+ use_shm: false
393
+ enable_gradient_checkpointing: true
394
+ enable_activation_offload: false
395
+ use_remove_padding: false
396
+ lora_rank: 0
397
+ lora_alpha: 16
398
+ target_modules: all-linear
399
+ _target_: verl.workers.config.FSDPCriticConfig
400
+ rollout_n: ${oc.select:actor_rollout_ref.rollout.n,1}
401
+ strategy: fsdp
402
+ enable: null
403
+ ppo_mini_batch_size: ${oc.select:actor_rollout_ref.actor.ppo_mini_batch_size,256}
404
+ ppo_micro_batch_size: null
405
+ ppo_micro_batch_size_per_gpu: ${oc.select:.ppo_micro_batch_size,null}
406
+ use_dynamic_bsz: ${oc.select:actor_rollout_ref.actor.use_dynamic_bsz,false}
407
+ ppo_max_token_len_per_gpu: 32768
408
+ forward_max_token_len_per_gpu: ${.ppo_max_token_len_per_gpu}
409
+ ppo_epochs: ${oc.select:actor_rollout_ref.actor.ppo_epochs,1}
410
+ shuffle: ${oc.select:actor_rollout_ref.actor.shuffle,false}
411
+ cliprange_value: 0.5
412
+ loss_agg_mode: ${oc.select:actor_rollout_ref.actor.loss_agg_mode,token-mean}
413
+ checkpoint:
414
+ _target_: verl.trainer.config.CheckpointConfig
415
+ save_contents:
416
+ - model
417
+ - optimizer
418
+ - extra
419
+ load_contents: ${.save_contents}
420
+ async_save: false
421
+ profiler:
422
+ _target_: verl.utils.profiler.ProfilerConfig
423
+ tool: ${oc.select:global_profiler.tool,null}
424
+ enable: false
425
+ all_ranks: false
426
+ ranks: []
427
+ save_path: ${oc.select:global_profiler.save_path,null}
428
+ tool_config:
429
+ nsys:
430
+ _target_: verl.utils.profiler.config.NsightToolConfig
431
+ discrete: ${oc.select:global_profiler.global_tool_config.nsys.discrete}
432
+ npu:
433
+ _target_: verl.utils.profiler.config.NPUToolConfig
434
+ contents: []
435
+ level: level1
436
+ analysis: true
437
+ discrete: false
438
+ torch:
439
+ _target_: verl.utils.profiler.config.TorchProfilerToolConfig
440
+ step_start: 0
441
+ step_end: null
442
+ torch_memory:
443
+ _target_: verl.utils.profiler.config.TorchMemoryToolConfig
444
+ trace_alloc_max_entries: ${oc.select:global_profiler.global_tool_config.torch_memory.trace_alloc_max_entries,100000}
445
+ stack_depth: ${oc.select:global_profiler.global_tool_config.torch_memory.stack_depth,32}
446
+ forward_micro_batch_size: ${oc.select:.ppo_micro_batch_size,null}
447
+ forward_micro_batch_size_per_gpu: ${oc.select:.ppo_micro_batch_size_per_gpu,null}
448
+ ulysses_sequence_parallel_size: 1
449
+ grad_clip: 1.0
450
+ reward_model:
451
+ enable: false
452
+ enable_resource_pool: false
453
+ n_gpus_per_node: 0
454
+ nnodes: 0
455
+ strategy: fsdp
456
+ model:
457
+ input_tokenizer: ${actor_rollout_ref.model.path}
458
+ path: ~/models/FsfairX-LLaMA3-RM-v0.1
459
+ external_lib: ${actor_rollout_ref.model.external_lib}
460
+ trust_remote_code: false
461
+ override_config: {}
462
+ use_shm: false
463
+ use_remove_padding: false
464
+ use_fused_kernels: ${actor_rollout_ref.model.use_fused_kernels}
465
+ fsdp_config:
466
+ _target_: verl.workers.config.FSDPEngineConfig
467
+ wrap_policy:
468
+ min_num_params: 0
469
+ param_offload: false
470
+ reshard_after_forward: true
471
+ fsdp_size: -1
472
+ forward_prefetch: false
473
+ micro_batch_size: null
474
+ micro_batch_size_per_gpu: null
475
+ max_length: null
476
+ use_dynamic_bsz: ${critic.use_dynamic_bsz}
477
+ forward_max_token_len_per_gpu: ${critic.forward_max_token_len_per_gpu}
478
+ reward_manager: naive
479
+ launch_reward_fn_async: false
480
+ sandbox_fusion:
481
+ url: null
482
+ max_concurrent: 64
483
+ memory_limit_mb: 1024
484
+ profiler:
485
+ _target_: verl.utils.profiler.ProfilerConfig
486
+ tool: ${oc.select:global_profiler.tool,null}
487
+ enable: false
488
+ all_ranks: false
489
+ ranks: []
490
+ save_path: ${oc.select:global_profiler.save_path,null}
491
+ tool_config: ${oc.select:actor_rollout_ref.actor.profiler.tool_config,null}
492
+ ulysses_sequence_parallel_size: 1
493
+ use_reward_loop: true
494
+ rollout:
495
+ _target_: verl.workers.config.RolloutConfig
496
+ name: ???
497
+ dtype: bfloat16
498
+ gpu_memory_utilization: 0.5
499
+ enforce_eager: true
500
+ cudagraph_capture_sizes: null
501
+ free_cache_engine: true
502
+ data_parallel_size: 1
503
+ expert_parallel_size: 1
504
+ tensor_model_parallel_size: 2
505
+ max_num_batched_tokens: 8192
506
+ max_model_len: null
507
+ max_num_seqs: 1024
508
+ load_format: auto
509
+ engine_kwargs: {}
510
+ limit_images: null
511
+ enable_chunked_prefill: true
512
+ enable_prefix_caching: true
513
+ disable_log_stats: true
514
+ skip_tokenizer_init: true
515
+ prompt_length: 512
516
+ response_length: 512
517
+ algorithm:
518
+ rollout_correction:
519
+ rollout_is: null
520
+ rollout_is_threshold: 2.0
521
+ rollout_rs: null
522
+ rollout_rs_threshold: null
523
+ rollout_rs_threshold_lower: null
524
+ rollout_token_veto_threshold: null
525
+ bypass_mode: false
526
+ use_policy_gradient: false
527
+ rollout_is_batch_normalize: false
528
+ _target_: verl.trainer.config.AlgoConfig
529
+ gamma: 1.0
530
+ lam: 1.0
531
+ adv_estimator: sgrpo
532
+ norm_adv_by_std_in_grpo: false
533
+ use_kl_in_reward: false
534
+ kl_penalty: kl
535
+ kl_ctrl:
536
+ _target_: verl.trainer.config.KLControlConfig
537
+ type: fixed
538
+ kl_coef: 0.001
539
+ horizon: 10000
540
+ target_kl: 0.1
541
+ use_pf_ppo: false
542
+ pf_ppo:
543
+ reweight_method: pow
544
+ weight_pow: 2.0
545
+ sgrpo:
546
+ num_truncations: 4
547
+ force_answer_prompt: Time is limited, stop thinking and start answering.\n</think>\n\n
548
+ answer_max_tokens: 256
549
+ min_truncation_ratio: 0.1
550
+ max_truncation_ratio: 0.9
551
+ custom_reward_function:
552
+ path: null
553
+ name: compute_score
554
+ trainer:
555
+ balance_batch: true
556
+ total_epochs: 15
557
+ total_training_steps: null
558
+ project_name: verl_sgrpo_example_gsm8k
559
+ experiment_name: qwen3_8b_sgrpo
560
+ logger:
561
+ - console
562
+ log_val_generations: 0
563
+ rollout_data_dir: ./outputs/sgrpo_samples
564
+ rollout_data_dump_freq: 20
565
+ validation_data_dir: null
566
+ nnodes: 1
567
+ n_gpus_per_node: 8
568
+ save_freq: 20
569
+ esi_redundant_time: 0
570
+ resume_mode: disable
571
+ resume_from_path: null
572
+ val_before_train: true
573
+ val_only: false
574
+ test_freq: 5
575
+ critic_warmup: 0
576
+ default_hdfs_dir: null
577
+ del_local_ckpt_after_load: false
578
+ default_local_dir: checkpoints/${trainer.project_name}/${trainer.experiment_name}
579
+ max_actor_ckpt_to_keep: null
580
+ max_critic_ckpt_to_keep: null
581
+ ray_wait_register_center_timeout: 300
582
+ device: cuda
583
+ use_legacy_worker_impl: auto
584
+ global_profiler:
585
+ _target_: verl.utils.profiler.ProfilerConfig
586
+ tool: null
587
+ steps: null
588
+ profile_continuous_steps: false
589
+ save_path: outputs/profile
590
+ global_tool_config:
591
+ nsys:
592
+ _target_: verl.utils.profiler.config.NsightToolConfig
593
+ discrete: false
594
+ controller_nsight_options:
595
+ trace: cuda,nvtx,cublas,ucx
596
+ cuda-memory-usage: 'true'
597
+ cuda-graph-trace: graph
598
+ worker_nsight_options:
599
+ trace: cuda,nvtx,cublas,ucx
600
+ cuda-memory-usage: 'true'
601
+ cuda-graph-trace: graph
602
+ capture-range: cudaProfilerApi
603
+ capture-range-end: null
604
+ kill: none
605
+ torch_memory:
606
+ trace_alloc_max_entries: 100000
607
+ stack_depth: 32
608
+ context: all
609
+ stacks: all
610
+ kw_args: {}
611
+ transfer_queue:
612
+ enable: false
613
+ ray_kwargs:
614
+ ray_init:
615
+ num_cpus: null
616
+ timeline_json_file: null
outputs/2026-01-14/21-41-16/.hydra/overrides.yaml ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ - algorithm.adv_estimator=sgrpo
2
+ - algorithm.norm_adv_by_std_in_grpo=False
3
+ - algorithm.use_kl_in_reward=False
4
+ - data.train_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/train.parquet
5
+ - data.val_files=/mnt/tidal-alsh01/usr/zhangruiqi1/my/verl/sgrpo/data/gsm8k/test.parquet
6
+ - data.train_batch_size=1024
7
+ - data.max_prompt_length=512
8
+ - data.max_response_length=1024
9
+ - data.filter_overlong_prompts=True
10
+ - data.truncation=error
11
+ - data.return_raw_chat=True
12
+ - actor_rollout_ref.model.path=/mnt/tidal-alsh01/dataset/redtrans/zhangruiqi/models/Qwen3-4B
13
+ - actor_rollout_ref.actor.optim.lr=1e-6
14
+ - actor_rollout_ref.model.use_remove_padding=True
15
+ - actor_rollout_ref.actor.ppo_mini_batch_size=256
16
+ - actor_rollout_ref.actor.ppo_micro_batch_size_per_gpu=32
17
+ - actor_rollout_ref.actor.use_kl_loss=True
18
+ - actor_rollout_ref.actor.kl_loss_coef=0.001
19
+ - actor_rollout_ref.actor.kl_loss_type=low_var_kl
20
+ - actor_rollout_ref.actor.entropy_coeff=0
21
+ - actor_rollout_ref.model.enable_gradient_checkpointing=True
22
+ - actor_rollout_ref.actor.fsdp_config.param_offload=False
23
+ - actor_rollout_ref.actor.fsdp_config.optimizer_offload=False
24
+ - actor_rollout_ref.rollout.log_prob_micro_batch_size_per_gpu=32
25
+ - actor_rollout_ref.rollout.tensor_model_parallel_size=2
26
+ - actor_rollout_ref.rollout.name=vllm
27
+ - actor_rollout_ref.rollout.gpu_memory_utilization=0.6
28
+ - actor_rollout_ref.rollout.n=1
29
+ - actor_rollout_ref.rollout.prompt_length=1536
30
+ - actor_rollout_ref.rollout.response_length=512
31
+ - actor_rollout_ref.ref.log_prob_micro_batch_size_per_gpu=32
32
+ - actor_rollout_ref.ref.fsdp_config.param_offload=False
33
+ - sgrpo.num_truncations=4
34
+ - sgrpo.force_answer_prompt="Time is limited, stop thinking and start answering.\n</think>\n\n"
35
+ - sgrpo.answer_max_tokens=256
36
+ - trainer.critic_warmup=0
37
+ - trainer.logger=["console"]
38
+ - trainer.project_name=verl_sgrpo_example_gsm8k
39
+ - trainer.experiment_name=qwen3_8b_sgrpo
40
+ - trainer.resume_mode=disable
41
+ - trainer.n_gpus_per_node=8
42
+ - trainer.nnodes=1
43
+ - trainer.save_freq=20
44
+ - trainer.test_freq=5
45
+ - trainer.total_epochs=15
46
+ - trainer.rollout_data_dir=./outputs/sgrpo_samples
47
+ - trainer.rollout_data_dump_freq=20