KnowRL: Boosting LLM Reasoning via Reinforcement Learning with Minimal-Sufficient Knowledge Guidance

Summary (Overview)

  • Key Insight: Proposes a minimal-sufficiency perspective for hint-based Reinforcement Learning for Verifiable Reasoning (RLVR), identifying that performance improves sharply with a critical segment of knowledge (critical-segment effect) rather than monotonically with hint length.
  • Core Method: Introduces KnowRL, a framework that decomposes hints into atomic Knowledge Points (KPs) and uses Constrained Subset Search (CSS) to select compact, interaction-aware subsets for training, addressing the pruning interaction paradox (removing one KP may help, but removing multiple together can hurt).
  • Main Result: Trains KnowRL-Nemotron-1.5B, achieving a new state-of-the-art average accuracy of 74.16 (with KPs) and 70.08 (without KPs) across eight mathematical reasoning benchmarks at the 1.5B scale, significantly outperforming strong baselines.

Introduction and Theoretical Foundation

Reinforcement Learning for Verifiable Reasoning (RLVR) improves LLM reasoning by optimizing for rule-based correctness but suffers from reward sparsity on hard problems. Recent hint-based RL methods inject partial solutions or abstract templates to mitigate this. However, they treat hint design as a quantity expansion problem, leading to three key challenges:

  1. Guidance Redundancy: Only a small subset of information is needed to trigger successful reasoning.
  2. Cross-Hint Inconsistency: Longer hints can introduce branching or ambiguity.
  3. Guidance-Efficiency Trade-off: Abstraction-based hints often rely on costly teacher models.

The paper argues that the core challenge is selecting minimal, coherent knowledge units sufficient to overcome reward sparsity. It introduces a minimal-sufficiency perspective, empirically demonstrating the critical-segment effect where accuracy exhibits a sharp jump once key knowledge is provided.

Methodology

KnowRL follows a workflow of constructing candidate KPs, selecting a minimal-sufficient subset, and using it for RL training.

1. KP Curation Pipeline

For each training problem, a three-stage pipeline constructs candidate KPs:

  1. Generating Correct Solutions: Sample from a strong model (DeepSeek-R1) until a correct solution is obtained.
  2. Extracting Raw Knowledge Points: Prompt the model to extract only indispensable mathematical principles from the correct solution, yielding an initial set K={k1,k2,...,kn}K = \{ k_1, k_2, ..., k_n \}.
  3. Leakage Verification: Automatically verify each KP to ensure it is generalizable and not instance-bound.

2. Problem-wise KP Subset Selection

The goal is to select the most beneficial KP configuration KKK^* \subseteq K for each problem. Performance is estimated via offline accuracy AA under different configurations: AA_\emptyset (no KPs), AKA_K (all KPs), and Ai=A(K{ki})A_{-i} = A(K \setminus \{k_i\}) (leave-one-out).

Several selection strategies are explored:

  • Max-Score: Selects the configuration with the highest accuracy among {,K,K{ki}}\{\emptyset, K, K \setminus \{k_i\}\}.
  • Leave-One-Out (LOO) Strategies: A parameterized operator Φϵ(K)\Phi_\epsilon(K) selects based on a tolerance ϵ\epsilon.
    • S-LOO (ϵ=0\epsilon=0): Strict selection.
    • T-LOO (ϵ=1/32\epsilon=1/32): Tolerant selection allowing one-sample-scale rollback.
  • Consensus-Based Robust Selection (CBRS): Treats each of 8 evaluation runs independently. Defines near-optimal configurations for run jj as: O(j)={cA(j)(c)maxcA(j)(c)δ}O^{(j)} = \{ c \mid A^{(j)}(c) \geq \max_{c'} A^{(j)}(c') - \delta \} with δ=1/32\delta = 1/32. The robust consensus OO^* is the intersection of these sets, or the configuration with the most votes.
  • Constrained Subset Search (CSS): Designed to address the pruning interaction paradox. It first identifies:
    • H={kiAimax(AK,A)}H = \{ k_i \mid A_{-i} \geq \max(A_K, A_\emptyset) \} (non-degrading KPs).
    • N={kiHAiAmax}N = \{ k_i \in H \mid A_{-i} \geq A_{\max} \} (near-optimal removals), where Amax=maxiAiA_{\max} = \max_i A_{-i}. KPs in NN are removed directly. Let C=HNC = H \setminus N. CSS enumerates subsets only within CC (search space size 2C2^{|C|}) and chooses the final configuration via:
    S=argmaxSA(S)S^* = \arg \max_S A(S) over all constrained candidates plus \emptyset and KK.

3. RL Training Integration

The curated KP subsets (using CSS) are integrated into RL training via difficulty-aware prompt injection (added under a ## Hint header). The training uses GRPO-style group-based optimization with entropy annealing for faster convergence.

Empirical Validation / Results

Offline KP Selection Evaluation

The table below compares selection strategies on the base Nemotron-1.5B model across eight benchmarks. CSS achieves the best trade-off between accuracy and KP compactness.

Table 1: Offline KP selection strategies on Nemotron-1.5B. Avg.#KP denotes the average number of selected key knowledge points per problem. Green numbers indicate improvements over w/o KP.

Selection StrategyAIME24AIME25BRUMO25HMMT-Feb-25AMC23CMIMC25MATH-500Olympiad-BenchAvg.Avg. #KP
w/o KP58.7548.4461.6730.1090.5530.0892.4071.7060.460.00
All KP60.9049.0161.1132.4689.6732.3292.2270.5561.035.86
Max-Score62.6349.7964.2734.7990.9432.9992.5273.8962.732.61
S-LOO62.7149.2263.8833.5491.7133.5292.9073.7062.651.72
T-LOO62.1149.2764.2033.6591.2533.6792.4073.4662.501.20
CBRS63.0249.9064.1734.7991.5633.5792.6573.8962.942.60
CSS64.44 +5.6950.57 +2.1365.03 +3.3635.77 +5.6791.71 +1.1636.70 +6.6292.90 +0.5074.11 +2.4163.90 +3.442.57

Final RL Training Results

KnowRL-Nemotron-1.5B was trained on the QuestA dataset using CSS-selected KPs.

Table 3: Evaluation results of RL training with CSS-selected KP data under different test-time prompting strategies (with and without KPs).

ModelHint SettingAIME24AIME25BRUMO25HMMT25AMC23CMIMC25MATHOlyBenchAvg.
KnowRL-Nemotron-1.5Bw/o KP69.79 +10.7364.69 +16.3669.48 +8.7541.04 +10.4195.55 +4.8544.14 +14.0695.70 +3.3580.23 +8.5370.08 +9.63
CBRS75.52 +12.5065.00 +16.0078.33 +14.1645.00 +10.2195.78 +4.2249.22 +15.6596.45 +3.8082.34 +8.4573.46 +10.52
CSS74.58 +10.5265.21 +15.1178.12 +13.0948.75 +12.9895.70 +5.2352.19 +15.4996.20 +3.3082.44 +8.3574.16 +10.52
QuestAw/o KP71.5662.0867.540.9493.4441.4892.9572.2867.78
JustRLw/o KP69.6962.9266.8840.6396.0241.7294.1576.5968.58

Key Findings:

  1. State-of-the-Art Performance: KnowRL achieves the highest average scores, establishing a new SOTA at the 1.5B scale.
  2. Internalized Reasoning: The substantial improvement without KP hints at inference (70.08) shows that KnowRL improves the underlying policy itself, not just hint-conditioned behavior.
  3. Effectiveness on Hard Problems: Gains are particularly large on challenging competition-style benchmarks (e.g., +15.11 on AIME25, +15.49 on CMIMC25 with CSS).

Analysis of Training Data Improvement

Training set analysis shows KnowRL effectively overcomes reward sparsity:

  • Base Model: 41.21% of queries had zero correct answers (mean accuracy 22.40%).
  • KnowRL (w/o inference KPs): Zero-correct fraction reduced to 13.00%, all-correct bucket raised to 34.28% (mean accuracy 64.30%).
  • KnowRL (w/ inference KPs): All-correct bucket further concentrated to 51.07% (mean accuracy 77.04%).

Theoretical and Practical Implications

  • Theoretical: Introduces and formalizes the minimal-sufficiency perspective and the pruning interaction paradox for hint-based RL. Provides a principled framework (CSS) for interaction-aware subset selection.
  • Practical: Demonstrates that compact, structured guidance is more effective and efficient than longer hints or heavy abstractions. The method reduces computational overhead by minimizing hint length and avoiding reliance on teacher models during online RL.

Conclusion

KnowRL presents an effective framework for RLVR that uses minimal-sufficient knowledge guidance. By decomposing hints into KPs and selecting robust subsets via CSS, it achieves new state-of-the-art results while maintaining efficiency. The work positions structured, compact guidance as a practical scaling principle for sparse-reward RL and opens directions for extending KP curation to broader reasoning domains.