Skill1: Unified Evolution of Skill-Augmented Agents via Reinforcement Learning

Summary (Overview)

  • Unified Co-evolution Framework: Skill1 is a framework that trains a single policy to co-evolve three core capabilities of skill-augmented agents—skill selection, utilization, and distillation—simultaneously towards a shared task-outcome objective.
  • Single-Signal Credit Assignment: It derives all learning signals from a single binary task-success reward r(τ){0,1}r(\tau) \in \{0, 1\}, decomposing it into a low-frequency trend (for skill selection) and high-frequency variation (for skill distillation), eliminating the need for auxiliary or conflicting reward sources.
  • State-of-the-Art Performance: Skill1 achieves a 97.5% average success rate on ALFWorld and strong performance on WebShop, outperforming prior skill-based and reinforcement learning baselines.
  • Mutually Reinforcing Dynamics: Empirical analysis confirms that the three capabilities improve simultaneously under the unified signal, and ablations show that removing any stage's credit signal degrades all capabilities, demonstrating their mutual dependence.
  • Efficient Library Management: The framework includes mechanisms for skill library maintenance (admission, retirement) and distillation, which compresses experience into concise skills, controlling computational overhead and improving library quality.

Introduction and Theoretical Foundation

Training Large Language Model (LLM) agents via Reinforcement Learning (RL) typically treats each task as an isolated episode, where successful strategies are only implicitly absorbed into policy parameters. A promising solution is to augment agents with a persistent skill library that accumulates reusable strategies, allowing the agent to draw on past successes. The workflow of such skill-augmented agents involves three coupled stages:

  1. Skill Selection: Choosing a relevant skill from the library for the current task.
  2. Skill Utilization: Executing the task while being guided by the selected skill.
  3. Skill Distillation: Deriving new reusable skills from the experience.

Existing methods often optimize these capabilities in isolation or with separate reward sources, leading to partial evolution and conflicting optimization pressures. This paper addresses two fundamental open questions:

  1. How can an agent evolve all three capabilities simultaneously?
  2. How can they co-evolve toward a shared objective?

The authors propose Skill1, a framework that achieves unified evolution by training a single policy to perform and optimize all three stages. The key innovation is credit assignment on a single task-outcome signal r(τ)r(\tau), decomposing it to provide targeted learning for each capability without auxiliary models.

The problem is formulated as a Partially Observable Markov Decision Process (POMDP) M=(S,A,O,T,Ω,R,γ)\mathcal{M} = (\mathcal{S}, \mathcal{A}, \mathcal{O}, T, \Omega, R, \gamma). A state S=(x,e,B)S = (x, e, \mathcal{B}) includes a task instruction xx, environment state ee, and a persistent skill library B={s1,s2,}\mathcal{B} = \{s_1, s_2, \dots\}. A skill sBs \in \mathcal{B} consists of a natural-language strategy s.strat (how to act) and a scenario description s.desc (when it applies). The overall training objective is:

maxθExD,τπθ(x)[r(τ)],\max_{\theta} \mathbb{E}_{x \sim \mathcal{D}, \tau \sim \pi_\theta(\cdot|x)} \left[ r(\tau) \right],

where πθ\pi_\theta is the policy optimized with RL algorithms like GRPO.

Methodology

Skill1 trains a single policy πθ\pi_\theta to perform a sequential three-stage workflow for each task xDx \sim \mathcal{D}, generating a complete trajectory τ=(q,z,a1,o1,,aT,oT,snew)\tau = (q, z, a_1, o_1, \dots, a_T, o_T, s_{\text{new}}).

3.1 Agent Workflow

  1. Skill Selection:

    • Query Generation: The policy generates a natural-language query qπθ(x)q \sim \pi_\theta(\cdot | x).
    • Retrieval: A frozen encoder EE retrieves the top-KK candidates by semantic similarity: BK=top-KsBsim(E(q),E(s.desc)).\mathcal{B}_K = \text{top-}K_{s \in \mathcal{B}} \text{sim}\left( E(q), E(s.\text{desc}) \right).
    • Re-ranking: The policy re-ranks these candidates by generating a permutation σπθ(x,BK)\sigma \sim \pi_\theta(\cdot | x, \mathcal{B}_K), and the top-ranked skill zz is selected for utilization.
  2. Skill Utilization: The policy interacts with the environment for up to TT turns conditioned on the selected skill's strategy: τπθ(x,z.strat,ot)\tau \sim \pi_\theta(\cdot | x, z.\text{strat}, o_{\leq t}). For each task, GG independent rollouts are sampled.

  3. Skill Distillation: After each rollout, the policy reflects on the trajectory to produce a new skill snews_{\text{new}}:

    • A reusable strategy snew.stratπθ(x,τ)s_{\text{new}}.\text{strat} \sim \pi_\theta(\cdot | x, \tau).
    • A scenario description snew.descπθ(x,τ)s_{\text{new}}.\text{desc} \sim \pi_\theta(\cdot | x, \tau). A new skill is admitted to B\mathcal{B} only if r(τ)=1r(\tau) = 1. When the library reaches capacity B=Nmax|\mathcal{B}| = N_{\max}, the skill with the lowest retirement score U(s)logn(s)U(s) \cdot \log n(s) is removed, where n(s)n(s) is the selection count.

3.2 Reward Assignment

The core challenge is assigning credit from the single task-outcome r(τ)r(\tau) to capabilities operating at different temporal scopes (episode, cross-episode, library improvement).

  • Crediting Utilization: The outcome directly measures execution quality:

    Riutil=r(τi).R_i^{\text{util}} = r(\tau_i).
  • Crediting Selection:

    • The query qq receives gradients through the utilization objective.
    • For re-ranking, a per-skill utility score U(s)U(s) is maintained as a low-frequency trend, updated via exponential moving average after each rollout: U(s)(1α)U(s)+αr(τi),sBKi.U(s) \leftarrow (1 - \alpha) \cdot U(s) + \alpha \cdot r(\tau_i), \quad \forall s \in \mathcal{B}_K^i. The best available utility U^i=maxsBKiU(s)\hat{U}_i = \max_{s \in \mathcal{B}_K^i} U(s) serves as the library baseline.
    • The trend supervises re-ranking by rewarding permutations σi\sigma_i that agree with the utility ordering, using Normalized Discounted Cumulative Gain (NDCG): Rirerank=NDCG(σi,argsort([U(BKi)])).R_i^{\text{rerank}} = \text{NDCG}\left( \sigma_i, \text{argsort}(-[U(\mathcal{B}_K^i)]) \right).
  • Crediting Distillation: The credit is the high-frequency variation of the current outcome relative to the library's trend, approximating whether the new skill improves future performance:

    Ridistill=r(τi)U^i.R_i^{\text{distill}} = r(\tau_i) - \hat{U}_i.

    A positive variation indicates the experience surpasses the current library boundary.

3.3 Joint Optimization

Each rollout τi\tau_i concatenates four generation segments (query qiq_i, permutation σi\sigma_i, actions a1:Ta_{1:T}, distilled skill snew,is_{\text{new},i}). They are optimized jointly in a single gradient step using GRPO, with a combined objective:

J(θ)=Jutil(θ)+λ1Jrerank(θ)+λ2Jdistill(θ).J(\theta) = J^{\text{util}}(\theta) + \lambda_1 J^{\text{rerank}}(\theta) + \lambda_2 J^{\text{distill}}(\theta).
  • Utilization & Query: Jutil(θ)J^{\text{util}}(\theta) uses group-relative advantages computed from RiutilR_i^{\text{util}}.
  • Re-ranking: Optimized independently per rollout with a REINFORCE-style objective: Jrerank(θ)=1NGiRireranklogπθ(σixi,BKi).J^{\text{rerank}}(\theta) = \frac{1}{N \cdot G} \sum_i R_i^{\text{rerank}} \cdot \log \pi_\theta(\sigma_i | x_i, \mathcal{B}_K^i).
  • Distillation: Jdistill(θ)J^{\text{distill}}(\theta) uses group-relative advantages computed from RidistillR_i^{\text{distill}}.

The full procedure is summarized in Algorithm 1.

Empirical Validation / Results

Experimental Setup: Evaluated on ALFWorld (text-based household environment) and WebShop (online shopping simulator). The base policy is Qwen2.5-7B-Instruct, trained with GRPO (G=16G=16). The skill library is initialized empty with a capacity of 5,000.

4.2 Main Results

Table 1: Main results on ALFWorld and WebShop (Success Rate, %). Bold denotes best results; ↑ indicates improvement over the previous best. “Avg.” stands for average success rate and “Succ.“ stands for success rate.

MethodPickLookCleanHeatCoolPick2Avg.WebShop ScoreWebShop Succ.
w/o Training
Zero-Shot33.421.619.36.92.83.214.826.47.8
ReAct48.535.434.313.218.217.631.246.219.5
Reflexion62.041.644.930.936.323.842.758.128.8
RL-Trained w/o Skills
PPO92.364.092.589.580.368.880.481.468.7
GRPO90.866.189.374.772.564.777.679.366.1
GiGPO97.782.798.883.789.379.290.884.472.8
RL-Trained w/ Skills
SkillRL97.971.490.090.095.587.589.985.272.7
RetroAgent97.990.999.292.985.391.094.988.982.3
Skill1 (Ours)100.0↑2.198.6↑7.797.399.2↑6.396.1↑0.696.0↑5.097.5↑2.689.782.9
  • Skill1 achieves a 97.5% average success rate on ALFWorld, surpassing the previous best (RetroAgent, 94.9%) by 2.6 points and ranking first on 5 out of 6 task types. It also demonstrates the best performance on WebShop.
  • Skill1 outperforms the strongest RL-only method (GiGPO, 90.8%) by 6.7 points, demonstrating the value of an explicit, reusable skill library.
  • Performance increases with the degree of co-evolution, as Skill1 optimizes all three stages while prior methods leave at least one unoptimized or use separate rewards.

4.3 Analysis

Ablation Study (Table 2):

Table 2: Ablation study on ALFWorld (Success Rate %). Upper block ablates workflow components; lower block ablates training objectives.

MethodPickLookCleanHeatCoolPick2Avg.
Skill1100.098.697.399.296.196.097.5
w/o Selection96.990.398.090.486.585.391.8
w/o Distillation97.488.598.196.187.689.592.4
w/o Library96.771.594.970.771.565.580.9
w/ λ1=0\lambda_1=099.580.598.8100.090.684.994.0
w/ λ2=0\lambda_2=0100.085.495.596.491.096.294.9
w/ λ1=λ2=0\lambda_1=\lambda_2=098.174.995.695.679.587.290.2
  • The skill library is foundational; removing it causes the largest drop (16.6 points).
  • Removing distillation or selection also significantly harms performance (5.1 and 5.7 point drops, respectively).
  • The auxiliary objectives (λ1,λ2\lambda_1, \lambda_2) are complementary; removing both causes a sharper decline than removing each individually.

Co-evolution Dynamics:

  • Under unified training, the three capabilities (selection precision, utilization success rate, distillation positive rate) exhibit mutual reinforcement and converge simultaneously.
  • Ablating any credit signal slows all three capabilities, evidencing their mutual dependence.

Evolution of Skill Management:

  • The policy learns to generate increasingly precise selection queries (task-skill similarity improves from 0.51 to 0.60).
  • The library ceiling U^\hat{U} rises (from ~0.5 to 0.91), indicating the policy distills increasingly effective skills due to the variation signal.

Skill Library Diversity:

  • Skill1 activates a broader set of skills more frequently and covers a diverse strategy space, unlike ablations where skill usage concentrates on a few popular skills.

Computational Overhead (Table 3):

Table276: Computational cost on ALFWorld training. We report wall-clock time per step (seconds) and library size (number of skills) at three checkpoints.

MethodTime / Step (s)Library Size
Step 20Step 60Step 100Step 20Step 60Step 100
GRPO (no library)301.3274.1296.7
SkillRL368.1319.0326.6607183
Skill1386.6444.3493.89153,8995,000
w/o Distill. Step508.8750.1738.42,2125,0005,000
  • Skill1 adds moderate overhead (1.3-1.7x slower than GRPO) due to the growing library context.
  • Distillation is crucial for controlling cost; without it, raw trajectories inflate the library, making selection slower (69% slower by step 60).

Theoretical and Practical Implications

  • Theoretical: Skill1 provides a principled framework for unified credit assignment in multi-stage, skill-based RL. It demonstrates how a single objective signal can be decomposed to guide optimization across different temporal scopes (episodic, cross-episodic, library-level).
  • Practical: The framework enables the development of more sample-efficient and capable LLM agents that can autonomously build and leverage a knowledge base of reusable strategies. It reduces the need for manually engineering separate reward functions or training pipelines for different agent capabilities.
  • Empirical: The results establish a new state-of-the-art for skill-augmented agents on benchmark environments and provide clear evidence that co-evolving all stages of the skill lifecycle is superior to optimizing them in isolation.

Conclusion

Skill1 presents a framework for the unified evolution of skill-augmented agents by training a single policy to co-evolve skill selection, utilization, and distillation toward a shared task-outcome objective. By decomposing the reward signal into low-frequency trend and high-frequency variation components, it achieves targeted credit assignment without auxiliary rewards. Experiments confirm significant performance gains, coupled evolution of capabilities, and the necessity of each credit signal.

Limitations:

  • Environment Coverage: Evaluation is limited to two text-based environments; generalization to visual or deeper search environments is unexplored.
  • Scalability: The fixed-size skill library (5,000 entries) may become a bottleneck for highly diverse tasks, requiring more sophisticated eviction or hierarchical organization strategies.
  • Safety & Oversight: Autonomous skill accumulation