Papers
arxiv:2604.27660

From Context to Skills: Can Language Models Learn from Context Skillfully?

Published on May 3
· Submitted by
ssz
on May 5
#2 Paper of the day
Authors:
,
,
,
,
,
,
,
,
,
,
,

Abstract

A self-evolving framework autonomously discovers and refines context-specific skills for language models through a multi-agent self-play loop with Challenger, Reasoner, and Judge components, improving context learning performance without human supervision.

AI-generated summary

Many real-world tasks require language models (LMs) to reason over complex contexts that exceed their parametric knowledge. This calls for context learning, where LMs directly learn relevant knowledge from the given context. An intuitive solution is inference-time skill augmentation: extracting the rules and procedures from context into natural-language skills. However, constructing such skills for context learning scenarios faces two challenges: the prohibitive cost of manual skill annotation for long, technically dense contexts, and the lack of external feedback for automated skill construction. In this paper, we propose Ctx2Skill, a self-evolving framework that autonomously discovers, refines, and selects context-specific skills without human supervision or external feedback. At its core, a multi-agent self-play loop has a Challenger that generates probing tasks and rubrics, a Reasoner that attempts to solve them guided by an evolving skill set, and a neutral Judge that provides binary feedback. Crucially, both the Challenger and the Reasoner evolve through accumulated skills: dedicated Proposer and Generator agents analyze failure cases and synthesize them into targeted skill updates for both sides, enabling automated skill discovery and refinement. To prevent adversarial collapse caused by increasingly extreme task generation and over-specialized skill accumulation, we further introduce a Cross-time Replay mechanism that identifies the skill set achieving the best balance across representative cases for the Reasoner side, ensuring robust and generalizable skill evolution. The resulting skills can be plugged into any language model to obtain better context learning capability. Evaluated on four context learning tasks from CL-bench, Ctx2Skill consistently improves solving rates across backbone models.

Community

Paper author Paper submitter

Ctx2Skill is a self-evolving framework that autonomously discovers, refines, and selects context-specific skills from complex contexts, requiring no human annotation and no external feedback. The resulting natural-language skills can be plugged into any language model at inference time to enhance context learning capability.

ctx2skill is cool in spirit, but i keep coming back to the binary judge as a potential bottleneck that could push the loop into brittle, overfit skills. the cross-time replay trick is clever, yet i worry how representative those historic skills and tasks really are for shifts in context or for out-of-domain data. an ablation where the judge provides richer feedback or where the proposer/generator are regularized to curb task gaming would help, because right now the signal path feels a bit fragile. arxivlens had a solid breakdown that helped me parse the method details; it covers the failure-driven skill edits and cross-time replay nicely, which is worth a look.

Sign up or log in to comment

Get this paper in your agent:

hf papers read 2604.27660
Don't have the latest CLI?
curl -LsSf https://hf.co/cli/install.sh | bash

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2604.27660 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2604.27660 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2604.27660 in a Space README.md to link it from this page.

Collections including this paper 4