Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

In-Context Learning for Pure Exploration

Created by
  • Haebom

Author

Alessio Russo, Ryan Welch, Aldo Pacchiano

Outline

This paper studies the problem of active sequential hypothesis testing (AST), where a learner adaptively gathers data from the environment to efficiently determine a fundamentally correct hypothesis for a new task. We present examples of the Best-Arm Identification (BAI) task in the multi-loss bandit problem and the generalized search problem. We introduce In-Context Pure Exploration (ICPE), which meta-trains a Transformer to map observation history to query behavior and predicted hypotheses. ICPE actively gathers evidence for a new task and infers the true hypothesis without parameter updates at inference time. On deterministic, probabilistic, and structured benchmarks, including BAI and generalized search, ICPE outperforms adaptive baselines without explicitly modeling the information structure.

Takeaways, Limitations

The ICPE model utilizing Transformer is presented as a practical architecture for pure search problems.
It demonstrates competitive performance with adaptive baselines on various benchmarks such as BAI and generalized search.
The advantage of being able to solve problems without explicit modeling of the information structure.
Demonstrate the model's generalization ability and adaptability to new tasks.
We present an efficient meta-learning and inference process for Transformer-based models.
The specific Limitations of the paper is not mentioned (but is not stated in the abstract).
👍