Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Cognitive Kernel-Pro: A Framework for Deep Research Agents and Agent Foundation Models Training

Created by
  • Haebom

Author

Tianqing Fang, Zhisong Zhang, Xiaoyang Wang, Rui Wang, Can Qin, Yuxuan Wan, Jun-Yu Ma, Ce Zhang, Jiaqi Chen, Xiyun Li, Hongming Zhang, Haitao Mi, Dong Yu

Outline

Cognitive Kernel-Pro is a fully open-source, free, and multi-modular agent framework for general AI agents, enabling complex reasoning, web interaction, coding, and autonomous research. This paper systematically examines the curation of high-quality training data for agent-based models, focusing on building queries, paths, and verifiable answers in four key domains: web, files, code, and general reasoning. Furthermore, we explore novel strategies for agent test-time reflection and voting to enhance agent robustness and performance. We evaluated Cognitive Kernel-Pro based on GAIA, achieving state-of-the-art results among open-source and free agents. Specifically, our 8-billion-parameter open-source model outperforms previous leading systems such as WebDancer and WebSailor, establishing a new performance standard for accessible and high-performance AI agents.

Takeaways, Limitations

Takeaways:
Providing a high-performance, open-source, and freely accessible AI agent framework.
Presenting a high-quality training data curation strategy and building datasets for four key areas: web, files, code, and general inference.
We propose a strategy to improve agent robustness and performance through agent testing time reflection and voting.
Achieve performance that surpasses existing top-performing open source agents.
Contributing to improving accessibility and reproducibility of AI agent research
Limitations:
The paper does not explicitly mention the specific Limitations. Further research is needed to improve performance and overcome limitations.
The performance of the 8 billion parameter model may be limited to a specific benchmark (GAIA), and its performance on other benchmarks or in real-world application environments requires further validation.
Although open source, access to the hardware and software resources required to run it may be limited.
👍