Search papers, labs, and topics across Lattice.
This survey paper examines the landscape of tool use in LLM-based agents, highlighting its importance for overcoming the limitations of static knowledge. It synthesizes existing research on agent architectures and tool invocation mechanisms, including function calling, dynamic tool retrieval, and autonomous tool creation. The paper identifies key challenges such as knowledge conflicts, performance degradation in long contexts, non-monotonic scaling, and security vulnerabilities, proposing a research agenda for future development.
Tool-augmented LLM agents, despite their promise, face critical challenges like knowledge conflicts and security vulnerabilities that demand a focused research agenda.
The emergence of Large Language Model (LLM)-based agents marks a significant step towards more capable Artificial Intelligence. However, the effectiveness of these agents is fundamentally constrained by the static nature of their internal knowledge. Tool use has become a critical paradigm to overcome these limitations, enabling agents to interact with dynamic data, execute complex computations, and act upon the world. This paper provides a comprehensive survey of the methods, challenges, and future directions in empowering LLM-based agents with tool-use capabilities. Through a systematic literature review, we synthesized the current state of the art, charting the evolution from foundational agent architectures and core invocation mechanisms like function calling to advanced strategies such as dynamic tool retrieval and autonomous tool creation. Our analysis revealed several critical challenges that impede the deployment of robust agents, including knowledge conflicts between internal priors and external evidence, significant performance degradation in long-context scenarios, non-monotonic scaling behaviors in compound systems, and novel security vulnerabilities. By mapping the current research landscape and identifying these key obstacles, this survey proposes a research agenda to guide future efforts in building more capable, secure, and reliable AI agents.