Publication Type
Journal Article
Version
acceptedVersion
Publication Date
2-2026
Abstract
Function calling is a fundamental capability of today's large language models, but sequential function calling posed efficiency problems. Recent studies have proposed to request function calls with parallelism support in order to alleviate this issue. However, they either delegate the concurrent function calls to users for execution which are conversely executed sequentially, or overlook the relations among various function calls, rending limited efficiency. This paper introduces LLMOrch, an advanced framework for automated, parallel function calling in large language models. The key principle behind LLMOrch is to identify an available processor to execute a function call while preventing any single processor from becoming overburdened. To this end, LLMOrch models the data relations (i.e., definition-use (def-use) dependencies among different function calls and coordinates their executions by their contro l relations (i.e., mutual-exclusion) as well as the working status of the underlying processors. When comparing with state-of-the-art techniques, LLMOrch demonstrated comparable efficiency improvements in orchestrating I/O-intensive functions, while significantly outperforming (2 & times;) them with compute-intensive functions. LLMOrch's performance even showed a linear correlation to the number of allocated processors. We believe that these results highlight the potential of LLMOrch as an efficient solution for parallel function orchestration in the context of large language models.
Keywords
Schedules, Large language models, Translation, Processor scheduling, Chatbots, Computational modeling, Scheduling, Python, Process control, Parallel processing, function call, parallel function call
Discipline
Software Engineering
Research Areas
Software and Cyber-Physical Systems
Publication
IEEE Transactions on Software Engineering
Volume
52
Issue
2
First Page
411
Last Page
427
ISSN
0098-5589
Identifier
10.1109/TSE.2025.3619112
Publisher
Institute of Electrical and Electronics Engineers
Citation
LIU, Xiaoxia; DI, Peng; LI, Cong; SUN, Jun; and WANG, Jingyi.
Efficient function orchestration for Large Language Models. (2026). IEEE Transactions on Software Engineering. 52, (2), 411-427.
Available at: https://ink.library.smu.edu.sg/sis_research/11019
Copyright Owner and License
Authors
Creative Commons License

This work is licensed under a Creative Commons Attribution-NonCommercial-No Derivative Works 4.0 International License.
Additional URL
https://doi.org/10.1109/TSE.2025.3619112