Publication Type
Journal Article
Version
acceptedVersion
Publication Date
2-2025
Abstract
Recently, numerous benchmarks have been developed to evaluate the logical reasoning abilities of large language models (LLMs). However, assessing the equally important creative capabilities of LLMs is challenging due to the subjective, diverse, and data-scarce nature of creativity, especially in multimodal scenarios. In this paper, we consider the comprehensive pipeline for evaluating the creativity of multimodal LLMs, with a focus on suitable evaluation platforms and methodologies. First, we find the Oogiri game—a creativity-driven task requiring humor, associative thinking, and the ability to produce unexpected responses to text, images, or both. This game aligns well with the input-output structure of modern multimodal LLMs and benefits from a rich repository of high-quality, human-annotated creative responses, making it an ideal platform for studying LLM creativity. Next, beyond using the Oogiri game for standard evaluations like ranking and selection, we propose LoTbench, an interactive, causality-aware evaluation framework, to further address some intrinsic risks in standard evaluations, such as information leakage and limited interpretability. The proposed LoTbench not only quantifies LLM creativity more effectively but also visualizes the underlying creative thought processes. Our results show that while most LLMs exhibit constrained creativity, the performance gap between LLMs and humans is not insurmountable. Furthermore, we observe a strong correlation between results from the multimodal cognition benchmark MMMU and LoTbench, but only a weak connection with traditional creativity metrics. This suggests that LoTbench better aligns with human cognitive theories, highlighting cognition as a critical foundation in the early stages of creativity and enabling the bridging of diverse concepts. Project Page.
Keywords
Creativity, Multimodal Large Language Models, Benchmark, Causal Intervention
Discipline
Artificial Intelligence and Robotics | Programming Languages and Compilers
Research Areas
Intelligent Systems and Optimization
Areas of Excellence
Digital transformation
Publication
IEEE Transactions on Pattern Analysis and Machine Intelligence
Volume
47
Issue
5
First Page
3830
Last Page
3846
ISSN
0162-8828
Identifier
10.1109/TPAMI.2025.3539433
Publisher
Institute of Electrical and Electronics Engineers
Citation
HUANG, Zhongzhan; ZHONG, Shanshan; ZHOU, Pan; GAO, Shanghua; ZITNIK, Marink; and LIN, Liang.
A causality-aware paradigm for evaluating creativity of multimodal large language models. (2025). IEEE Transactions on Pattern Analysis and Machine Intelligence. 47, (5), 3830-3846.
Available at: https://ink.library.smu.edu.sg/sis_research/10486
Creative Commons License

This work is licensed under a Creative Commons Attribution-NonCommercial-No Derivative Works 4.0 International License.
Additional URL
https://doi.org/10.1109/TPAMI.2025.3539433
Included in
Artificial Intelligence and Robotics Commons, Programming Languages and Compilers Commons