Wed 8 Oct 2025 14:54 - 15:12 at DCIH 102 - Session 3: Large Language Models and Modeling Chair(s): Bentley Oakes

Recently, large language models (LLMs) have achieved widespread application across various fields. Despite their impressive capabilities, LLMs suffer from a lack of structured reasoning ability, particularly for complex tasks requiring domain-specific best practices, which are often unavailable in the training data. Although multi-step prompting methods incorporating human best practices, such as chain-of-thought and tree-of-thought, have gained popularity, they lack a general mechanism to control LLM behavior. In this paper, we propose SHERPA, a model-driven framework to improve the LLM performance on complex tasks by explicitly incorporating domain-specific best practices into hierarchical state machines. By structuring the LLM execution processes using state machines, SHERPA enables more fine-grained control over their behavior via rules or decisions driven by machine learning-based approaches, including LLMs. We show that SHERPA is applicable to a wide variety of tasks—specifically, code generation, class name generation, and question answering—replicating previously proposed approaches while further improving the performance. We demonstrate the effectiveness of SHERPA for the aforementioned tasks using various LLMs. Our systematic evaluation compares different state machine configurations against baseline approaches without state machines. Results show that integrating well-designed state machines significantly improves the quality of LLM outputs. Additionally, SHERPA is particularly beneficial for complex tasks with well-established human best practices but lacking data used for training LLMs.

Wed 8 Oct

Displayed time zone: Eastern Time (US & Canada) change

14:00 - 15:30
Session 3: Large Language Models and ModelingResearch Papers / New Ideas and Emerging Results (NIER) at DCIH 102
Chair(s): Bentley Oakes Polytechnique Montréal

Hybrid

14:00
18m
Talk
MCeT: Behavioral Model Correctness Evaluation using Large Language ModelsFT@In Person
Research Papers
Khaled Ahmed Huawei Research Canada, University of British Columbia (UBC), Jialing Song Huawei Technologies Canada, Boqi Chen McGill University, Ou Wei Huawei Technologies Canada, Bingzhou Zheng Huawei Technologies Canada
Pre-print
14:18
18m
Talk
Model-Driven Quantum Code Generation Using Large Language Models and Retrieval-Augmented GenerationIn Person
New Ideas and Emerging Results (NIER)
Nazanin Siavash University of Colorado Colorado Springs (UCCS), Armin Moin University of Colorado Colorado Springs
14:36
18m
Talk
Towards LLM-enhanced Conflict Detection and Resolution in Model VersioningRemote
New Ideas and Emerging Results (NIER)
Martin Eisenberg Johannes Kepler University, Linz, Stefan Klikovits Johannes Kepler University, Linz, Manuel Wimmer JKU Linz, Konrad Wieland LieberLieber Software GmbH
14:54
18m
Talk
SHERPA: A Model-Driven Framework for Large Language Model Execution@RemoteFTArtifacts Available
Research Papers
Boqi Chen McGill University, Kua Chen McGill University, José Antonio Hernández López Department of Computer Science and Systems, University of Murcia, Gunter Mussbacher McGill University, Daniel Varro Linköping University / McGill University, Amir Feizpour Aggregate Intellect
Pre-print
15:12
18m
Talk
Accurate and Consistent Graph Model Generation from Text with Large Language Models@RemoteFT
Research Papers
Boqi Chen McGill University, Ou Wei Huawei Technologies Canada, Bingzhou Zheng Huawei Technologies Canada, Gunter Mussbacher McGill University
Pre-print