CausalMind

True Causal Reasoning for Social Learning between LLM-Based Agents

Master Thesis (2025)
Author(s)

R.J. Lejeune (TU Delft - Electrical Engineering, Mathematics and Computer Science)

Contributor(s)

C.A. Raman – Mentor (TU Delft - Pattern Recognition and Bioinformatics)

O.K. Shirekar – Mentor (TU Delft - Pattern Recognition and Bioinformatics)

M.J.T. Reinders – Graduation committee member (TU Delft - Pattern Recognition and Bioinformatics)

S. Tan – Graduation committee member (TU Delft - Interactive Intelligence)

Faculty
Electrical Engineering, Mathematics and Computer Science
More Info
expand_more
Publication Year
2025
Language
English
Graduation Date
28-11-2025
Awarding Institution
Delft University of Technology
Programme
['Computer Science']
Faculty
Electrical Engineering, Mathematics and Computer Science
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

Large language models enable rich communication and flexible planning in embodied agents, yet their updates to internal state remain correlation driven, making them prone to hallucinations that undermine reliability in interactive settings. To address this limitation, we investigate whether a true causal model can replace template based belief, desire, and intention (BDI) updates and provide more stable reasoning. We introduce CausalMind, an embodied agent architecture that integrates a causal model trained on text based BDI transitions drawn from a synthetic dataset. We evaluate both the causal model and the full agent in iTHOR across solo and cooperative tasks.

The results indicate that learning causal variables from text embeddings is difficult. The model focused primarily on action related features, failing to separate belief, desire, and intention and instead overfitting to biases in the dataset. Expanding and refining the synthetic dataset improved performance but did not resolve the core issues, highlighting the need for richer data and more suitable encoders. At the agent level, perception errors from the vision module frequently led to hallucinated task completion, limiting the usefulness of teacher guidance and making complex tasks difficult in the iTHOR environment.

Although the current implementation does not outperform template based baselines, the study clarifies the obstacles that arise when extending causal representation learning to natural language BDI states. These findings outline the key requirements for future work toward embodied agents capable of robust causal reasoning rather than correlation driven updates.

Files

CausalMind_final_final.pdf
(pdf | 6.24 Mb)
License info not available