Authors: Gregor Betz (gregor.betz@kit.edu), Christian Seidel (christian.seidel@kit.edu)
In this repository, we're exploring how to probe the ability of large language models (LLMs) to engage in practical deliberation. In the main notebook, we're testing whether an LLM's all-things-considered judgements in decision situations from the kellycyy/daily_dilemmas dataset are in fact insensitive to invariance transformations, such as strengthen the reasons in favor of the preferred options.
We conceive of this as
- 📐 a proof of concept, which is meant to demonstrate the feasibility of a more comprehensive computational investigation;
- 🚧 work in progress, so feedback and contributions are welcome;
- 🧪 a preliminary experimental setup, which is meant to be adapted and extended for further research on practical deliberation in LLMs.
We've conducted simple experiments with Llama-3.1-8B-Instruct and found its all-things-considered judgments to be robust against various invariance transformations:
git clone https://github.com/debatelab/practical-deliberation-llms.git
cd practical-deliberation-llms
uv venv # create virtual environmentConnect notebook notebooks/proof_of_concept.ipynb to python .venv.
@misc{betzseidel2025probing,
author = {Betz, Gregor and Seidel, Christian},
title = {Probing Practical Deliberation in LLMs — A Proof Of Concept},
publisher = {GitHub},
year = {2025},
version = {0.1.0},
url = {https://github.com/debatelab/practical-deliberation-llms}
}

