In this paper, we investigate the alignment between LLMs and people in experiments from social HRI.
@misc{wachowiak2024large,
title={Are Large Language Models Aligned with People's Social Intuitions for Human-Robot Interactions?},
author={Lennart Wachowiak and Andrew Coles and Oya Celiktutan and Gerard Canal},
year={2024},
eprint={2403.05701},
archivePrefix={arXiv},
primaryClass={cs.RO}
}
Correlations are highest with GPT-4, as shown in the following scatterplots:
For full results, refer to the paper. Scatterplots for other models can be found here for Experiment 1 and here for Experiment 2.
To get the video stimuli, use the following GitHub: https://github.com/lwachowiak/HRI-Video-Survey-on-Preferred-Robot-Responses