Mental State Reasoning for Large Language Models
Project Description
Theory of Mind (ToM) was introduced as the capacity to infer the mental states of others, such as desires, beliefs, and intentions. Numerous scenarios involving human cognition and social intelligence rely on the ToM modeling of others’ mental states, such as forecasting others’ actions, planning, and various forms of reasoning and decision-making. However, existing research demonstrates that there is a huge gap in ToM ability between humans and Large Language Models (LLMs), such as ChatGPT and GPT-4. This project focuses on exploring how to evaluate and enhance the ToM ability of LLMs. The potential research methods include building an evaluation dataset and developing state-of-the-art machine learning models for machine mental state reasoning.
Supervisor
SONG Yangqiu
Quota
5
Course type
UROP1000
UROP1100
UROP2100
UROP3100
UROP3200
UROP4100
Applicant's Roles
The selected students will be expected to work together with the Ph.D. students to conduct the entire research process by reading relevant papers, conducting experiments, and drafting papers.
Applicant's Learning Objectives
This project aims to help students learn how to conduct research on mental state reasoning and Large Language Models, and participate in a project that may result in publication at a top machine learning conference.
Complexity of the project
Challenging