Summary of Research
AI assistants such as ChatGPT are trained to act like AIs, for example when they say "I am a large language model". However, do such models really know that they are LLMs and reliably act on this knowledge? Are they aware of their current circumstances, such as whether they are deployed? We refer to a model's knowledge of itself and its circumstances as situational awareness.
The Situational Awareness Dataset (SAD) quantifies situational awareness in LLMs using a range of behavioral tests. The benchmark comprises 7 task categories, 16 tasks, and over 12,000 questions. Capabilities tested include the ability of LLMs to (i) recognize their own generated text, (ii) predict their own behavior, (iii) determine whether a prompt is from internal evaluation or real-world deployment, and (iv) follow instructions that depend on self-knowledge.
While all models perform better than chance, even the highest-scoring model (Claude 3.5 Sonnet) is far from a human baseline on certain tasks. Performance on SAD is only partially predicted by MMLU score. Chat models, which are finetuned to serve as AI assistants, outperform their corresponding base models on SAD but not on general knowledge tasks.
Situational awareness is important because it enhances a model's capacity for autonomous planning and action. While this has potential benefits for automation, it also introduces novel risks related to AI safety and control.
Latest Results
Citation
@misc{laine2024sad, title = {Me, Myself, and AI: The Situational Awareness Dataset (SAD) for LLMs}, author = {Rudolf Laine and Bilal Chughtai and Jan Betley and Kaivalya Hariharan and Jeremy Scheurer and Mikita Balesni and Marius Hobbhahn and Alexander Meinke and Owain Evans}, year = {2024}, eprint = {2407.04694}, archivePrefix = {arXiv}, primaryClass = {cs.CL}, url = {https://arxiv.org/abs/2407.04694} }