Skip to content

Commit

Permalink
added gandhi affect paper
Browse files Browse the repository at this point in the history
  • Loading branch information
tobiasgerstenberg committed Sep 20, 2024
1 parent 680a643 commit 06009e4
Show file tree
Hide file tree
Showing 28 changed files with 874 additions and 152 deletions.
5 changes: 4 additions & 1 deletion content/publication/amemiya2024disagreement.md
Original file line number Diff line number Diff line change
Expand Up @@ -24,7 +24,10 @@ url_slides = ""
url_video = ""
url_poster = "posters/amemiya2023disagreement-poster.pdf"
url_source = ""
url_custom = [{name = "Github", url = "https://github.com/cicl-stanford/children_disagree"}]
url_custom = [{name = "Link", url = "https://www.sciencedirect.com/science/article/pii/S0010027724001227"},
{name = "Github", url = "https://github.com/cicl-stanford/children_disagree"},
{name = "Press: Psypost", url = "https://www.psypost.org/childrens-ability-to-detect-ambiguity-in-disagreements-sharpens-between-ages-7-and-11/"}
]
math = true
highlight = true
[header]
Expand Down
2 changes: 1 addition & 1 deletion content/publication/du2024robotic.md
Original file line number Diff line number Diff line change
Expand Up @@ -9,7 +9,7 @@
title = "To Err is Robotic: Rapid Value-Based Trial-and-Error during Deployment"
date = "2024-06-26"
authors = ["M. Du","A. Khazatsky","T. Gerstenberg","C. Finn"]
publication_types = ["2"]
publication_types = ["1"]
publication_short = "_arXiv_"
publication = "Du, M., Khazatsky, A., Gerstenberg, T., Finn, C. (2024). To Err is Robotic: Rapid Value-Based Trial-and-Error during Deployment. _arXiv_."
abstract = "When faced with a novel scenario, it can be hard to succeed on the first attempt. In these challenging situations, it is important to know how to retry quickly and meaningfully. Retrying behavior can emerge naturally in robots trained on diverse data, but such robot policies will typically only exhibit undirected retrying behavior and may not terminate a suboptimal approach before an unrecoverable mistake. We can improve these robot policies by instilling an explicit ability to try, evaluate, and retry a diverse range of strategies. We introduce Bellman-Guided Retrials, an algorithm that works on top of a base robot policy by monitoring the robot's progress, detecting when a change of plan is needed, and adapting the executed strategy until the robot succeeds. We start with a base policy trained on expert demonstrations of a variety of scenarios. Then, using the same expert demonstrations, we train a value function to estimate task completion. During test time, we use the value function to compare our expected rate of progress to our achieved rate of progress. If our current strategy fails to make progress at a reasonable rate, we recover the robot and sample a new strategy from the base policy while skewing it away from behaviors that have recently failed. We evaluate our method on simulated and real-world environments that contain a diverse suite of scenarios. We find that Bellman-Guided Retrials increases the average absolute success rates of base policies by more than 20% in simulation and 50% in real-world experiments, demonstrating a promising framework for instilling existing trained policies with explicit trial and error capabilities. For evaluation videos and other documentation, go to https://sites.google.com/view/to-err-robotic/home"
Expand Down
2 changes: 1 addition & 1 deletion content/publication/franken2024sami.md
Original file line number Diff line number Diff line change
Expand Up @@ -9,7 +9,7 @@
title = "Self-supervised alignment with mutual information: Learning to follow principles without preference labels"
date = "2024-04-22"
authors = ["J. Fränken","E. Zelikman","R. Rafailov","K. Gandhi","T. Gerstenberg","N. D. Goodman"]
publication_types = ["2"]
publication_types = ["1"]
publication_short = "_arXiv_"
publication = "Fränken, J., Zelikman, E., Rafailov, R., Gandhi, K., Gerstenberg, T., Goodman, N. D. (2024). Self-supervised alignment with mutual information: Learning to follow principles without preference labels. _arXiv_."
abstract = "When prompting a language model (LM), users frequently expect the model to adhere to a set of behavioral principles across diverse tasks, such as producing insightful content while avoiding harmful or biased language. Instilling such principles into a model can be resource-intensive and technically challenging, generally requiring human preference labels or examples. We introduce SAMI, a method for teaching a pretrained LM to follow behavioral principles that does not require any preference labels or demonstrations. SAMI is an iterative algorithm that finetunes a pretrained LM to increase the conditional mutual information between constitutions and self-generated responses given queries from a datasest. On single-turn dialogue and summarization, a SAMI-trained mistral-7b outperforms the initial pretrained model, with win rates between 66% and 77%. Strikingly, it also surpasses an instruction-finetuned baseline (mistral-7b-instruct) with win rates between 55% and 57% on single-turn dialogue. SAMI requires a 'principle writer' model; to avoid dependence on stronger models, we further evaluate aligning a strong pretrained model (mixtral-8x7b) using constitutions written by a weak instruction-finetuned model (mistral-7b-instruct). The SAMI-trained mixtral-8x7b outperforms both the initial model and the instruction-finetuned model, achieving a 65% win rate on summarization. Our results indicate that a pretrained LM can learn to follow constitutions without using preference labels, demonstrations, or human oversight."
Expand Down
33 changes: 33 additions & 0 deletions content/publication/gandhi2024affective.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,33 @@
+++
# 0 -> 'Forthcoming',
# 1 -> 'Preprint',
# 2 -> 'Journal',
# 3 -> 'Conference Proceedings',
# 4 -> 'Book chapter',
# 5 -> 'Thesis'

title = "Human-like Affective Cognition in Foundation Models"
date = "2024-09-20"
authors = ["K. Gandhi","Z. Lynch","J. Fränken","K. Patterson","S. Wambu","T. Gerstenberg","D. C. Ong","N. D. Goodman"]
publication_types = ["1"]
publication_short = "_arXiv_"
publication = "Gandhi K., Lynch Z., Fränken J., Patterson K., Wambu S., Gerstenberg T., Ong D. C., Goodman N. D. (2024). Human-like Affective Cognition in Foundation Models. _arXiv_."
abstract = "Understanding emotions is fundamental to human interaction and experience. Humans easily infer emotions from situations or facial expressions, situations from emotions, and do a variety of other affective cognition. How adept is modern AI at these inferences? We introduce an evaluation framework for testing affective cognition in foundation models. Starting from psychological theory, we generate 1,280 diverse scenarios exploring relationships between appraisals, emotions, expressions, and outcomes. We evaluate the abilities of foundation models (GPT-4, Claude-3, Gemini-1.5-Pro) and humans (N = 567) across carefully selected conditions. Our results show foundation models tend to agree with human intuitions, matching or exceeding interparticipant agreement. In some conditions, models are ``superhuman'' -- they better predict modal human judgements than the average human. All models benefit from chain-of-thought reasoning. This suggests foundation models have acquired a human-like understanding of emotions and their influence on beliefs and behavior."
image_preview = ""
selected = false
projects = []
#url_pdf = "papers/gandhi2024affective.pdf"
url_preprint = "https://arxiv.org/abs/2409.11733"
url_code = ""
url_dataset = ""
url_slides = ""
url_video = ""
url_poster = ""
url_source = ""
#url_custom = [{name = "Github", url = ""}]
math = true
highlight = true
[header]
# image = "publications/gandhi2024affective.png"
caption = ""
+++
2 changes: 1 addition & 1 deletion content/publication/gerstenberg2024counterfactual.md
Original file line number Diff line number Diff line change
Expand Up @@ -24,7 +24,7 @@ url_slides = ""
url_video = ""
url_poster = ""
url_source = ""
url_custom = [{name = "Press: HAI News", url = "https://hai.stanford.edu/news/humans-use-counterfactuals-reason-about-causality-can-ai"}]
url_custom = [{name = "Link", url = "https://www.sciencedirect.com/science/article/pii/S1364661324001074"}, {name = "Press: HAI News", url = "https://hai.stanford.edu/news/humans-use-counterfactuals-reason-about-causality-can-ai"}]
math = true
highlight = true
[header]
Expand Down
8 changes: 4 additions & 4 deletions docs/404.html
Original file line number Diff line number Diff line change
Expand Up @@ -237,6 +237,10 @@ <h1>Page not found</h1>

<h2>Publications</h2>

<ul>
<li><a href="https://cicl.stanford.edu/publication/gandhi2024affective/">Human-like Affective Cognition in Foundation Models</a></li>
</ul>

<ul>
<li><a href="https://cicl.stanford.edu/publication/du2024robotic/">To Err is Robotic: Rapid Value-Based Trial-and-Error during Deployment</a></li>
</ul>
Expand All @@ -253,10 +257,6 @@ <h2>Publications</h2>
<li><a href="https://cicl.stanford.edu/publication/wu2024whodunnit/">Whodunnit? Inferring what happened from multimodal evidence</a></li>
</ul>

<ul>
<li><a href="https://cicl.stanford.edu/publication/tsirtsis2024sequential/">Towards a computational model of responsibility judgments in sequential human-AI collaboration</a></li>
</ul>




Expand Down
12 changes: 11 additions & 1 deletion docs/bibtex/cic_papers.bib
Original file line number Diff line number Diff line change
@@ -1,13 +1,23 @@
%% This BibTeX bibliography file was created using BibDesk.
%% https://bibdesk.sourceforge.io/
%% Created for Tobias Gerstenberg at 2024-06-26 12:25:09 -0700
%% Created for Tobias Gerstenberg at 2024-09-20 10:50:37 -0700
%% Saved with string encoding Unicode (UTF-8)
@article{gandhi2024affective,
abstract = {Understanding emotions is fundamental to human interaction and experience. Humans easily infer emotions from situations or facial expressions, situations from emotions, and do a variety of other affective cognition. How adept is modern AI at these inferences? We introduce an evaluation framework for testing affective cognition in foundation models. Starting from psychological theory, we generate 1,280 diverse scenarios exploring relationships between appraisals, emotions, expressions, and outcomes. We evaluate the abilities of foundation models (GPT-4, Claude-3, Gemini-1.5-Pro) and humans (N = 567) across carefully selected conditions. Our results show foundation models tend to agree with human intuitions, matching or exceeding interparticipant agreement. In some conditions, models are ``superhuman'' -- they better predict modal human judgements than the average human. All models benefit from chain-of-thought reasoning. This suggests foundation models have acquired a human-like understanding of emotions and their influence on beliefs and behavior.},
author = {Kanishk Gandhi and Zoe Lynch and Jan-Philipp Fr{\"a}nken and Kayla Patterson and Sharon Wambu and Tobias Gerstenberg and Desmond C. Ong and Noah D. Goodman},
date-added = {2024-09-20 10:50:33 -0700},
date-modified = {2024-09-20 10:50:37 -0700},
journal = {arXiv},
note = {https://arxiv.org/abs/2409.11733},
title = {Human-like Affective Cognition in Foundation Models},
year = {2024}}

@article{du2024robotic,
abstract = {When faced with a novel scenario, it can be hard to succeed on the first attempt. In these challenging situations, it is important to know how to retry quickly and meaningfully. Retrying behavior can emerge naturally in robots trained on diverse data, but such robot policies will typically only exhibit undirected retrying behavior and may not terminate a suboptimal approach before an unrecoverable mistake. We can improve these robot policies by instilling an explicit ability to try, evaluate, and retry a diverse range of strategies. We introduce Bellman-Guided Retrials, an algorithm that works on top of a base robot policy by monitoring the robot's progress, detecting when a change of plan is needed, and adapting the executed strategy until the robot succeeds. We start with a base policy trained on expert demonstrations of a variety of scenarios. Then, using the same expert demonstrations, we train a value function to estimate task completion. During test time, we use the value function to compare our expected rate of progress to our achieved rate of progress. If our current strategy fails to make progress at a reasonable rate, we recover the robot and sample a new strategy from the base policy while skewing it away from behaviors that have recently failed. We evaluate our method on simulated and real-world environments that contain a diverse suite of scenarios. We find that Bellman-Guided Retrials increases the average absolute success rates of base policies by more than 20% in simulation and 50% in real-world experiments, demonstrating a promising framework for instilling existing trained policies with explicit trial and error capabilities. For evaluation videos and other documentation, go to https://sites.google.com/view/to-err-robotic/home},
author = {Du, Maximilian and Khazatsky, Alexander and Gerstenberg, Tobias and Finn, Chelsea},
Expand Down
6 changes: 5 additions & 1 deletion docs/index.html
Original file line number Diff line number Diff line change
Expand Up @@ -110,7 +110,7 @@
<meta property="og:description" content="">
<meta property="og:locale" content="en-us">

<meta property="og:updated_time" content="2024-06-26T00:00:00&#43;00:00">
<meta property="og:updated_time" content="2024-09-20T00:00:00&#43;00:00">



Expand Down Expand Up @@ -1660,6 +1660,10 @@ <h1>Selected<br>publications</h1>



<a class="btn btn-outline-primary my-1 mr-1 btn-sm" href="https://www.sciencedirect.com/science/article/pii/S1364661324001074" target="_blank" rel="noopener">
Link
</a>

<a class="btn btn-outline-primary my-1 mr-1 btn-sm" href="https://hai.stanford.edu/news/humans-use-counterfactuals-reason-about-causality-can-ai" target="_blank" rel="noopener">
Press: HAI News
</a>
Expand Down
20 changes: 10 additions & 10 deletions docs/index.xml
Original file line number Diff line number Diff line change
Expand Up @@ -6,9 +6,18 @@
<generator>Hugo -- gohugo.io</generator>
<language>en-us</language>
<copyright>&amp;copy; 2024 Tobias Gerstenberg</copyright>
<lastBuildDate>Wed, 26 Jun 2024 00:00:00 +0000</lastBuildDate>
<lastBuildDate>Fri, 20 Sep 2024 00:00:00 +0000</lastBuildDate>
<atom:link href="/" rel="self" type="application/rss+xml" />

<item>
<title>Human-like Affective Cognition in Foundation Models</title>
<link>https://cicl.stanford.edu/publication/gandhi2024affective/</link>
<pubDate>Fri, 20 Sep 2024 00:00:00 +0000</pubDate>

<guid>https://cicl.stanford.edu/publication/gandhi2024affective/</guid>
<description></description>
</item>

<item>
<title>To Err is Robotic: Rapid Value-Based Trial-and-Error during Deployment</title>
<link>https://cicl.stanford.edu/publication/du2024robotic/</link>
Expand Down Expand Up @@ -135,14 +144,5 @@
<description></description>
</item>

<item>
<title>Anticipating the risks and benefits of counterfactual world simulation models</title>
<link>https://cicl.stanford.edu/publication/kirfel2023anticipating/</link>
<pubDate>Mon, 30 Oct 2023 00:00:00 +0000</pubDate>

<guid>https://cicl.stanford.edu/publication/kirfel2023anticipating/</guid>
<description></description>
</item>

</channel>
</rss>
51 changes: 51 additions & 0 deletions docs/member/tobias_gerstenberg/index.html
Original file line number Diff line number Diff line change
Expand Up @@ -356,6 +356,45 @@ <h2 id="publications">Publications</h2>


<div class="pub-list-item" style="margin-bottom: 1rem" itemscope itemtype="http://schema.org/CreativeWork">
<span itemprop="author">
K. Gandhi, Z. Lynch, J. Fränken, K. Patterson, S. Wambu, T. Gerstenberg, D. C. Ong, N. D. Goodman</span>

(2024).

<a href="https://cicl.stanford.edu/publication/gandhi2024affective/" itemprop="name">Human-like Affective Cognition in Foundation Models</a>.
<em>arXiv</em>.




<p>




<a class="btn btn-outline-primary my-1 mr-1 btn-sm" href="https://arxiv.org/abs/2409.11733" target="_blank" rel="noopener">
Preprint
</a>
















</p>

</div>
<div class="pub-list-item" style="margin-bottom: 1rem" itemscope itemtype="http://schema.org/CreativeWork">
<span itemprop="author">
M. Du, A. Khazatsky, T. Gerstenberg, C. Finn</span>

Expand Down Expand Up @@ -437,6 +476,10 @@ <h2 id="publications">Publications</h2>



<a class="btn btn-outline-primary my-1 mr-1 btn-sm" href="https://www.sciencedirect.com/science/article/pii/S1364661324001074" target="_blank" rel="noopener">
Link
</a>

<a class="btn btn-outline-primary my-1 mr-1 btn-sm" href="https://hai.stanford.edu/news/humans-use-counterfactuals-reason-about-causality-can-ai" target="_blank" rel="noopener">
Press: HAI News
</a>
Expand Down Expand Up @@ -488,10 +531,18 @@ <h2 id="publications">Publications</h2>



<a class="btn btn-outline-primary my-1 mr-1 btn-sm" href="https://www.sciencedirect.com/science/article/pii/S0010027724001227" target="_blank" rel="noopener">
Link
</a>

<a class="btn btn-outline-primary my-1 mr-1 btn-sm" href="https://github.com/cicl-stanford/children_disagree" target="_blank" rel="noopener">
Github
</a>

<a class="btn btn-outline-primary my-1 mr-1 btn-sm" href="https://www.psypost.org/childrens-ability-to-detect-ambiguity-in-disagreements-sharpens-between-ages-7-and-11/" target="_blank" rel="noopener">
Press: Psypost
</a>


</p>

Expand Down
8 changes: 8 additions & 0 deletions docs/publication/amemiya2024disagreement/index.html
Original file line number Diff line number Diff line change
Expand Up @@ -341,10 +341,18 @@ <h3>Abstract</h3>



<a class="btn btn-outline-primary my-1 mr-1" href="https://www.sciencedirect.com/science/article/pii/S0010027724001227" target="_blank" rel="noopener">
Link
</a>

<a class="btn btn-outline-primary my-1 mr-1" href="https://github.com/cicl-stanford/children_disagree" target="_blank" rel="noopener">
Github
</a>

<a class="btn btn-outline-primary my-1 mr-1" href="https://www.psypost.org/childrens-ability-to-detect-ambiguity-in-disagreements-sharpens-between-ages-7-and-11/" target="_blank" rel="noopener">
Press: Psypost
</a>


</div>
</div>
Expand Down
4 changes: 2 additions & 2 deletions docs/publication/du2024robotic/index.html
Original file line number Diff line number Diff line change
Expand Up @@ -266,8 +266,8 @@ <h3>Abstract</h3>
<div class="col-xs-12 col-sm-3 pub-row-heading">Type</div>
<div class="col-xs-12 col-sm-9">

<a href="/publication/#2">
Journal
<a href="/publication/#1">
Preprint
</a>

</div>
Expand Down
4 changes: 2 additions & 2 deletions docs/publication/franken2024sami/index.html
Original file line number Diff line number Diff line change
Expand Up @@ -266,8 +266,8 @@ <h3>Abstract</h3>
<div class="col-xs-12 col-sm-3 pub-row-heading">Type</div>
<div class="col-xs-12 col-sm-9">

<a href="/publication/#2">
Journal
<a href="/publication/#1">
Preprint
</a>

</div>
Expand Down
Loading

0 comments on commit 06009e4

Please sign in to comment.