Skip to content

Commit

Permalink
updated sami paper datea
Browse files Browse the repository at this point in the history
  • Loading branch information
tobiasgerstenberg committed Oct 4, 2024
1 parent 17b599e commit 63dd19e
Show file tree
Hide file tree
Showing 13 changed files with 49 additions and 203 deletions.
2 changes: 1 addition & 1 deletion content/publication/franken2024sami.md
Original file line number Diff line number Diff line change
Expand Up @@ -7,7 +7,7 @@
# 5 -> 'Thesis'

title = "Self-supervised alignment with mutual information: Learning to follow principles without preference labels"
date = "2024-04-22"
date = "2024-10-08"
authors = ["J. Fränken","E. Zelikman","R. Rafailov","K. Gandhi","T. Gerstenberg","N. D. Goodman"]
publication_types = ["3"]
publication_short = "_Advances in Neural Information Processing Systems_"
Expand Down
16 changes: 8 additions & 8 deletions docs/index.xml
Original file line number Diff line number Diff line change
Expand Up @@ -127,20 +127,20 @@
</item>

<item>
<title>Self-supervised alignment with mutual information: Learning to follow principles without preference labels</title>
<link>https://cicl.stanford.edu/publication/franken2024sami/</link>
<pubDate>Mon, 22 Apr 2024 00:00:00 +0000</pubDate>
<title>Procedural dilemma generation for evaluating moral reasoning in humans and language models</title>
<link>https://cicl.stanford.edu/publication/franken2024rails/</link>
<pubDate>Wed, 17 Apr 2024 00:00:00 +0000</pubDate>

<guid>https://cicl.stanford.edu/publication/franken2024sami/</guid>
<guid>https://cicl.stanford.edu/publication/franken2024rails/</guid>
<description></description>
</item>

<item>
<title>Procedural dilemma generation for evaluating moral reasoning in humans and language models</title>
<link>https://cicl.stanford.edu/publication/franken2024rails/</link>
<pubDate>Wed, 17 Apr 2024 00:00:00 +0000</pubDate>
<title>STaR-GATE: Teaching Language Models to Ask Clarifying Questions</title>
<link>https://cicl.stanford.edu/publication/andukuri2024stargate/</link>
<pubDate>Sun, 31 Mar 2024 00:00:00 +0000</pubDate>

<guid>https://cicl.stanford.edu/publication/franken2024rails/</guid>
<guid>https://cicl.stanford.edu/publication/andukuri2024stargate/</guid>
<description></description>
</item>

Expand Down
47 changes: 0 additions & 47 deletions docs/member/tobias_gerstenberg/index.html
Original file line number Diff line number Diff line change
Expand Up @@ -967,53 +967,6 @@ <h2 id="publications">Publications</h2>



</p>

</div>
<div class="pub-list-item" style="margin-bottom: 1rem" itemscope itemtype="http://schema.org/CreativeWork">
<span itemprop="author">
J. Fränken, E. Zelikman, R. Rafailov, K. Gandhi, T. Gerstenberg, N. D. Goodman</span>

(2024).

<a href="https://cicl.stanford.edu/publication/franken2024sami/" itemprop="name">Self-supervised alignment with mutual information: Learning to follow principles without preference labels</a>.
<em>Advances in Neural Information Processing Systems</em>.




<p>




<a class="btn btn-outline-primary my-1 mr-1 btn-sm" href="https://arxiv.org/abs/2404.14313" target="_blank" rel="noopener">
Preprint
</a>


<a class="btn btn-outline-primary my-1 mr-1 btn-sm" href="https://cicl.stanford.edu/papers/franken2024sami.pdf" target="_blank" rel="noopener">
PDF
</a>














<a class="btn btn-outline-primary my-1 mr-1 btn-sm" href="https://github.com/janphilippfranken/sami" target="_blank" rel="noopener">
Github
</a>


</p>

</div>
Expand Down
72 changes: 0 additions & 72 deletions docs/publication/index.html
Original file line number Diff line number Diff line change
Expand Up @@ -1597,19 +1597,6 @@ <h1>Publications</h1>



















Expand Down Expand Up @@ -2517,65 +2504,6 @@ <h1>Publications</h1>



<div class='grid-sizer col-md-12 isotope-item pubtype-3 year-2024 author-'>

<div class="pub-list-item" style="margin-bottom: 1rem" itemscope itemtype="http://schema.org/CreativeWork">
<span itemprop="author">
J. Fränken, E. Zelikman, R. Rafailov, K. Gandhi, T. Gerstenberg, N. D. Goodman</span>

(2024).

<a href="https://cicl.stanford.edu/publication/franken2024sami/" itemprop="name">Self-supervised alignment with mutual information: Learning to follow principles without preference labels</a>.
<em>Advances in Neural Information Processing Systems</em>.




<p>




<a class="btn btn-outline-primary my-1 mr-1 btn-sm" href="https://arxiv.org/abs/2404.14313" target="_blank" rel="noopener">
Preprint
</a>


<a class="btn btn-outline-primary my-1 mr-1 btn-sm" href="https://cicl.stanford.edu/papers/franken2024sami.pdf" target="_blank" rel="noopener">
PDF
</a>














<a class="btn btn-outline-primary my-1 mr-1 btn-sm" href="https://github.com/janphilippfranken/sami" target="_blank" rel="noopener">
Github
</a>


</p>

</div>


</div>







<div class='grid-sizer col-md-12 isotope-item pubtype-3 year-2024 author-'>

<div class="pub-list-item" style="margin-bottom: 1rem" itemscope itemtype="http://schema.org/CreativeWork">
Expand Down
9 changes: 0 additions & 9 deletions docs/publication/index.xml
Original file line number Diff line number Diff line change
Expand Up @@ -129,15 +129,6 @@
<description></description>
</item>

<item>
<title>Self-supervised alignment with mutual information: Learning to follow principles without preference labels</title>
<link>https://cicl.stanford.edu/publication/franken2024sami/</link>
<pubDate>Mon, 22 Apr 2024 00:00:00 +0000</pubDate>

<guid>https://cicl.stanford.edu/publication/franken2024sami/</guid>
<description></description>
</item>

<item>
<title>Procedural dilemma generation for evaluating moral reasoning in humans and language models</title>
<link>https://cicl.stanford.edu/publication/franken2024rails/</link>
Expand Down
8 changes: 4 additions & 4 deletions docs/publication_types/3/index.html
Original file line number Diff line number Diff line change
Expand Up @@ -311,19 +311,19 @@ <h2><a href="https://cicl.stanford.edu/publication/wu2024resource/">Resource-rat
</div>

<div>
<h2><a href="https://cicl.stanford.edu/publication/franken2024sami/">Self-supervised alignment with mutual information: Learning to follow principles without preference labels</a></h2>
<h2><a href="https://cicl.stanford.edu/publication/franken2024rails/">Procedural dilemma generation for evaluating moral reasoning in humans and language models</a></h2>
<div class="article-style">

When prompting a language model (LM), users frequently expect the model to adhere to a set of behavioral principles across diverse tasks, such as producing insightful content while avoiding harmful or biased language. Instilling such principles into
As AI systems like language models are increasingly integrated into decision-making processes affecting people's lives, it's critical to ensure that these systems have sound moral reasoning. To test whether they do, we need to develop systematic

</div>
</div>

<div>
<h2><a href="https://cicl.stanford.edu/publication/franken2024rails/">Procedural dilemma generation for evaluating moral reasoning in humans and language models</a></h2>
<h2><a href="https://cicl.stanford.edu/publication/kirfel2023anticipating/">Anticipating the risks and benefits of counterfactual world simulation models</a></h2>
<div class="article-style">

As AI systems like language models are increasingly integrated into decision-making processes affecting people's lives, it's critical to ensure that these systems have sound moral reasoning. To test whether they do, we need to develop systematic
This paper examines the transformative potential of Counterfactual World Simulation Models (CWSMs). CWSMs use pieces of multi-modal evidence, such as the CCTV footage or sound recordings of a road accident, to build a high-fidelity 3D reconstruction

</div>
</div>
Expand Down
9 changes: 0 additions & 9 deletions docs/publication_types/3/index.xml
Original file line number Diff line number Diff line change
Expand Up @@ -84,15 +84,6 @@
<description></description>
</item>

<item>
<title>Self-supervised alignment with mutual information: Learning to follow principles without preference labels</title>
<link>https://cicl.stanford.edu/publication/franken2024sami/</link>
<pubDate>Mon, 22 Apr 2024 00:00:00 +0000</pubDate>

<guid>https://cicl.stanford.edu/publication/franken2024sami/</guid>
<description></description>
</item>

<item>
<title>Procedural dilemma generation for evaluating moral reasoning in humans and language models</title>
<link>https://cicl.stanford.edu/publication/franken2024rails/</link>
Expand Down
18 changes: 9 additions & 9 deletions docs/publication_types/3/page/2/index.html
Original file line number Diff line number Diff line change
Expand Up @@ -238,15 +238,6 @@ <h1 class="pt-3">3</h1>



<div>
<h2><a href="https://cicl.stanford.edu/publication/kirfel2023anticipating/">Anticipating the risks and benefits of counterfactual world simulation models</a></h2>
<div class="article-style">

This paper examines the transformative potential of Counterfactual World Simulation Models (CWSMs). CWSMs use pieces of multi-modal evidence, such as the CCTV footage or sound recordings of a road accident, to build a high-fidelity 3D reconstruction …

</div>
</div>

<div>
<h2><a href="https://cicl.stanford.edu/publication/franken2023rails/">Off The Rails: Procedural Dilemma Generation for Moral Reasoning</a></h2>
<div class="article-style">
Expand Down Expand Up @@ -328,6 +319,15 @@ <h2><a href="https://cicl.stanford.edu/publication/zhang2023llm/">You are what y
</div>
</div>

<div>
<h2><a href="https://cicl.stanford.edu/publication/cao2023semantics/">A Semantics for Causing, Enabling, and Preventing Verbs Using Structural Causal Models</a></h2>
<div class="article-style">

When choosing how to describe what happened, we have a number of causal verbs at our disposal. In this paper, we develop a model-theoretic formal semantics for nine causal verbs that span the categories of CAUSE, ENABLE, and PREVENT. We use …

</div>
</div>



<nav>
Expand Down
18 changes: 9 additions & 9 deletions docs/publication_types/3/page/3/index.html
Original file line number Diff line number Diff line change
Expand Up @@ -238,15 +238,6 @@ <h1 class="pt-3">3</h1>



<div>
<h2><a href="https://cicl.stanford.edu/publication/cao2023semantics/">A Semantics for Causing, Enabling, and Preventing Verbs Using Structural Causal Models</a></h2>
<div class="article-style">

When choosing how to describe what happened, we have a number of causal verbs at our disposal. In this paper, we develop a model-theoretic formal semantics for nine causal verbs that span the categories of CAUSE, ENABLE, and PREVENT. We use …

</div>
</div>

<div>
<h2><a href="https://cicl.stanford.edu/publication/gonzalez2023agents/">Causal Reasoning Across Agents and Objects</a></h2>
<div class="article-style">
Expand Down Expand Up @@ -328,6 +319,15 @@ <h2><a href="https://cicl.stanford.edu/publication/vodrahalli2022uncalibrated/">
</div>
</div>

<div>
<h2><a href="https://cicl.stanford.edu/publication/davis2021fishing/">Who went fishing? Inferences from social evaluations</a></h2>
<div class="article-style">

Humans have a remarkable ability to go beyond the observable. From seeing the current state of our shared kitchen, we can infer what happened and who did it. Prior work has shown how the physical state of the world licenses inferences about the …

</div>
</div>



<nav>
Expand Down
18 changes: 9 additions & 9 deletions docs/publication_types/3/page/4/index.html
Original file line number Diff line number Diff line change
Expand Up @@ -238,15 +238,6 @@ <h1 class="pt-3">3</h1>



<div>
<h2><a href="https://cicl.stanford.edu/publication/davis2021fishing/">Who went fishing? Inferences from social evaluations</a></h2>
<div class="article-style">

Humans have a remarkable ability to go beyond the observable. From seeing the current state of our shared kitchen, we can infer what happened and who did it. Prior work has shown how the physical state of the world licenses inferences about the …

</div>
</div>

<div>
<h2><a href="https://cicl.stanford.edu/publication/beller2020language/">The language of causation</a></h2>
<div class="article-style">
Expand Down Expand Up @@ -328,6 +319,15 @@ <h2><a href="https://cicl.stanford.edu/publication/yildirim2017problem/">Physica
</div>
</div>

<div>
<h2><a href="https://cicl.stanford.edu/publication/niemi2016implicit/">Implicit measurement of motivated causal attribution</a></h2>
<div class="article-style">

Moral judgment often involves pinning causation for harm to a particular person. Since it reveals “who one sides with”, expression of moral judgment can be a costly social act that people may be motivated to conceal. Here, we demonstrate that a …

</div>
</div>



<nav>
Expand Down
18 changes: 9 additions & 9 deletions docs/publication_types/3/page/5/index.html
Original file line number Diff line number Diff line change
Expand Up @@ -238,15 +238,6 @@ <h1 class="pt-3">3</h1>



<div>
<h2><a href="https://cicl.stanford.edu/publication/niemi2016implicit/">Implicit measurement of motivated causal attribution</a></h2>
<div class="article-style">

Moral judgment often involves pinning causation for harm to a particular person. Since it reveals “who one sides with”, expression of moral judgment can be a costly social act that people may be motivated to conceal. Here, we demonstrate that a …

</div>
</div>

<div>
<h2><a href="https://cicl.stanford.edu/publication/bramley2016natural/">Natural science: Active learning in dynamic physical microworlds</a></h2>
<div class="article-style">
Expand Down Expand Up @@ -328,6 +319,15 @@ <h2><a href="https://cicl.stanford.edu/publication/bramley2014order/">The order
</div>
</div>

<div>
<h2><a href="https://cicl.stanford.edu/publication/gerstenberg2014replacement/">Wins above replacement: Responsibility attributions as counterfactual replacements</a></h2>
<div class="article-style">

In order to be held responsible, a person’s action has to have made some sort of difference to the outcome. In this paper, we propose a counterfactual replacement model according to which people attribute responsibility by comparing their prior …

</div>
</div>



<nav>
Expand Down
9 changes: 0 additions & 9 deletions docs/publication_types/3/page/6/index.html
Original file line number Diff line number Diff line change
Expand Up @@ -238,15 +238,6 @@ <h1 class="pt-3">3</h1>



<div>
<h2><a href="https://cicl.stanford.edu/publication/gerstenberg2014replacement/">Wins above replacement: Responsibility attributions as counterfactual replacements</a></h2>
<div class="article-style">

In order to be held responsible, a person’s action has to have made some sort of difference to the outcome. In this paper, we propose a counterfactual replacement model according to which people attribute responsibility by comparing their prior …

</div>
</div>

<div>
<h2><a href="https://cicl.stanford.edu/publication/gerstenberg2013back/">Back on track: Backtracking in counterfactual reasoning</a></h2>
<div class="article-style">
Expand Down
8 changes: 0 additions & 8 deletions docs/sitemap.xml
Original file line number Diff line number Diff line change
Expand Up @@ -150,14 +150,6 @@



<url>

<loc>https://cicl.stanford.edu/publication/franken2024sami/</loc>
<lastmod>2024-04-22T00:00:00+00:00</lastmod>
</url>



<url>

<loc>https://cicl.stanford.edu/publication/franken2024rails/</loc>
Expand Down

0 comments on commit 63dd19e

Please sign in to comment.