diff --git a/content/publication/cao2023semantics.md b/content/publication/cao2023semantics.md
index f982904..99104c1 100644
--- a/content/publication/cao2023semantics.md
+++ b/content/publication/cao2023semantics.md
@@ -8,10 +8,10 @@
title = "A Semantics for Causing, Enabling, and Preventing Verbs Using Structural Causal Models"
date = "2023-05-10"
-authors = ["A. Cao","A. Geiger","E. Kreiss","T. Icard","T. Gerstenberg"]
+authors = ["A. Cao\\*","A. Geiger\\*","E. Kreiss\\*","T. Icard","T. Gerstenberg"]
publication_types = ["3"]
publication_short = "_Proceedings of the 45th Annual Conference of the Cognitive Science Society_"
-publication = "Cao A., Geiger A., Kreiss E., Icard T., Gerstenberg T. (2023). A Semantics for Causing, Enabling, and Preventing Verbs Using Structural Causal Models. In _Proceedings of the 45th Annual Conference of the Cognitive Science Society_."
+publication = "Cao A.\\*, Geiger A.\\*, Kreiss E.\\*, Icard T., Gerstenberg T. (2023). A Semantics for Causing, Enabling, and Preventing Verbs Using Structural Causal Models. In _Proceedings of the 45th Annual Conference of the Cognitive Science Society_."
abstract = "When choosing how to describe what happened, we have a number of causal verbs at our disposal. In this paper, we develop a model-theoretic formal semantics for nine causal verbs that span the categories of CAUSE, ENABLE, and PREVENT. We use structural causal models (SCMs) to represent participants' mental construction of a scene when assessing the correctness of causal expressions relative to a presented context. Furthermore, SCMs enable us to model events relating both the physical world as well as agents' mental states. In experimental evaluations, we find that the proposed semantics exhibits a closer alignment with human evaluations in comparison to prior accounts of the verb families."
image_preview = ""
selected = false
diff --git a/content/publication/gershman2016goals.md b/content/publication/gershman2016goals.md
index c9a565a..e6b0e20 100644
--- a/content/publication/gershman2016goals.md
+++ b/content/publication/gershman2016goals.md
@@ -9,10 +9,10 @@
title = "Plans, habits, and theory of mind"
date = "2016-01-01"
-authors = ["S. J. Gershman ","T. Gerstenberg", "C. L. Baker", "F. Cushman"]
+authors = ["S. J. Gershman","T. Gerstenberg", "C. L. Baker", "F. Cushman"]
publication_types = ["2"]
publication_short = "_PLoS ONE_"
-publication = "Gershman*, S. J., Gerstenberg*, T., Baker, C. L., & Cushman, F. (2016). Plans, habits, and theory of mind. _PLoS ONE_, 11(9), e0162246. "
+publication = "Gershman, S. J.\\*, Gerstenberg, T.\\*, Baker, C. L., & Cushman, F. (2016). Plans, habits, and theory of mind. _PLoS ONE_, 11(9), e0162246. "
image_preview = ""
selected = false
projects = []
diff --git a/content/publication/mccoy2012probabilistic.md b/content/publication/mccoy2012probabilistic.md
index 474a493..05df6e2 100644
--- a/content/publication/mccoy2012probabilistic.md
+++ b/content/publication/mccoy2012probabilistic.md
@@ -9,10 +9,10 @@
title = "Why blame Bob? Probabilistic generative models, counterfactual reasoning, and blame attribution"
date = "2012-01-01"
-authors = ["J. McCoy* ","T. D. Ullman* ","A. Stuhlmüller","T. Gerstenberg","J. B. Tenenbaum"]
+authors = ["J. McCoy\\*","T. D. Ullman\\*","A. Stuhlmüller","T. Gerstenberg","J. B. Tenenbaum"]
publication_types = ["3"]
publication_short = "_Proceedings of the 34th Annual Conference of the Cognitive Science Society_"
-publication = "McCoy, J.*, Ullman, T.*, Stuhlmüller, A., Gerstenberg, T., & Tenenbaum, J. B. (2012). Why blame Bob? Probabilistic generative models, counterfactual reasoning, and blame attribution. In _Proceedings of the 34th Annual Conference of the Cognitive Science Society_ (pp. 1996-2001). Austin, TX: Cognitive Science Society."
+publication = "McCoy, J.\\*, Ullman, T.\\*, Stuhlmüller, A., Gerstenberg, T., & Tenenbaum, J. B. (2012). Why blame Bob? Probabilistic generative models, counterfactual reasoning, and blame attribution. In _Proceedings of the 34th Annual Conference of the Cognitive Science Society_ (pp. 1996-2001). Austin, TX: Cognitive Science Society."
image_preview = ""
selected = false
projects = []
diff --git a/content/publication/outa2022stop.md b/content/publication/outa2022stop.md
index 3068ae5..5bc5a21 100644
--- a/content/publication/outa2022stop.md
+++ b/content/publication/outa2022stop.md
@@ -8,10 +8,10 @@
title = "Stop, children what's that sound? Multi-modal inference through mental simulation"
date = "2022-05-12"
-authors = ["J. Outa","X. Jia Zhou","H. Gweon","T. Gerstenberg"]
+authors = ["J. Outa\\*","X. Jia Zhou\\*","H. Gweon","T. Gerstenberg"]
publication_types = ["3"]
publication_short = "_Cognitive Science Proceedings_"
-publication = "Outa* J., Zhou* X. J., Gweon H., Gerstenberg T. (2022). Stop, children what's that sound? Multi-modal inference through mental simulation. In _Cognitive Science Proceedings_."
+publication = "Outa J.\\*, Zhou X. J.\\*, Gweon H., Gerstenberg T. (2022). Stop, children what's that sound? Multi-modal inference through mental simulation. In _Cognitive Science Proceedings_."
image_preview = ""
abstract = "Human adults can figure out what happened by combining evidence from different sensory modalities, such as vision and sound. How does the ability to integrate multi-modal information develop in early childhood? Inspired by prior computational work and behavioral studies with adults, we examined 3- to 8-year-old children's ability to reason about the physical trajectory of a ball that was dropped into an occluded Plinko box. Children had to infer in which one of three holes the ball was dropped based on visual information (i.e., where the ball landed) and auditory information (i.e., the sounds of the ball colliding with parts of the box). We compare children's responses to the predictions of four computational models. The results suggest that although even the youngest children make systematic judgments rather than randomly guessing, children's ability to integrate visual and auditory evidence continues to develop into late childhood."
selected = false
diff --git a/content/publication/wu2024whodunnit.md b/content/publication/wu2024whodunnit.md
new file mode 100644
index 0000000..5fb8087
--- /dev/null
+++ b/content/publication/wu2024whodunnit.md
@@ -0,0 +1,33 @@
++++
+# 0 -> 'Forthcoming',
+# 1 -> 'Preprint',
+# 2 -> 'Journal',
+# 3 -> 'Conference Proceedings',
+# 4 -> 'Book chapter',
+# 5 -> 'Thesis'
+
+title = "Whodunnit? Inferring what happened from multimodal evidence"
+date = "2024-05-13"
+authors = ["S. A Wu\\*","E. Brockbank\\*","H. Cha","J. Fränken","E. Jin","Z. Huang","W. Liu","R. Zhang","J. Wu","T. Gerstenberg"]
+publication_types = ["3"]
+publication_short = "_Proceedings of the 46th Annual Conference of the Cognitive Science Society_"
+publication = "Wu*, S. A., Brockbank*, E., Cha, H., Fränken, J., Jin, E., Huang, Z., Liu, W., Zhang, R., Wu, J., Gerstenberg, T. (2024). Whodunnit? Inferring what happened from multimodal evidence. In _Proceedings of the 46th Annual Conference of the Cognitive Science Society_."
+abstract = "Humans are remarkably adept at inferring the causes of events in their environment; doing so often requires incorporating information from multiple sensory modalities. For instance, if a car slows down in front of us, inferences about why they did so are rapidly revised if we also hear sirens in the distance. Here, we investigate the ability to reconstruct others' actions and events from the past by integrating multimodal information. Participants were asked to infer which of two agents performed an action in a household setting given either visual evidence, auditory evidence, or both. We develop a computational model that makes inferences by generating multimodal simulations, and also evaluate our task on a large language model (GPT-4) and a large multimodal model (GPT-4V). We find that humans are relatively accurate overall and perform best when given multimodal evidence. GPT-4 and GPT-4V performance comes close overall, but is very weakly correlated with participants across individual trials. Meanwhile, the simulation model captures the pattern of human responses well. Multimodal event reconstruction represents a challenge for current AI systems, and frameworks that draw on the cognitive processes underlying people's ability to reconstruct events offer a promising avenue forward."
+image_preview = ""
+selected = false
+projects = []
+url_pdf = "papers/wu2024whodunnit.pdf"
+url_preprint = ""
+url_code = ""
+url_dataset = ""
+url_slides = ""
+url_video = ""
+url_poster = ""
+url_source = ""
+url_custom = [{name = "Github", url = "https://github.com/cicl-stanford/whodunnit_multimodal_inference"}]
+math = true
+highlight = true
+[header]
+# image = "publications/wu2024whodunnit.png"
+caption = ""
++++
\ No newline at end of file
diff --git a/content/publication/zhang2023llm.md b/content/publication/zhang2023llm.md
index 112e0a9..d7d5c68 100644
--- a/content/publication/zhang2023llm.md
+++ b/content/publication/zhang2023llm.md
@@ -8,10 +8,10 @@
title = "You are what you're for: Essentialist categorization in large language models"
date = "2023-05-11"
-authors = ["S. Zhang","J. S. She","T. Gerstenberg","D. Rose"]
+authors = ["S. Zhang\\*","J. S. She\\*","T. Gerstenberg","D. Rose"]
publication_types = ["3"]
publication_short = "_Proceedings of the 45th Annual Conference of the Cognitive Science Society_"
-publication = "Zhang S., She J. S., Gerstenberg T., Rose D. (2023). You are what you're for: Essentialist categorization in large language models. In _Proceedings of the 45th Annual Conference of the Cognitive Science Society_."
+publication = "Zhang S.\\*, She J. S.\\*, Gerstenberg T., Rose D. (2023). You are what you're for: Essentialist categorization in large language models. In _Proceedings of the 45th Annual Conference of the Cognitive Science Society_."
abstract = "How do essentialist beliefs about categories arise? We hypothesize that such beliefs are transmitted via language. We subject large language models (LLMs) to vignettes from the literature on essentialist categorization and find that they align well with people when the studies manipulated teleological information - information about what something is for. We examine whether in a classic test of essentialist categorization - the transformation task - LLMs prioritize teleological properties over information about what something looks like, or is made of. Experiments 1 and 2 find that telos and what something is made of matter more than appearance. Experiment 3 manipulates all three factors and finds that what something is for matters more than what it's made of. Overall, these studies suggest that language alone may be sufficient to give rise to essentialist beliefs, and that information about what something is for matters more."
image_preview = ""
selected = false
diff --git a/docs/404.html b/docs/404.html
index 453cb26..f145e22 100644
--- a/docs/404.html
+++ b/docs/404.html
@@ -237,6 +237,10 @@
+
+ S. A Wu*, E. Brockbank*, H. Cha, J. Fränken, E. Jin, Z. Huang, W. Liu, R. Zhang, J. Wu, T. Gerstenberg
+
+ (2024).
+
+ Whodunnit? Inferring what happened from multimodal evidence.
+ Proceedings of the 46th Annual Conference of the Cognitive Science Society.
+
+
+
+
+
S. Tsirtsis, M. Gomez-Rodriguez, T. Gerstenberg
@@ -1463,7 +1506,7 @@
Publications
- S. Zhang, J. S. She, T. Gerstenberg, D. Rose
+ S. Zhang*, J. S. She*, T. Gerstenberg, D. Rose
(2023).
@@ -1514,7 +1557,7 @@
Publications
- A. Cao, A. Geiger, E. Kreiss, T. Icard, T. Gerstenberg
+ A. Cao*, A. Geiger*, E. Kreiss*, T. Icard, T. Gerstenberg
(2023).
@@ -2002,7 +2045,7 @@
Publications
- J. Outa, X. Jia Zhou, H. Gweon, T. Gerstenberg
+ J. Outa*, X. Jia Zhou*, H. Gweon, T. Gerstenberg
(2022).
@@ -3781,7 +3824,7 @@
Publications
- S. J. Gershman , T. Gerstenberg, C. L. Baker, F. Cushman
+ S. J. Gershman, T. Gerstenberg, C. L. Baker, F. Cushman
(2016).
@@ -4759,7 +4802,7 @@
Publications
- J. McCoy* , T. D. Ullman* , A. Stuhlmüller, T. Gerstenberg, J. B. Tenenbaum
+ J. McCoy*, T. D. Ullman*, A. Stuhlmüller, T. Gerstenberg, J. B. Tenenbaum
(2012).
diff --git a/docs/papers/wu2024whodunnit.pdf b/docs/papers/wu2024whodunnit.pdf
new file mode 100644
index 0000000..bf2aafe
Binary files /dev/null and b/docs/papers/wu2024whodunnit.pdf differ
diff --git a/docs/publication/cao2023semantics/index.html b/docs/publication/cao2023semantics/index.html
index 07efbe0..a37315b 100644
--- a/docs/publication/cao2023semantics/index.html
+++ b/docs/publication/cao2023semantics/index.html
@@ -239,7 +239,7 @@
A Semantics for Causing, Enabling, and Preventing Verbs Using Structural Causal Models
- A. Cao, A. Geiger, E. Kreiss, T. Icard, T. Gerstenberg
+ A. Cao*, A. Geiger*, E. Kreiss*, T. Icard, T. Gerstenberg
@@ -283,7 +283,7 @@
Abstract
Publication
-
Cao A., Geiger A., Kreiss E., Icard T., Gerstenberg T. (2023). A Semantics for Causing, Enabling, and Preventing Verbs Using Structural Causal Models. In Proceedings of the 45th Annual Conference of the Cognitive Science Society.
+
Cao A.*, Geiger A.*, Kreiss E.*, Icard T., Gerstenberg T. (2023). A Semantics for Causing, Enabling, and Preventing Verbs Using Structural Causal Models. In Proceedings of the 45th Annual Conference of the Cognitive Science Society.
+
+ S. A Wu*, E. Brockbank*, H. Cha, J. Fränken, E. Jin, Z. Huang, W. Liu, R. Zhang, J. Wu, T. Gerstenberg
+
+ (2024).
+
+ Whodunnit? Inferring what happened from multimodal evidence.
+ Proceedings of the 46th Annual Conference of the Cognitive Science Society.
+
+
+
+
+
- S. Zhang, J. S. She, T. Gerstenberg, D. Rose
+ S. Zhang*, J. S. She*, T. Gerstenberg, D. Rose
(2023).
@@ -3142,7 +3210,7 @@
Publications
- A. Cao, A. Geiger, E. Kreiss, T. Icard, T. Gerstenberg
+ A. Cao*, A. Geiger*, E. Kreiss*, T. Icard, T. Gerstenberg
(2023).
@@ -3750,7 +3818,7 @@
Publications
- J. Outa, X. Jia Zhou, H. Gweon, T. Gerstenberg
+ J. Outa*, X. Jia Zhou*, H. Gweon, T. Gerstenberg
(2022).
@@ -5973,7 +6041,7 @@
Publications
- S. J. Gershman , T. Gerstenberg, C. L. Baker, F. Cushman
+ S. J. Gershman, T. Gerstenberg, C. L. Baker, F. Cushman
(2016).
@@ -7215,7 +7283,7 @@
Publications
- J. McCoy* , T. D. Ullman* , A. Stuhlmüller, T. Gerstenberg, J. B. Tenenbaum
+ J. McCoy*, T. D. Ullman*, A. Stuhlmüller, T. Gerstenberg, J. B. Tenenbaum
(2012).
diff --git a/docs/publication/index.xml b/docs/publication/index.xml
index 7b3091b..dd8041b 100644
--- a/docs/publication/index.xml
+++ b/docs/publication/index.xml
@@ -12,6 +12,15 @@
+
+ Whodunnit? Inferring what happened from multimodal evidence
+ https://cicl.stanford.edu/publication/wu2024whodunnit/
+ Mon, 13 May 2024 00:00:00 +0000
+
+ https://cicl.stanford.edu/publication/wu2024whodunnit/
+
+
+
Towards a computational model of responsibility judgments in sequential human-AI collaboration
https://cicl.stanford.edu/publication/tsirtsis2024sequential/
diff --git a/docs/publication/mccoy2012probabilistic/index.html b/docs/publication/mccoy2012probabilistic/index.html
index 70fbece..6ca5c40 100644
--- a/docs/publication/mccoy2012probabilistic/index.html
+++ b/docs/publication/mccoy2012probabilistic/index.html
@@ -239,7 +239,7 @@
- J. McCoy* , T. D. Ullman* , A. Stuhlmüller, T. Gerstenberg, J. B. Tenenbaum
+ J. McCoy*, T. D. Ullman*, A. Stuhlmüller, T. Gerstenberg, J. B. Tenenbaum
@@ -283,7 +283,7 @@
Abstract
Publication
-
McCoy, J., Ullman, T., Stuhlmüller, A., Gerstenberg, T., & Tenenbaum, J. B. (2012). Why blame Bob? Probabilistic generative models, counterfactual reasoning, and blame attribution. In Proceedings of the 34th Annual Conference of the Cognitive Science Society (pp. 1996-2001). Austin, TX: Cognitive Science Society.
+
McCoy, J.*, Ullman, T.*, Stuhlmüller, A., Gerstenberg, T., & Tenenbaum, J. B. (2012). Why blame Bob? Probabilistic generative models, counterfactual reasoning, and blame attribution. In Proceedings of the 34th Annual Conference of the Cognitive Science Society (pp. 1996-2001). Austin, TX: Cognitive Science Society.
Stop, children what's that sound? Multi-modal inference through mental simulation
- J. Outa, X. Jia Zhou, H. Gweon, T. Gerstenberg
+ J. Outa*, X. Jia Zhou*, H. Gweon, T. Gerstenberg
@@ -292,7 +292,7 @@
Abstract
Publication
-
Outa* J., Zhou* X. J., Gweon H., Gerstenberg T. (2022). Stop, children what’s that sound? Multi-modal inference through mental simulation. In Cognitive Science Proceedings.
+
Outa J.*, Zhou X. J.*, Gweon H., Gerstenberg T. (2022). Stop, children what’s that sound? Multi-modal inference through mental simulation. In Cognitive Science Proceedings.
Whodunnit? Inferring what happened from multimodal evidence
+
+
+ S. A Wu*, E. Brockbank*, H. Cha, J. Fränken, E. Jin, Z. Huang, W. Liu, R. Zhang, J. Wu, T. Gerstenberg
+
+
+
+
+
+
+
+
+
+
+
+
Abstract
+
Humans are remarkably adept at inferring the causes of events in their environment; doing so often requires incorporating information from multiple sensory modalities. For instance, if a car slows down in front of us, inferences about why they did so are rapidly revised if we also hear sirens in the distance. Here, we investigate the ability to reconstruct others’ actions and events from the past by integrating multimodal information. Participants were asked to infer which of two agents performed an action in a household setting given either visual evidence, auditory evidence, or both. We develop a computational model that makes inferences by generating multimodal simulations, and also evaluate our task on a large language model (GPT-4) and a large multimodal model (GPT-4V). We find that humans are relatively accurate overall and perform best when given multimodal evidence. GPT-4 and GPT-4V performance comes close overall, but is very weakly correlated with participants across individual trials. Meanwhile, the simulation model captures the pattern of human responses well. Multimodal event reconstruction represents a challenge for current AI systems, and frameworks that draw on the cognitive processes underlying people’s ability to reconstruct events offer a promising avenue forward.
Wu, S. A., Brockbank, E., Cha, H., Fränken, J., Jin, E., Huang, Z., Liu, W., Zhang, R., Wu, J., Gerstenberg, T. (2024). Whodunnit? Inferring what happened from multimodal evidence. In Proceedings of the 46th Annual Conference of the Cognitive Science Society.
You are what you're for: Essentialist categorization in large language models
- S. Zhang, J. S. She, T. Gerstenberg, D. Rose
+ S. Zhang*, J. S. She*, T. Gerstenberg, D. Rose
@@ -283,7 +283,7 @@
Abstract
Publication
-
Zhang S., She J. S., Gerstenberg T., Rose D. (2023). You are what you’re for: Essentialist categorization in large language models. In Proceedings of the 45th Annual Conference of the Cognitive Science Society.
+
Zhang S.*, She J. S.*, Gerstenberg T., Rose D. (2023). You are what you’re for: Essentialist categorization in large language models. In Proceedings of the 45th Annual Conference of the Cognitive Science Society.
+
+ Humans are remarkably adept at inferring the causes of events in their environment; doing so often requires incorporating information from multiple sensory modalities. For instance, if a car slows down in front of us, inferences about why they did so …
+
+
-
- Human commonsense understanding of the physical and social world is organized around intuitive theories. These theories support making causal and moral judgments. When something bad happens, we naturally ask: who did what, and why? A rich literature …
-
-