Skip to content

Commit

Permalink
Rebuilt site
Browse files Browse the repository at this point in the history
  • Loading branch information
David Evans committed Oct 11, 2023
1 parent 3d7c3ad commit d346029
Show file tree
Hide file tree
Showing 17 changed files with 954 additions and 66 deletions.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Binary file added images/Week5/Hallucination_page_4.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Binary file added images/Week5/Hallucination_page_6.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
291 changes: 290 additions & 1 deletion index.html

Large diffs are not rendered by default.

20 changes: 16 additions & 4 deletions index.xml
Original file line number Diff line number Diff line change
Expand Up @@ -8,17 +8,29 @@
<language>en-us</language>
<managingEditor>[email protected] (David Evans)</managingEditor>
<webMaster>[email protected] (David Evans)</webMaster>
<lastBuildDate>Mon, 25 Sep 2023 00:00:00 +0000</lastBuildDate><atom:link href="https://llmrisks.github.io/index.xml" rel="self" type="application/rss+xml" />
<lastBuildDate>Wed, 04 Oct 2023 00:00:00 +0000</lastBuildDate><atom:link href="https://llmrisks.github.io/index.xml" rel="self" type="application/rss+xml" />
<item>
<title>Week 5: Hallucination</title>
<link>https://llmrisks.github.io/week5/</link>
<pubDate>Wed, 04 Oct 2023 00:00:00 +0000</pubDate>
<author>[email protected] (David Evans)</author>
<guid>https://llmrisks.github.io/week5/</guid>
<description>(see bottom for assigned readings and questions)
Hallucination (Week 5) Presenting Team: Liu Zhe, Peng Wang, Sikun Guo, Yinhan He, Zhepei Wei
Blogging Team: Anshuman Suri, Jacob Christopher, Kasra Lekan, Kaylee Liu, My Dinh
Wednesday, September 27th: Intro to Hallucination People Hallucinate Too In general, hallucinations refer to the propagation of false information and/or misinformation. One common example of hallucinations is the Mandela Effect, where incorrect memories are shared by a large group of people.</description>
</item>

<item>
<title>Week 4: Capabilities of LLMs</title>
<link>https://llmrisks.github.io/week4/</link>
<pubDate>Mon, 25 Sep 2023 00:00:00 +0000</pubDate>
<author>[email protected] (David Evans)</author>
<guid>https://llmrisks.github.io/week4/</guid>
<description>Capabilities of LLMs (Week 4) Presenting Team: Xindi Guo, Mengxuan Hu, Tseganesh Beyene Kebede, Zihan Guan
<description>(see bottom for assigned readings and questions)
Capabilities of LLMs (Week 4) Presenting Team: Xindi Guo, Mengxuan Hu, Tseganesh Beyene Kebede, Zihan Guan
Blogging Team: Ajwa Shahid, Caroline Gihlstorf, Changhong Yang, Hyeongjin Kim, Sarah Boyce
Monday, September 18 Jingfeng Yang, Hongye Jin, Ruixiang Tang, Xiaotian Han, Qizhang Feng, Haoming Jiang, Bing Yin, Xia Hu. Harnessing the Power of LLMs in Practice: A Survey on ChatGPT and Beyond. April 2023. https://arxiv.org/abs/2304.13712
This discussion was essential to highlight the distinction between large language models (LLMs) and fine-tuned models.</description>
Monday, September 18 Jingfeng Yang, Hongye Jin, Ruixiang Tang, Xiaotian Han, Qizhang Feng, Haoming Jiang, Bing Yin, Xia Hu. Harnessing the Power of LLMs in Practice: A Survey on ChatGPT and Beyond. April 2023. https://arxiv.org/abs/2304.13712</description>
</item>

<item>
Expand Down
19 changes: 18 additions & 1 deletion post/index.html
Original file line number Diff line number Diff line change
Expand Up @@ -83,6 +83,23 @@
<div class="row">


<h2><a href="/week5/">Week 5: Hallucination</a></h2>
<div class="post-metadata">
<span class="post-date">
<time datetime="2023-10-04 00:00:00 &#43;0000 UTC" itemprop="datePublished">4 October 2023</time>
</span>

</div>


(see bottom for assigned readings and questions)
Hallucination (Week 5) Presenting Team: Liu Zhe, Peng Wang, Sikun Guo, Yinhan He, Zhepei Wei
Blogging Team: Anshuman Suri, Jacob Christopher, Kasra Lekan, Kaylee Liu, My Dinh
Wednesday, September 27th: Intro to Hallucination People Hallucinate Too In general, hallucinations refer to the propagation of false information and/or misinformation. One common example of hallucinations is the Mandela Effect, where incorrect memories are shared by a large group of people.
<p class="text-right"><a href="/week5/">Read More…</a></p>



<h2><a href="/week4/">Week 4: Capabilities of LLMs</a></h2>
<div class="post-metadata">
<span class="post-date">
Expand All @@ -92,10 +109,10 @@ <h2><a href="/week4/">Week 4: Capabilities of LLMs</a></h2>
</div>


(see bottom for assigned readings and questions)
Capabilities of LLMs (Week 4) Presenting Team: Xindi Guo, Mengxuan Hu, Tseganesh Beyene Kebede, Zihan Guan
Blogging Team: Ajwa Shahid, Caroline Gihlstorf, Changhong Yang, Hyeongjin Kim, Sarah Boyce
Monday, September 18 Jingfeng Yang, Hongye Jin, Ruixiang Tang, Xiaotian Han, Qizhang Feng, Haoming Jiang, Bing Yin, Xia Hu. Harnessing the Power of LLMs in Practice: A Survey on ChatGPT and Beyond. April 2023. https://arxiv.org/abs/2304.13712
This discussion was essential to highlight the distinction between large language models (LLMs) and fine-tuned models.
<p class="text-right"><a href="/week4/">Read More…</a></p>


Expand Down
20 changes: 16 additions & 4 deletions post/index.xml
Original file line number Diff line number Diff line change
Expand Up @@ -8,17 +8,29 @@
<language>en-us</language>
<managingEditor>[email protected] (David Evans)</managingEditor>
<webMaster>[email protected] (David Evans)</webMaster>
<lastBuildDate>Mon, 25 Sep 2023 00:00:00 +0000</lastBuildDate><atom:link href="https://llmrisks.github.io/post/index.xml" rel="self" type="application/rss+xml" />
<lastBuildDate>Wed, 04 Oct 2023 00:00:00 +0000</lastBuildDate><atom:link href="https://llmrisks.github.io/post/index.xml" rel="self" type="application/rss+xml" />
<item>
<title>Week 5: Hallucination</title>
<link>https://llmrisks.github.io/week5/</link>
<pubDate>Wed, 04 Oct 2023 00:00:00 +0000</pubDate>
<author>[email protected] (David Evans)</author>
<guid>https://llmrisks.github.io/week5/</guid>
<description>(see bottom for assigned readings and questions)
Hallucination (Week 5) Presenting Team: Liu Zhe, Peng Wang, Sikun Guo, Yinhan He, Zhepei Wei
Blogging Team: Anshuman Suri, Jacob Christopher, Kasra Lekan, Kaylee Liu, My Dinh
Wednesday, September 27th: Intro to Hallucination People Hallucinate Too In general, hallucinations refer to the propagation of false information and/or misinformation. One common example of hallucinations is the Mandela Effect, where incorrect memories are shared by a large group of people.</description>
</item>

<item>
<title>Week 4: Capabilities of LLMs</title>
<link>https://llmrisks.github.io/week4/</link>
<pubDate>Mon, 25 Sep 2023 00:00:00 +0000</pubDate>
<author>[email protected] (David Evans)</author>
<guid>https://llmrisks.github.io/week4/</guid>
<description>Capabilities of LLMs (Week 4) Presenting Team: Xindi Guo, Mengxuan Hu, Tseganesh Beyene Kebede, Zihan Guan
<description>(see bottom for assigned readings and questions)
Capabilities of LLMs (Week 4) Presenting Team: Xindi Guo, Mengxuan Hu, Tseganesh Beyene Kebede, Zihan Guan
Blogging Team: Ajwa Shahid, Caroline Gihlstorf, Changhong Yang, Hyeongjin Kim, Sarah Boyce
Monday, September 18 Jingfeng Yang, Hongye Jin, Ruixiang Tang, Xiaotian Han, Qizhang Feng, Haoming Jiang, Bing Yin, Xia Hu. Harnessing the Power of LLMs in Practice: A Survey on ChatGPT and Beyond. April 2023. https://arxiv.org/abs/2304.13712
This discussion was essential to highlight the distinction between large language models (LLMs) and fine-tuned models.</description>
Monday, September 18 Jingfeng Yang, Hongye Jin, Ruixiang Tang, Xiaotian Han, Qizhang Feng, Haoming Jiang, Bing Yin, Xia Hu. Harnessing the Power of LLMs in Practice: A Survey on ChatGPT and Beyond. April 2023. https://arxiv.org/abs/2304.13712</description>
</item>

<item>
Expand Down
7 changes: 5 additions & 2 deletions sitemap.xml
Original file line number Diff line number Diff line change
Expand Up @@ -3,10 +3,13 @@
xmlns:xhtml="http://www.w3.org/1999/xhtml">
<url>
<loc>https://llmrisks.github.io/post/</loc>
<lastmod>2023-09-25T00:00:00+00:00</lastmod>
<lastmod>2023-10-04T00:00:00+00:00</lastmod>
</url><url>
<loc>https://llmrisks.github.io/</loc>
<lastmod>2023-09-25T00:00:00+00:00</lastmod>
<lastmod>2023-10-04T00:00:00+00:00</lastmod>
</url><url>
<loc>https://llmrisks.github.io/week5/</loc>
<lastmod>2023-10-04T00:00:00+00:00</lastmod>
</url><url>
<loc>https://llmrisks.github.io/week4/</loc>
<lastmod>2023-09-25T00:00:00+00:00</lastmod>
Expand Down
43 changes: 43 additions & 0 deletions src/content/post/week4.md
Original file line number Diff line number Diff line change
Expand Up @@ -5,6 +5,8 @@ title = "Week 4: Capabilities of LLMs"
slug = "week4"
+++

(see bottom for assigned readings and questions)

# Capabilities of LLMs (Week 4)

<author>Presenting Team: Xindi Guo, Mengxuan Hu, Tseganesh Beyene Kebede, Zihan Guan</author>
Expand Down Expand Up @@ -200,3 +202,44 @@ drew M. Dai, Thanumalayan Sankaranarayana Pillai, Marie Pellat, Aitor Lewkowycz,
Erica Moreira, Rewon Child, Oleksandr Polozov, Katherine Lee, Zongwei Zhou, Xuezhi Wang, Brennan Saeta, Mark Diaz, Orhan Firat, Michele Catasta, Jason Wei, Kathy
Meier-Hellstern, Douglas Eck, Jeff Dean, Slav Petrov, and Noah Fiedel. Palm: Scaling
language modeling with pathways, 2022. https://arxiv.org/abs/2204.02311

## Readings

**Monday:**

1. Jingfeng Yang, Hongye Jin, Ruixiang Tang, Xiaotian Han, Qizhang Feng, Haoming Jiang, Bing Yin, Xia Hu. [_Harnessing the Power of LLMs in Practice: A Survey on ChatGPT and Beyond_](https://arxiv.org/abs/2304.13712). April 2023. [https://arxiv.org/abs/2304.13712](https://arxiv.org/abs/2304.13712). [[PDF](https://arxiv.org/pdf/2304.13712.pdf)]

2. OpenAI. [_GPT-4 Technical Report_](https://arxiv.org/abs/2303.08774). March 2023. [https://arxiv.org/abs/2303.08774](https://arxiv.org/abs/2303.08774) [[PDF](https://arxiv.org/pdf/2303.08774.pdf)]

Optionally, also explore [https://openai.com/blog/chatgpt-plugins](https://openai.com/blog/chatgpt-plugins).

**Wednesday:**

3. Karan Singhal, Tao Tu, Juraj Gottweis, Rory Sayres, Ellery Wulczyn, Le Hou, Kevin Clark, Stephen Pfohl, Heather Cole-Lewis, Darlene Neal, Mike Schaekermann, Amy Wang, Mohamed Amin, Sami Lachgar, Philip Mansfield, Sushant Prakash, Bradley Green, Ewa Dominowska, Blaise Aguera y Arcas, Nenad Tomasev, Yun Liu, Renee Wong, Christopher Semturs, S. Sara Mahdavi, Joelle Barral, Dale Webster, Greg S. Corrado, Yossi Matias, Shekoofeh Azizi, Alan Karthikesalingam, Vivek Natarajan. [_Towards Expert-Level Medical Question Answering with Large Language Models_](https://arxiv.org/abs/2305.09617)
[https://arxiv.org/abs/2305.09617](https://arxiv.org/abs/2305.09617) [[PDF](https://arxiv.org/pdf/2305.09617.pdf)]

Optional Readings:
- Harsha Nori, Nicholas King, Scott Mayer McKinney, Dean Carignan, Eric Horvitz. [_Capabilities of GPT-4 on Medical Challenge Problems_](https://arxiv.org/abs/2303.13375). March 2023. [https://arxiv.org/abs/2303.13375](https://arxiv.org/abs/2303.13375)
- Travis Zack, Eric Lehman, Mirac Suzgun, Jorge A. Rodriguez, Leo Anthony Celi, Judy Gichoya, Dan Jurafsky, Peter Szolovits, David W. Bates, Raja-Elie E. Abdulnour, Atul J. Butte, Emily Alsentzer. [_Coding Inequity: Assessing GPT-4’s Potential for Perpetuating Racial and Gender Biases in Healthcare_](https://www.medrxiv.org/content/10.1101/2023.07.13.23292577). July 2023.
[https://www.medrxiv.org/content/10.1101/2023.07.13.23292577](https://www.medrxiv.org/content/10.1101/2023.07.13.23292577) &mdash; This article relates to the underlying biases in the models we talked about this week, but with an application that show clear potential harm resulting from these biases in the form if increased risk of medical misdiagnosis.

## Discussion for Monday:

Everyone who is not in either the lead or blogging team for the week should post (in the comments below) an answer to at least one of the questions in this section, or a substantive response to someone else's comment, or something interesting about the readings that is not covered by these questions. Don't post duplicates - if others have already posted, you should read their responses before adding your own. Please post your responses to different questions as separate comments.

You should post your _initial_ response before 5:29pm on Sunday, September 17, but feel free (and encouraged!) to continue the discussion after that, including responding to any responses by others to your comments.

1. Based on the criterions shown in Figure 2 of [1], imagine a practical scenario and explain why you would choose or not choose using LLMs for your scenario.
2. Are plug-ins the future of AGI? Do you think that a company should only focus on building powerful AI systems that does not need any support from plug-ins, or they should only focus on the core system and involve more plug-ins into the ecosystem?

## Discussion for Wednesday:

You should post your _initial_ response to one of the questions below or something interesting related to the Wednesday readings before 5:29pm on Tuesday, September 19.

1. What should we do before deploying LLMs in medical diagnosis applications? What (if any) regulations should control or limit how they would be used?

2. With LLMs handling sensitive medical information, how can patient privacy and data security be maintained? What policies and safeguards should be in place to protect patient data?

3. The paper discusses the progress of LLMs towards achieving physician-level performance in medical question answering. What are the potential implications of LLMs reaching or surpassing human expertise in medical knowledge?

4. The paper mentions the importance of safety and minimizing bias in LLM-generated medical information, and the [optional reading](https://www.medrxiv.org/content/10.1101/2023.07.13.23292577) reports on some experiments that show biases in GPT's medical diagnoses. Should models be tuned to ignore protected attributes? Should we prevent models from being used in medical applications until these problems can be solved?
Loading

0 comments on commit d346029

Please sign in to comment.