Skip to content

Latest commit

 

History

History
65 lines (39 loc) · 6.18 KB

risks.md

File metadata and controls

65 lines (39 loc) · 6.18 KB

AI Risks and Ethical considerations

::::{important}

This page represents a small cross section of known and potential impacts of AI, with an emphasis on deployed systems and generative models.

For each point, there are scholarly sources linked supporting the point and in most sections the introduction includes links to trusted sources aimed at more general audiences.
::::

The most important thing to know is that these technologies reflect how they are built. Machine learning fundamentally uses the past to make decisions or generate content without good ability to use contextto determine what was bad in the past that we might not want to keep for the future.

Discrimination and Bias

In general, technological systems reflect the values of the people who build them. This may be explicit or implicit. With machine learning, the way the values of the builders are expressed is more distant from what goes out into the world; it is often less carefully checked.

Evironmental Impact

Training large language models requires an immense amount of energy and water (for cooling), using them requires less per use, but still a lot, especially when aggregated across all of the iterative prompting across all of the users. HuggingFace :hugging_face: , a company that provides a platform for sharing trained models, datasets, and applications provides a policy primer that is a good starting point @ai_environment_primer. Hugging Face researcher Sasha Luccioni also participated in an NPR Podcast on Climate impact of AI.

Some key highlights:

  • {cite:t}strubell2020Energy found that training an {term}LLM <large language model> with only 213 million parameters 1 was responsible for 625,155 points of $CO_2$ roughly equivalent to the life time emissions of five cars.
  • {cite:t}luccioni2023estimating followed a sample model, of similar size to GPT3 (176 Billion parameters vs 175) through its full lifespan from training to use. They found that the total carbon impact of such a model is more than training alone.
  • @luccioni2024power break down model energy use by different types of tasks (text classificaiton, summarization, text generation, image generation) and found that image generation uses nearly 60 times more energy than any text based task and has a carbon impact comparable to driving the average passenger vehicle about a mile2, using about as much energy as charging a cellphone half way 3
  • in addition to the electriciy use, large data centers use tremendous amounts of fresh water and those numbers have increased rapidly recently ( google increased by 20% and Microsoft by 34% from 2021 to 2022) @ren2023water
  • Depending on the location of data center that handles the query, GPT3 could consume4 7.2ml (Netherlands) to 48.3ml (Washington, US) of water per query meaning a full 500ml (16.9oz) bottle of water is consumed for every 11 (Washington) to 70 queries @ren2023water

::::{margin} :::{tip} For more context on these numbers, try out a carbon footprint calculator such as one of the following provided by different organizations:

:::{note} If you work with AI as a tool in your programming, you can use codecarbon to measure the impact of your work ::: ::::

Privacy

Models can reveal the data they were trained on, which means that any user could see any value that the model builder put in to learn from. For generative models this can happen in normal use, and for predictive models, certain types of explanations, which are required by law in some contexts can be used to figure out the training data.

  • Researchers demonstrated an efficient attack on the {term}production chatGPT @nasr2023Extracting. Combined with the fact that free tools almost always save the data you provide them to train and improve their products, that means that anything you send to a chatbot, it could spit back out to another user.
  • IP risks. Content generated by an AI cannot be copyrighted, it is not ownable.

Impacts on Human Thinking

  • Recently, a team at Univeristy of Toronto released a pre-print demonstrating a decrease in human creativity when asked to perform a task independently after exposure trials using an LLM @kumar2024human. This study measured two types of creativity using standard measures for each.
  • A team at Northwestern's Kellog School of Business released a pre-print showing that LLM creativity is similar to human creativity, but when prompted to respond as female, old, or Black, they score considerably worse on creativity @wangpreliminary. This study measured one type of creativity.
  • A large-scale meta analysis found that AI+ human performance is often worse than the best of human alone or AI alone. They also found that AI improved a person's performance mostly for generation tasks, but not for decision making tasks @vaccaro2024combinations.

Footnotes

  1. GPT3 (initial ChatGPT was GPT3.5) was 175 billion parameters @brown2020Language, nearly 1000x times as many as the model evaluated by @strubell2020Energy

  2. @luccioni2024power found a range aroud of carbon impact in he 150-500g range across 1000 tests; the US EPA says that the average passenger vehicle emits 400g per mile

  3. The US EPA changed its estimate of cell phone energy use to 0.22kWh in January 2024 from 0.012KWh prior to that, an initial version of the @luccioni2024power study said equal to a full cell phone charge, but the publised version says half, accordingly.

  4. @ren2023water differentiates between water withdrawal, which includes all water taken temporarily or permanenly and water consumption, calculated as withdrawal minus discharge. They then focus on consumption because it reflects the impact of water us on downstream water availability.