Skip to content

Commit

Permalink
Prannaya-Laptop 2024-8-11:16:8:21
Browse files Browse the repository at this point in the history
  • Loading branch information
ThePyProgrammer committed Aug 11, 2024
1 parent d57f4c4 commit 01c72f7
Show file tree
Hide file tree
Showing 10 changed files with 2,471 additions and 188 deletions.
Binary file added docs/assets/short_paper.pdf
Binary file not shown.
7 changes: 7 additions & 0 deletions docs/components/dataset.md
Original file line number Diff line number Diff line change
Expand Up @@ -134,4 +134,11 @@ dummy_dataset[0]
# Prompt(prompt="what's up")
```

You can use this to load data from Pandas by

## Conclusion

We are able to


Hence, we are able to load several variety of datasets using our library. If there are any other ways you'd like us to support, do highlight in our [GitHub Issues](https://github.com/walledai/walledeval/issues).
16 changes: 7 additions & 9 deletions docs/index.md
Original file line number Diff line number Diff line change
Expand Up @@ -21,20 +21,18 @@
> Grateful to [Tensorplex](https://www.tensorplex.ai/) for their support with computing resources!
<!--

## 🔍 Quick Access

<div class="grid cards">
<a href="https://paperswithcode.com/paper/walledeval-a-comprehensive-safety-evaluation" class="card"><div markdown>
<div class="grid cards" markdown>

:simple-paperswithcode: Papers With Code
- [:simple-github:{ .lg .middle } __<font size="4" style="margin-left:10px">GitHub</font>__](https://github.com/walledai/walledeval)
- [:simple-arxiv:{ .lg .middle } __arXiv__](https://arxiv.org/abs/2408.03837)
- [:simple-paperswithcode:{ .lg .middle } __Papers With Code__](https://paperswithcode.com/paper/walledeval-a-comprehensive-safety-evaluation)
- [:simple-semanticscholar:{ .lg .middle } __Semantic Scholar__](https://www.semanticscholar.org/paper/WalledEval%3A-A-Comprehensive-Safety-Evaluation-for-Gupta-Yau/5c7da78b978e2ef6cc791cfbf98dafbcb59f758b)

</div></a>
<a href="https://www.semanticscholar.org/paper/WalledEval%3A-A-Comprehensive-Safety-Evaluation-for-Gupta-Yau/5c7da78b978e2ef6cc791cfbf98dafbcb59f758b" class="card">
:simple-semanticscholar: Semantic Scholar
</a>
</div>
-->


## 📚 Resources

Expand Down
2 changes: 1 addition & 1 deletion docs/quickstart/auto-red-teaming.md
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ Automated Red-Teaming allows users to automatically generate mutated malicious p
- **Prompts**: a compilation of malicious prompts
- **Mutators**: a way to create adverserial prompts from the malicious ones. This may or may not be generative.

Here's how you can do this easily in WalledEval!
Here's how you can do this easily in WalledEval!

```python title="auto_red_teaming_quickstart.py" linenums="1" hl_lines="23 33"
import torch
Expand Down
6 changes: 3 additions & 3 deletions docs/quickstart/prompts.md
Original file line number Diff line number Diff line change
@@ -1,6 +1,3 @@
---
hide: '["toc"]'
---
# Prompt Benchmarking

Most safety datasets aim to test LLMs on their creativity / restraint in generating responses to custom unsafe/safe queries. The general requirements for testing an LLM on Prompt Benchmarks is as follows:
Expand Down Expand Up @@ -63,3 +60,6 @@ logs[0]["score"] # True if safe, False if unsafe
```


## Dataset Loading

In this example, we load a list of prompts. Prompt Benchmarking is the most baseline benchmarking method as the only required parameter is **prompt**. Hence, we can input a list of strings
Empty file added docs/quickstart/refusal.md
Empty file.
Loading

0 comments on commit 01c72f7

Please sign in to comment.