Picture for Peter Henderson

Peter Henderson

FLawN-T5: An Empirical Examination of Effective Instruction-Tuning Data Mixtures for Legal Reasoning

Add code
Apr 02, 2024
Viaarxiv icon

What's in Your "Safe" Data?: Identifying Benign Data that Breaks Safety

Add code
Apr 01, 2024
Viaarxiv icon

A Safe Harbor for AI Evaluation and Red Teaming

Add code
Mar 07, 2024
Figure 1 for A Safe Harbor for AI Evaluation and Red Teaming
Figure 2 for A Safe Harbor for AI Evaluation and Red Teaming
Figure 3 for A Safe Harbor for AI Evaluation and Red Teaming
Figure 4 for A Safe Harbor for AI Evaluation and Red Teaming
Viaarxiv icon

On the Societal Impact of Open Foundation Models

Add code
Feb 27, 2024
Figure 1 for On the Societal Impact of Open Foundation Models
Figure 2 for On the Societal Impact of Open Foundation Models
Viaarxiv icon

Assessing the Brittleness of Safety Alignment via Pruning and Low-Rank Modifications

Add code
Feb 07, 2024
Viaarxiv icon

Fine-tuning Aligned Language Models Compromises Safety, Even When Users Do Not Intend To!

Add code
Oct 05, 2023
Figure 1 for Fine-tuning Aligned Language Models Compromises Safety, Even When Users Do Not Intend To!
Figure 2 for Fine-tuning Aligned Language Models Compromises Safety, Even When Users Do Not Intend To!
Figure 3 for Fine-tuning Aligned Language Models Compromises Safety, Even When Users Do Not Intend To!
Figure 4 for Fine-tuning Aligned Language Models Compromises Safety, Even When Users Do Not Intend To!
Viaarxiv icon

LegalBench: A Collaboratively Built Benchmark for Measuring Legal Reasoning in Large Language Models

Add code
Aug 20, 2023
Figure 1 for LegalBench: A Collaboratively Built Benchmark for Measuring Legal Reasoning in Large Language Models
Figure 2 for LegalBench: A Collaboratively Built Benchmark for Measuring Legal Reasoning in Large Language Models
Figure 3 for LegalBench: A Collaboratively Built Benchmark for Measuring Legal Reasoning in Large Language Models
Figure 4 for LegalBench: A Collaboratively Built Benchmark for Measuring Legal Reasoning in Large Language Models
Viaarxiv icon

Where's the Liability in Harmful AI Speech?

Add code
Aug 16, 2023
Figure 1 for Where's the Liability in Harmful AI Speech?
Figure 2 for Where's the Liability in Harmful AI Speech?
Figure 3 for Where's the Liability in Harmful AI Speech?
Figure 4 for Where's the Liability in Harmful AI Speech?
Viaarxiv icon

Cheaply Evaluating Inference Efficiency Metrics for Autoregressive Transformer APIs

Add code
May 03, 2023
Figure 1 for Cheaply Evaluating Inference Efficiency Metrics for Autoregressive Transformer APIs
Figure 2 for Cheaply Evaluating Inference Efficiency Metrics for Autoregressive Transformer APIs
Figure 3 for Cheaply Evaluating Inference Efficiency Metrics for Autoregressive Transformer APIs
Figure 4 for Cheaply Evaluating Inference Efficiency Metrics for Autoregressive Transformer APIs
Viaarxiv icon

Foundation Models and Fair Use

Add code
Mar 28, 2023
Figure 1 for Foundation Models and Fair Use
Figure 2 for Foundation Models and Fair Use
Figure 3 for Foundation Models and Fair Use
Figure 4 for Foundation Models and Fair Use
Viaarxiv icon