Picture for Sang Michael Xie

Sang Michael Xie

A Survey on Data Selection for Language Models

Add code
Mar 08, 2024
Viaarxiv icon

DoReMi: Optimizing Data Mixtures Speeds Up Language Model Pretraining

Add code
May 24, 2023
Figure 1 for DoReMi: Optimizing Data Mixtures Speeds Up Language Model Pretraining
Figure 2 for DoReMi: Optimizing Data Mixtures Speeds Up Language Model Pretraining
Figure 3 for DoReMi: Optimizing Data Mixtures Speeds Up Language Model Pretraining
Figure 4 for DoReMi: Optimizing Data Mixtures Speeds Up Language Model Pretraining
Viaarxiv icon

Reward Design with Language Models

Add code
Feb 27, 2023
Figure 1 for Reward Design with Language Models
Figure 2 for Reward Design with Language Models
Figure 3 for Reward Design with Language Models
Figure 4 for Reward Design with Language Models
Viaarxiv icon

Data Selection for Language Models via Importance Resampling

Add code
Feb 06, 2023
Figure 1 for Data Selection for Language Models via Importance Resampling
Figure 2 for Data Selection for Language Models via Importance Resampling
Figure 3 for Data Selection for Language Models via Importance Resampling
Figure 4 for Data Selection for Language Models via Importance Resampling
Viaarxiv icon

Holistic Evaluation of Language Models

Add code
Nov 16, 2022
Figure 1 for Holistic Evaluation of Language Models
Figure 2 for Holistic Evaluation of Language Models
Figure 3 for Holistic Evaluation of Language Models
Figure 4 for Holistic Evaluation of Language Models
Viaarxiv icon

Same Pre-training Loss, Better Downstream: Implicit Bias Matters for Language Models

Add code
Oct 25, 2022
Figure 1 for Same Pre-training Loss, Better Downstream: Implicit Bias Matters for Language Models
Figure 2 for Same Pre-training Loss, Better Downstream: Implicit Bias Matters for Language Models
Figure 3 for Same Pre-training Loss, Better Downstream: Implicit Bias Matters for Language Models
Figure 4 for Same Pre-training Loss, Better Downstream: Implicit Bias Matters for Language Models
Viaarxiv icon

Connect, Not Collapse: Explaining Contrastive Learning for Unsupervised Domain Adaptation

Add code
Apr 01, 2022
Figure 1 for Connect, Not Collapse: Explaining Contrastive Learning for Unsupervised Domain Adaptation
Figure 2 for Connect, Not Collapse: Explaining Contrastive Learning for Unsupervised Domain Adaptation
Figure 3 for Connect, Not Collapse: Explaining Contrastive Learning for Unsupervised Domain Adaptation
Figure 4 for Connect, Not Collapse: Explaining Contrastive Learning for Unsupervised Domain Adaptation
Viaarxiv icon

Extending the WILDS Benchmark for Unsupervised Adaptation

Add code
Dec 09, 2021
Figure 1 for Extending the WILDS Benchmark for Unsupervised Adaptation
Figure 2 for Extending the WILDS Benchmark for Unsupervised Adaptation
Figure 3 for Extending the WILDS Benchmark for Unsupervised Adaptation
Figure 4 for Extending the WILDS Benchmark for Unsupervised Adaptation
Viaarxiv icon

An Explanation of In-context Learning as Implicit Bayesian Inference

Add code
Nov 14, 2021
Figure 1 for An Explanation of In-context Learning as Implicit Bayesian Inference
Figure 2 for An Explanation of In-context Learning as Implicit Bayesian Inference
Figure 3 for An Explanation of In-context Learning as Implicit Bayesian Inference
Figure 4 for An Explanation of In-context Learning as Implicit Bayesian Inference
Viaarxiv icon

No True State-of-the-Art? OOD Detection Methods are Inconsistent across Datasets

Add code
Sep 12, 2021
Figure 1 for No True State-of-the-Art? OOD Detection Methods are Inconsistent across Datasets
Figure 2 for No True State-of-the-Art? OOD Detection Methods are Inconsistent across Datasets
Figure 3 for No True State-of-the-Art? OOD Detection Methods are Inconsistent across Datasets
Figure 4 for No True State-of-the-Art? OOD Detection Methods are Inconsistent across Datasets
Viaarxiv icon