Picture for Tim Dettmers

Tim Dettmers

Distributed Inference and Fine-tuning of Large Language Models Over The Internet

Add code
Dec 13, 2023
Viaarxiv icon

MatFormer: Nested Transformer for Elastic Inference

Add code
Oct 11, 2023
Figure 1 for MatFormer: Nested Transformer for Elastic Inference
Figure 2 for MatFormer: Nested Transformer for Elastic Inference
Figure 3 for MatFormer: Nested Transformer for Elastic Inference
Figure 4 for MatFormer: Nested Transformer for Elastic Inference
Viaarxiv icon

SpQR: A Sparse-Quantized Representation for Near-Lossless LLM Weight Compression

Add code
Jun 05, 2023
Figure 1 for SpQR: A Sparse-Quantized Representation for Near-Lossless LLM Weight Compression
Figure 2 for SpQR: A Sparse-Quantized Representation for Near-Lossless LLM Weight Compression
Figure 3 for SpQR: A Sparse-Quantized Representation for Near-Lossless LLM Weight Compression
Figure 4 for SpQR: A Sparse-Quantized Representation for Near-Lossless LLM Weight Compression
Viaarxiv icon

QLoRA: Efficient Finetuning of Quantized LLMs

Add code
May 23, 2023
Figure 1 for QLoRA: Efficient Finetuning of Quantized LLMs
Figure 2 for QLoRA: Efficient Finetuning of Quantized LLMs
Figure 3 for QLoRA: Efficient Finetuning of Quantized LLMs
Figure 4 for QLoRA: Efficient Finetuning of Quantized LLMs
Viaarxiv icon

Towards A Unified View of Sparse Feed-Forward Network in Pretraining Large Language Model

Add code
May 23, 2023
Figure 1 for Towards A Unified View of Sparse Feed-Forward Network in Pretraining Large Language Model
Figure 2 for Towards A Unified View of Sparse Feed-Forward Network in Pretraining Large Language Model
Figure 3 for Towards A Unified View of Sparse Feed-Forward Network in Pretraining Large Language Model
Figure 4 for Towards A Unified View of Sparse Feed-Forward Network in Pretraining Large Language Model
Viaarxiv icon

Stable and low-precision training for large-scale vision-language models

Add code
Apr 25, 2023
Figure 1 for Stable and low-precision training for large-scale vision-language models
Figure 2 for Stable and low-precision training for large-scale vision-language models
Figure 3 for Stable and low-precision training for large-scale vision-language models
Figure 4 for Stable and low-precision training for large-scale vision-language models
Viaarxiv icon

SWARM Parallelism: Training Large Models Can Be Surprisingly Communication-Efficient

Add code
Jan 27, 2023
Figure 1 for SWARM Parallelism: Training Large Models Can Be Surprisingly Communication-Efficient
Figure 2 for SWARM Parallelism: Training Large Models Can Be Surprisingly Communication-Efficient
Figure 3 for SWARM Parallelism: Training Large Models Can Be Surprisingly Communication-Efficient
Figure 4 for SWARM Parallelism: Training Large Models Can Be Surprisingly Communication-Efficient
Viaarxiv icon

The case for 4-bit precision: k-bit Inference Scaling Laws

Add code
Dec 19, 2022
Figure 1 for The case for 4-bit precision: k-bit Inference Scaling Laws
Figure 2 for The case for 4-bit precision: k-bit Inference Scaling Laws
Figure 3 for The case for 4-bit precision: k-bit Inference Scaling Laws
Figure 4 for The case for 4-bit precision: k-bit Inference Scaling Laws
Viaarxiv icon

BLOOM: A 176B-Parameter Open-Access Multilingual Language Model

Add code
Nov 09, 2022
Viaarxiv icon

Petals: Collaborative Inference and Fine-tuning of Large Models

Add code
Sep 02, 2022
Figure 1 for Petals: Collaborative Inference and Fine-tuning of Large Models
Figure 2 for Petals: Collaborative Inference and Fine-tuning of Large Models
Figure 3 for Petals: Collaborative Inference and Fine-tuning of Large Models
Figure 4 for Petals: Collaborative Inference and Fine-tuning of Large Models
Viaarxiv icon