From 0e83ff17a34b126c516e5e6ada0cf3ec4a7a3121 Mon Sep 17 00:00:00 2001
From: Maxime Labonne <81252890+mlabonne@users.noreply.github.com>
Date: Sat, 22 Jul 2023 20:19:15 +0100
Subject: [PATCH] add 4-bit LLM quantization using GPTQ
---
README.md | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
diff --git a/README.md b/README.md
index d788897..8c25372 100644
--- a/README.md
+++ b/README.md
@@ -10,7 +10,7 @@ A list of notebooks and articles related to large language models.
| Introduction to Weight Quantization | Large language model optimization using 8-bit quantization. | [Article](https://mlabonne.github.io/blog/posts/Introduction_to_Weight_Quantization.html) |
|
| Visualizing GPT-2's Loss Landscape | 3D plot of the loss landscape based on weight pertubations. | [Tweet](https://twitter.com/maximelabonne/status/1667618081844219904) |
|
| Improve ChatGPT with Knowledge Graphs | Augment ChatGPT's answers with knowledge graphs. | [Article](https://mlabonne.github.io/blog/posts/Article_Improve_ChatGPT_with_Knowledge_Graphs.html) |
|
-| Create a 4-bit GPT-2 model using AutoGPTQ | Quick notebook to quantize LLMs and upload them to the HF Hub. | [Tweet](https://twitter.com/maximelabonne/status/1672205530717122561) |
|
+| 4-bit LLM Quantization using GPTQ | Quantize your own open-source LLMs to run them on consumer hardware. | [Tweet](https://twitter.com/maximelabonne/status/1672205530717122561) |
|
| Fine-tune Llama 2 in Google Colab | Fine-tune a Llama 2 model on an HF dataset and upload it to the HF Hub. | [Tweet](https://twitter.com/maximelabonne/status/1681791164083576833) |
|