From 0e83ff17a34b126c516e5e6ada0cf3ec4a7a3121 Mon Sep 17 00:00:00 2001 From: Maxime Labonne <81252890+mlabonne@users.noreply.github.com> Date: Sat, 22 Jul 2023 20:19:15 +0100 Subject: [PATCH] add 4-bit LLM quantization using GPTQ --- README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/README.md b/README.md index d788897..8c25372 100644 --- a/README.md +++ b/README.md @@ -10,7 +10,7 @@ A list of notebooks and articles related to large language models. | Introduction to Weight Quantization | Large language model optimization using 8-bit quantization. | [Article](https://mlabonne.github.io/blog/posts/Introduction_to_Weight_Quantization.html) | Open In Colab | | Visualizing GPT-2's Loss Landscape | 3D plot of the loss landscape based on weight pertubations. | [Tweet](https://twitter.com/maximelabonne/status/1667618081844219904) | Open In Colab | | Improve ChatGPT with Knowledge Graphs | Augment ChatGPT's answers with knowledge graphs. | [Article](https://mlabonne.github.io/blog/posts/Article_Improve_ChatGPT_with_Knowledge_Graphs.html) | Open In Colab | -| Create a 4-bit GPT-2 model using AutoGPTQ | Quick notebook to quantize LLMs and upload them to the HF Hub. | [Tweet](https://twitter.com/maximelabonne/status/1672205530717122561) | Open In Colab | +| 4-bit LLM Quantization using GPTQ | Quantize your own open-source LLMs to run them on consumer hardware. | [Tweet](https://twitter.com/maximelabonne/status/1672205530717122561) | Open In Colab | | Fine-tune Llama 2 in Google Colab | Fine-tune a Llama 2 model on an HF dataset and upload it to the HF Hub. | [Tweet](https://twitter.com/maximelabonne/status/1681791164083576833) | Open In Colab |