@ -13,6 +13,7 @@ A list of notebooks and articles related to large language models.
| Introduction to Weight Quantization | Large language model optimization using 8-bit quantization. | [Article](https://mlabonne.github.io/blog/posts/Introduction_to_Weight_Quantization.html) | <ahref="https://colab.research.google.com/drive/1DPr4mUQ92Cc-xf4GgAaB6dFcFnWIvqYi?usp=sharing"><imgsrc="images/colab.svg"alt="Open In Colab"></a> |
| 4-bit LLM Quantization using GPTQ | Quantize your own open-source LLMs to run them on consumer hardware. | [Article](https://mlabonne.github.io/blog/4bit_quantization/) | <ahref="https://colab.research.google.com/drive/1lSvVDaRgqQp_mWK_jC9gydz6_-y6Aq4A?usp=sharing"><imgsrc="images/colab.svg"alt="Open In Colab"></a> |
| Quantize Llama models with GGML and llama.cpp | Quantize Llama 2 models with llama.cpp and upload GGUF to the HF Hub. | [Article](https://mlabonne.github.io/blog/posts/Quantize_Llama_2_models_using_ggml.html) | <ahref="https://colab.research.google.com/drive/1pL8k7m04mgE5jo2NrjGi8atB0j_37aDD?usp=sharing"><imgsrc="images/colab.svg"alt="Open In Colab"></a> |
| ExLlamaV2: The Fastest Library to RunLLMs | Quantize and run EXL2models and upload them to the HF Hub. | [Article]() | <ahref="https://colab.research.google.com/drive/1yrq4XBlxiA0fALtMoT2dwiACVc77PHou?usp=sharing"><imgsrc="images/colab.svg"alt="Open In Colab"></a> |