From c7ce2a58b77021b50bc6ecf48dfba1d1915be2f9 Mon Sep 17 00:00:00 2001 From: Younes Belkada <49240599+younesbelkada@users.noreply.github.com> Date: Thu, 25 May 2023 16:56:53 +0200 Subject: [PATCH] Add embedded space (#1143) --- 4bit-transformers-bitsandbytes.md | 14 +++++++++++++- 1 file changed, 13 insertions(+), 1 deletion(-) diff --git a/4bit-transformers-bitsandbytes.md b/4bit-transformers-bitsandbytes.md index 305b244956..e03d249bc4 100644 --- a/4bit-transformers-bitsandbytes.md +++ b/4bit-transformers-bitsandbytes.md @@ -30,7 +30,7 @@ This blogpost and release come with several resources to get started with 4bit m - [Basic usage Google Colab notebook](https://colab.research.google.com/drive/1ge2F1QSK8Q7h0hn3YKuBCOAS0bK8E0wf?usp=sharing) - This notebook shows how to use 4bit models in inference with all their variants, and how to run GPT-neo-X (a 20B parameter model) on a free Google Colab instance 🤯 - [Fine tuning Google Colab notebook](https://colab.research.google.com/drive/1VoYNfYDKcKRQRor98Zbf2-9VQTtGJ24k?usp=sharing) - This notebook shows how to fine-tune a 4bit model on a downstream task using the Hugging Face ecosystem. We show that it is possible to fine tune GPT-neo-X 20B on a Google Colab instance! - [Original repository for replicating the paper's results](https://github.com/artidoro/qlora) -- [Guanaco 33b playground](https://huggingface.co/spaces/uwnlp/guanaco-playground-tgi) +- [Guanaco 33b playground](https://huggingface.co/spaces/uwnlp/guanaco-playground-tgi) - or check the playground section below ## Introduction @@ -229,6 +229,18 @@ We have also made some benchmarks on the impact of this quantization method on t We have used the recent `SFTTrainer` from TRL library, and the benchmarking script can be found [here](https://gist.github.com/younesbelkada/f48af54c74ba6a39a7ae4fd777e72fe8) +## Playground + +Try out the Guananco model cited on the paper on [the playground](https://huggingface.co/spaces/uwnlp/guanaco-playground-tgi) or directly below + + + + + + ## Acknowledgements The HF team would like to acknowledge all the people involved in this project from University of Washington, and for making this available to the community.