From c7ce2a58b77021b50bc6ecf48dfba1d1915be2f9 Mon Sep 17 00:00:00 2001
From: Younes Belkada <49240599+younesbelkada@users.noreply.github.com>
Date: Thu, 25 May 2023 16:56:53 +0200
Subject: [PATCH] Add embedded space (#1143)
---
4bit-transformers-bitsandbytes.md | 14 +++++++++++++-
1 file changed, 13 insertions(+), 1 deletion(-)
diff --git a/4bit-transformers-bitsandbytes.md b/4bit-transformers-bitsandbytes.md
index 305b244956..e03d249bc4 100644
--- a/4bit-transformers-bitsandbytes.md
+++ b/4bit-transformers-bitsandbytes.md
@@ -30,7 +30,7 @@ This blogpost and release come with several resources to get started with 4bit m
- [Basic usage Google Colab notebook](https://colab.research.google.com/drive/1ge2F1QSK8Q7h0hn3YKuBCOAS0bK8E0wf?usp=sharing) - This notebook shows how to use 4bit models in inference with all their variants, and how to run GPT-neo-X (a 20B parameter model) on a free Google Colab instance 🤯
- [Fine tuning Google Colab notebook](https://colab.research.google.com/drive/1VoYNfYDKcKRQRor98Zbf2-9VQTtGJ24k?usp=sharing) - This notebook shows how to fine-tune a 4bit model on a downstream task using the Hugging Face ecosystem. We show that it is possible to fine tune GPT-neo-X 20B on a Google Colab instance!
- [Original repository for replicating the paper's results](https://github.com/artidoro/qlora)
-- [Guanaco 33b playground](https://huggingface.co/spaces/uwnlp/guanaco-playground-tgi)
+- [Guanaco 33b playground](https://huggingface.co/spaces/uwnlp/guanaco-playground-tgi) - or check the playground section below
## Introduction
@@ -229,6 +229,18 @@ We have also made some benchmarks on the impact of this quantization method on t
We have used the recent `SFTTrainer` from TRL library, and the benchmarking script can be found [here](https://gist.github.com/younesbelkada/f48af54c74ba6a39a7ae4fd777e72fe8)
+## Playground
+
+Try out the Guananco model cited on the paper on [the playground](https://huggingface.co/spaces/uwnlp/guanaco-playground-tgi) or directly below
+
+
+
+
+
+
## Acknowledgements
The HF team would like to acknowledge all the people involved in this project from University of Washington, and for making this available to the community.