diff --git a/README.md b/README.md index 7e37de71..1253b422 100644 --- a/README.md +++ b/README.md @@ -2,7 +2,7 @@ - 🤗 **Try the pretrained model out [here](https://huggingface.co/spaces/tloen/alpaca-lora), courtesy of a GPU grant from Huggingface!** - Users have created a Discord server for discussion and support [here](https://discord.gg/prbq284xX5) -- 4/6: Repo has been updated with Microsoft Research's [LLaMA-GPT4 dataset](https://github.com/Instruction-Tuning-with-GPT-4/GPT-4-LLM). +- 4/14: Chansung Park's GPT4-Alpaca adapters: https://github.com/tloen/alpaca-lora/issues/340 This repository contains code for reproducing the [Stanford Alpaca](https://github.com/tatsu-lab/stanford_alpaca) results using [low-rank adaptation (LoRA)](https://arxiv.org/pdf/2106.09685.pdf). We provide an Instruct model of similar quality to `text-davinci-003` that can run [on a Raspberry Pi](https://twitter.com/miolini/status/1634982361757790209) (for research), @@ -158,8 +158,10 @@ docker-compose down --volumes --rmi all - [dolly-15k-instruction-alpaca-format](https://huggingface.co/datasets/c-s-ale/dolly-15k-instruction-alpaca-format), an Alpaca-compatible version of [Databricks' Dolly 15k human-generated instruct dataset](https://github.com/databrickslabs/dolly/tree/master/data) (see [blog](https://www.databricks.com/blog/2023/04/12/dolly-first-open-commercially-viable-instruction-tuned-llm)) - Various adapter weights (download at own risk): - 7B: - - - - + - 3️⃣ + - 3️⃣ + - **4️⃣ ** + - 🚀 - 🇧🇷 - 🇨🇳 - 🇨🇳 @@ -174,10 +176,11 @@ docker-compose down --volumes --rmi all - 🇺🇦 - 🇮🇹 - 13B: - - - - - - - - + - 3️⃣ + - 3️⃣ + - 3️⃣ + - 3️⃣ + - **4️⃣ ** - 🇯🇵 - 🇰🇷 - 🇨🇳 @@ -185,8 +188,9 @@ docker-compose down --volumes --rmi all - 🇪🇸 - 🇮🇹 - 30B: - - - - + - 3️⃣ + - 3️⃣ + - **4️⃣ ** - 🇯🇵 - 65B -