Skip to content

Commit

Permalink
Merge pull request #2 from zqypku/patch-1
Browse files Browse the repository at this point in the history
Update README.md
  • Loading branch information
cnut1648 authored Jul 11, 2024
2 parents 0a3fc50 + 84ea6bd commit 4ae5e8a
Showing 1 changed file with 4 additions and 4 deletions.
8 changes: 4 additions & 4 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -40,9 +40,9 @@ We explore six downstream datasets. This is NOT needed if you only need to finge

Alpaca 52k is in [Alpaca](https://github.com/tatsu-lab/stanford_alpaca) repo already. For the rest of dataset:
```shell
python preapre_ni.py # natural instruction v2
python preapre_dolly.py # dolly
python preapre_sharegpt.py # share GPT
python prepare_ni.py # natural instruction v2
python prepare_dolly.py # dolly
python prepare_sharegpt.py # share GPT
```
`Alpaca-GPT4` can be downloaded in [their repo](https://github.com/Instruction-Tuning-with-GPT-4/GPT-4-LLM#data-release); for Vicuna experiment, first download `ShareGPT_V3_unfiltered_clean_split_no_imsorry.json` from [here](https://huggingface.co/datasets/anon8231489123/ShareGPT_Vicuna_unfiltered/tree/main) and use Vicuna's [offical processing script](https://github.com/lm-sys/FastChat/blob/main/docs/commands/data_cleaning.md) to generate the dataset.
```shell
Expand Down Expand Up @@ -252,4 +252,4 @@ If you find our project helpful, please cite our paper:
archivePrefix={arXiv},
primaryClass={cs.CR}
}
```
```

0 comments on commit 4ae5e8a

Please sign in to comment.