gpt-2/README.md

49 lines
1.4 KiB
Markdown
Raw Normal View History

2019-02-10 20:22:00 -08:00
# gpt-2
2019-02-14 08:43:50 -08:00
Code and samples from the paper ["Language Models are Unsupervised Multitask Learners"](https://d4mucfpksywv.cloudfront.net/better-language-models/language-models.pdf).
For now, we have only released a smaller (117M parameter) version of GPT-2.
See more details in our [blog post](https://blog.openai.com/better-language-models/).
2019-02-10 20:22:00 -08:00
## Installation
2019-02-14 08:43:50 -08:00
Download the model data (needs [gsutil](https://cloud.google.com/storage/docs/gsutil_install)):
2019-02-10 20:22:00 -08:00
```
2019-02-14 09:12:05 -08:00
sh download_model.sh 117M
2019-02-10 20:22:00 -08:00
```
Install python packages:
```
2019-02-14 09:12:05 -08:00
pip3 install -r requirements.txt
2019-02-10 20:22:00 -08:00
```
## Unconditional sample generation
2019-02-10 20:22:00 -08:00
| WARNING: Samples are unfiltered and may contain offensive content. |
| --- |
To generate unconditional samples from the small model:
```
python3 src/generate_unconditional_samples.py | tee samples
2019-02-10 20:22:00 -08:00
```
There are various flags for controlling the samples:
```
python3 src/generate_unconditional_samples.py --top_k 40 --temperature 0.7 | tee samples
2019-02-10 20:22:00 -08:00
```
2019-02-14 00:17:55 -08:00
While we have not yet released GPT-2 itself, you can see some unconditional samples from it (with default settings of temperature 1 and no truncation) in `gpt2-samples.txt`.
## Conditional sample generation
To give the model custom prompts, you can use:
```
python3 src/interactive_conditional_samples.py
```
2019-02-14 00:17:55 -08:00
## Future work
We may release code for evaluating the models on various benchmarks.
We are still considering release of the larger models.