mirror of
https://github.com/openai/gpt-2
synced 2025-08-22 02:02:01 +00:00
30 lines
580 B
Markdown
30 lines
580 B
Markdown
|
|
# gpt-2
|
||
|
|
|
||
|
|
Code and samples from the paper "Language Models are Unsupervised Multitask Learners"
|
||
|
|
|
||
|
|
## Installation
|
||
|
|
|
||
|
|
Download the model data:
|
||
|
|
```
|
||
|
|
gsutil rsync -r gs://gpt-2/models/ models/
|
||
|
|
```
|
||
|
|
|
||
|
|
Install python packages:
|
||
|
|
```
|
||
|
|
pip install -r requirements.txt
|
||
|
|
```
|
||
|
|
|
||
|
|
## Sample generation
|
||
|
|
|
||
|
|
| WARNING: Samples are unfiltered and may contain offensive content. |
|
||
|
|
| --- |
|
||
|
|
|
||
|
|
To generate unconditional samples from the small model:
|
||
|
|
```
|
||
|
|
python3 src/main.py | tee samples
|
||
|
|
```
|
||
|
|
There are various flags for controlling the samples:
|
||
|
|
```
|
||
|
|
python3 src/main.py --top_k 40 --temperature 0.7 | tee samples
|
||
|
|
```
|