Skip to content
This repository has been archived by the owner on Oct 31, 2022. It is now read-only.

Added datasize paramater; changed imports #42

Open
wants to merge 65 commits into
base: finetuning
Choose a base branch
from

Conversation

theurerjohn3
Copy link

I could not train on smaller data sets, so I added an argument sample_batch_size to the train.py file, to make it possible to specify the sample size.

Additionally, I changed the file imports so that you no longer need to move the train.py to the src directory. I additionally changed the encode.py file and the train-horovod.py in the same manner, providing a path to look at the src file.

WuTheFWasThat and others added 30 commits February 17, 2019 17:24
…ve LF line endings and all files stay unix on commit
Add note about setting PYTHONIOENCODING=UTF-8 env var for running
examples
Example will `tee` stdout to `/tmp/samples` from conditional and
unconditional generation scripts.
added python download script and modified requirements to add the modules needed. Tested in Windows Version 10.0.17134 Build 17134  and Ubuntu 18.04.1 LTS
This write-up was loosely inspired in part by Mitchell et al.’s work on
[Model Cards for Model Reporting](https://arxiv.org/abs/1810.03993).
Adding such model usage sections could be good practice in general for
open source research projects with potentially broad applications.
This enables multi-GPU or distributed training using Horovod
Neil Shepperd and others added 29 commits March 20, 2019 10:46
Added the medium blog link "Beginner’s Guide to Retrain GPT-2 (117M) to Generate Custom Text Content"
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

10 participants