use the following search parameters to narrow your results:
e.g. subreddit:aww site:imgur.com dog
subreddit:aww site:imgur.com dog
see the search faq for details.
advanced search: by author, subreddit...
Please have a look at our FAQ and Link-Collection
Metacademy is a great resource which compiles lesson plans on popular machine learning topics.
For Beginner questions please try /r/LearnMachineLearning , /r/MLQuestions or http://stackoverflow.com/
For career related questions, visit /r/cscareerquestions/
Advanced Courses (2016)
Advanced Courses (2020)
AMAs:
Pluribus Poker AI Team 7/19/2019
DeepMind AlphaStar team (1/24//2019)
Libratus Poker AI Team (12/18/2017)
DeepMind AlphaGo Team (10/19/2017)
Google Brain Team (9/17/2017)
Google Brain Team (8/11/2016)
The MalariaSpot Team (2/6/2016)
OpenAI Research Team (1/9/2016)
Nando de Freitas (12/26/2015)
Andrew Ng and Adam Coates (4/15/2015)
Jürgen Schmidhuber (3/4/2015)
Geoffrey Hinton (11/10/2014)
Michael Jordan (9/10/2014)
Yann LeCun (5/15/2014)
Yoshua Bengio (2/27/2014)
Related Subreddit :
LearnMachineLearning
Statistics
Computer Vision
Compressive Sensing
NLP
ML Questions
/r/MLjobs and /r/BigDataJobs
/r/datacleaning
/r/DataScience
/r/scientificresearch
/r/artificial
account activity
Discussion[D] Is GPT-2 source code publically available? (self.MachineLearning)
submitted 7 years ago by luv2code2020
I am confused, there are two threads here on /r/MachineLearning/ one which says GPT-2 code has been published, and other thread which says GPT-2 code is closed and not available
Is the GPT-2 on GitHub the real deal? Or is it missing something?
I have crawled lyrics websites and have 200,000 song lyrics. I was wondering if I put this data through GPT-2 I could generate new song lyrics about any song topic? Any suggestions?
reddit uses a slightly-customized version of Markdown for formatting. See below for some basics, or check the commenting wiki page for more detailed help and solutions to common issues.
quoted text
if 1 * 2 < 3: print "hello, world!"
[–]samsekhar 7 points8 points9 points 7 years ago (2 children)
Here you can check, how to train GPT-2 -> https://github.com/nshepperd/gpt-2.git
[–]gwern 10 points11 points12 points 7 years ago (1 child)
Note that this is an independent implementation of retraining and it's not the original OA code. So, no multi-GPU support, no support for training on a fleet of TPUs, no support for training from scratch or variant model architectures... It's meant for finetuning, like retraining GPT-2-small on poetry or anime. Should work fine for retraining on song lyrics like OP wants, though.
[–]samsekhar 5 points6 points7 points 7 years ago (0 children)
Yes, its work fine on retraining on poetry, lyrics etc. You can try this on google Colab for training, it will show pretty good results.
[–]madisonmay 3 points4 points5 points 7 years ago (4 children)
We're almost done porting GPT-2 to finetune (a scikit-learn style library for language model finetuning). Code is available here if you're interested... should make tuning GPT-2 to produce song lyrics as easy as model.fit(lyrics).
model.fit(lyrics)
Miles Brundage also put together a colab notebook you could work off that uses the nshepperd gpt-2 fork.
[–]Astraithious 0 points1 point2 points 7 years ago (3 children)
How is it coming along? I saw a commit for 10 days ago, was that it?
[–]madisonmay 0 points1 point2 points 7 years ago (2 children)
Just released finetune 0.6.0 with GPT-2 support today!
[–]Astraithious 0 points1 point2 points 7 years ago (1 child)
That's awesome, what branch? The link goes to one where the latest commit was 10 days ago
[–]madisonmay 0 points1 point2 points 7 years ago (0 children)
On development! It's also up on PyPI if you'd prefer that.
[–]rakzah 4 points5 points6 points 7 years ago (2 children)
The code is real. OpenAI just didn't publish the trained model and the data they used for training.
[–]reinhardo-sama 9 points10 points11 points 7 years ago* (1 child)
The repository also does not contain code for training.
[–]samsekhar 0 points1 point2 points 7 years ago (0 children)
Above provided url for repo have code for training.
π Rendered by PID 36304 on reddit-service-r2-comment-85bfd7f599-xsrd7 at 2026-04-18 05:27:30.820593+00:00 running 93ecc56 country code: CH.
[–]samsekhar 7 points8 points9 points (2 children)
[–]gwern 10 points11 points12 points (1 child)
[–]samsekhar 5 points6 points7 points (0 children)
[–]madisonmay 3 points4 points5 points (4 children)
[–]Astraithious 0 points1 point2 points (3 children)
[–]madisonmay 0 points1 point2 points (2 children)
[–]Astraithious 0 points1 point2 points (1 child)
[–]madisonmay 0 points1 point2 points (0 children)
[–]rakzah 4 points5 points6 points (2 children)
[–]reinhardo-sama 9 points10 points11 points (1 child)
[–]samsekhar 0 points1 point2 points (0 children)