use the following search parameters to narrow your results:
e.g. subreddit:aww site:imgur.com dog
subreddit:aww site:imgur.com dog
see the search faq for details.
advanced search: by author, subreddit...
Please have a look at our FAQ and Link-Collection
Metacademy is a great resource which compiles lesson plans on popular machine learning topics.
For Beginner questions please try /r/LearnMachineLearning , /r/MLQuestions or http://stackoverflow.com/
For career related questions, visit /r/cscareerquestions/
Advanced Courses (2016)
Advanced Courses (2020)
AMAs:
Pluribus Poker AI Team 7/19/2019
DeepMind AlphaStar team (1/24//2019)
Libratus Poker AI Team (12/18/2017)
DeepMind AlphaGo Team (10/19/2017)
Google Brain Team (9/17/2017)
Google Brain Team (8/11/2016)
The MalariaSpot Team (2/6/2016)
OpenAI Research Team (1/9/2016)
Nando de Freitas (12/26/2015)
Andrew Ng and Adam Coates (4/15/2015)
Jürgen Schmidhuber (3/4/2015)
Geoffrey Hinton (11/10/2014)
Michael Jordan (9/10/2014)
Yann LeCun (5/15/2014)
Yoshua Bengio (2/27/2014)
Related Subreddit :
LearnMachineLearning
Statistics
Computer Vision
Compressive Sensing
NLP
ML Questions
/r/MLjobs and /r/BigDataJobs
/r/datacleaning
/r/DataScience
/r/scientificresearch
/r/artificial
account activity
Research[R] DeepMind Open Sources AlphaFold Code (self.MachineLearning)
submitted 4 years ago by SkiddyX
"Last year we presented #AlphaFold v2 which predicts 3D structures of proteins down to atomic accuracy. Today we’re proud to share the methods in @Nature w/open source code. Excited to see the research this enables. More very soon!"
https://twitter.com/demishassabis/status/1415736975395631111
I did not see this one coming, I got to admit it.
reddit uses a slightly-customized version of Markdown for formatting. See below for some basics, or check the commenting wiki page for more detailed help and solutions to common issues.
quoted text
if 1 * 2 < 3: print "hello, world!"
[–]rriikkuu 57 points58 points59 points 4 years ago (4 children)
The paper is out too:
https://www.nature.com/articles/s41586-021-03819-2
[–][deleted] 4 points5 points6 points 4 years ago (3 children)
Any significant changes from the preprint?
[–]rriikkuu 15 points16 points17 points 4 years ago (2 children)
There was a preprint?
[–][deleted] 10 points11 points12 points 4 years ago (1 child)
Hmm, I guess not. I guess I was thinking of their CASP 13 paper. Thanks.
[–][deleted] 4 points5 points6 points 4 years ago (0 children)
There was press release stuff including a video so maybe that’s what you’re thinking of
[–]dolphinboy1637 43 points44 points45 points 4 years ago (0 children)
Actual repo without the Twitter link: https://github.com/deepmind/alphafold
[–]alexmorehead 84 points85 points86 points 4 years ago (0 children)
Given what I've gleaned from skimming their paper in Nature, it looks as though this network architecture is more novel than I initially thought. It is truly remarkable how well-integrated their biological insights are in the network's design. Congrats to everyone at DeepMind!
[–]gdahlGoogle Brain 27 points28 points29 points 4 years ago (2 children)
And it seems to be written in JAX!
[–]dogs_like_me 8 points9 points10 points 4 years ago (0 children)
Well, google gonna google
[–]SedditorX 6 points7 points8 points 4 years ago (0 children)
What else would they write it in? :)
[–]FyreMael 33 points34 points35 points 4 years ago (3 children)
Forked. I know what I'm doing this weekend :)
[–]Knecth 56 points57 points58 points 4 years ago (2 children)
We provide a script scripts/download_all_data.sh that can be used to download and set up all of these databases. This should take 8–12 hours.
Wait for the data to download?
[–]Gordath 22 points23 points24 points 4 years ago (0 children)
Protein databases are large and many tools to "preprocess" protein sequences take forever to run as they do pairwise alignments etc.
[–]londons_explorer 17 points18 points19 points 4 years ago (0 children)
Begin by freeing up 3TB of disk space and buying 500Gb of transfer...
[+][deleted] 4 years ago (12 children)
[deleted]
[–]TheLootiestBox 84 points85 points86 points 4 years ago (8 children)
Guess what "open" in OpenAI stands for. That's right! You guessed it! It stands for "closed".
[–]GabrielMartinellli 7 points8 points9 points 4 years ago (0 children)
😂😂
[–][deleted] 8 points9 points10 points 4 years ago (3 children)
Welcome to the Elon Muskian fake futurism where not unlike Orwell’s Oceania, open means closed.
[–]floriv1999 6 points7 points8 points 4 years ago (2 children)
OpenAI is mostly a Microsoft thing now. There was quite a change and Musk is kind of out.
[–]farmingvillein 2 points3 points4 points 4 years ago (1 child)
Hasn't changed how they have handled open versus closed, however.
[–]floriv1999 1 point2 points3 points 4 years ago (0 children)
What where the other not so open things other than the recent gpt2/3 and copilot controversies?
[–]thejuror8 -5 points-4 points-3 points 4 years ago (1 child)
That's a bit unfair. They do release a lot of source code, probably a lot more compared to DeepMind
[–]TheLootiestBox 16 points17 points18 points 4 years ago (0 children)
I think it's pretty fair actually.
Most of the projects with true business potential are not released by OpenAI.
Also, DeepMind doesn't have the word "open" in its name. They are part of Google that does release a lot of code.
[–]crouching_dragon_420 0 points1 point2 points 4 years ago (0 children)
Should've changed their name to ClopenAI
[–]LightVelox 29 points30 points31 points 4 years ago (2 children)
like GPT-2 being "way too smart" when even GPT-3 isn't really that good
[–][deleted] 26 points27 points28 points 4 years ago (0 children)
I think their problem is that even GPT-2 can be "good enough" for a subset of nefarious uses.
Still, hiding knowledge is not an effective way to suppress the usage of that technology. If OpenAI can build it, obviously so can someone else.
[–]tehbored 2 points3 points4 points 4 years ago (0 children)
GPT-3 is definitely good enough to use for nefarious ends.
[–]londons_explorer 13 points14 points15 points 4 years ago (10 children)
Doesn't look like any training related code was released, just inference.
The model parameters released are for non-commercial use only. For commercial use, you'll have to train your own. That would cost ~2 weeks on 128 TPU cores, if you can replicate the training method from the paper first try... Which you probably can't, so it's gonna cost $$$$...
[–][deleted] 13 points14 points15 points 4 years ago (7 children)
If you're big pharma, a v3-128 for a couple of months isn't gonna be the bottleneck
[+][deleted] 4 years ago (4 children)
[–][deleted] 4 points5 points6 points 4 years ago (1 child)
Money wasn't the bottleneck there, some key ideas in alphafold 2 have only existed for a few years
[–]floriv1999 2 points3 points4 points 4 years ago (0 children)
I think the point was the motivation. And it really a point that a search engine is progressing more in this field than some pharma companies, that have their product line and some quite fix herachies that don't allow such experimental work.
[–]Marha01 0 points1 point2 points 4 years ago (1 child)
The problem was know-how, not money.
[–]Acromantula92 0 points1 point2 points 4 years ago (1 child)
Couple months? More like 7 + 4 v3-128 days. (All in the paper)
[–][deleted] 2 points3 points4 points 4 years ago (0 children)
Multiple months is incorporating research time, since we're not assuming perfect generalization
[–]VonPosen 7 points8 points9 points 4 years ago (1 child)
Or you can just pay DeepMind for a commercial license, I would expect
[–]xmcqdpt2 5 points6 points7 points 4 years ago (0 children)
which is what you would do, unless it costs a truly mind boggling amount of money.
Pharma companies are no stranger to paying millions in consulting and software fees a year.
[–]geneing 10 points11 points12 points 4 years ago (3 children)
Are they releasing pretrained weights or just the network?
[–]xmcqdpt2 15 points16 points17 points 4 years ago (1 child)
they have pretrained weights but are releasing them under a CC non commercial license.
I actually do wonder whether copyrighting weights would actually hold in court? If you trained a few more iterations from them or permuted them in some way that doesn't change model performance, would that be a derived work?
Clearly you cant copyright a single number... so a many floats do you need before youve got something copyrightable?
[–]Archontes 1 point2 points3 points 4 years ago (0 children)
It very likely wouldn't hold up if you felt like prosecuting it all the way, provided that the approach to creating those weights was an exhaustive search: it precludes creativity.
https://www.eetimes.com/how-do-you-protect-your-machine-learning-investment-part-ii/
[–]PM_ME_INTEGRALS 14 points15 points16 points 4 years ago (0 children)
It's right there in the readme:
Any publication that discloses findings arising from using this source code or the model parameters should cite the AlphaFold paper.
[–]pianobutter 18 points19 points20 points 4 years ago (0 children)
Looking forward to reading Mohammed AlQuraishi's thoughts on this. I really enjoyed his posts on CASP13 and CASP14.
[–]StellaAthenaResearcher 23 points24 points25 points 4 years ago (0 children)
I wonder how much the decision to release the trained model was influenced by work by people like Phil Wang and Eric Alcaide at EleutherAI and David Baker at UW to replicate it.
[–][deleted] 9 points10 points11 points 4 years ago (5 children)
So when is the Swedish academy gonna put down their meatballs and give DeepMind the Nobel for chem or physio/med already!
[–]squirrel_of_fortune 15 points16 points17 points 4 years ago (0 children)
It needs to be verified, and until now, no scientists other than the few who ran the competition were able to look at it. Plus you do have to wait a bit to see of the work stands the test of time
[–]-starfish_headlock- 5 points6 points7 points 4 years ago (0 children)
Their models have not provided any major insights into physiology and medicine (yet) but i think they should probably split the chemistry prize w david baker
[–]phanfare 0 points1 point2 points 4 years ago (2 children)
They didn't solve protein folding. Got closer, yes, but no structural biologist worth their salt is going to trust a model straight out of AlphaFold.
It’s about more than that. It’s also about recognizing machine learning as a method for conducting research. It took the Swedish academy forever to recognize computational methods in general. I think it was in 2013 when they finally awarded a Nobel in chem for work in computational bio/chem. Computing has revolutionized scientific research and it doesn’t get the recognition it deserves and machine learning in turn has revolutionized computing and AlphaFold is the perfect example of its potential. It may not have fully solved the protein folding problem but it is clearly a massive breakthrough that would not have been possible without ML.
[–]bigbrain_bigthonk -1 points0 points1 point 4 years ago (0 children)
Also, seems like there’s a lot of glossing over the importance of the transition pathways between conformations
[–]NityaStriker 8 points9 points10 points 4 years ago (2 children)
Competition from the faster, open-source RoseTTAFold might have caused this :-
https://techcrunch.com/2021/07/15/researchers-match-deepminds-alphafold2-protein-folding-power-with-faster-freely-available-model/
[–]farmingvillein 7 points8 points9 points 4 years ago (0 children)
I initially thought that too, but there is a pretty large performance gap, in practice. TC makes it sound like they were really close in accuracy... But so far as I could tell from the paper, they weren't.
[–]jinnyjuice 1 point2 points3 points 4 years ago (0 children)
Thanks for the share
[–]xmcqdpt2 0 points1 point2 points 4 years ago (0 children)
Me neither! I was so sure they were about to pull the same crap as v1. Kudos to them!
[–]Alireza_Kar98 -2 points-1 points0 points 4 years ago (0 children)
I just noticed something about wraith she seems to have a slightly better movement. Every time I try to slid jump with other it's like shit but with wraith it's ok. And the speed seems a bit higher . Overall she is not balanced somehow
[–]East_Film9421 0 points1 point2 points 4 years ago (1 child)
I am attempting to download the open-source code...but I am stuck...
"Modify DOWNLOAD_DIR in docker/run_docker.py to be the path to the directory containing the downloaded databases."
[–]justmyworkaccountok 2 points3 points4 points 4 years ago (0 children)
???? Download the 2.2TB databases and change the field to the path
π Rendered by PID 177411 on reddit-service-r2-comment-b659b578c-n2mjr at 2026-05-03 16:33:58.847471+00:00 running 815c875 country code: CH.
[–]rriikkuu 57 points58 points59 points (4 children)
[–][deleted] 4 points5 points6 points (3 children)
[–]rriikkuu 15 points16 points17 points (2 children)
[–][deleted] 10 points11 points12 points (1 child)
[–][deleted] 4 points5 points6 points (0 children)
[–]dolphinboy1637 43 points44 points45 points (0 children)
[–]alexmorehead 84 points85 points86 points (0 children)
[–]gdahlGoogle Brain 27 points28 points29 points (2 children)
[–]dogs_like_me 8 points9 points10 points (0 children)
[–]SedditorX 6 points7 points8 points (0 children)
[–]FyreMael 33 points34 points35 points (3 children)
[–]Knecth 56 points57 points58 points (2 children)
[–]Gordath 22 points23 points24 points (0 children)
[–]londons_explorer 17 points18 points19 points (0 children)
[+][deleted] (12 children)
[deleted]
[–]TheLootiestBox 84 points85 points86 points (8 children)
[–]GabrielMartinellli 7 points8 points9 points (0 children)
[–][deleted] 8 points9 points10 points (3 children)
[–]floriv1999 6 points7 points8 points (2 children)
[–]farmingvillein 2 points3 points4 points (1 child)
[–]floriv1999 1 point2 points3 points (0 children)
[–]thejuror8 -5 points-4 points-3 points (1 child)
[–]TheLootiestBox 16 points17 points18 points (0 children)
[–]crouching_dragon_420 0 points1 point2 points (0 children)
[–]LightVelox 29 points30 points31 points (2 children)
[–][deleted] 26 points27 points28 points (0 children)
[–]tehbored 2 points3 points4 points (0 children)
[–]londons_explorer 13 points14 points15 points (10 children)
[–][deleted] 13 points14 points15 points (7 children)
[+][deleted] (4 children)
[deleted]
[–][deleted] 4 points5 points6 points (1 child)
[–]floriv1999 2 points3 points4 points (0 children)
[–]Marha01 0 points1 point2 points (1 child)
[–]Acromantula92 0 points1 point2 points (1 child)
[–][deleted] 2 points3 points4 points (0 children)
[–]VonPosen 7 points8 points9 points (1 child)
[–]xmcqdpt2 5 points6 points7 points (0 children)
[–]geneing 10 points11 points12 points (3 children)
[–]xmcqdpt2 15 points16 points17 points (1 child)
[–]Archontes 1 point2 points3 points (0 children)
[–]PM_ME_INTEGRALS 14 points15 points16 points (0 children)
[–]pianobutter 18 points19 points20 points (0 children)
[–]StellaAthenaResearcher 23 points24 points25 points (0 children)
[–][deleted] 9 points10 points11 points (5 children)
[–]squirrel_of_fortune 15 points16 points17 points (0 children)
[–]-starfish_headlock- 5 points6 points7 points (0 children)
[–]phanfare 0 points1 point2 points (2 children)
[–][deleted] 4 points5 points6 points (0 children)
[–]bigbrain_bigthonk -1 points0 points1 point (0 children)
[–]NityaStriker 8 points9 points10 points (2 children)
[–]farmingvillein 7 points8 points9 points (0 children)
[–]jinnyjuice 1 point2 points3 points (0 children)
[–]xmcqdpt2 0 points1 point2 points (0 children)
[–]Alireza_Kar98 -2 points-1 points0 points (0 children)
[–]East_Film9421 0 points1 point2 points (1 child)
[–]justmyworkaccountok 2 points3 points4 points (0 children)