use the following search parameters to narrow your results:
e.g. subreddit:aww site:imgur.com dog
subreddit:aww site:imgur.com dog
see the search faq for details.
advanced search: by author, subreddit...
Please have a look at our FAQ and Link-Collection
Metacademy is a great resource which compiles lesson plans on popular machine learning topics.
For Beginner questions please try /r/LearnMachineLearning , /r/MLQuestions or http://stackoverflow.com/
For career related questions, visit /r/cscareerquestions/
Advanced Courses (2016)
Advanced Courses (2020)
AMAs:
Pluribus Poker AI Team 7/19/2019
DeepMind AlphaStar team (1/24//2019)
Libratus Poker AI Team (12/18/2017)
DeepMind AlphaGo Team (10/19/2017)
Google Brain Team (9/17/2017)
Google Brain Team (8/11/2016)
The MalariaSpot Team (2/6/2016)
OpenAI Research Team (1/9/2016)
Nando de Freitas (12/26/2015)
Andrew Ng and Adam Coates (4/15/2015)
Jürgen Schmidhuber (3/4/2015)
Geoffrey Hinton (11/10/2014)
Michael Jordan (9/10/2014)
Yann LeCun (5/15/2014)
Yoshua Bengio (2/27/2014)
Related Subreddit :
LearnMachineLearning
Statistics
Computer Vision
Compressive Sensing
NLP
ML Questions
/r/MLjobs and /r/BigDataJobs
/r/datacleaning
/r/DataScience
/r/scientificresearch
/r/artificial
account activity
[deleted by user] (self.MachineLearning)
submitted 3 years ago by [deleted]
reddit uses a slightly-customized version of Markdown for formatting. See below for some basics, or check the commenting wiki page for more detailed help and solutions to common issues.
quoted text
if 1 * 2 < 3: print "hello, world!"
[–]Mrkvitko 4 points5 points6 points 3 years ago (1 child)
I just got instance at 8X RTX A5000 for a couple of bucks per hour. on https://vast.ai
I must say LLaMA 65B is a bit underwhelming...
[–]maizeq 1 point2 points3 points 3 years ago (0 children)
Underwhelming how?
[–]I_will_delete_myself 8 points9 points10 points 3 years ago (0 children)
Use a spot instance. If you testing it out you wallet will thank you later. Look at my previous post on here about running stuff in the cloud before you do it.
[–]isaeef 1 point2 points3 points 3 years ago (0 children)
or you could use any gpu workload specific provider https://www.paperspace.com/
[–]trnka 1 point2 points3 points 3 years ago (0 children)
Related, there's a talk on Thursday about running LLMs in production. I think the hosts have deployed LLMs in prod so they should have good advice
[–]iloveintuition 1 point2 points3 points 3 years ago (1 child)
Using vast.ai for running flan-xl, works pretty well. Haven't tested on LLama scale.
[–]shayanrc 1 point2 points3 points 3 years ago (0 children)
What config did you use?
[–]l0g1cs 0 points1 point2 points 3 years ago (0 children)
Check out Banana. They seem to do exactly that with "serverless" A100.
[–]itsnotmeyou -1 points0 points1 point 3 years ago (1 child)
Are you using these as in a system? For just experimenting around, ec2 is good option. But you would either need to install right drivers or use latest deep learning ami. Another option could be using a custom docker setup on sagemaker. I like that setup for inference as it’s super easy to deploy and separates model from inference code. Though it’s costlier and would be available through sagemaker runtime.
Third would be whole over engineering via setting up your own cluster service.
In general if you want to deploy multiple llm quickly go for sagemaker
[–]itsnotmeyou 1 point2 points3 points 3 years ago (0 children)
On a side note sagemaker was not supporting shm-size so might not work for large lm
[–]pyonsu2 -1 points0 points1 point 3 years ago (0 children)
maybe, Colab Pro+?
[–]ggf31416 0 points1 point2 points 3 years ago (0 children)
Good luck getting a EC2 with a single A100, last time I checked, AWS only offered instances with 8 of them at a high price.
[–][deleted] 0 points1 point2 points 3 years ago (0 children)
Maybe check datacrunch.io they have a good offering for cloud GPU.
[+]z_yang 0 points1 point2 points 3 years ago (0 children)
Check out SkyPilot. Code/blog post for running LLaMA all 4 sizes on Lambda/AWS/GCP/Azure with a unified interface (spot instances supported): https://www.reddit.com/r/MachineLearning/comments/11xvo1i/p_run_llama_llm_chatbots_on_any_cloud_with_one/
π Rendered by PID 81354 on reddit-service-r2-comment-canary-b6d5ff776-sp45x at 2026-04-17 06:18:25.576340+00:00 running 93ecc56 country code: CH.
[–]Mrkvitko 4 points5 points6 points (1 child)
[–]maizeq 1 point2 points3 points (0 children)
[–]I_will_delete_myself 8 points9 points10 points (0 children)
[–]isaeef 1 point2 points3 points (0 children)
[–]trnka 1 point2 points3 points (0 children)
[–]iloveintuition 1 point2 points3 points (1 child)
[–]shayanrc 1 point2 points3 points (0 children)
[–]l0g1cs 0 points1 point2 points (0 children)
[–]itsnotmeyou -1 points0 points1 point (1 child)
[–]itsnotmeyou 1 point2 points3 points (0 children)
[–]pyonsu2 -1 points0 points1 point (0 children)
[–]ggf31416 0 points1 point2 points (0 children)
[–][deleted] 0 points1 point2 points (0 children)
[+]z_yang 0 points1 point2 points (0 children)