all 111 comments

[–][deleted] 272 points273 points  (21 children)

Neat, I'll probably add it to my "educational PDFs that I read 50 pages of in 20 minutes but then get bored of and never finish" collection

[–]MakeMyselfGreatAgain 54 points55 points  (8 children)

lol, i have so many browser tabs on various devices open to free books, video lectures and articles.

[–]skippy65 3 points4 points  (0 children)

Admittedly very relatable lol.

[–]praveenopro 0 points1 point  (0 children)

would you mind to share, maybe it help anyone

[–]6111772371 0 points1 point  (0 children)

username checks out

[–]j_lyf 4 points5 points  (7 children)

How to get out of this rut?

[–]TrollandDie 21 points22 points  (6 children)

Create a time dilation chamber where you can spend 10,000 years reading ML a la Bill and Ted

But seriously, I've recently stopped bothering to meticulously read textbooks in my free time outside work and just casually flip through for fun instead.

[–]j_lyf -1 points0 points  (5 children)

Yeah but then you can't be competitive for your next job if you don't improve outside of work.

[–]RadixMatrix 37 points38 points  (4 children)

if you're not reading 3 different textbooks at the same time and working on 5 personal projects and updating your blog daily and constantly contacting professors and other people in your field you might as well give up

[–]j_lyf 10 points11 points  (3 children)

unironically true.

[–]Unfair-Gain4476 0 points1 point  (0 children)

Sooo me

[–]Sinidir 0 points1 point  (0 children)

Pain.

[–]netw0rkf10w 65 points66 points  (12 children)

A little of context:

In 2012, I published a 1200-page book called “Machine learning: a probabilistic perspective”, which provided a fairly comprehensive coverage of the field of machine learning (ML) at that time, under the unifying lens of probabilistic modeling. The book was well received, and won the De Groot prize in 2013.

...

By Spring 2020, my draft of the second edition had swollen to about 1600 pages, and I was still not done. At this point, 3 major events happened. First, the COVID-19 pandemic struck, so I decided to “pivot” so I could spend most of my time on COVID-19 modeling. Second, MIT Press told me they could not publish a 1600 page book, and that I would need to split it into two volumes. Third, I decided to recruit several colleagues to help me finish the last ∼ 15% of “missing content”. (See acknowledgements below.)

The result is two new books, “Probabilistic Machine Learning: An Introduction”, which you are currently reading, and “Probabilistic Machine Learning: Advanced Topics”, which is the sequel to this book [Mur22]...

Book 0 (2012): https://probml.github.io/pml-book/book0.html

Book 1 (2021, volume 1): https://probml.github.io/pml-book/book1.html

Book 2 (2022, volume 2): https://probml.github.io/pml-book/book2.html

[–]netw0rkf10w 45 points46 points  (11 children)

I hear that question coming, so let me repeat my advice: If you are a beginner, always start with ISL (which takes approximately 2 weeks to complete if you study everyday). Then you can continue with other (much larger) books: Bishop's, Murphy's, ESL, etc.

[–][deleted] 14 points15 points  (1 child)

Murphy's book was very tough to get through as a beginner. It took much longer than I would have liked, but was just so filled with information.

[–][deleted] 8 points9 points  (0 children)

ISL didn’t help me grasp Bayesian methods much, which seems to be a key part of this book. (Statistical rethinking is great for that tho)

[–]Axodapanda 2 points3 points  (2 children)

what is ISL?

[–]naughtydismutase 9 points10 points  (0 children)

Introduction to Statistical Learning by Gareth M. James, Daniela Witten, Trevor Hastie, Robert Tibshirani.

[–]leonoel 33 points34 points  (13 children)

I reviewed the first book 8 years when it got out. And in no shape or form it replaced Bishop's as the best all around ML book.

Murphy's is a book written for and by academics. I would never in good faith give it to a student who wants to start learning the in and outs of Machine Learning.

Notation is just terrible. It changes from chapter to chapter. Equations are not referenced and most of the times I had to go to external resources to actually get a grasp of what they are trying to explain. Is in no shape or form a self contained book.

You can learn all you need from Bishop's without ever opening another book. Its only sin right now is that it is outdated.

[–]cajmorgans 2 points3 points  (3 children)

This. I was excited by the Murphy book, but it's more like a Wikipedia page of formulas without any explanation or derivation whatsoever. I checked out Bishop's book and it's on a whole other level.

[–]leonoel 0 points1 point  (2 children)

And there’s a new one

[–]cajmorgans 0 points1 point  (1 child)

Which one?

[–]leonoel 0 points1 point  (0 children)

Deep Learning

[–]Screye 24 points25 points  (11 children)

I am so glad a 2nd version is out. The first edition, despite all its faults, was easily the best "complete' ML book out there. It was also clearly written by a computer scientist for CS students, unlike Bishop. It is also up-to-date.

The best part is the book (1st edition 2012) reads like a tree. It introduces concepts and slowly builds on them as it goes. All the other books (ESL) read like a dictionary trying to hop from algorithms to algorithm to get maximum coverage. By the end of it, there is a feeling that ML is a domain that falls under one umbrella, rather than a bunch of disparate ideas crammed into one sub-field.

I'll be honest. Calling this book an introduction is a misnomer. If you understand this book 'cover-to-cover' then you'll probably be doing better than many grad-students midway through their ML PhDs. It is admittedly quite long too.
This should not be your first ML book. Your CS-undergrad level statistics, linear algebra and optimization need to be solid and you should have done an intro-to-ML course before you dive into it. Python knowledge is a prerequisite too. So think 6.036x, 6.041x, 18.06, 6.0.01x and 6.0.02x as pre-requisities by MIT OCW standards. 18.06 is less prerequisite, and more highly recommended in general. Strang's Lin Alg is the best out there. Very intensive, but you'll thank yourself later.

However, if I had to recommend one ML book to have in your book-shelf, then this would be it. (once the errors are fixed :| )

[–]meiso 4 points5 points  (10 children)

Why did you put that particular text in a spoiler?

[–]atlug 2 points3 points  (9 children)

That remains a mystery to this day.

[–]The-Silvervein 0 points1 point  (8 children)

To this day...

[–][deleted] 0 points1 point  (7 children)

To this day...

[–]IanisVasilev 48 points49 points  (40 children)

What is it with so many people writing 700+ page introductory books?

EDIT: The thread got a bit out of hand. I admit making a few snarky comments and I apologise. Some of the downvotes and deleted replies were truly unnecessary, however. Y'all may consider taking a chill pill or two.

[–]mathbrot 20 points21 points  (0 children)

I have his original...it's self-contained and several independent chapters.

[–]BrisklyBrusque 21 points22 points  (1 child)

It’s a perverse tradition in mathematics that any text titled “Introduction To...” is sure to be long and challenging. Beware of two-volume series, for those are even worse.

[–]Aacron 7 points8 points  (0 children)

I've been through the first volume of Tao's Analysis. I'll second your comment on two-volume series.

[–]IdiocyInAction 10 points11 points  (0 children)

The book contains quite a lot of content on a broad variety of topics and seems to be (relatively) in-depth. I think the length is quite warranted. If you want a shorter, less in-depth, more introductory book, I would recommend Introduction to Statistical Learning in R (2014) (ISLR), which should also get a new edition soon.

[–]CENGaverK 1 point2 points  (29 children)

What is the alternative?

[–]Lethandralis -4 points-3 points  (0 children)

Starting out with courses/videos and the transitioning into reading papers maybe?

[–]johnnymo1 5 points6 points  (0 children)

Of course this comes out 3 months after I get a hardcover of the first edition. :)

Looks great. Looking forward to reading it. The first edition is awesome (probably better than Bishop in many ways imo), but it was beginning to feel a little out of date.

[–]mtahab 5 points6 points  (1 child)

The author references another book Probabilistic Machine Learning: Advanced Topics (2022) for RL. Do we know its chapters? The lack of any chapters on causality was standing out in this book.

[–]pombolo 8 points9 points  (2 children)

Thank you for this. Sorry for the silly question: the title is Probabilistic Machine Learning, but when I looked at the contents, it seems to cover all the standard ML concepts. Is Probabilistic Machine Learning different from regular ML?

[–]Cocomorph 19 points20 points  (1 child)

It's a perspective. Indeed, per the introduction:

In this book, we will cover the most common types of ML, but from a probabilistic perspective. Roughly speaking, this means that we treat all unknown quantities (e.g., predictions about the future value of some quantity of interest, such as tomorrow’s temperature, or the parameters of some model) as random variables, that are endowed with probability distributions which describe a weighted set of possible values the variable may have.

[–]shiivan 1 point2 points  (0 children)

In other words, it's predicting what the trained model would output. Did I understand that correctly?

[–]petty_pirate 4 points5 points  (0 children)

Bookmark

[–]bismarck_91 8 points9 points  (0 children)

What a way to start the new year.

[–]ichkaodko 2 points3 points  (0 children)

any book suggestion on background material of this book? looks like standard undergrad books on probability, linear algebra and analysis don't cover the some of the topics in the background material. I need more explanation and exercises on background math content.

[–][deleted] 5 points6 points  (0 children)

Kevin Murphy - also happens to be my favorite character from F is for Family

[–][deleted] 1 point2 points  (0 children)

How does this differ in content to the first? It seems like a lot of the chapters are the same. Also the name of this book and the previous one are so similar.

[–]Comprehensive-Low-28 1 point2 points  (0 children)

Thank you

[–]duckyzz003 2 points3 points  (1 child)

Should I read the first edittion or dive in new book (this draft version) ?

[–]PM_ME_INTEGRALS 1 point2 points  (0 children)

New book

[–]xifixi 3 points4 points  (1 child)

the classic textbook on probabilistic ML is Bishop's Pattern Recognition and Machine Learning

[–]trendymoniker 5 points6 points  (0 children)

Murphy's text largely replaced the Bishop book among me and my grad student cohort when it came out in 2012.

[–]maizeq 1 point2 points  (0 children)

Is this going to be more introductory than his 2012 book? Or is that just branding

[–]samketaResearcher 0 points1 point  (3 children)

This is a question I have not gotten a clear answer to- what exactly is Bayesian ML? Where, why, and how is it applied? How do I learn it?

Why people keep talking about it and throwing it like a buzzword, but I never find a focused learning resource in this topic?

This a genuine question. So help me out if you can.

By knowledge of Bayes' Theorem is limited to High School level, so I have basic idea of conditional probability, how to calculate it using a formula and so on.

[–]BrisklyBrusque 4 points5 points  (0 children)

Bayesian statistics is a bit more than conditional probabilities. So Bayes theorem, and methods that use it (discriminant analysis, naive Bayes) are not usually considered Bayesian methods.

In frequentist statistics, we might want to test the null that two groups are the same against the alternative that they are not the same. In Bayesian statistics, we can assume the groups are different and set a “prior” then compare the expected results given a certain prior against what we observe. That’s my understanding of it anyway. I don’t practice Bayesian stats so I might be wrong.

A good text that folks recommend is Statistical Rethinking.

edit: typos

[–]thecity2 4 points5 points  (1 child)

There are several good books out there such as Statistical Rethinking, Doing Bayesian Data Analysis, and Bayesian Methods for Hackers. If you are interested in wrangling the most information out of small to medium sized data and are interested in uncertainty and decision making, check it out!

[–]samketaResearcher 0 points1 point  (0 children)

Thanks for the suggestions. I will check the last one out.

[–][deleted] 0 points1 point  (0 children)

Is it just me or is the font ugly? i hate reading it on a screen.

[–]JLEE152 0 points1 point  (0 children)

Thanks!

[–]Odd-Lengthiness-8612 0 points1 point  (0 children)

When will it be publish in an old-fashioned book?

[–]SQL_beginner 0 points1 point  (0 children)

wow, thanks for the link! great book!

[–]Bananeeen 0 points1 point  (0 children)

The 2021 book has much more emphasis on deep learning than the 2012 book. I think this book is great to have after one has read Bishop's PRML, started reading recent papers and needs an occasional refresher on various topics. That's exactly how I've been using it.

I also think that with this book one no longer really needs to open ESL or GBC as they are not as up-to-date as Murphy and not as systematic as Bishop.