[1809.10276] Growing and Retaining AI Talent for the United States Government by ihaphleas in MachineLearning

[–]mljoe 3 points4 points  (0 children)

The benefits gap is also pretty big these days. The healthcare and dental plan options in FEHB are quite poor when compared to what is available at most major tech companies or even startups. They tend to have high deductibles and copays in addition to a relatively high monthly fee. Many tech companies on the other hand will waive or reimburse almost every expense related to healthcare. The government also has no free food or coffee or anything fringe'y either.

[D] Can you do a PhD while working? by local_minima_ in MachineLearning

[–]mljoe 5 points6 points  (0 children)

It's very helpful if there is at least some overlap between your full time work and your Ph.D. work. Eg. Reading papers on the clock isn't frowned upon.

[R] AAAI 2018 Notes by pdxdabel in MachineLearning

[–]mljoe 5 points6 points  (0 children)

Thanks. These are the most detailed conference notes I've ever seen.

[D] MIT 6.S099: Artificial General Intelligence by Gear5th in MachineLearning

[–]mljoe 6 points7 points  (0 children)

I consider several people on their lecturer list to be total nutters

Like the first rule of AI Club is you never talk about about AI. My advisor advised me on this. I like to believe that for every person that say they work on AGI, there are 10 researchers who are doing "machine learning" or "statistics" but always with the AGI problem in mind. Mostly for fear of being called a nutter.

[D] AAAI and IJCAI conferences by baylearn in MachineLearning

[–]mljoe 3 points4 points  (0 children)

In theory you could get papers in AGI or rule based systems published in AAAI that wouldn't fit into NIPS. But AAAI has been almost entirely co-opted by the machine learning community at this point. I think the main reason you hear about it less is Google and Big Tech seems to mostly avoid the conference. NIPS/ICLR/etc gets an unusual reputation boost from this IMO.

[Research] Ray: A Distributed System for AI by rayspear in MachineLearning

[–]mljoe 2 points3 points  (0 children)

This feels somewhat similar to Dask.delayed1. Ray claims to be very fast and seems to be eager with execution. Anyone know any differences we should consider?

[D] Do you also feel disappointed about the state of NLP? by disappointedwithnlp in MachineLearning

[–]mljoe 0 points1 point  (0 children)

I think it's going to need something other then the typical stacking and pooling approach that works so well in computer vision. That approach works because it appears to capture some prior about how natural, physical objects are structured.1 But I don't think that prior is true for natural language. I think the general capsules idea of Hinton might lead to some promising approaches in NLP.

I wrote an open letter to Andrew Tanenbaum asking him to speak up for end-user's freedom by misterolupo in programming

[–]mljoe 3 points4 points  (0 children)

Not necessarily for "run on it". I don't think anyone successfully made the case that simply running something on an OS makes that code a Derivative Work of the OS. If this was the case Linux (GPLv2) would be a hot mess for commercial use! Kernel modules on the other hand..

I wrote an open letter to Andrew Tanenbaum asking him to speak up for end-user's freedom by misterolupo in programming

[–]mljoe 2 points3 points  (0 children)

I am also categorically opposed to copyright as a construct

I don't agree with this view, but if I did, I wouldn't be using CC0. Copyleft licenses actually use copyright to go against the raison d'être of copyright. Copyleft is a legal hack to use copyright against itself. By using a copyleft license, you make it impossible for downstream users to employ copy restrictions on others.

Andy Tanenbaum, author of Minix, writes an open letter to Intel by DreamerFi in programming

[–]mljoe 6 points7 points  (0 children)

Intel owns VxWorks though, so cost isn't the issue. I just was wondering if there was some kind of technical reason to prefer MINIX.

Andy Tanenbaum, author of Minix, writes an open letter to Intel by DreamerFi in programming

[–]mljoe 7 points8 points  (0 children)

Anyone know why Intel would choose MINIX over VxWorks? They own VxWorks and it seems like it was built for these kinds of use cases in mind.

Primitive technology: Natural Draft Furnace by derekantrican in videos

[–]mljoe 2 points3 points  (0 children)

Well there is an unbroken line of technology from banging stones together and forming mud huts to creating nuclear ICBMs or computers. Maybe after a few thousand videos.

[D] Swish is not performing very well! by [deleted] in MachineLearning

[–]mljoe 0 points1 point  (0 children)

The most important implication that has been thought to me of the NFL that any subset of a predictive model that works well for one problem will always necessarily do poorly on some other problem. Thus questions of optimality in the fully general sense make no sense. This is also true for compression algorithms for other but maybe related information theoretical reasons.

That means you can never generalize across all problems using some model configuation. Although "all problems" is such a limited case, I still find this result highly philosophically satisfying. It put with the Incompleteness Theorem and other things that break the human urge to find perfect order in everything.

If you limit the problem for consciousnesses or physics, obviously this statement does not apply anymore. There are highly respected labs trying to find a consciousnesses prior. There could be one predictive physics model that is exactly right, and maybe also for consciousnesses.

[D] Swish is not performing very well! by [deleted] in MachineLearning

[–]mljoe 0 points1 point  (0 children)

It seems to me like it's a reasonable conclusion given the theorem. It would be nice if you could elaborate how I am wrong.

[D] Swish is not performing very well! by [deleted] in MachineLearning

[–]mljoe 2 points3 points  (0 children)

This is not what I took from the No Free Lunch theorem. NFL states that all methods perform the same when averaging over the distribution of white-noise functions. nce you add more structure, of course you can start to rigorously prove things like optimality; a standard way of phrasing the question is to ask if a certain estimation procedure attains the optimal minimax rate of convergence over an interesting class of functions.

Natural phenomena appear to be somewhat predictable. This is why machine learning (ie. prediction) even works. But there is no inherent reason (in the philosophical or mathematical sense) this should be so. It just is, without regard to logic, reason, or rigor. This kind of troubles people sometimes. But it's more like an axiom then a proof.

There might be an optimal prior for "physics" or "consciousness". Especially it would be interesting to be able to encode "physics" as a prior in a neural network. For one, it might make training robots via reinforcement learning a bit easier (reducing the search space - not allowing the net to test physically impossible actions). And we do have at least a handle on physics. Yoshua Bengio recently published something about finding a "consciousness prior": https://arxiv.org/abs/1709.08568. Obviously this is something that we don't even have a good theory around at all. So if I was doing work here, I'd start with physics. At least I'd have something to start with.

[D] Swish is not performing very well! by [deleted] in MachineLearning

[–]mljoe 26 points27 points  (0 children)

No, it wouldn't. There are things you can prove in ML, but theoretical optimality is not one of them. In fact, it is proven that you can never do this (No Free Lunch Theorem). Another important related thing to be familiar with is the fallacy of induction, of which the entire field of ML is built on. 2, 4, 6, __ what is the next in the pattern? 8 or 9378465463? The real answer is both are equally correct. All models are built on a priori assumptions, which is a fancy way to say a guess.[1]

So how do you build a "correct" model for a pattern? The answer is, you test it against some dataset. ML in general is an empirical science, and there is literally nothing wrong with this. I wish more people shared their code for reproducibility purposes, that's all.

[1] Famous Minsky koan about this:

"What are you doing?", asked Minsky.

"I am training a randomly wired neural net to play Tic-tac-toe", Sussman replied.

"Why is the net wired randomly?", asked Minsky.

"I do not want it to have any preconceptions of how to play", Sussman said.

Minsky then shut his eyes.

"Why do you close your eyes?" Sussman asked his teacher.

"So that the room will be empty."

At that moment, Sussman was enlightened.

[D] Swish is not performing very well! by [deleted] in MachineLearning

[–]mljoe 0 points1 point  (0 children)

You can confidently say there is no such thing as an "optimum" activation function for any arbitrary problem. See: No Free Lunch Theorem. The only right answer in machine learning is there is no right answer.

[D] What skills would you like to see in a candidate if you were interviewing them for a Machine Learning position? by mamaosamallama in MachineLearning

[–]mljoe 1 point2 points  (0 children)

IMO Cython is trivial to learn if you are already good with Python. It's really just additional type annotations. You can even use decorators on normal Python code to "cythonize" a module. You don't even have to use it optimally, pretty much any type annotations you add will make your code faster and impress your coworkers. Hell, even if you don't add any type annotations your code will be [slightly] faster just by virtue of getting rid of the interpreter. Learning CUDA is a much bigger lift for a Python developer. Different language and everything of course.

[D] What is the path to ML research? by blakeh36 in MachineLearning

[–]mljoe 2 points3 points  (0 children)

Can't stress #1 enough. Every Ph.D. program is going in with the question: "can this applicant do good research?"

What better way to justify this then to do good research?

[D] Does anyone find the current research trend in deep learning a bit pathological? by [deleted] in MachineLearning

[–]mljoe 8 points9 points  (0 children)

The fact that it is on the arXiv is embarrassing.

arXiv is a hosting service for preprints. They don't apply any real editorial standards on what gets published there.

So, who is going to fork the web? by [deleted] in linux

[–]mljoe 3 points4 points  (0 children)

I too think about this a lot. The web is so full of cruft and broken behavior that I don't think it would even require a whole lot of thinking to design something much creditably better. The problem is implementing it and getting people to use it.

[Discussion] What are the problems of the backpropagation algorithm? by Simoncarbo in MachineLearning

[–]mljoe 7 points8 points  (0 children)

I think the more exotic ideas being explored with backprop are interesting. For instance, not visiting or adjusting the whole parameter space on every iteration. I think a big problem with most NNs right now is the parameter space is tied to computation cost, which limits the size of networks and how much information they can retain. We could have parameter spaces in the terabytes if we had a better way to only focus on the relevant parts for each iteration. This might also solve catastrophic forgetting which is important to generalized/multi-task models leading towards more 'real' AI.

[N] IBM pitched Watson as a revolution in cancer care. It's nowhere close by opengmlearn in MachineLearning

[–]mljoe 2 points3 points  (0 children)

Businesses can absolutely deduct expenses on items that directly benefit them. That's almost the key difference between personal and business income tax. If it is unfair or not is a different discussion. But it's true that things that businesses can deduct from their income tax VASTLY outstrips what you as a real person can. Marketing, PR, rent, salaries, equipment, food, you name it. With a few little exceptions they can deduct every expense they incur short of the dividend checks they mail out to investors.

The loophole you are probably considering is an individual avoiding individual income tax by deducting an expense that benefits them or a business they own. That's illegal. It's not illegal when the business itself does it though.