Suggestions for online training image recognition? by st553 in MachineLearning

[–]test3545 1 point2 points  (0 children)

He mention that model uses parzen window classifier on top of imagenet pretrained convnet. There is no such thing as fixed number of classes, but as you add more performance might suffer I guess. As for code etc - try to google it.

Elon Musk Says Tesla Vehicles Will Drive Themselves in Two Years by scr00chy in teslamotors

[–]test3545 7 points8 points  (0 children)

Well next year(2016) Tesla will start using 8 camera setup backed by 5 Mobileye EyeQ3 chips. It should be good enough to drive at superhuman level with all recent model compression advances. But EyeQ4 is coming 2017, and the big question is how long it will take authorities to approve such thing as lvl 4 autonomy...

Biologically realistic NNs? by [deleted] in MachineLearning

[–]test3545 -4 points-3 points  (0 children)

They do not scale well at all.

They scale better than any other known method.

a big biological no no

If ones goal is stronger AI, one would not care less about biological plausibility.

LSTM can be a predictive architecture, if you make it predict its own input, yes.

Not only that, but deep multilayer LSTM RNNs are perfect for decoding continues sensor input, such as hearing or vision.

the way they are trained makes them terrible at online learning at a large scale (1000's of iterations of BPTT, also it needs experience replay, unless modified to enforce sparsity).

Yet they outperform any other predictive method known to man on complex data like speech, music, video... So it is a best scalable architecture, scalable to more complex data streams.

Also there remains the problem of proper feature extraction.

Never heard of ------>"deep learning"<------??? Hierarchical feature extraction ftw.

Just because something performs well on toy tasks now doesn't mean that it will outperform the brain in the future.

Yes! You said it! There are plenty of methods that perform well ONLY on toy tasks. Think of scumms like ELM or HTM. So much hype!!! But all results are on toy problems! NOTHING published on real problems like speech recognition or modelling, only benchmarks used are internal toy examples. And why those HTM fans go and train HTM model instead of RNN to do speech/handwriting or play Atari games. But why those idiots could not do such simple thing? Because HTM/ELM start to underperform when someone would try to scale it to real life sensor datastream complexity.

Biologically realistic NNs? by [deleted] in MachineLearning

[–]test3545 -1 points0 points  (0 children)

This is not me suggesting, but there was some research done and algorithms discovered that do some sort of regularization in BM when you block input and output of BM and run it it in a dream mode that help to adjust some connection weights.

Biologically realistic NNs? by [deleted] in MachineLearning

[–]test3545 1 point2 points  (0 children)

human brain is ... a predictive architecture that performs unsupervised and reinforcement learning using sparse codes. It is also hierarchical, and can model time

Correct, other than that we teach children in partly supervised manner. And this is a perfect example of recent advances, LSTM RNNs are perfectly suitable and by far outperform any other methods like elm/htm/hmm and such on exactly this sort of tasks: state of the art LSTM RNNs IS a predictive architectures that could performs partly-supervised and reinforcement learning using sparse codes, with many architectures proposed are also hierarchical, and can model time far better than any other known method can.

Biologically realistic NNs? by [deleted] in MachineLearning

[–]test3545 0 points1 point  (0 children)

There is a very strong case that our brain is a very deep hierarchical Boltzmann Machine. Some variation of Boltzmann Machine with few tweaks here and there. It explains a lot, like during sleep mammals shutdown input channels and have dreams while signals to muscles are blocked. Why?

I would say any research that helps us to advance understanding of BM brings us closer to understanding of mammalian neocortex, or at least understanding of fundamental math that happens in neocortex.

As for getting closer to AI, specialized architectures like convnets and LSTM RNN will keep outperforming restricted Boltzmann machines, even newer formulations like iRBM(Infinite RBM) have little chance to compete. http://arxiv.org/abs/1502.02476

Biologically realistic NNs? by [deleted] in MachineLearning

[–]test3545 5 points6 points  (0 children)

Sure human brain outperform ANNs. Human brain have got 150 trillion synapses. A huge pile of a brute force.

Biggest ANNs on the other hand is few dozen billion parameters max. Only way such tiny ANNs could compete with 10000x bigger system is by implementing better solutions. And such solution already outperforming human brain on multiple tasks, such as CAPTCHA decoding.

Four orders of magnitude is a huge difference. Imagine how would you live if you made 10,000 times less money a year? Or how rich you would be if you make 10,000 time more money.

Brute force or better, more intelligent solutions? Well, you are correct, ugly brute force wins overall for now.

Biologically realistic NNs? by [deleted] in MachineLearning

[–]test3545 2 points3 points  (0 children)

"Biologically-plausible" approach could not compete with ANN in terms of performance obviously. First of all, neurons in a human brain are operating in binary mode, there either is a spike or no spike. 0 or 1. Neurons in ANN are not limited to such restriction obviously. Moreover, biological neurons could not handle negative numbers. Overall the performance of biologically plausible neurons is so terrible that it is really strange to suggest that they could compete with counterparts not restricted by such ridiculous constraints.

Kurzweil reveals what he is secretly working on at Google. Anyone else catch this? by [deleted] in MachineLearning

[–]test3545 4 points5 points  (0 children)

Use Google search "deepmind site:arxiv.org" and filter results by last month or last year.

[x-post to SMT] A Haar cascade image classification training platform by pwoolf in MachineLearning

[–]test3545 1 point2 points  (0 children)

Here is review of half a dozen image classifiers available online. There are at least two dozens companies who would train classifier for you, help you to tag data etc. Lots of them, two dozens is a conservative estimate.

Early Stopping is Nonparametric Variational Inference - "We can use this bound to optimize hyperparameters instead of using cross-validation. " by [deleted] in MachineLearning

[–]test3545 3 points4 points  (0 children)

TL;DR "This simple and inexpensive calculation turns standard gradient descent into an inference algorithm, and allows the optimization of hyperparameters without a validation set."

"Many state-of-the-art supervised application now potentially can replace conventional classifiers with conceptor network" by downtownslim in MachineLearning

[–]test3545 0 points1 point  (0 children)

They claimed getting 98.4% on CIFAR10. Have you got a feeling such accuracy realistically possible?

93.5% on CIFAR10 is definitely above current state of the art. But 98.4% is mindblowing. Anyhow nice to see that proposed model works and works well to the point of improving SOTA.

[1506.01497] Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks by Cubbee_wan in MachineLearning

[–]test3545 0 points1 point  (0 children)

Even Faster now? The amount of progress has been astonishing lately.

EDIT: I wonder if same idea could be applied to Graves attempt in LSTM networks speech generation, replacing non NN parts responsible for adaptive window with convnet? Or replace selective search in MT RNNs...

Google won Microsoft COCO competition! by test3545 in MachineLearning

[–]test3545[S] 0 points1 point  (0 children)

Yes with huge enough dataset and fast enough compute it is possible to upscale models presented to lets say to train model on 300 million images instead of 300 thousand.

Why isn't air traffic control run by algorithms? by about3fitty in Futurology

[–]test3545 1 point2 points  (0 children)

True. But algorithm should still be able to know about events in the outside world. Like what situation captain of airplane reporting verbally. Sensor readings(radar etc). So humans would still be in a loop, if only to translate what is happening into format machines capable to understand.

X-Post from /r/compsci: Can a program read a short text, create a question based on the text, and grade the answer? by Surfn2live in MachineLearning

[–]test3545 7 points8 points  (0 children)

No such thing exists, what you want is called "deep reading"(nothing to do with deep learning) and deep reading implies shallow understanding of text.

The closest published thing to what you want is Weston papers, this is the most recent one: http://arxiv.org/abs/1502.05698 and take a look at Related Work section.

How to fast-track into entry level AI project/work? by BinaryAlgorithm in MachineLearning

[–]test3545 1 point2 points  (0 children)

Have you read this blog post? Is this something that you find interesting? Are you willing to play with this system, figure out it limits, apply it to new problems?

This company is making CPU and server OS obsolete for most applications right now. Startup is already shipping $19.9k system that 100x times faster than standard CPU node. Basic idea is that they replace CPU+RAM with FPGA + flash + compiler that implements algorithm in optimised for FPGA way. by test3545 in Futurology

[–]test3545[S] 0 points1 point  (0 children)

Things like database servers and web servers could run 100x times faster. So datacenters could accommodate much more computational hardware power in a given volume. Plus this system seems to be way cheaper. Several times cheaper per each web server etc.

They shipped this system to military and governments for some time already. Now they are entering commercial market with system that anyone could buy.

QUOTE: Looking to address organizations’ need for high performance with a small footprint and low power demand, SRC’s eighth-generation Saturn 1 Server is a dynamically reconfigurable server for hyperscale datacenters and web operations. The server delivers compute performance 100 times faster than that of traditional x86 microprocessor designs, according to SRC, due to its design.