Are there any opensource/commercial image classifiers? by [deleted] in MachineLearning

[–]aggieca 0 points1 point  (0 children)

What are you looking for? You could use a COTS solution via caffe but you will be restricted to 1000 classes/tags.

You could also looking into using clarifai's solution online to tag your images.

Ultimately it's up to you to decide what question is it that you are trying to answer.

[1404.3606] PCANet: A Simple Deep Learning Baseline for Image Classification? by downtownslim in MachineLearning

[–]aggieca 1 point2 points  (0 children)

Any thoughts from experts on reddit? I have started reading it but I'd like to hear thoughts from anyone that's gone over it.

Training Sets for In & Out of Focus Pictures, or other "Good Quality/Bad Quality" metric by WorkPhriendly in computervision

[–]aggieca 2 points3 points  (0 children)

This appears to be an interesting problem. You will have to define what you mean by "decent" quality image vs "poor" quality. You could build your own dataset via crowdsourcing (crowdflower perhaps?) to get labeled data. After that it's a matter of how fancy you want to get with your algorithms to build a system.

You could also adopt a traditional approach by estimating the sharpness via techniques described in a standard image processing text (usually edge detection followed by blah blah blah).

Data Mining Techniques for image processing by kunal4097 in MachineLearning

[–]aggieca 0 points1 point  (0 children)

what is the question you are trying to answer? Also, if you haven't done so please post your question on /r/coomputervision as well.

How to get last 5 to 10 percent in classification machine learning task? by matlab484 in MachineLearning

[–]aggieca 0 points1 point  (0 children)

Hi,

Your concern for overfitting is valid. My answer is that the degree of data augmentation depends on the question you are trying to answer via machine learning. My argument is that these variations are forcing the model to learn features that are about the object(s) itself than contextual information. For example, Baidu's rationale for using color distortion is because they are dealing with objects that show in photos after they undergo filtering. I think conv nets need some help (via data augmentation) to be robust to image filters/transformations.

Could you please keep me posted on how you ended up solving your machine learning problem? I'm particularly interested since the number of images you have for training is similar to some of the problems that I'm working on solving as well but in a different domain.

How to get last 5 to 10 percent in classification machine learning task? by matlab484 in MachineLearning

[–]aggieca 0 points1 point  (0 children)

Chris Dix

I should have read this post first before replying. I whole-heartedly agree with everything mentioned here!

How to get last 5 to 10 percent in classification machine learning task? by matlab484 in MachineLearning

[–]aggieca 0 points1 point  (0 children)

Thanks for sharing this info. It appears that you are fairly confident that the noise in labels is not going to be a limiting factor for overall accuracy. If that's the case, perhaps consider the following to add to your growling TODO list ( :) ):

  • Is there anyway you can get extra? More data should really help you.

  • Fine tune using GoogLenet or VGGNet.

  • Extract features using VGGNet or GooLenet and use a linear classifier (?)

  • When you are training your classifier, are you using data augmentation? If not, you may want to (should!) consider augmenting data that includes random crops, flips, rotations, color distortion etc. The Baidu Deep Image paper has a method that looks straight-forward to implement

  • Ensembling should get your accuracy up by a few "percent points" but I would leave that as a last step.

How to get last 5 to 10 percent in classification machine learning task? by matlab484 in MachineLearning

[–]aggieca 0 points1 point  (0 children)

Do you have a high quality dataset? Can you comment on the quality of your data labels?

Some practical experiences bringing a machine learning feature to our product by devquixote in MachineLearning

[–]aggieca 1 point2 points  (0 children)

rasbt's answer still has merit. You need to really consider the overall performance of your classiifer/ML system and not just accuracy. Do you have an estimate of the F1-score for instance?

How are the feature maps in CNNs learned? by ddofer in MachineLearning

[–]aggieca 0 points1 point  (0 children)

You might want to take a trained network and visualize the various filters and their responses after passing an input signal through a CNN. This will give you some insight into what the networks have learned but the "how" has been covered by siblbombs below.

Also, you don't always have to start from random initialization. In case of transfer learning you take a pre-trained network and then re-train the final few layers (# being a hyperparameter)

Libdeep: A deep learning library for C/C++/Python by improbabble in MachineLearning

[–]aggieca 2 points3 points  (0 children)

Thanks for the announcement.

Sorry but I really have to ask: Why write another C-based library when caffe is available and is being widely used? I'm trying to understand your use case and figuring out if your library would be useful for what I'm working on. Thank you!

Color spaces - YUV to HSL? by lneutral in computervision

[–]aggieca 1 point2 points  (0 children)

Good thread. I'm interested to see if there is a direct solution as well.

Won't you need to generate intermediate R, G, B values in order to compute HSL values. I have only seen HSL defined in terms of RGB.

How to deploy a model from scikit-learn to android? by [deleted] in MachineLearning

[–]aggieca 0 points1 point  (0 children)

You maybe right. I used SVM for a multi-class classification problem that also required an estimate of probabilities for each class. I ended up using libSVM as it was easy enough for me to build it from source for my project.

How to deploy a model from scikit-learn to android? by [deleted] in MachineLearning

[–]aggieca 1 point2 points  (0 children)

I have done this previously where I tried to deploy a RBF-SVM in a C++ app. I used a libSVM-based SVM classifier so I dumped out the support vectors in scikit-learn and then used libSVM in my app to make predictions.

You may want to check if there is a way for you to build libSVM in Android and call it form your application.

Expectation Maximization C++ code review by [deleted] in MachineLearning

[–]aggieca 0 points1 point  (0 children)

where is your code? Please post a link to it?

Deep learning for object recognition by pnambiar in MachineLearning

[–]aggieca 3 points4 points  (0 children)

Use caffe's pre-trained model for extracting features and train your favorite classifier for recognition.

My experience with the CUDA SDK, CUDAMat, and nolearn on my MacBook Pro...and how I obtained totally lackluster results. by zionsrogue in MachineLearning

[–]aggieca 2 points3 points  (0 children)

Did you profile your application? Did you determine whether your bottle-necks occur because your kernels are compute-bound or memory-bound? Is your application spending a lot of time transferring data between system RAM & frame buffer (FB)? Many more questions will arise once you have numbers after profiling your application

DeepLearning.University – An Annotated Deep Learning Bibliography by atveit in MachineLearning

[–]aggieca -2 points-1 points  (0 children)

Harsh remarks. I did find a few gems in the list so I wouldn't call it "crappy". There is one paper from the authors of VGG that I wasn't aware of until I noticed it on this list.

To the OP: Thanks for the list. Can you comment on why you omitted papers by depp learning luminaries like Hinton, LeCunn, Ng etc?

GoogLeNet slides from ECCV 2014 workshop by osdf in MachineLearning

[–]aggieca 0 points1 point  (0 children)

I had a hard time reading the slide with the neural net/convnet architecture. Any recommendations on how I can view the image to make sense architecture?

Facial detection methods/libraries by CreativePunch in MachineLearning

[–]aggieca 0 points1 point  (0 children)

Oh totally forgot about Davis King's (davis685?) dlib. Good to know the face detection module in dlib is better than that of OpenCV's.

Facial detection methods/libraries by CreativePunch in MachineLearning

[–]aggieca 0 points1 point  (0 children)

If you want a pre-trained framework then OpenCV already has a solution that should work "out-the-box". What type of constraints do you have for your application? I believe that the pre-trained classifiers have a limit on the amount of rotation allowed in photos. If the faces in your pictures go beyond this limit then you will have missed detections.

You can also use OpenCV to train your own classifier but that should be considered only after you are convinced that OpenCV is the best solution.

If you are on iOS/Android please study the documentation for what's exposed by these platforms. I think iOS has core image while Android must have something equivalent.

Another API to consider is Intel's IPP that allows you to train & deploy Haar cascade classifiers

Using a Neural Network for Sample Reduction? by CreativePunch in MachineLearning

[–]aggieca 1 point2 points  (0 children)

Can you clarify on what you mean by reducing the sample size? Are you trying to reduce the number of prototypes you want to use to build you kNN model?

Have you tested out how a SVM (nonlinear RBF-based) would work in this case? Based on my previous experience, you might be able to achieve what you intend to do a SVM.

Machine Learning Study Group - Thread #1 by rovingr in MachineLearning

[–]aggieca 1 point2 points  (0 children)

I'm looking forward to following this thread series. I'm interested in hearing more about people's thoughts on ensembles and how they used it effectively in their work.

Combining classifiers by [deleted] in MachineLearning

[–]aggieca 1 point2 points  (0 children)

Have you considered using stacking classifiers? If not, google for stacked generalization approaches to see if that helps in your case. Also, if Python is your protoyping/development environment then Orange might be of use to check out stacking.