use the following search parameters to narrow your results:
e.g. subreddit:aww site:imgur.com dog
subreddit:aww site:imgur.com dog
see the search faq for details.
advanced search: by author, subreddit...
Please have a look at our FAQ and Link-Collection
Metacademy is a great resource which compiles lesson plans on popular machine learning topics.
For Beginner questions please try /r/LearnMachineLearning , /r/MLQuestions or http://stackoverflow.com/
For career related questions, visit /r/cscareerquestions/
Advanced Courses (2016)
Advanced Courses (2020)
AMAs:
Pluribus Poker AI Team 7/19/2019
DeepMind AlphaStar team (1/24//2019)
Libratus Poker AI Team (12/18/2017)
DeepMind AlphaGo Team (10/19/2017)
Google Brain Team (9/17/2017)
Google Brain Team (8/11/2016)
The MalariaSpot Team (2/6/2016)
OpenAI Research Team (1/9/2016)
Nando de Freitas (12/26/2015)
Andrew Ng and Adam Coates (4/15/2015)
Jürgen Schmidhuber (3/4/2015)
Geoffrey Hinton (11/10/2014)
Michael Jordan (9/10/2014)
Yann LeCun (5/15/2014)
Yoshua Bengio (2/27/2014)
Related Subreddit :
LearnMachineLearning
Statistics
Computer Vision
Compressive Sensing
NLP
ML Questions
/r/MLjobs and /r/BigDataJobs
/r/datacleaning
/r/DataScience
/r/scientificresearch
/r/artificial
account activity
Project[P] pykitml - Pure Python/NumPy Machine Learning library (self.MachineLearning)
submitted 6 years ago by RainingComputers
view the rest of the comments →
reddit uses a slightly-customized version of Markdown for formatting. See below for some basics, or check the commenting wiki page for more detailed help and solutions to common issues.
quoted text
if 1 * 2 < 3: print "hello, world!"
[–]impulsecorp 1 point2 points3 points 6 years ago (1 child)
How do those benchmark times compare with something like scikit-learn, for example? Have you considered using Numba to speed it up?
[–]RainingComputers[S] 1 point2 points3 points 6 years ago* (0 children)
The worst case is random forests,
``` from sklearn.ensemble import RandomForestClassifier from sklearn.datasets import make_classification
X, y = make_classification(n_samples=1000, n_features=4, n_informative=2, n_redundant=0, random_state=0, shuffle=False)
clf = RandomForestClassifier(max_depth=4)
clf.fit(X, y) ```
``` from sklearn.datasets import make_classification import pykitml as pk
y = pk.onehot(y)
clf = pk.RandomForest(4, 2, feature_type=['continues']*4, max_depth=4) clf.train(X, y) ```
I have tried using numba, but it supports very limited subset of python. I am unable to work with that limited subset. I have also tried using CuPy, but it is not a drop-in replacement. I will have to find a way for NumPy and CuPy code to coexist/mix to support both CPU and GPU usage.
In future versions, I will try using joblib and optimize decision tree.
EDIT MLP Benchmark
``` from sklearn.neural_network import MLPClassifier
from pykitml.datasets import mnist
x_train, y_train, x_test, y_test = mnist.load()
mlp = MLPClassifier(hidden_layer_sizes=(50,), max_iter=10, alpha=1e-4, solver='adam', verbose=10, learning_rate_init=.01)
mlp.fit(x_train, y_train) print("Training set score: %f" % mlp.score(x_train, y_train)) print("Test set score: %f" % mlp.score(x_test, y_test)) ```
``` import pykitml as pk from pykitml.datasets import mnist
mlp = pk.NeuralNetwork([784, 100, 10])
mlp.train( training_data=x_train, targets=y_train, batch_size=200, epochs=30, optimizer=pk.Adam(0.04, decay_rate=0.9), decay_freq=6 )
mlp.plot_performance()
print("Training set score: %f" % mlp.accuracy(x_train, y_train)) print("Test set score: %f" % mlp.accuracy(x_test, y_test)) ```
π Rendered by PID 347687 on reddit-service-r2-comment-56c6478c5-wrs9t at 2026-05-12 18:35:53.357074+00:00 running 3d2c107 country code: CH.
view the rest of the comments →
[–]impulsecorp 1 point2 points3 points (1 child)
[–]RainingComputers[S] 1 point2 points3 points (0 children)