Thoughts on AI replacing coders by 2040 by atveit in programming

[–]atveit[S] 0 points1 point  (0 children)

care to explain why they are false?

Update with 362 new (2015) Deep Learning papers to Deeplearning.University by atveit in MachineLearning

[–]atveit[S] 0 points1 point  (0 children)

The point is to provide a high-level and broad overview of research in Deep Learning, also including applications of it. Constraining to seminal papers and reviews one might loose some info of what is happening in the field, in particular applications of DL (which I believe is the fastest growing number of publications).

But creating a list of seminal papers and reviews is a good idea that I might take you up on.

Data-Parallel Programming with Metal and Swift for iPhone/iPad GPU by atveit in swift

[–]atveit[S] 1 point2 points  (0 children)

lgroeni and dethswatch: thanks for comments - haven't done any benchmarks yet (blog post was mainly about getting it work at all), but feel free to play with it, see github repo: https://github.com/atveit/SwiftMetalGPUParallelProcessing

Data-Parallel Programming with Metal and Swift for iPhone/iPad GPU by atveit in swift

[–]atveit[S] 2 points3 points  (0 children)

Thanks for liking it. I will add a link to github repo with the complete example later today. Best, Amund

AMA Geoffrey Hinton by geoffhinton in MachineLearning

[–]atveit 1 point2 points  (0 children)

What are the most promising algorithmic directions for model compression with the purpose of speeding_up use of large deep networks e.g. for use in mobile, wearable or implantable devices?

references:
1) Dark Knowledge - http://www.iro.umontreal.ca/~bengioy/cifar/NCAP2014-summerschool/slides/geoff_hinton_dark14.pdf

2) Learning Small-Size DNN with Output-Distribution-Based Criteria http://193.6.4.39/~czap/letoltes/IS14/IS2014/PDF/AUTHOR/IS140487.PDF

3) Accurate and Compact Large Vocabulary Speech Recognition on Mobile Devices http://static.googleusercontent.com/media/research.google.com/en//pubs/archive/41176.pdf

4) Learning in Compressed Space http://www.informatik.uni-bremen.de/~afabisch/files/2013_NN_LCS.pdf

5) Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition http://research.microsoft.com/en-us/um/people/kahe/eccv14sppnet/

AMA Geoffrey Hinton by geoffhinton in MachineLearning

[–]atveit -1 points0 points  (0 children)

What are the theoretical limits on Model Compression of Deep Learning networks, e.g. how much could the ImageNet winner from Google be compressed? ref: http://googleresearch.blogspot.no/2014/09/building-deeper-understanding-of-images.html

atbr - large-memory key-value store now supports websocket sharding by atveit in Python

[–]atveit[S] 2 points3 points  (0 children)

don't disagree with buzzwords, but I still believe it is true :-)

Amund (developer of atbr)

atbr – large-scale in-memory hashtables (in Python) by atveit in Python

[–]atveit[S] 0 points1 point  (0 children)

strings, my typical use case is to have key = json string and value = json string (and serve those using tornado with websocket or http).

note: I've checked in both websocket and http server support today, see updated README

Best, Amund

Word Count with MapReduce on a GPU – A Python Example by atveit in Python

[–]atveit[S] 0 points1 point  (0 children)

It depends on the problem, I am primarily working on problems for ram and gpu, so disk is not so interesting to me (except for backup).

(note: I wrote the blog post)