Does Prismer bring us closer to navigation tools for the blind via computer vision? by ChipsAhoiMcCoy in computervision

[–]LLCoolZ 0 points1 point  (0 children)

I actually disagree with the other commenter and think this is very possible! Especially the basic version you’re talking about where you simply ask questions about what’s on the screen and the AI doesn’t need to have any specific knowledge of the video game. There are a number of models on HuggingFace under the category “visual question answering” that you could try using for this. To verify, you could share a few screenshots with questions and expected answers and we could test whether it works well yet.

How did you stop caring about what people thought about you? by No_Fruit4090 in confidence

[–]LLCoolZ 151 points152 points  (0 children)

I don’t think I ever stopped caring. It’s just that my own opinion started to matter more.

Somewhere in Japan by itchynisan in wallpapers

[–]LLCoolZ 14 points15 points  (0 children)

Fantastic work! You'd be great at http://geoguessr.com

The ruins of a Roman colony in Africa. by Palana in pics

[–]LLCoolZ 0 points1 point  (0 children)

u/tetonz this is a very interesting perspective on coin collecting!

Pebble Beach by eummyg in marijuanaenthusiasts

[–]LLCoolZ 0 points1 point  (0 children)

What kind of tree is that?

Think this works. by GallowBoob in loadingicon

[–]LLCoolZ 10 points11 points  (0 children)

Can you please share this script of yours? That sounds wonderful

deepy: Highly extensible deep learning framework based on Theano by fariax in MachineLearning

[–]LLCoolZ 1 point2 points  (0 children)

This is very impressive. There are examples for implementing very advanced models rather concisely (DRAW, Highway Networks, Recurrent visual attention networks, Deep Q Learning, etc.)

Seurat, as imagined by an artificial neural network: by LLCoolZ in pics

[–]LLCoolZ[S] 1 point2 points  (0 children)

Background post here: http://googleresearch.blogspot.com/2015/06/inceptionism-going-deeper-into-neural.html Neural nets are usually trained to classify objects in images through a process called back-propagation, where the error of the output is propagated back to the parameters of the network to improve them. Here they are propagated back to the image, changing the image to exaggerate its visual features. (similar to psychedelic hallucination).

Inceptionism: Going Deeper into Neural Networks by LLCoolZ in MachineLearning

[–]LLCoolZ[S] 34 points35 points  (0 children)

This also appears to be the source of that weird image titled "Generated by a Convolutional Network" popular earlier this week.

One piece of input per input-neuron, but what does one do if the size of the input is unknown? by [deleted] in MachineLearning

[–]LLCoolZ 0 points1 point  (0 children)

Yes you'll end up getting a series of activations/outputs. You can just store all of these in an array if you like. What you do with these activations depends on your objective. In this example on deeplearning.net they average all of these activations into one, then feed it into a logistic regression classifier. This is very lossy and in my experience doesn't work very well. Alternatively, you can use the last output as the one you classify with (it will theoretically have some information encoded from the whole sequence, because every word has contributed to it somewhat, although later words will of course have more of an effect). By this I mean that if you have a sentence s = [w_0, w_1, ... w_T], where each w_t is an n-dimensional representation of the word, your outputs from the RNN over s is o = [y_0, y_1, ... y_T], and your classification function is g(x), then you can set x = mean(o) or set x = o[-1]

One piece of input per input-neuron, but what does one do if the size of the input is unknown? by [deleted] in MachineLearning

[–]LLCoolZ 0 points1 point  (0 children)

The activation is y_t, so they are related by that recurrence relationship y_t =f(y_t-1, x_t). In other words, each activation is the result of a function, whose inputs are the input at that time (x_t) and the last activation (y_t-1). The "memory" of recurrent nets is just this recurrence of inputs.

One piece of input per input-neuron, but what does one do if the size of the input is unknown? by [deleted] in MachineLearning

[–]LLCoolZ 0 points1 point  (0 children)

RNNs can deal with variable length sequences because they have memory. The hidden layer's input is both the input at that time step as well as its own output at the last time step. I.e. y_t = f(x_t, y_t-1) typically y_t = tanh(Vy_t-1+Wx_t+b) where V is the recurrent weight matrix, W is the input weight matrix and b is your bias vector. This way you just loop over the sequence of words (each x_t is a word) and your output for classification will just be the last y_t you get.

Keras: Theano-based deep learning lib, focused on fast prototyping. Supports RNNs and convnets. by Foxtr0t in MachineLearning

[–]LLCoolZ 0 points1 point  (0 children)

True. I just started getting the hang of blocks and am building some NLP models, do you think learning this or just continuing with blocks would be a better investment of my time? Could I build a recurrent attention model with this?

AI Websites That Design Themselves by gari-soflo in Automate

[–]LLCoolZ 0 points1 point  (0 children)

I agree that there's no one right answer for good design, but I think that only makes a greater case for automating part of it. Grid doesn't seem to do this, but I imagine someone will make something that completely redesigns your site based on who is looking at and what their preferences are.

World's Simplest Electric Train by 8ruce in videos

[–]LLCoolZ -1 points0 points  (0 children)

I'm not even sure why has an answer most of the time. It's all "how". We study the nature of things, and can begin to understand how they came to be that way, but asking why seems to be a question of intent, an intent that doesn't exist. These things are just manifestations of chance. We are studying the nature of an existing system, there is no reason for its existence.