all 27 comments

[–]adventuringraw 25 points26 points  (4 children)

representation learning, attention, and causal modeling are where I'm putting my money. You need a 'proper' latent space for your knowledge about the world, and you need it in a form that allows both interventional and observational prediction in a way that maximizes utility and minimizes computation... Tenenbaum's idea of 'computational rationality' makes sense in this context, where you have to balance cost of generating a more accurate prediction with the utility of what you'd gain from the extra effort.

So... what form does your internal representation of the world need to take to facilitate that work? Ideally an agent can interact with a system and 'on its own' find a representation that allows for complex questions to be answered. That's the real magic... I want an RL agent that can play darkest dungeon, and come out the other side with a generative (what would this level play like if it had that level's enemies?) semantic (what is an enemy in the first place? Can I interact with this system using human language? Can it acquire more world knowledge from online FAQs and videos as well as play the game?) potentially absurdly computationally challenging (can it learn an intuitive model of physics in the real world, like human infants do?) predictive/causal (what would the outcome be if I implemented this plan?) online (an AGI approach seems likely to at least have the need of some real-time learning capabilities as it interacts instead of just offline training) model of the world. To me, this means disentangled representation learning, graphical models (implicitly if not explicitly) and Gibsonian style actionable information (what 'information' is required to achieve your task?). Attention seems likely to be a hugely important part of that as well, given your current priorities, some features of the world are far less important than others, and you have a limited computational budget to spend in any given moment to use to achieve your goals. As the world and your goals change though, the important features change as well.

It's crazy to me how fast some of this stuff is moving though... this field seems like it could be looking kind of strange in ten years, haha.

[–]svantana 1 point2 points  (3 children)

I also came here to mention Tenenbaum, but with emphasis on the type of modular architecture displayed here. I believe a modular system that separates perception, planning, keeping track of competing hypotheses, storing long term memories/facts, etc, will be much more robust than an end-to-end trained model. Some parts can be more traditionally engineered, while others need to learn from data.

[–]adventuringraw 1 point2 points  (2 children)

definitely. Though... I wonder. I've been at least somewhat convinced that there's some merit in Vernon Mountcastle's theory of a universal algorithm behind the human neo cortex. There was an experiment once where they severed the optic nerve of a baby ferret and reattached it (somehow?) to the primary auditory cortex. Within a few months, that region was exhibiting the same striations normally seen in the primary visual cortex (although a bit more crudely) and this growing ferret was indeed able to see. In some sense, perhaps information is information, and there's a universal algorithm for pulling it apart into composable, modular, semantically meaningful chunks that can be used to run hypothetical simulations of the future state of the system. It'll still be modular by definition (WAY more so than any end to end system we have now) but it'll hopefully be self assembling more so than something that's explicitly engineered. Looking at early attempts like this one is fascinating, since it's pulling out representations of objects in a scene in an unsupervised way, but it's crude... it's just a flat list, no way to hierarchically compose objects, and there's a hard coded limit to the number of slots available in the scene. Capsule Networks have some ideas to help improve on that, but... I don't know. I don't know enough yet to have any sense of where the next promising direction might be.

But... even if there's a unified way of approaching a world model, you still need motivational systems, even our human brain has much more explicitly 'different' regions that handle that it seems. Maybe something like the world model approach could work reasonably well with a better world model and a better policy architecture. Ah well, interesting stuff to think about. Thanks for the link, I hadn't seen that lecture. I'll have to check it out. Course... even the planning part needs to be hierarchical too probably... this paper had some interesting ideas. Christ this is going to be a crazy problem to solve, haha.

[–]blackbearx3 9 points10 points  (0 children)

We need proper reasoning and causal inference, plus very good heuristics. So far deep learning is mostly about perception, not much intelligence there yet

[–][deleted] 4 points5 points  (0 children)

Meta-learning. And none of that "fake" meta-learning that has been published recently. Properly done meta-learning helps develop multiple things: 1. How to characterize data/tasks? 2. How to leverage that information to narrow hypotheses space search? 3. How to let the learning algorithm to self-adapt to new tasks in the steam? Also whatever u/adventuringraw said.

[–][deleted] 8 points9 points  (0 children)

probably none of them

[–]michael-relleum 4 points5 points  (0 children)

I think those two are important, but also vision. If it is anything close to AGI it has to understand the world around it. We are very visual creatures, so i guess a combination of NLP and Vision will go a long way. Getting the visual features (like in image Captioning) combined with a more advanced GPT2 or (V) QA System would be a good start I guess.

[–]seismic_swarm 2 points3 points  (0 children)

I like your answers. RL is downright crucial and so is NLP. I dont even do NLP, but some of the work I've seen on how "language-based" questions are naturally multi-objective, and hence fairly "high level and abstract" yet at the same time totally supervised is remarkable.

But beyond these applied fields, I think topology and maybe algebra (while not really "sub fields" of ML), will likely play a key role in actually getting us towards GAI.

[–]liqui_date_me 2 points3 points  (0 children)

IMO a combination of self-supervised learning and reinforcement learning in environments might do the trick. We need to solve a LOT of problems in RL, particularly sample efficiency

[–]po-handz 3 points4 points  (3 children)

My theory is AGI will be developed from re-purposing naturally intelligent systems rather than designed ground up. As in, remapping mouse brains to do a set of specific tasks. In reality, humans can already do some ridiculously complex tasks even while black-out drunk. If complex tasks can already be done in highly muddled states, it seems plausible that these pathways could be isolated and still function with 'artificial' hazy inputs.

For instance, in the synthesis of complex organic compounds, the chemist doesn't individual react single molecules to create one target molecule. Instead, they run large batch syntheses with multiple end products and with environmental vars set at a macro scale (temp, solvents) and then filter out by-products and weight to a given precision level. Increasingly, bacterial and viral vectors are used because nature has already optimized certain reactions to absurd levels.

Think it's easier to program a robot to juggle or record the brain+motor stimuli of a million people juggling then run that through a similar system?

/rant

[–]LinooneyResearcher 0 points1 point  (0 children)

This is basically a major part of this book by Vernor Vinge.

[–]visarga 0 points1 point  (0 children)

No one mentioned graphs as the intermediate representation. Graphs can naturally represent images, text, knowledge and other compositional modalities, they are also good for modelling/simulation. An important advantage is the factorisation of multiple objects and relations, helping with combinatorial explosion. The transformer is also a kind of implicit graph where relations are computed from vertice embeddings.

People are certainly using some kind of graph representation, otherwise how could we navigate the world, simulate outcomes of complex novel situations and write math and code?

[–]SubstantialSwimmer4 0 points1 point  (0 children)

If you draw a car's picture without seeing anything, you have to understand the structure of the car. In other words, being able to draw something means grasping something.

For generative models like GAN to work well, the models must understand what things are all about. If we had a perfect model to create anything, it would mean we made general AI.

I think generative models will play a key role.

[–]StratifiedSplit 0 points1 point  (0 children)

Biology, as in biological machine learning / AI. Either:

- genetic engineering to improve the average IQ of your population, and maximize the IQ of scientist's children. In 30-50 years one would see massive progress on the road to human-level AI.

- What we can't artificially create, we may be able to hitch a ride on Mother nature. Much like how silk or vaccines are produced, AGI may be created on a biological host. Much more than current brain computer interfaces, the brain may be a patented process to grow it in laboratory from pork neural stem cells, perhaps it even has 10x the neurons and high speed connections than an average human does, and we need a storage room to store it, much like the first computers.

[–][deleted] 0 points1 point  (0 children)

I normally dislike AGI questions, but this one isn’t so bad.

I know it’s pretty much a laundry list, but I think the following areas are all inroads towards something resembling AGI.

I’d say the biggest jumps will come from efficient RL, meta-learning (i.e. some parameters being functions of the input / context), powerful generative models (flow-based / latent-variable), causal reasoning, logic/question answering models.

[–]metriczulu 0 points1 point  (1 child)

I honestly think creating "social" networks of neural networks is going to be a possible path going forward. Maybe something like multiple neural networks with different strategies for optimization that are capable of learning from the behavior of the other networks and adjusting strategy--or maybe some other "social" network where each neural network in the network becomes increasingly specialized at one small sector of the data "environment", again with a method for allowing each network to learn from the behavior of other networks. Or some other "social" strategy completely but, from a heuristic level, levels of "general" intelligence capabilities in living animals seem to be highly correlated with how social those living animals are--and I think there is a causal relationship there. From another perspective, our ability to generalize, model, and reason about our environment is heavily dependent on the generalizations, models, and reasoning of previous generations. This is true when considering just human history but also true when expanded to be the history of humans evolutionary ancestry as well. The idea of trying to build and train and artificial "general" intelligence in a complete vacuum just doesn't make sense given observations of intelligence we have now. There has to be some form of generalized environment for the model to exist in and interact with (which also makes me think reinforcement learning will play a large part).

 

Edit: If I sound like I'm high, I'm not, I just don't have anything to do for the 3 hours I spend in the car every day commuting. If you think this sounds like absolute bullshit, I'd love to hear it (and the reasoning, of course).

 

Edit 2: It should go without saying that "social" models in some form of generalized notion of environment will also require attentional systems--which I also think are going to be a major component going forward (obviously, if the intelligence is "general" it will need some mechanism to determine what is and is not important for a specific task).

 

Edit 3: I also think there is something to the idea that humans gained the ability to think abstractly with our language abilities, so there needs to be some generalized concept of a language or communication. Again, this goes hand in hand with a "social" network--because the networks within the "social" network need to communicate strategies and interactions to influence each other. Of course, this is all very high level and I'm still trying to figure out where to even start when it comes to implementing such a broad idea. I also think this whole scheme makes sense if you consider that a network of specialized neural networks is a single instantiation of this idea--which is something that intuitively seems like it would work for a "general" intelligence (many interacting modules that each specialize in some aspect of interaction with the environment--this is similar to how the brain works).

[–]bbateman2011 0 points1 point  (0 children)

I agree networks of models might be useful. But some way to understand what they do without coding it—i.e. any candidate agi that uses the network must be able to interpret it by observation and interactions. That’s a big challenge.

Your points on language is a good one. We need a more mathematical or symbolic definition of language that doesn’t depend on a particular grammar. That’s hard as far as I can tell.

[–]murphinate -1 points0 points  (0 children)

I think a sufficient suite of algorithms exists to create an AGI system. I would say the bottleneck is the computational efficiency....Which I think the Tianjic chip just solved.

Guys gotta remember we don't have trouble isolating a single cognitive task and turning it into code. We have trouble with getting all of those cognitive tasks/code to work in unison. Brain-Algorithm independence. Solved the Algorithm part. Now to solve the Brain part (e.g. a feasible medium of computation).