Hacker News new | past | comments | ask | show | jobs | submit login

The MCTS part of the algorithm per se doesn't seem super-important AFAIK from my brief read of the paper, except insofar as it shows that the 'neural' part of the algorithm was meshing well with a broadly consequentialist idiom, which is important. Another way of looking it at is that the neural algorithm was able to well 'steer' the MCTS part, and it's the neural steering which is important.



For what it's worth, there's a recent PLOS paper that I found to be a real eye-opener. If you haven't already seen it I'd give it a read: http://journals.plos.org/plosone/article?id=10.1371/journal....

It's quite a low-key affair (no GPU clusters and so on) but what I think is remarkable is that it described a fairly complex cognitive (neural) architecture that was able to mimic certain kinds of child-level cognition with TINY amounts of training data. Instead, human supervisors guided the evolution of its cognitive strategy in a white-box kind of fashion to encourage it to answer questions as children did.

In many respects (like training data volume, layer depth, and so on) it couldn't be further from the current deep learning trends, and yet it seemed much more along the lines of what I imagine an actual AGI would be doing, especially one we hope to control.


Great share, this is quite impressive. The system doesn't require much training input to go from blank slate to conversation-capable (1587 input sentences -> 521 output sentences; see Appendix S1 for examples). This high learning efficiency might imply that it resembles language-processing architecture in humans. At the low level it's neurons, but organised and connected in a specific, planned way.

A key point is that the central executive (the core, which controls the flow of data between slave systems) is a trainable neural network itself, which learns to generate the "mental actions" that control the flow of data between slave systems (like short-term and long-term memory), rather than rely on fixed rules to control the flow. This allows the system to generalise.

(Some of the training itself doesn't seem to be that similar to training a human child, though. It's instead tailored to the system's architecture.)

I'm excited to see how much richer this and similar systems will become, as researchers improve the neural architecture and processing efficiency. Will we see truly human-like language and reasoning systems sooner than expected?




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: