Hacker News new | past | comments | ask | show | jobs | submit login
Learning to Imitate (stanford.edu)
30 points by ghgr on Nov 8, 2022 | hide | past | favorite | 5 comments



"Another challenge with current AI systems is that they require explicit programming or hand-designing of reward functions in-order to make correct decisions"

I read the exact same thing 20 years ago.


Personally I think calling randomized fine tuning of large statistical models "Artificial Intelligence" is incredibly pretentious. But then again, humans pretend to be so much more then the naked monkeys we really are and isn't that the whole point of this brief 'civilization' phenomenon we seem to bloom into shortly before killing our host planet?


Has it changed since then?


Personally, I would argue that David Silver's body of work at DeepMind is making a very strong case in favor of "simple reward + lots of compute". Richard Sutton wrote about this back in 2019 as The Bitter Lesson[1]. An important note is that David Silver did his PhD under Sutton.

Quite frankly, it's been a while since I've read through Silver's papers, but the original Deep Q-Networks paper[2] is probably a good start.

1: https://www.cs.utexas.edu/~eunsol/courses/data/bitter_lesson... 2: https://daiwk.github.io/assets/dqn.pdf


Can this be used to imitate actual work on the computer like data entry , web scrapping or writing code?




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: