horselover_f4t

horselover_f4t t1_j368cx1 wrote

>There are also AI that can improve themselves more than the human given data. The AlphaGo project started off with human Go matches as training data, and evolved into tabula-rasa training by self play. By the end, the AI beats the best human.

https://www.reddit.com/r/MachineLearning/comments/103694n/comment/j2ycihi/?utm_source=share&utm_medium=web2x&context=3

​

>YOU brought up the vanilla version, in a response to someone else who was talking about the zero version.

... who responded to someone who talked about the vanilla version. In my first response to you, I did not realize you were not actually the person I responded to in the first place. Apparently you have not read what they responded to, which seems to be the reason you're missing the context.

I assume they must be laughing if they see us still talking about this.

1

horselover_f4t t1_j31e8tm wrote

>The system's neural networks were initially bootstrapped from human
gameplay expertise. AlphaGo was initially trained to mimic human play by
attempting to match the moves of expert players from recorded
historical games, using a database of around 30 million moves.[21]
Once it had reached a certain degree of proficiency, it was trained
further by being set to play large numbers of games against other
instances of itself, using reinforcement learning to improve its play.

https://en.wikipedia.org/wiki/AlphaGo#Algorithm

1