This, says Togelius, is a bit like an AI developing a superstition: "You got a reward or a punishment for something – but why did you get it?" This is one of the pitfalls of "reinforcement learning", in which an AI ends up devising a wrong-headed strategy based on what it encounters in its environment. The AI doesn't know why it succeeded, it can only base its actions on learned associations. A bit like early human cultures that began to associate rituals with changes in the weather, for example.
อ่านต่อได้ที่ : โรงเรียนบ้านสันดอน
สาระน่ารู้ : work