Seems plausible. But another possibility: mediation is just a good way to clarify our thoughts, so that we can better achieve happiness along the same lines, and constraints, evolution set out? See what I mean?
-
-
Hmm, can i give a clear example of that, in ai? It just seems like whenever u write a learning algorithm, u give the reward function as an input, and never allow the agent to ever touch this function. I don't see why that would be something the agent would ever "learn" to hack.
-
Because as a generally intelligent robot, it can reverse engineer its own design, and eventually it will figure out how to hold a soldering iron to its DRM chip. The only way to prevent that is to limit its intelligence.
- 4 more replies
New conversation -
-
-
Why not allow it to intelligently decide what is rewarding? Task it with "creating the best possible reality" and let let it ponder on what that really means. Let it soak up the knowledge from humanity.
-
That’s fine. But the end goal needs to be pre-specified. In the learning literature that end goal is called the “reward.” Although intermediate goals and values (which may be subjectively felt as “rewarding” or “pleasant”) are subject to the agent. Important to distinguish. Imo.
- 11 more replies
New conversation -
-
-
But at that point, what would differentiate it from us individuals? The only fundamental difference then might be the learning and improvement rate, right?
-
That is going to be a dramatic difference. Human minds are tiny, slow and noisy, crash every few hours, and worst of all burn out after only 32 Billion clock cycles!
- 1 more reply
New conversation -
-
-
Amodei and Clark [2016] trained an RL agent on a boat racing game. The agent found a way to get high observed reward by repeatedly going in a circle in a small lagoon and hitting the same targets, while losing every race. https://blog.openai.com/faulty-reward-functions/ …
Thanks. Twitter will use this to make your timeline better. UndoUndo
-
-
-
It's not really a problem. In fact, I would say that helping everyone understand what makes them happiest, in the most lasting way, is the solution. The faster we go through the experimental/mistake-making stage, the faster we mature.
Thanks. Twitter will use this to make your timeline better. UndoUndo
-
-
-
Perhaps this is a reasonable test of a mature intelligence. We can ask an individual animal/vegetable/mineral how it's own happiness relates to the happiness of others. If it can see a connection, then perhaps it has reached the self-transcendent stage of 3D thinking.
Thanks. Twitter will use this to make your timeline better. UndoUndo
-
Loading seems to be taking a while.
Twitter may be over capacity or experiencing a momentary hiccup. Try again or visit Twitter Status for more information.