Loweren's question on Mesa-Optimizers
|Reviewed as Unreviewed (edit)|
|Tags:||None (add tags)|
OriginWhere was this question originally asked
|YouTube (Comment link)|
|On video:||The OTHER AI Alignment Problem: Mesa-Optimizers and Inner Alignment|
|Asked on Discord?||No|
Great explanation! I heard about these concepts before, but never really grasped them. So on 19:45, is this kind of scenario a realistic concern for a superintelligent AI? How would a superintelligent AI know that it's still in training? How can it distinguish between training and real data if it never seen real data? I assume programmers won't just freely provide the fact that AI is still being trained.