How can I update my emotional state regarding the urgency of AI safety?

It can be hard to know how to feel in the face of potentially world-changing events. People generally cannot easily use their thoughts to change how they feel, but we may nonetheless ask how we would prefer to feel. It can be useful to try to strike a balance between ‘feeling motivated’ and ‘not feeling overwhelmed’.

A person who is completely unaffected emotionally by a risk of catastrophe is liable to pay lip service to the ideas without actually trying to solve the problem. For certain intellectual stances, including contemplating the extinction of humanity, the proper mood is a tragic one.

Each person has to find the sweet spot, where the problem is emotionally salient enough to motivate them, but not so much as to cause them panic - leading to rushed, unthinking action[1]. You should think about your productivity, mental health and wellbeing before deciding how far you want to get involved emotionally. It can also help to discuss this with someone to help you gain some objectivity about your likely emotional response.

If you decide that you want to make the risk more real to your psyche, one technique is to “zoom in” on a concrete example. Instead of the abstract thought, “Humanity would be killed”, imagine actually real people in your life, your family, your friends, dying.

This challenge of making things real to a person comes up in many areas. For an example from a different existential risk, President Ronald Reagan found the movie The Day After effective for changing his attitude towards nuclear war. He found that a detailed story about how it would take place can also help bridge the psychological gap. You can use a similar technique here as well, developing your own detailed story of how it might happen.

More links:


  1. Unfortunately, some people have been so overtaken with the risk that they have proposed dangerous solutions that won’t actually solve the problem and would just make things worse. ↩︎